Giter Club home page Giter Club logo

omkarcloud / g2-scraper Goto Github PK

View Code? Open in Web Editor NEW
35.0 1.0 12.0 34.65 MB

G2 Scraper helps you collect G2 product data, including names, product descriptions, reviews, ratings, comparisons, alternatives, and more.

Home Page: https://www.omkar.cloud/

License: MIT License

Python 8.40% Dockerfile 0.14% JavaScript 4.10% TypeScript 87.36%
g2 g2-python g2-scraper g2-scraping lead-generation lead-generator leadgeneration g2-crowd-scraper g2-data g2-data-scraping

g2-scraper's Introduction

G2 Scraper Featured Image

โœจ G2 Scraper ๐Ÿš€

๐Ÿ’ฆ Find G2 Product Details ๐Ÿ’ฆ

(Programming Language - Python 3)

g2-scraper forks Repo stars g2-scraper License issues

View

Open in Gitpod


Disclaimer for G2 Scraper Project

By using G2 Scraper, you agree to comply with all applicable local and international laws related to data scraping, copyright, and privacy. The developers of G2 Scraper will not be held liable for any misuse of this software. It is the user's sole responsibility to ensure adherence to all relevant laws regarding data scraping, copyright, and privacy, and to use G2 Scraper in an ethical and legal manner, in line with both local and international regulations.

We take concerns related to the G2 Scraper Project very seriously. If you have any inquiries or issues, please contact Chetan Jain at [email protected]. We will take prompt and necessary action in response to your emails.

๐Ÿ‘‰ Explore Our Other Awesome Products

  • โœ… Botasaurus: The All-in-One Web Scraping Framework with Anti-Detection, Parallelization, Asynchronous, and Caching Superpowers.

G2 Scraper helps you scrape G2 Products. ๐Ÿš€

โšก Benefits

  1. Absolutely, 0 risk as we offer a generous 90-Day No Question Asked Refund Policy!

  2. Easily access G2 results through our user-friendly dashboard by product links or categories, and also get all product reviews.

  3. Get 101% accurate, fresh (updated within the last 30 days), and comprehensive data, allowing you to use the results to their full potential.

In just 5 minutes, you'll extract 400+ products from G2.

G2 Result

๐Ÿ“ฆ Requirements

To use this tool, you'll need:

  • Node.js version 16 or later to run the UI Dashboard (please check your Node.js version by running node -v)
  • Python for running the scraper

Don't have Node.js or Python? No problem!

You can easily run this tool within Gitpod, a cloud-based development environment, by following these steps.

๐Ÿš€ Getting Started

Let's get started by following these super simple steps:

1๏ธโƒฃ Clone the Magic ๐Ÿง™โ€โ™€๏ธ:

git clone https://github.com/omkarcloud/g2-scraper
cd g2-scraper

2๏ธโƒฃ Install Dependencies ๐Ÿ“ฆ:

python -m pip install -r requirements.txt
python run.py install

3๏ธโƒฃ Launch the UI Dashboard ๐Ÿš€:

python run.py

4๏ธโƒฃ Open your browser and go to http://localhost:3000, then press the Run button to see G2 Products. ๐Ÿ˜Ž

G2 Result

Note: If you don't have Node.js 16+ and Python installed or you are facing errors, follow this Simple FAQ here, and you will have your search results in the next 5 Minutes

โšก Enlightening Questions

โ“ How to Get Results for My Queries?

  1. Visit http://localhost:3000 and enter either the

Queries

  1. Now, simply press the Run button.

Run

โ“ What Data Points Are Scraped?

We scrape over 45+ data points, important among them are:

  • Name
  • Product Description
  • Reviews
  • Rating
  • Category
  • Comparisons
  • Alternatives
  • All Reviews upto 2500
  • And many more...

โ“ How many products, categories, and reviews are there on G2?

  • Products: Over 170K+ products are listed on G2. You can view the complete list here.
  • Categories: Over 2,051 categories are listed on G2. You can view the complete list here.
  • Reviews: Over 2.9M+ reviews are listed on G2.

โ“ Can I see a sample of the scraped data?

You can find sample overviews and reviews of 1000 products below:

โ“ How to Scrape More Products?

Kindly follow these steps to use our Rapid API and easily get the g2 product details:

  1. Sign up on RapidAPI by visiting this link.

Sign Up on RapidAPI

  1. Subscribe to the Free Plan by visiting this link.

Subscribe to Free Plan

  1. Copy the API key. Copy the API Key

  2. Put the Key in the "Rapid API Key" textbox and simply run it:

api-key

The first 50 products are free to scrape with the API. After that, you can upgrade to the Pro Plan to scrape 4,000 products for $16, and access 101% accurate, fresh, and comprehensive data points with 45+ fields.

โ“ Do You Offer Discounts for Complete or Partial Datasets?

If you want to:

  • Get the complete G2 dataset or 40K+ products
  • Get the complete G2 reviews dataset or 400K+ reviews

Please contact us, as

  • we offer a 25% discount compared to our API pricing.
  • The data will be freshly scraped, and we will deliver the JSON/CSV files of the requested dataset to you within 7 days.
  • Absolutely, 0 risk as we offer a generous 90-Day No Question Asked Refund Policy!

Contacting us will only take a few minutes but will save you a lot of time and money.

Contact Us on WhatsApp about G2 Scraper

Contact Us on Email about G2 Scraper

โ“ What if I Don't Get Value from It?

We wholeheartedly โค๏ธ believe in the value our product brings for you, especially since it has successfully worked for hundreds of people like you.

But, we also understand the reservations you might have.

That's why we've put the ball in your court: If, within the next 90 days, you feel that our product hasn't met your expectations, don't hesitate. Reach out to us, and within 24 hours, we will gladly refund your money, no questions and no hassles.

The risk is entirely on us! because we're that confident in what we've created!

โ“ How Do I Get a Refund?

We are ethical and honest, and we will not keep your money if you are not satisfied with our product. Requesting a refund is a simple process that should only take about 5 minutes.

For API:

  • You need to submit a support request here, and we will issue the refund.

request-refund

For Dataset Purchase:

  • It's even easier to get a refund in case of a dataset purchase. Simply ask for a refund via WhatsApp/Email and provide your PayPal email address. We will issue the complete refund within 24 hours, without any questions and without any hidden charges.

โ“ This Scraper is Truly One-of-a-Kind, Something I've Never Seen Before. How Did You Build It?

Thank you! We used Botasaurus, which is the secret behind our awesome G2 Scraper.

Botasaurus is a web scraping framework that makes life a lot easier for web scrapers.

It handled the hardest parts of our scraper, such as:

  • Creating a gorgeous UI dashboard with task management features
  • Sorting, filtering, and exporting data as CSV, JSON, Excel, etc.
  • Caching, parallel and asynchronous scraping

If you're a web scraper, I really recommend learning about Botasaurus here ๐Ÿš€.

Trust me, learning Botasaurus will only take 20 minutes, but I guarantee it will definitely save you thousands of hours in your life as a web scraper.

botasaurus

โ“ Advanced Questions

Having read this page, you have all the knowledge needed to effectively use the tool.

You may choose to read the following questions based on your interests:

  1. I Don't Have Python, or I'm Facing Errors When Setting Up the Scraper on My PC. How to Solve It?
  2. How to Update the Scraper to the Latest Version?

โ“ Need More Help or Have Additional Questions?

For further help, feel free to reach out to us through:

  • WhatsApp: If you prefer WhatsApp, simply send a message here. Also, to help me provide the best possible answer, please include as much detail as possible.

    Contact Us on WhatsApp about G2 Scraper

  • Email: If you prefer email, kindly send your queries to [email protected]. Also, to help me provide the best possible answer, please include as much detail as possible.

    Contact Us on Email about G2 Scraper

We look forward to helping you and will respond to emails and WhatsApp messages within 24 hours.

Good Luck!

Become one of our amazing stargazers by giving us a star โญ on GitHub!

It's just one click, but it means the world to me.

Stargazers for @omkarcloud/g2-scraper

Made with โค๏ธ using Botasaurus Web Scraping Framework

g2-scraper's People

Contributors

chetan11-dev avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar

g2-scraper's Issues

This does not work

The default targets are all located in folder cache\get_products
Adding new products, deleting the cache folder and running the script will error

Error: 500 {'status': 500, 'message': 'Server Error'}
Could not get data for 1 product due to Unknown Error.

Getting issue of opening and closing Chrome Web Browser repeatedly

Getting Error in terminal

Traceback (most recent call last):
File "/Users/akash/PycharmProjects/SCRAPPER/g2-scraper/main.py", line 5, in
launch_tasks(*tasks_to_be_run)
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/bose/launch_tasks.py", line 54, in launch_tasks
current_output = task.begin_task(current_data, task_config)
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/bose/base_task.py", line 249, in begin_task
final = run_task(False, 0)
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/bose/base_task.py", line 196, in run_task
driver = self.create_driver(config)
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/bose/base_task.py", line 73, in create_driver
driver = create_driver(config)
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/bose/create_driver.py", line 288, in create_driver
driver = retry_if_is_error(
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/bose/utils.py", line 96, in retry_if_is_error
raise e
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/bose/utils.py", line 81, in retry_if_is_error
created_result = func()
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/bose/create_driver.py", line 254, in run
driver = BoseUndetectedDriver(
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/undetected_chromedriver/init.py", line 453, in init
super(Chrome, self).init(
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/selenium/webdriver/chrome/webdriver.py", line 69, in init
super().init(DesiredCapabilities.CHROME['browserName'], "goog",
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/selenium/webdriver/chromium/webdriver.py", line 92, in init
super().init(
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/selenium/webdriver/remote/webdriver.py", line 272, in init
self.start_session(capabilities, browser_profile)
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/undetected_chromedriver/init.py", line 716, in start_session
super(selenium.webdriver.chrome.webdriver.WebDriver, self).start_session(
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/selenium/webdriver/remote/webdriver.py", line 364, in start_session
response = self.execute(Command.NEW_SESSION, parameters)
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/selenium/webdriver/remote/webdriver.py", line 429, in execute
self.error_handler.check_response(response)
File "/Users/akash/PycharmProjects/SCRAPPER/venv/lib/python3.9/site-packages/selenium/webdriver/remote/errorhandler.py", line 243, in check_response
raise exception_class(message, screen, stacktrace)
selenium.common.exceptions.WebDriverException: Message: unknown error: cannot connect to chrome at 127.0.0.1:56374
from session not created: This version of ChromeDriver only supports Chrome version 114
Current browser version is 119.0.6045.105
Stacktrace:
0 undetected_chromedriver 0x000000010dcfc6b8 undetected_chromedriver + 4937400
1 undetected_chromedriver 0x000000010dcf3b73 undetected_chromedriver + 4901747
2 undetected_chromedriver 0x000000010d8b1616 undetected_chromedriver + 435734
3 undetected_chromedriver 0x000000010d8e3d10 undetected_chromedriver + 642320
4 undetected_chromedriver 0x000000010d8daf98 undetected_chromedriver + 606104
5 undetected_chromedriver 0x000000010d922a08 undetected_chromedriver + 899592
6 undetected_chromedriver 0x000000010d921ebf undetected_chromedriver + 896703
7 undetected_chromedriver 0x000000010d918de3 undetected_chromedriver + 859619
8 undetected_chromedriver 0x000000010d8e6d7f undetected_chromedriver + 654719
9 undetected_chromedriver 0x000000010d8e80de undetected_chromedriver + 659678
10 undetected_chromedriver 0x000000010dcb82ad undetected_chromedriver + 4657837
11 undetected_chromedriver 0x000000010dcbd130 undetected_chromedriver + 4677936
12 undetected_chromedriver 0x000000010dcc3def undetected_chromedriver + 4705775
13 undetected_chromedriver 0x000000010dcbe05a undetected_chromedriver + 4681818
14 undetected_chromedriver 0x000000010dc9092c undetected_chromedriver + 4495660
15 undetected_chromedriver 0x000000010dcdb838 undetected_chromedriver + 4802616
16 undetected_chromedriver 0x000000010dcdb9b7 undetected_chromedriver + 4802999
17 undetected_chromedriver 0x000000010dcec99f undetected_chromedriver + 4872607
18 libsystem_pthread.dylib 0x00007ff815e56202 _pthread_start + 99
19 libsystem_pthread.dylib 0x00007ff815e51bab thread_start + 15

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.