Tools

The 4 Best Web Scraping Tools – From $0

What is Web Scraping?

Web scraping, sometimes called data scraping, data extraction, or web harvesting is merely the process of collecting data from websites and storing it on your local database or spreadsheets. Today, web scraping tools are necessary for the modern marketer.

For the uninitiated, web scraping may sound like one of these scary tech buzzwords, but it’s not that big a deal, technically speaking. To do any web scraping though, you need the right tools. Web scraping tools come in handy not only for recruitment purposes but also for marketing, finance, e-commerce, and many other industries.

The tools we are going to examine are scraping-bot.io,import.io, webscraper.io, and data-miner.io.

Before we move any further, I can hear you asking “Why should I spend my precious time learning how to web scrape”?

Keep reading to find

Why Learn Web Scraping?

Lead Generation

Leads can be either direct customers or influencers who will help you represent your brand. You can search for leads by yourself, looking through websites and social media.

But all this research takes time. So what if you could leave this research to the machines while you were focusing on more on strategic and vital tasks?

Web scraping is the cutting-edge technology that is aimed to gather your leads’ contact information from millions of web pages within the shortest time.

Customer knowledge

This is an information age, and a lot of buyers base their judgment on online reviews. So, it is very important to find out what people say about your brand, your products, your services, and those of your competitors. Web scraping tools can help you gather all this data and help you know how to improve.

Price Comparison & Optimization

Aka spying on your competitors.

Usually, the biggest challenge for a small business is how to increase the prices without losing clients. However, without raising prices, it will be impossible to get more profit.

This is where you can use web scraping tools to increase profit:

– keep you informed of any competitors’ price changes to quickly react and optimize your prices.

– track the success of promotions and campaigns made by your competitors to know what works best.

1. Import.Io

About

First on our web scraping tools list is Import.io.

Import.io is an enterprise-ready platform that extracts and transforms data. With Import.io you can first extract the data you need, wrangle the data to the format you want and gain insight through data visualization. This tool allows people to convert unstructured web data into a structured format for use in Machine Learning, Artificial Intelligence, Retail Price Monitoring, Store Locators as well as academic and other research.

So how does web scraping work with this tool?

How it works

Step 1: First, find the page where your data is located. For instance, a product page on Amazon.com.

Step 2: Copy and paste the URL from that page into Import.io, to create an extractor that will attempt to get the right data.

Step 3: Click Go and Import.io will query the page. It will use machine learning to try to determine what data you want.

Step 4: Once it’s done, you can decide if the extracted data is what you need. In this case, we want to extract the images as well as the product names and prices into columns. We trained the extractor by clicking on the top three items in each column, which then outlines all items belonging to that column in green.

Step 5: Import.io then populates the rest of the column for the product names and prices.

Step 6: Next, click on ‘Extract data from the website’.

Step 7: Import.io has detected that the product listing data spans more than one page, so you can add as many pages as needed to ensure that you get every product in this category into your spreadsheet.

Step 8: Now, you can download the images, product names, and prices.

Step 9: First, download the product name and price into an Excel spreadsheet.

Step 10: Next, download the images as files and you are ready to put them to good use!

What else can you do with Import.Io?

What we just examined was how to transform a basic list page of data into a spreadsheet.

There’s much more you can do, such as:

  • Link a listing page to data contained on the detail pages for each product.
  • Schedule a ‘change report’ to run daily to track when prices change, or items are removed or added to the category.
  • Compare product prices on Amazon to other online retailers, such as Walmart, Target, etc.
  • Visualize the data in charts and graphs using Import.io Insights.
  • Feed this data into your internal processes or analysis tools via the Import.io APIs.

2. Webscraper.io

About

Using the webscrape.io  extension, you can create a plan (sitemap) about how a website should be inspected and what should be extracted. Using these sitemaps, Web Scraper will navigate the site accordingly and extract all data. Scraped data later can be exported as CSV.

In the demo below we will scrape the data from Macy’s website and export the data to a CSV file. We will concentrate on how to get product detail information from their Home category.

Let’s get started!

How it works

Step 1: You’ll need to download the Chrome browser if you don’t already have it along with WebScraper.io, which is a Chrome extension. After downloading the extension, you should see a spider web icon on the right side of the browser toolbar. If it isn’t there then, try restarting your browser.

Step 2: Select the hamburger menu at the far right side of your toolbar, go to “More Tools” and then select “Developer Tools”. This will open up a developer tools window at the bottom of the browser.

Step 3: Select “Web Scraper” from the developer tools and then select “Create new sitemap” from the options. The sitemap is the blueprint for how you want the scraper to navigate through the website and obtain the data you want. Give your sitemap a name along with the URL where you want the scraper to start.

For our Macy’s example, we will be starting at the Homepage. Also, make sure to go to the page in the browser. Since this tool works in the browser, we navigate through the site while setting up our sitemap.

Step 4: After setting up the initial starting point, you’ll be able to add the first selector. A selector is essentially what you want the web scraper to do next. This is where we will tell the web scraper to navigate from the Macy’s Homepage to their Home category where we’ll look to get product detail data from their home goods section. Make sure that you’re inside the root selector and select the add new selector button.

Step 5: This step is about setting up the Home link selector. We want the scraper to select the Home category from Macy’s Homepage so we can then enter the home goods section of the website. The type of selector will be a link since selecting this button will link us to that section of the site. We give the selector an ID and choose the type of the Type field.In our case, this will be a Link type.

What we do then is click the Select button, which brings the Element preview, which is a toolbar that will obtain the information (link and the href attribute of the link) of the element you are selecting on the page. If you select “Enable key events” on the toolbar, you’ll see an “S”, “P”, and “C”. If you hold down S on your keyboard while hovering over the area you want to select, it’ll get the information needed from that link. Here, we will hover over the Home category button, which gets the element’s info and places it in the Element preview.

Then select the “Done selecting!” button, which will take that element info and place it in the Selector field of your selector creation window. Then save the selector.

Step 6: Next, we have to create a selector for the subcategories of the Home category. This selector will allow the scraper to get the product detail from each subcategory as it iterates over each one. Similar to the last step, this will be creating a link selector, but this will be for multiple links. In our sitemap be sure to select the previous selector, “home_link”.

We do this because it is a hierarchical setup in which we navigate the site so this new selector will be a child of the previous one. Once we’re inside of “home_link” we add a new selector. In the browser navigate to the Home category, and you should see on the left side of the page subcategories under “Home Categories”.

Fill out the ID field (I call it home_categories_links) and the Type field as Link. Select “Multiple” underneath the Selector field and then enable key events in the Element preview. Then hold S on your keyboard and start selecting the links under Home Categories.

After you select two of the links, the scraper is smart enough to detect the pattern and select the rest of the links that have common element information for every link. This way when the scraper is navigating the site, it’ll know it has to go through all of those subcategories and get product info. Be sure that the element’s info is in the Selector field and then save the selector.

Step 7: Select the product links under the subcategories. To get a product’s details, we need the scraper to select the product from each subcategory. Once again similar, to the last step, let’s make sure we are now inside of our previous selector “home_categories_links” and then add a new selector. In the browser, select one of the subcategories, so we are on that page.

Give the selector a name (I called it “item_links”). We will be selecting multiple links again. So set up the selector in the same way as the previous step. In this case, you can either link to choose the product’s title or the image since both links to the product detail page.

I choose to select the image. Once you start to select multiple product images while holding down the S on your keyboard, you’ll notice that similarly to the previous step, all of the image boxes will be selected and the common element info will be in the element preview toolbar. Verify that this info is in the Selector field and save that selector.

Note: A quick recap of what the parent/child relationship of our sitemap graph would look like at this point:

Step 8: Select the product detail information you want. Make sure you’re inside the “home_categories_links” selector and create a new selector. In the browser select one of the products from the subcategory so that you‘re on that product’s detail page. We are going to get a product’s name, price, color and image, which will be a URL to the image.

We will create four selectors for these, which will all be children of the “home_categories_links” selector. To add a new selector for the product name and give it an ID (I called it “item_name”). The selector type will be Text.

Bring up the Element preview and select the text of the item’s name to obtain the element info and then save this selector. The same steps will apply for the price, colour and image selectors. Regarding the image, the only difference will be the selector’s type, which will be Image instead of Text.

Step 9: Verify your sitemap navigation. You can view your sitemap’s parent/child relationship by selecting the sitemap’s drop-down and then selecting “Selector Graph”.

Step 10: Scrape dat_data! To start scraping select “Scrape ” under the sitemap’s drop-down. This will take you to a screen that allows you to adjust the request interval and page load delay times in milliseconds. The default is 2000, but in the case of Macy’s, I noticed that it had to be increased to 3000, to allow the product detail page to fully load in time before trying to get the information from the page. Once you select “Start Scraping”,  a secondary browser will launch that allows you to observe the scraping in progress.

Step 11: Export as CSV. While the website scraping is in progress, you can refresh to show the data collected so far. After the scraping is finished, you can then export the data into a CSV, which is located under the sitemap dropdown, and then you’re done!

3. Parsehub

About

ParseHub is a powerful visual interfaced web scraping tool, that allows for easy web scraping without the use of code. ParseHub makes it easy for anyone to automate their data extraction from any website. This web scraping tool is often used by sales teams, data researchers, and analysts, to name a few.

ParseHub’s free plan can be used without a credit card, and is worth $99 a month. The free plan includes 200 pages of scraping per run and 5 public projects. This is more than enough for someone trying to scrape a small project or as a trial.

The most notable features of ParseHub are its IP Rotation and Scheduling, which are included with the Standard and Professional plans. With IP Rotation, you will no longer be blocked by websites that block web scrapers, such as Yelp, e-commerce websites and other directories. With scheduling, you can make sure all your data is up to date with specific intervals.

ParseHub also has an expansive amount of help documents, blog posts, and YouTube videos on scraping multiple websites. Chances are the website you want to scrape has a tutorial already! They also provide live chat support to make sure your data gets extracted smoothly.

How it works

Step 1: Begin by visiting ParseHub.com and clicking “Download ParseHub for Free”. You can download ParseHub for Windows, Mac and Linux.

Step 2: Open ParseHub and create an account if you haven’t already. Then login into the application. You will then be directed to the home screen:

Step 3: Create a new project by clicking the “New Project” button.

Step 4: Enter the URL you want to scrape from, in this example we will use Amazon. Click the first product name and it should turn green. The AI will show yellow borders around the rest of the product names. Click the next product name and all the products will be extracted!

Step 5: To scrape the rest of the pages, you need to add pagination. Start by scrolling down to the next page button, and create a new select for it. Then you want to add a click command to that selection. You will see this popup which you should click “Yes” and choose 0 pages to get data from all the pages:

Step 6: You are ready to export your data, click the green “Get Data” button and then “Run”. ParseHub will now begin to scrape data on the ParseHub servers. Once the data is extracted, you can download it as CSV, JSON or API:

Overall, ParseHub is an amazing free web scraper that allows you to scrape data with an easy-to-use interface. ParseHub comes with many tutorials within it, which can help you get started. For more ParseHub scraping tutorials and tips, visit their blog and YouTube channel.

4. Data Miner

About

Date Miner is our second web scraping tool for today. It is a chrome extension software that assists you in extracting data that you see in your browser and saves it into an Excel spreadsheet file.

Data Miner is a personal browser extension that helps you transform HTML data in your browser window into a clean table format.

When using Data Miner, the data you scrape is always private. Your data or your credentials never leave your browser and never touch Data Miner’s server. Only you have access to the data you scrape whether you have the tool.

Data miner features a function called ‘recipes’. Recipes are data extraction instructions that Data Miner uses to extract data from websites. Recipes contain name and position of HTML elements on a web page. Furthermore, Recipes do not include actual data or private information about you.

When you visit a website, Data Miner automatically filters thousands of recipes that users have created and shared and shows only the ones that are appropriate for the site you are currently viewing.

How it works

Step 1: Visit the site you want, launch Recipe Creator, and pick your page type.

Step 2: Listing pages require rows and have multiple pages while detail pages only have one page and only need columns.

Step 3: Starting with a listing page, hover your mouse over the data until a highlighted box encloses all the information you are looking to scrape.

Step 4: Once the Row is highlighted, press shift, then on the tool select one of the suggested classes to lock in the selection.

Step 5: You can now start selecting your individual data. Click on the Column tab and select “col1”.

Step 6: Give the column a name, hover over the data you wish to extract in this column, and press Shift.

Step 7: Pick the class that highlights the data the best. (Helpful tip – use the Up Parent button for more options if the data isn’t selected correctly).

Step 8: Once the data is highlighted correctly, click on the class name confirm it and click Data double check your work.

Step 9: Continue creating by clicking “+ Column”.

Step 10: Once you have all columns, finish by clicking the Save tab at the top. Give the recipe a name and click Wave.

Step 11: Recipes will save over each other unless you start a new recipe. Start a new recipe by clicking the new recipe button on the save tab.

Basic Selectors

Advanced Selectors and Selector Combos

Before you go

There are many web scraping tools to choose from. I encourage you to research which tool or service to use based on your needs and budget.

To sum up, web scraping can be used in any sphere of business: e-commerce, education, healthcare, real estate, research, marketing, etc. Thanks to web scraping you can:

– generate new leads

– find new ways to step forward among competitors

– increase your profits

– finding reliable partners

If you use any other web scraping tools, please let me know in the comments. I’d love to hear them!

Bye-bye – for now!

View Comments

  • Hi, great article, but I don't find any reference to code your own web scraper ? Thanks

    • Hi Arnaud! Like the title suggests, this article is about 'The 3 best web scraping tools', not 'How to code your own web scraper'. Thanks for your feedback!

  • Thanks, Anna Moraglia for sharing the best article. I know the importance of scraping social media. All the mentioned Scraping tool are good as well as very useful.

  • Great Article- but you have overlooked one of the key players http://promptcloud.com/. Allow me to introduce and what we do.

    PromptCloud is a cloud-based web scraping tool that enables businesses to extract and transform data from any web or cloud source through advanced automation and intelligent mining technology. By using PromptCloud one can get clean data from any website in their desired format, frequency, and format without the technical hassle. Once data is extracted, PromptCloud helps users transform and combine it into a dataset.

    PromptCloud helps banking, retail, government, researcher, students tech industries and more to conduct background checks, monitor brands and perform research by providing data from various categories such as eCommerce, Travel, Job, Healthcare, Restaurants, Classified and many more.

    We offer a free trial to all our users so check it out for yourself and experience one of the most powerful and advanced web scraper solutions in the market. Our support team is always available and happy to assist.

Share
Published by
Theodore Moulos

Recent Posts

Top 12 Performance Marketing Agencies in 2024

A list with some of the best performance marketing agencies that are experts in PPC,…

2 days ago

How Visual Hierarchy in UX Design Can Transform Your CRO

Page load speed, a compelling headline, social proof, and a clear call to action. All…

1 week ago

Building an AI-First Mindset

AI knowledge or prompt engineering alone does not equip teams to lead in the AI…

3 days ago

Democratizing Innovation with AI

Democratizing Innovation is the potential ability of users to develop what they need themselves. AI…

4 days ago

B2B vs B2C Marketing: 7 Differences & 3 Similarities

B2B vs B2C Marketing: are they that different? In some ways, yes. From price sensitivity…

1 week ago

The Top 10 B2B Marketing Agencies [2024 Edition]

Are you looking to work with one of the top B2B marketing agencies? Here's a…

1 day ago