Web Scraper - Top features
Our Easy Web Extract software contains a lot of advanced features to enable users to scrape content from simple
to complicate websites but doesn't require any effort to setup a web scraping project. In this page, we will show you only must-known
features which makes our web scraping tool so easy-to-use as its names.
Easily Create an Extracting Project
Creating a new project in a wizard-based window is never simpler for any user. The project setup wizard will drive you
step-by-step until all neccessary tasks done. Here are some main steps:
- Step 1: Enter a start URL that is the starting page that web scraper will load. It is often a link to an scraped product list
- Step 2: Input keywords to submit form to get results if the website requires. This step can be skipped in most of cases
- Step 3: Select an item in the list, select scraped properties of the item as data-columns
- Step 4: Select next-page-urls to reach other pages
Scraping Data in Multiple Threads
In a web scraping project, hundred thousands of links need to be crawled for harvesting.
Conventional scrapers may take you serveral hours or days. However, Easy Web Extract can run
multiple threads to browse simultaneously upto 24 different webpages to save your precious time
for waiting harvested results. Thereby Easy Web Extract can utilize your system at the best performance.
The beside animated image demonstrates extraction in 8 simultaneous threads.
Scrape Data from All Sorts of Data Loading
Some highly dynamic websites employ data loading techniques based on client-side to create asynchronous requests like AJAX.
This is really a challenge for not only primitive web scapers but also professional web scraping tools since the web content is not
embedded in the HTML source. However, Easy Web Extract has very powerful techniques to enable even novice users to harvest data from those kinds of websites.
Moreover, our web scraper even can simulate scrolling down to the bottom of the page to load more data in some particular websites like
Linkedin Contact list. In this challange, most of web scrapers keep harvesting a huge duplicated information and
quickly become plodding. However, don't worry about this nightmare because Easy Web Extract has smart
functions to avoid it.
Auto Execute Project Anytime
With embedded autorun scheduler of Easy Web Extract, you can schedule to run your web scraping
project at any time without any manipulation.
The scheduled task will run and export scraped results to a destination without any always-running-background service to save
your system resources. Moreover, all duplications can be removed from the harvested results to make sure only new data is kept.
Supported schedule types:
- Run project every hours
- Run project every days
- Run project at a specific time
Export Data to Any Format
Our best web scraper
supports various formats for exporting scraped web data such as: CSV, Access, XML, HTML,
SQL Server, MySQL
. You also can commit results directly to any sorts of Database destinations via ODBC connection.
If your website has a submit form, you even can submit them via HTTP Post or Get.
Data is exported automatically to your pre-defined format once a data extraction is done.
Or you can export scraped results manually any time.
And much more others...
Easy Web Extract has much more other functions to list, however you may interest some few more:
- Robust transformation scripts to tranform scraped data in any kind of forms. Our web scraping tool provides some
ready-to-use scripts such as (sub-string selection, replace, split...) or you can write your own scripts if you
have some programing knowledges of C# or VB.NET
- Random extracting delay to avoid blocking by remote hosts. Some website will block your IP address from access to protect its data if
you send multiple requets at short time. This feature may help you in many cases.
- Support Image List type to download all product images from a web region. Very useful for scraping all product image gallery.
- Execute project by command line. Enable calling our web scraping tool from your own application easily
- Login to secure panel by your given credentials to extract private data
- And much more...