Compare Products

Ficstar App Scrapy App

Features

* Custom Web Crawler - All web crawlers will be custom designed for your business needs. Whether you want to track prices or in-stock status, the custom web crawler is the ultimate tool to help you find exactly the data you want. * Web Data Collection - We specialize in advanced web data collection for crawling difficult sites with challenging scripts and collecting hard-to-get embedded web content. Our web crawlers scan millions of web pages in hours and save billions of records a day. * Data Processing - The system is able to intelligently match and compare data from different sources and refine all results in the exact output format you want. * Output Reporting - All data output is presented in high-quality reports in an aggregated format based on your requirements. Data files will be delivered on your required schedule through secure web servers. * Easy Data Input - Our no-hassle importing process gets your input data into the system. Or you can simply send us your input data, and we’ll take care of it. * Simple Scheduling Steps - Set up schedules the way you want to receive data deliveries and updates: daily, weekly, monthly, or however you want. * Automatic Email Notifications - Add your email address to the system to receive notifications any time you like for important messages such as data delivery or system errors.

Features

* Fast and powerful - write the rules to extract the data and let Scrapy do the rest. * Easily extensible - extensible by design, plug new functionality easily without having to touch the core. * Portable, Python - written in Python and runs on Linux, Windows, Mac and BSD. * Built-in support for selecting and extracting data from HTML/XML sources using extended CSS selectors and XPath expressions, with helper methods to extract using regular expressions. * An interactive shell console (IPython aware) for trying out the CSS and XPath expressions to scrape data, very useful when writing or debugging your spiders. * Built-in support for generating feed exports in multiple formats (JSON, CSV, XML) and storing them in multiple backends (FTP, S3, local filesystem) * Robust encoding support and auto-detection, for dealing with foreign, non-standard and broken encoding declarations. * Strong extensibility support, allowing you to plug in your own functionality using signals and a well-defined API (middlewares, extensions, and pipelines). * Wide range of built-in extensions and middlewares for handling: cookies and session handling HTTP features like compression, authentication, caching, user-agent spoofing, robots.txt, crawl depth restriction * A Telnet console for hooking into a Python console running inside your Scrapy process, to introspect and debug your crawler Plus other goodies like reusable spiders to crawl sites from Sitemaps and XML/CSV feeds, a media pipeline for automatically downloading images (or any other media) associated with the scraped items, a caching * DNS resolver, and much more!

Languages

Other

Languages

Python

Source Type

Closed

Source Type

Open

License Type

Proprietary

License Type

Proprietary

OS Type

OS Type

Pricing

  • By Quotation

Pricing

  • free
X

Compare Products

Select up to three two products to compare by clicking on the compare icon () of each product.

{{compareToolModel.Error}}

Now comparing:

{{product.ProductName | createSubstring:25}} X
Compare Now