Meet the Scrapy pros
The following companies offer support contracts and consultancy services for Scrapy, and can also develop bespoke crawlers to meet your needs:
Scrapinghub is currently the largest company sponsoring Scrapy development. It specializes in web crawling, it was founded by Scrapy creators and employs crawling experts including many Scrapy core developers.
Arbisoft scours massive websites several layers deep to collect valuable data powering leading firms around the world. It offers realtime crawling and custom-built fully-managed spiders. Over 6 years of quality service, their Python engineers have come to trust Scrapy as their tool of choice.
Datahut provides Scrapy consulting services across different business verticals like e-commerce, content discovery, lead generation, opinion mining etc. They provide clean ready to use data in the most common formats.
SayOne uses Scrapy to power their web crawling and visualization services. They have a strong team of crawling experts who specialize in crawling, information extraction and application integration. They also offer web & mobile app development. Read more about their data scraping services here.
Tryolabs is a Boutique dev Shop specialized in building Python apps with Machine Learning components. They embed Scrapy into their customers applications as well as into their own products.
Intoli uses Scrapy to provide customized web scraping solutions, delivering data that is used by clients to power their core products, for lead generation, and for competitor research. They specialize in advanced services such as cross-site data aggregation, user logins, and bypassing captchas.
Companies using Scrapy
Check who is using Scrapy to do business and make an impact on the world.
(Should you be on this list? Fork this Github repo, add yourself, and send a pull request!)
Scrapinghub: From the creators of Scrapy, Scrapinghub is a leading technology and professional services company, providing successful web crawling and data processing solutions.
Parsely: Uses Scrapy to scrape articles from hundreds of news sites. Its CTO talks about Scrapy in this talk.
DirectEmployers Foundation: Uses Scrapy to scrape job postings from many websites, which are published on the My.jobs site.
Flax: Is a search consulting company based in Cambridge (UK) that uses Scrapy to power the crawling needs of their solutions (blog post).
Médialab Sciences Po: In Paris is using Scrapy to develop a web mining tool for Social Sciences researchers (announcement here).
ScraperWiki: Is a data services company based in Liverpool providing bespoke solutions for data scraping and aggregation using Scrapy as a core technology (blog post).
Data.Gov.Uk: UK government data aggregation site (tweet).
Oposicionesaldia: Uses Scrapy to collect data from jobs postings, scholarships and online free courses in Spain.
Iberestudios: Uses Scrapy to collect data from masters degrees, doctorates and postgradute degrees in Spain.
GrabLab: Is a Russian company which specializes in web scraping, data collection and web automation tasks.
the urge is a Fashion Search Engine focussed on using Artificial Intelligence to help shoppers find the fashion they’re looking for. They use Scrapy at scale to crawl retailers websites for Fashion products.
Alistek: Uses Scrapy for updating partner related information in their OpenERP based back-office system, by scraping various data sources, both on the web and off-line.
Zhitongba: Is a company trying to help people better commute within big cities in China. They use Scrapy to scrape ride-sharing information from multiple sources.
Stilivo: Is a discovery shopping site that uses Scrapy to collect product information from e-commerce sites.
Zopper: Uses Scrapy to crawl hundreds of ecommerce portals for classifying the products sold online.
Competera: Is a price intelligence service that uses Scrapy to collect price, availability and promo data from over the million product pages every day.
Jobijoba: Uses Scrapy to scrape job offers daily from many job boards. Operate in France, several European countries, Russia, Mexico and Australia.
Data Quarry: Designs data scrapers specifically to address the needs of e-commerce users and offers custom scrapy devlopment.
El Útero de Marita: Peruvian leading news blog uses Scrapy to download public documents from governmental institutions in Peru for data journalism.
Shimply: Building world’s largest online marketplace connecting sellers and buyers. Parsing over 500 large and small sites daily. (tweet).
Allclasses: Uses Scrapy to collect over 100,000 online and local classes from recreational through advanced education sites - providing more accessibility to education.
MonkeyLearn: Is a cloud platform that allows any company to extract relevant data from unstructured text using machine learning. It uses Scrapy to get data to train its algorithms.
neu.land GmbH: Uses Scrapy to crawl client websites, allowing to identify possibe optimization measures with the aim to make websites faster, more accessible, and user-friendly.
SayOne: uses Scrapy to crawl data for their clients and thereby develop more customer-centric applications for them.
Videdressing.com: Uses Scrapy to crawl and collect data on fashion products, clothes and accessories.
Up Hail: Does for taxi services what so many websites have done for other travel costs — compares them in real time to find the best deal. Uses Scrapy to crawl and scrape taxi and transportation sites to gather rates, coverage zones, and deals (blog post).
Allogarage: Allogarage is the leading company in France to compare auto mechanics. Uses Scrapy to crawl and scrape car manufacturer sites to gather name, address, phone and timetables for car dealers. (tweet)
TheWebFollower: Uses Scrapy to crawl latest Malayalam language blogs, news, articles, stories published in different websites. (tweet)
Cocon.Se: A tool that shows your website like a crawler sees it. Uses Scrapy to prototype data gathering algorithms, generate specific datasets for machine learning and datavizualisation. (blog post)
WatchPatrol: Uses Scrapy for aggregating private sales listings for used and new watches from the web’s most popular watch enthusiast forums
PolitePol: Online tool for creation of RSS feeds for any web page. Uses scrapy for parsing and processing source pages. This site have open source version you can check out and use absolutely for free.