A comprehensive collection of production-ready web scrapers for major e-commerce and marketplace websites
Scraper Bank is an open-source organization providing battle-tested, production-ready scrapers for extracting data from popular websites. All scrapers are built with multiple framework implementations (Python & Node.js) and optimized for reliability, performance, and anti-bot evasion.
- Amazon Scrapers - Product data, reviews, sellers, and category scrapers
- Alibaba Scrapers - Product category, product data, and product search scrapers
- Aliexpress Scrapers - Global marketplace product category, product data, and product search scrapers
- Best Buy Scrapers - Product category, product data, and product search scrapers
- Costco Scrapers - Product category, product data, and product search scrapers
- Crate & Barrel Scrapers - Home decor and furniture product category, product data, and product search scrapers
- Decathlon Scrapers - Sports and outdoor product category, product data, and product search scrapers
- Depop Scrapers - Social resale product category, product data, and product search scrapers
- eBay Scrapers - Product category, product data, and product search scrapers
- Etsy Scrapers - Product category, product data, and product search scrapers
- Flipkart Scrapers - Product category, product data, and product search scrapers
- Groupon Scrapers - Deals, coupons, and offers scrapers
- Home Depot Scrapers - Product category, product data, and product search scrapers
- IKEA Scrapers - Product category, product data, and product search scrapers
- Kroger Scrapers - Grocery product category, product data, and product search scrapers
- Macy's Scrapers - Department store product category, product data, and product search scrapers
- Nike Scrapers - Product category, product data, and product search scrapers
- Pottery Barn Scrapers - Home decor and furniture product category, product data, and product search scrapers
- REI Scrapers - Outdoor gear product category, product data, and product search scrapers
- StockX Scrapers - Sneaker and resale marketplace product data and search scrapers
- Target Scrapers - Product category, product data, and product search scrapers
- Walmart Scrapers - Product data, reviews, sellers, and category scrapers
- Wayfair Scrapers - Product category, product data, and product search scrapers
- World Market Scrapers - Home decor and specialty product category, product data, and product search scrapers
- Zappos Scrapers - Product category, product data, and product search scrapers
- Beautylish Scrapers - Beauty and cosmetics product data, search, and category scrapers
- Dermstore Scrapers - Skincare and beauty product data, search, and category scrapers
- Ulta Scrapers - Cosmetics and beauty product data, search, and category scrapers
- Booking.com Scrapers - Hotel listings, product data, and search scrapers
- ProductHunt Scrapers - Product listings, category, and search scrapers
- AppSumo Scrapers - Product category, product data, and product search scrapers
We're continuously expanding our collection! More scrapers for additional websites are in development. Stay tuned for updates.
Browse our repositories above and select the website you want to scrape.
Each repository offers multiple implementations:
Python:
- BeautifulSoup - Fast, lightweight HTML parsing
- Playwright - Modern browser automation with excellent JavaScript support
- Selenium - Industry-standard browser automation
- Scrapy - High-performance scraping framework (where available)
Node.js:
- Cheerio & Axios - Fast server-side HTML parsing
- Playwright - Modern browser automation
- Puppeteer - Chrome/Chromium automation
Go:
- HTTP Client - High-performance HTTP scraping (where available)
All scrapers integrate with ScrapeOps for:
- ✅ Proxy Rotation - Distribute requests across multiple IP addresses
- ✅ Request Header Optimization - Reduce bot detection
- ✅ Rate Limiting Management - Built-in retry logic and rate limiting
- ✅ CAPTCHA Handling - Advanced anti-bot evasion
Get your free API key: https://scrapeops.io/app/register/ai-builder
💡 Free Tier Available: ScrapeOps offers a generous free tier perfect for testing and small-scale scraping.
Each scraper includes detailed documentation:
- Installation instructions
- Usage examples
- Configuration options
- Output format specifications
- Troubleshooting guides
All scrapers in Scraper Bank are pre-configured to work with ScrapeOps. Here's how to use them:
Sign up at https://scrapeops.io/app/register/ai-builder and get your free API key.
No environment variables are required. Add your ScrapeOps API key inside the scraper file where the script reads it (e.g. at the top or in the config section).
Python — in the scraper file:
# Add your ScrapeOps API key directly in the scraper (e.g. at the top)
SCRAPEOPS_API_KEY = 'your-api-key-here'Node.js — in the scraper file:
// Add your ScrapeOps API key directly in the scraper (e.g. at the top)
const SCRAPEOPS_API_KEY = 'your-api-key-here';Each repository's scraper files have a clear place to paste your key. Check the framework-specific README in the repo for the exact variable name and location.
The scrapers automatically use ScrapeOps proxy endpoints when configured:
Python Example:
import requests
# ScrapeOps proxy endpoint
proxy_url = f"http://scrapeops.headless_browser_mode=true:{SCRAPEOPS_API_KEY}@proxy.scrapeops.io:5353"
proxies = {
'http': proxy_url,
'https': proxy_url
}
response = requests.get(url, proxies=proxies)Node.js Example:
const axios = require('axios');
const proxyUrl = `http://scrapeops.headless_browser_mode=true:${SCRAPEOPS_API_KEY}@proxy.scrapeops.io:5353`;
const response = await axios.get(url, {
proxy: {
host: 'proxy.scrapeops.io',
port: 5353,
auth: {
username: 'scrapeops.headless_browser_mode=true',
password: SCRAPEOPS_API_KEY
}
}
});- Residential Proxies: Use residential IPs for better success rates
- Geolocation Targeting: Target specific countries/regions
- Browser Fingerprinting: Rotate browser fingerprints
- CAPTCHA Solving: Automatic CAPTCHA resolution (premium feature)
Documentation: https://docs.scrapeops.io
Create custom scrapers in seconds with ScrapeOps AI-powered code generator!
🔗 Visit: https://scrapeops.io/ai-web-scraping-assistant/scraper-builder
- Enter Target URL: Paste the website URL you want to scrape
- Get Generated Code: Receive production-ready scraper code instantly
- Customize & Deploy: Fine-tune the code and integrate with ScrapeOps
- ✅ Multi-Language Support: Generate code in Python, Node.js, or other languages
- ✅ Framework Selection: Choose from BeautifulSoup, Playwright, Selenium, Puppeteer, and more
- ✅ Automatic Selector Detection: AI identifies the best CSS/XPath selectors
- ✅ Anti-Bot Integration: Built-in ScrapeOps proxy and header optimization
- ✅ Error Handling: Includes retry logic and error handling
- ✅ Output Formatting: Structured JSON output ready to use
1. Go to scrapeops.io/ai-web-scraping-assistant/scraper-builder
2. Enter: "https://example.com/products"
3. Select: Python + Playwright
4. Click "Generate Code"
5. Copy the generated scraper code
6. Add your ScrapeOps API key
7. Run and enjoy! 🎉
- Be Specific: Clearly describe the data fields you need
- Provide Examples: Share example URLs or HTML snippets if possible
- Iterate: Refine your prompts for better results
- Test Locally: Always test generated code before production use
Depending on the repository, you can extract:
- Product Information: Names, prices, descriptions, images, ratings
- Product Categories: Category listings, navigation structures
- Search Results: Search query results, filters, pagination
- Reviews & Ratings: Customer reviews, ratings, helpful votes
- Seller Information: Seller profiles, ratings, store information
- Inventory Data: Stock status, availability, variants
- Pricing Data: Current prices, historical prices, discounts
Each repository's README provides detailed information about available data fields.
Modern websites employ sophisticated anti-bot measures. All Scraper Bank scrapers are designed to work with ScrapeOps to handle:
- ✅ IP Rotation: Distribute requests across multiple IPs
- ✅ Header Optimization: Mimic real browser headers
- ✅ Rate Limiting: Respectful request rates
- ✅ CAPTCHA Solving: Automatic CAPTCHA resolution (premium)
- ✅ Browser Fingerprinting: Rotate browser signatures
Important: Anti-bot measures vary by site and may change over time. CAPTCHA challenges may occur and cannot be guaranteed to be resolved automatically. Using proxies and browser automation can help reduce blocking, but effectiveness depends on the target site's specific anti-bot measures.
All scrapers output structured JSON data. Example output formats are included in each repository's example/ directories.
We welcome contributions! Whether it's:
- 🐛 Bug fixes
- ✨ New features
- 📝 Documentation improvements
- 🆕 New scraper implementations
- 🌐 New website scrapers
Please check each repository's contributing guidelines.
Each repository may have its own license. Please check individual repository LICENSE files.
- ScrapeOps Homepage: https://scrapeops.io
- ScrapeOps Documentation: https://docs.scrapeops.io
- AI Scraper Builder: https://scrapeops.io/ai-web-scraping-assistant/scraper-builder
- ScrapeOps API Dashboard: https://app.scrapeops.io
These scrapers are provided for educational and research purposes. Always:
- ✅ Respect websites' Terms of Service
- ✅ Follow robots.txt guidelines
- ✅ Use reasonable request rates
- ✅ Comply with applicable laws and regulations
- ✅ Respect website owners' intellectual property rights
The maintainers of Scraper Bank are not responsible for any misuse of these tools.
If you find Scraper Bank useful, please consider giving our repositories a ⭐ on GitHub!
Built with ❤️ using ScrapeOps