How to Scrape Amazon Product Data
Amazon is one of the most commonly scraped e-commerce sites. With ScrapingLab, you can extract product data without writing code or managing proxies.
What You Can Extract
- Product title and description
- Current price and discount percentage
- Star rating and review count
- ASIN (Amazon Standard Identification Number)
- Seller information
- Product images URLs
- Availability status
Step-by-Step with ScrapingLab
1. Create a New Task
Enter the Amazon product URL or search results page you want to scrape. For example: https://www.amazon.com/s?k=wireless+headphones
2. Build Your Workflow
Add these steps in the visual builder:
- Navigate — Go to the target URL
- Wait — Wait for the product grid to load (2-3 seconds)
- Extract — Select the product card container and map fields:
- Title →
.s-title-instructions-style span - Price →
.a-price .a-offscreen - Rating →
.a-icon-star-small .a-icon-alt - Link →
a.a-link-normal(href attribute)
- Title →
- Screenshot — Capture the page for verification
3. Handle Pagination
To scrape multiple pages of results, add a Loop step:
- Click — Click the “Next” pagination button
- Wait — Wait for the new page to load
- Extract — Repeat the extraction step
- Set the loop to run for a specific number of pages
4. Schedule and Export
- Set the task to run daily or weekly to track price changes
- Export to CSV for spreadsheet analysis or JSON for API integration
- Send to a webhook to trigger alerts when prices drop
Common Challenges
Anti-bot Protection
Amazon actively blocks automated access. ScrapingLab handles this with built-in proxy rotation — requests come from different IP addresses automatically.
Dynamic Content
Amazon loads some content via JavaScript. ScrapingLab uses a real browser engine, so dynamically loaded content is captured just like a human would see it.
Layout Variations
Amazon shows different layouts for different product categories. Build separate workflows for search results pages vs. individual product pages.
Best Practices
- Respect rate limits — Schedule runs during off-peak hours and avoid scraping thousands of pages per minute
- Use specific URLs — Target specific categories or search queries rather than crawling the entire site
- Monitor for changes — Amazon frequently updates their page structure. Check your workflows weekly
- Store historical data — Export to CSV regularly to build a price history database