Hi Tavily community,
I’m currently working on a project where I need to scrape product data from an eCommerce website with multiple pages (e.g., a product category with pagination). I want to avoid manually inputting each page URL and automate the process of scraping all products across multiple pages.
My key questions are:
- Does Tavily’s API support automated crawling of multiple pages from a base URL, like a category or product listing page, or do I need to manually provide each page URL?
- If Tavily supports crawling, how can I set it up to handle pagination automatically (e.g.,
?page=2
,?page=3
, etc.) for eCommerce websites? - If Tavily does not support automated crawling, do you have any suggestions or best practices for automating link discovery (e.g., extracting product links from paginated pages) before passing them to Tavily for scraping?
- For JavaScript-heavy websites (e.g., infinite scrolling or dynamically loaded pages), can Tavily handle these scenarios, or should I integrate a tool like Selenium/Playwright to retrieve the full content before using Tavily?
Any guidance or advice on best practices for scaling eCommerce scraping with Tavily would be greatly appreciated!
Thanks in advance!