I'm trying to scrape a site to get product details, I have 2 functions 1) getURL() 2) parseData(), the getURL functions works fine, I can get product URL of 30 pages with 8 products per page, my trouble is with parseData() function, in this function I got a blank page after click on cookie consent page. I suspect that the site knows it's robot and not human, so it blocked/stopped sending date. How can I bypass this hurdle?
- getURL() this function walk through a product summary page with links to product detail, this function gathers the URL of each product and save to a list
- getData() this function uses each of the link
gathered by previous getURL() function and get details of each
product and saves to a dictionary.
ps: during troubleshooting, I can see both
functions showed cookie consent form when starts, the difference
is after clicking "Accept" in the getURL(), the product summary
page shows up with data, but in getData() function, it just
showed a blank page. Going to the site manually, the cookie
consent form shows only once, in the product summary page, not
in the product details page.
--
You received this message because you are subscribed to the Google Groups "Selenium Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to selenium-user...@googlegroups.com.
To view this discussion visit https://groups.google.com/d/msgid/selenium-users/63a79075-6ffa-465c-9241-c41fb84c30f2%40gmail.com.