wise-white•2y ago
how to stop skipping of urls due to error handling
when i am scraping product data from product urls, if i am trying to either see whether a tag is available and if not to use a different tag or if a tag simply isn't found, i don't want it to give a full error for not finding that certain element i want and not scrape and save the rest of the data
how do i avoid this "skipping" over by overriding or changing the natural response of the crawler
i even have tried try catch statements and if else statements to handle a product not being found and nothing works
2 Replies
I recommend using the $ cheerio parser with https://crawlee.dev/api/puppeteer-crawler/interface/PuppeteerCrawlingContext#parseWithCheerio for nicer parsing API.
Otherwise, you can do something like
wise-whiteOP•2y ago
thank you, the error catching built into the scraping tag finally worked, thanks again!