Bug: Crawler reached the maxRequestsPerCrawl limit of 1 requests and will shut down soon #2152
Replies: 4 comments 1 reply
-
Can you try following the steps defined here: #2031 (comment), specifically providing a new instance of Configuration per crawler you make, with persistStorage set to false? |
Beta Was this translation helpful? Give feedback.
-
Hi @vladfrangu
I'm happy to try it out. However, there is now
How can I import the |
Beta Was this translation helpful? Give feedback.
-
You have to import things from crawlee, not from Cheerio! |
Beta Was this translation helpful? Give feedback.
-
@vladfrangu |
Beta Was this translation helpful? Give feedback.
-
Which package is this bug report for? If unsure which one to select, leave blank
@crawlee/cheerio (CheerioCrawler)
Issue description
Cheerio crawler is not crawling when set
maxRequestPerCrawl
to 1.Even when I set
maxRequestPerCrawl
to 10 or 100, after the 10th or 100th request nothing will be crawled again anymore.I use a new instance of Cheerio for any single request, no parallel requests necessary in my usecases.
However, it counts requests on a global basis, no matter if I use a new instance of Cheerio for every request or if I use a shared instance.
Once the count of all requests is reaching the value of
maxRequestPerCrawl
, it will deny all further requests. The only solution is to shutdown the full process and start it again.Code sample
Package version
3.4.0
Node.js version
18.17.1
Operating system
MacOS
Apify platform
I have tested this on the
next
releaseNo response
Other context
Log:
Beta Was this translation helpful? Give feedback.
All reactions