Replies: 1 comment 2 replies
-
This happens with a higher initial concurrency and is expected behavior, we never ignore requests that were started (in fact it's usually not possible to cancel them in the first place), as you can see yourself in the logs (the message starting with This fact is described in several places in the docs, e.g. https://crawlee.dev/docs/introduction/adding-urls#limit-your-crawls-with-maxrequestspercrawl
This is not very helpful, reproductions need to be complete and specific. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Which package is this bug report for? If unsure which one to select, leave blank
@crawlee/playwright (PlaywrightCrawler)
Issue description
the actual scrape page count is more exceed than expect, even 2x(13:5)
Code sample
Package version
@builder.io/[email protected] /Users/zheng/Workbench/github/gpt-crawler └── [email protected] -> ./node_modules/.pnpm/[email protected][email protected]/node_modules/crawlee
Node.js version
v18
Operating system
MacOS
Apify platform
I have tested this on the
next
releaseNo response
Other context
No response
Beta Was this translation helpful? Give feedback.
All reactions