You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
All 680 product URLs passed to crawler.arun_many() should produce a corresponding result.extracted_content if the crawling and extraction process succeeds. Each result should be saved to MongoDB as a new document. I have 680 unique URL's that should output 680 unique MongoDB documents.
Current Behavior
I have 680 unique URL's that output 540 MongoDB documents. The MemoryAdaptiveDispatcher is not indicating any errors processing the URL's and all URL's were displayed to be processed. Is this a concurrent processing issue or a race condition issue?
Is this reproducible?
Yes
Inputs Causing the Bug
- A list of 680 product URLs loaded from product_links.json
- This bug happens with any large URL set that I run concurrent processing on.
crawl4ai version
0.5.0
Expected Behavior
All 680 product URLs passed to crawler.arun_many() should produce a corresponding result.extracted_content if the crawling and extraction process succeeds. Each result should be saved to MongoDB as a new document. I have 680 unique URL's that should output 680 unique MongoDB documents.
Current Behavior
I have 680 unique URL's that output 540 MongoDB documents. The MemoryAdaptiveDispatcher is not indicating any errors processing the URL's and all URL's were displayed to be processed. Is this a concurrent processing issue or a race condition issue?
Is this reproducible?
Yes
Inputs Causing the Bug
- A list of 680 product URLs loaded from product_links.json - This bug happens with any large URL set that I run concurrent processing on.
Steps to Reproduce
Code snippets
OS
macOS
Python version
3.13.2
Browser
Chrome
Browser version
No response
Error logs & Screenshots (if applicable)
No Error Logs.
The text was updated successfully, but these errors were encountered: