fix(crawl): throttle concurrent CrawlJobs and relax fpw/proxyscrape HTTP
- CrawlJob waits on crawl_slot before JobExecutor semaphore so crawl-all does not fill slots while queued - BaseHTTPPlugin: longer connect budget for slow international links - proxyscrape: jsDelivr mirror + longer GitHub/API phases - fpw_*: higher timeouts/retries; lower internal concurrency on heavy multi-URL plugins Made-with: Cursor
This commit is contained in:
@@ -130,7 +130,7 @@ class FpwSpysOnePlugin(BaseHTTPPlugin):
|
||||
|
||||
async def _one(proto: str, url: str, xf5: str) -> Tuple[str, str]:
|
||||
data = {**form_base, "xf5": xf5}
|
||||
html = await self.fetch_post(url, data=data, timeout=14, retries=1)
|
||||
html = await self.fetch_post(url, data=data, timeout=25, retries=2)
|
||||
return proto, html or ""
|
||||
|
||||
pairs = await asyncio.gather(
|
||||
|
||||
Reference in New Issue
Block a user