fix(plugins): fpw parsers for JSON API, mirrors, and looser HTML
- fpw_proxy_list_download: parse JSON list/proxies bodies; jsDelivr monosans tier; crawl timeout 300s - fpw_socks_ssl: try parse_html_table before regex - fpw_hidemy: loose row scan when fixed columns fail - fpw_proxynova: plain IP/port row fallback - fpw_spys_one: HTTPS endpoints; crawl timeout 180s - fpw_gatherproxy: HTTPS + extra JSON key patterns - fpw_checkerproxy: lower min HTML length for parse - fpw_premproxy: ip:port regex fallback when few table rows Made-with: Cursor
This commit is contained in:
@@ -15,8 +15,8 @@ class FpwGatherproxyPlugin(BaseHTTPPlugin):
|
||||
def __init__(self):
|
||||
super().__init__()
|
||||
self.urls = [
|
||||
"http://www.gatherproxy.com/proxylist/anonymity/?t=Elite",
|
||||
"http://www.gatherproxy.com/proxylist/country/?c=United%20States",
|
||||
"https://www.gatherproxy.com/proxylist/anonymity/?t=Elite",
|
||||
"https://www.gatherproxy.com/proxylist/country/?c=United%20States",
|
||||
]
|
||||
|
||||
def _extract_from_text(self, text: str) -> List[ProxyRaw]:
|
||||
@@ -42,6 +42,17 @@ class FpwGatherproxyPlugin(BaseHTTPPlugin):
|
||||
results.append(ProxyRaw(ip, int(port), "http"))
|
||||
except ValueError:
|
||||
continue
|
||||
for m in re.finditer(
|
||||
r'"(?:proxy_)?ip"\s*:\s*"([\d.]+)"\s*,\s*"(?:proxy_)?port"\s*:\s*"?(\d+)"?',
|
||||
text,
|
||||
re.I,
|
||||
):
|
||||
ip, port = m.group(1), m.group(2)
|
||||
if port.isdigit() and 1 <= int(port) <= 65535:
|
||||
try:
|
||||
results.append(ProxyRaw(ip, int(port), "http"))
|
||||
except ValueError:
|
||||
continue
|
||||
return results
|
||||
|
||||
async def crawl(self) -> List[ProxyRaw]:
|
||||
|
||||
Reference in New Issue
Block a user