- 删除 ValidationQueue 双轨持久化队列,替换为纯内存 AsyncWorkerPool - 引入统一后台任务框架 JobExecutor(Job/CrawlJob/ValidateAllJob) - 新增 PluginRunner 统一插件执行(超时、重试、健康检查、统计) - 重构 SchedulerService 职责收敛为仅定时触发 ValidateAllJob - 使用 AsyncExitStack 重构 lifespan,安全管理长生命周期资源 - 路由层瘦身 50%+,业务异常上抛由全局中间件统一处理 - 实现设置全热更新(WorkerPool 并发、Validator 超时即时生效) - 前端 Store 强制写后重新拉取,消除乐观更新数据不同步 - 删除 queue.py / task_repo.py / task_service.py - 新增 execution 单元测试,全部 85 个测试通过
55 lines
1.3 KiB
Python
55 lines
1.3 KiB
Python
"""领域模型 - 纯数据结构,不依赖任何框架"""
|
|
from dataclasses import dataclass, field
|
|
from datetime import datetime
|
|
from typing import List, Optional
|
|
|
|
|
|
@dataclass
|
|
class ProxyRaw:
|
|
"""爬虫爬取的原始代理数据"""
|
|
ip: str
|
|
port: int
|
|
protocol: str = "http"
|
|
|
|
def __post_init__(self):
|
|
self.protocol = self.protocol.lower().strip()
|
|
if self.protocol not in ("http", "https", "socks4", "socks5"):
|
|
self.protocol = "http"
|
|
if not isinstance(self.port, int) or not (1 <= self.port <= 65535):
|
|
raise ValueError(f"port must be between 1 and 65535, got {self.port}")
|
|
|
|
|
|
@dataclass
|
|
class Proxy:
|
|
"""数据库中的代理实体"""
|
|
ip: str
|
|
port: int
|
|
protocol: str
|
|
score: int
|
|
response_time_ms: Optional[float] = None
|
|
last_check: Optional[datetime] = None
|
|
created_at: Optional[datetime] = None
|
|
|
|
|
|
@dataclass
|
|
class PluginInfo:
|
|
"""插件元数据"""
|
|
id: str
|
|
name: str
|
|
display_name: str
|
|
description: str
|
|
enabled: bool
|
|
last_run: Optional[datetime] = None
|
|
success_count: int = 0
|
|
failure_count: int = 0
|
|
|
|
|
|
@dataclass
|
|
class CrawlResult:
|
|
"""插件爬取结果"""
|
|
plugin_name: str
|
|
proxies: List[ProxyRaw] = field(default_factory=list)
|
|
success_count: int = 0
|
|
failure_count: int = 0
|
|
error: Optional[str] = None
|