Files
ProxyPool/app/models/domain.py
祀梦 b972b64616 refactor: 全面重构核心架构,消除反复修改的根因
- 删除 ValidationQueue 双轨持久化队列,替换为纯内存 AsyncWorkerPool
- 引入统一后台任务框架 JobExecutor(Job/CrawlJob/ValidateAllJob)
- 新增 PluginRunner 统一插件执行(超时、重试、健康检查、统计)
- 重构 SchedulerService 职责收敛为仅定时触发 ValidateAllJob
- 使用 AsyncExitStack 重构 lifespan,安全管理长生命周期资源
- 路由层瘦身 50%+,业务异常上抛由全局中间件统一处理
- 实现设置全热更新(WorkerPool 并发、Validator 超时即时生效)
- 前端 Store 强制写后重新拉取,消除乐观更新数据不同步
- 删除 queue.py / task_repo.py / task_service.py
- 新增 execution 单元测试,全部 85 个测试通过
2026-04-04 22:36:57 +08:00

55 lines
1.3 KiB
Python

"""领域模型 - 纯数据结构,不依赖任何框架"""
from dataclasses import dataclass, field
from datetime import datetime
from typing import List, Optional
@dataclass
class ProxyRaw:
"""爬虫爬取的原始代理数据"""
ip: str
port: int
protocol: str = "http"
def __post_init__(self):
self.protocol = self.protocol.lower().strip()
if self.protocol not in ("http", "https", "socks4", "socks5"):
self.protocol = "http"
if not isinstance(self.port, int) or not (1 <= self.port <= 65535):
raise ValueError(f"port must be between 1 and 65535, got {self.port}")
@dataclass
class Proxy:
"""数据库中的代理实体"""
ip: str
port: int
protocol: str
score: int
response_time_ms: Optional[float] = None
last_check: Optional[datetime] = None
created_at: Optional[datetime] = None
@dataclass
class PluginInfo:
"""插件元数据"""
id: str
name: str
display_name: str
description: str
enabled: bool
last_run: Optional[datetime] = None
success_count: int = 0
failure_count: int = 0
@dataclass
class CrawlResult:
"""插件爬取结果"""
plugin_name: str
proxies: List[ProxyRaw] = field(default_factory=list)
success_count: int = 0
failure_count: int = 0
error: Optional[str] = None