feat: Phase 1+2 — project setup and core infrastructure
- requirements.txt, config.yaml, .env, Dockerfile, docker-compose.yml
- app/core: config (YAML+env override), logging (JSON structured),
exceptions (typed hierarchy), json_utils (Markdown fence stripping)
- app/clients: LLMClient ABC + ZhipuAIClient (run_in_executor),
StorageClient ABC + RustFSClient (boto3 head_object for size check)
- app/main.py: FastAPI app with health endpoint and router registration
- app/core/dependencies.py: lru_cache singleton factories
- tests/conftest.py: mock_llm, mock_storage, test_app, client fixtures
- pytest.ini: asyncio_mode=auto
- 11 unit tests passing
2026-04-10 15:22:45 +08:00
|
|
|
import asyncio
|
|
|
|
|
|
|
|
|
|
from zhipuai import ZhipuAI
|
|
|
|
|
|
|
|
|
|
from app.clients.llm.base import LLMClient
|
|
|
|
|
from app.core.exceptions import LLMCallError
|
|
|
|
|
from app.core.logging import get_logger
|
|
|
|
|
|
|
|
|
|
logger = get_logger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class ZhipuAIClient(LLMClient):
|
|
|
|
|
def __init__(self, api_key: str) -> None:
|
|
|
|
|
self._client = ZhipuAI(api_key=api_key)
|
|
|
|
|
|
|
|
|
|
async def chat(self, model: str, messages: list[dict]) -> str:
|
|
|
|
|
return await self._call(model, messages)
|
|
|
|
|
|
|
|
|
|
async def chat_vision(self, model: str, messages: list[dict]) -> str:
|
|
|
|
|
return await self._call(model, messages)
|
|
|
|
|
|
2026-04-10 16:43:28 +08:00
|
|
|
async def submit_finetune(self, jsonl_url: str, base_model: str, hyperparams: dict) -> str:
|
|
|
|
|
loop = asyncio.get_running_loop()
|
|
|
|
|
try:
|
|
|
|
|
resp = await loop.run_in_executor(
|
|
|
|
|
None,
|
|
|
|
|
lambda: self._client.fine_tuning.jobs.create(
|
|
|
|
|
training_file=jsonl_url,
|
|
|
|
|
model=base_model,
|
|
|
|
|
hyperparameters=hyperparams,
|
|
|
|
|
),
|
|
|
|
|
)
|
|
|
|
|
return resp.id
|
|
|
|
|
except Exception as exc:
|
|
|
|
|
raise LLMCallError(f"微调任务提交失败: {exc}") from exc
|
|
|
|
|
|
|
|
|
|
async def get_finetune_status(self, job_id: str) -> dict:
|
|
|
|
|
loop = asyncio.get_running_loop()
|
|
|
|
|
try:
|
|
|
|
|
resp = await loop.run_in_executor(
|
|
|
|
|
None,
|
|
|
|
|
lambda: self._client.fine_tuning.jobs.retrieve(job_id),
|
|
|
|
|
)
|
|
|
|
|
return {
|
|
|
|
|
"job_id": resp.id,
|
|
|
|
|
"status": resp.status,
|
|
|
|
|
"progress": int(resp.progress) if getattr(resp, "progress", None) is not None else None,
|
|
|
|
|
"error_message": getattr(resp, "error_message", None),
|
|
|
|
|
}
|
|
|
|
|
except Exception as exc:
|
|
|
|
|
raise LLMCallError(f"查询微调任务失败: {exc}") from exc
|
|
|
|
|
|
feat: Phase 1+2 — project setup and core infrastructure
- requirements.txt, config.yaml, .env, Dockerfile, docker-compose.yml
- app/core: config (YAML+env override), logging (JSON structured),
exceptions (typed hierarchy), json_utils (Markdown fence stripping)
- app/clients: LLMClient ABC + ZhipuAIClient (run_in_executor),
StorageClient ABC + RustFSClient (boto3 head_object for size check)
- app/main.py: FastAPI app with health endpoint and router registration
- app/core/dependencies.py: lru_cache singleton factories
- tests/conftest.py: mock_llm, mock_storage, test_app, client fixtures
- pytest.ini: asyncio_mode=auto
- 11 unit tests passing
2026-04-10 15:22:45 +08:00
|
|
|
async def _call(self, model: str, messages: list[dict]) -> str:
|
2026-04-10 16:43:28 +08:00
|
|
|
loop = asyncio.get_running_loop()
|
feat: Phase 1+2 — project setup and core infrastructure
- requirements.txt, config.yaml, .env, Dockerfile, docker-compose.yml
- app/core: config (YAML+env override), logging (JSON structured),
exceptions (typed hierarchy), json_utils (Markdown fence stripping)
- app/clients: LLMClient ABC + ZhipuAIClient (run_in_executor),
StorageClient ABC + RustFSClient (boto3 head_object for size check)
- app/main.py: FastAPI app with health endpoint and router registration
- app/core/dependencies.py: lru_cache singleton factories
- tests/conftest.py: mock_llm, mock_storage, test_app, client fixtures
- pytest.ini: asyncio_mode=auto
- 11 unit tests passing
2026-04-10 15:22:45 +08:00
|
|
|
try:
|
|
|
|
|
response = await loop.run_in_executor(
|
|
|
|
|
None,
|
|
|
|
|
lambda: self._client.chat.completions.create(
|
|
|
|
|
model=model,
|
|
|
|
|
messages=messages,
|
|
|
|
|
),
|
|
|
|
|
)
|
|
|
|
|
content = response.choices[0].message.content
|
|
|
|
|
logger.info("llm_call", extra={"model": model, "response_len": len(content)})
|
|
|
|
|
return content
|
|
|
|
|
except Exception as exc:
|
|
|
|
|
logger.error("llm_call_error", extra={"model": model, "error": str(exc)})
|
|
|
|
|
raise LLMCallError(f"大模型调用失败: {exc}") from exc
|