Initial implementation of xtrm-agent multi-agent system
Multi-agent AI automation system with shared message bus, specialized roles (coder/researcher/reviewer), and deny-by-default security. - Config system with Pydantic validation and YAML loading - Async message bus with inter-agent delegation - LLM providers: Anthropic (Claude) and LiteLLM (DeepSeek/Kimi/MiniMax) - Tool system: registry, builtins (file/bash/web), approval engine, MCP client - Agent engine with tool-calling loop and orchestrator for multi-agent management - CLI channel (REPL) and Discord channel - Docker + Dockge deployment config - Typer CLI: chat, serve, status, agents commands Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
92
xtrm_agent/llm/litellm.py
Normal file
92
xtrm_agent/llm/litellm.py
Normal file
@@ -0,0 +1,92 @@
|
||||
"""LiteLLM provider — DeepSeek, Kimi, MiniMax, and more."""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import json
|
||||
from typing import Any
|
||||
|
||||
import litellm
|
||||
from json_repair import repair_json
|
||||
|
||||
from xtrm_agent.llm.provider import LLMProvider, LLMResponse, ToolCallRequest
|
||||
|
||||
|
||||
class LiteLLMProvider(LLMProvider):
|
||||
"""Multi-provider via LiteLLM."""
|
||||
|
||||
def __init__(self, model: str = "deepseek/deepseek-chat-v3.1") -> None:
|
||||
self.model = model
|
||||
litellm.drop_params = True
|
||||
|
||||
async def complete(
|
||||
self,
|
||||
messages: list[dict[str, Any]],
|
||||
tools: list[dict[str, Any]] | None = None,
|
||||
model: str | None = None,
|
||||
max_tokens: int = 8192,
|
||||
temperature: float = 0.3,
|
||||
) -> LLMResponse:
|
||||
model = model or self.model
|
||||
|
||||
kwargs: dict[str, Any] = {
|
||||
"model": model,
|
||||
"messages": messages,
|
||||
"max_tokens": max_tokens,
|
||||
"temperature": temperature,
|
||||
}
|
||||
if tools:
|
||||
kwargs["tools"] = tools
|
||||
kwargs["tool_choice"] = "auto"
|
||||
|
||||
response = await litellm.acompletion(**kwargs)
|
||||
return self._parse_response(response)
|
||||
|
||||
def get_default_model(self) -> str:
|
||||
return self.model
|
||||
|
||||
def _parse_response(self, response: Any) -> LLMResponse:
|
||||
"""Parse LiteLLM (OpenAI-format) response."""
|
||||
choice = response.choices[0]
|
||||
message = choice.message
|
||||
|
||||
content = message.content or ""
|
||||
tool_calls: list[ToolCallRequest] = []
|
||||
|
||||
if message.tool_calls:
|
||||
for tc in message.tool_calls:
|
||||
args = self._parse_arguments(tc.function.arguments)
|
||||
tool_calls.append(
|
||||
ToolCallRequest(
|
||||
id=tc.id,
|
||||
name=tc.function.name,
|
||||
arguments=args,
|
||||
)
|
||||
)
|
||||
|
||||
usage_data = {}
|
||||
if hasattr(response, "usage") and response.usage:
|
||||
usage_data = {
|
||||
"input_tokens": getattr(response.usage, "prompt_tokens", 0),
|
||||
"output_tokens": getattr(response.usage, "completion_tokens", 0),
|
||||
}
|
||||
|
||||
return LLMResponse(
|
||||
content=content,
|
||||
tool_calls=tool_calls,
|
||||
finish_reason=choice.finish_reason or "",
|
||||
usage=usage_data,
|
||||
)
|
||||
|
||||
def _parse_arguments(self, raw: str | dict) -> dict[str, Any]:
|
||||
"""Parse tool call arguments, using json-repair for malformed JSON."""
|
||||
if isinstance(raw, dict):
|
||||
return raw
|
||||
try:
|
||||
return json.loads(raw)
|
||||
except (json.JSONDecodeError, TypeError):
|
||||
try:
|
||||
repaired = repair_json(raw)
|
||||
result = json.loads(repaired)
|
||||
return result if isinstance(result, dict) else {}
|
||||
except Exception:
|
||||
return {}
|
||||
Reference in New Issue
Block a user