nanoClaw/backend/services/chat.py

515 lines
20 KiB
Python

"""Chat completion service"""
import json
import logging
import uuid
from typing import Optional, Union
from flask import current_app, Response, request as flask_request
from werkzeug.exceptions import ClientDisconnected
import requests
from backend import db
from backend.models import Conversation, Message
from backend.tools import registry, ToolExecutor
from backend.utils.helpers import (
record_token_usage,
build_messages,
)
from backend.services.llm_client import LLMClient
from backend.config import config as _cfg
logger = logging.getLogger(__name__)
def _client_disconnected():
"""Check if the client has disconnected."""
try:
stream = flask_request.input_stream
if stream is None:
return False
return stream.closed
except Exception:
return False
def _sse_event(event: str, data: dict) -> str:
"""Format a Server-Sent Event string."""
return f"event: {event}\ndata: {json.dumps(data, ensure_ascii=False)}\n\n"
class ChatService:
"""Chat completion service with tool support"""
def __init__(self, llm: LLMClient):
self.llm = llm
def stream_response(
self,
conv: Conversation,
tools_enabled: bool = True,
project_id: str = None,
tool_choice: Optional[Union[str, dict]] = None,
):
"""Stream response with tool call support.
Uses 'process_step' events to send thinking and tool calls in order,
allowing them to be interleaved properly in the frontend.
Args:
conv: Conversation object
tools_enabled: Whether to enable tools
project_id: Project ID for workspace isolation
tool_choice: Optional tool_choice override (e.g. "auto", "required", or dict)
"""
conv_id = conv.id
conv_model = conv.model
conv_max_tokens = conv.max_tokens
conv_temperature = conv.temperature
conv_thinking_enabled = conv.thinking_enabled
app = current_app._get_current_object()
tools = registry.list_all() if tools_enabled else None
initial_messages = build_messages(conv, project_id)
executor = ToolExecutor(registry=registry)
context = {
"model": conv_model,
"max_tokens": conv_max_tokens,
"temperature": conv_temperature,
}
if project_id:
context["project_id"] = project_id
elif conv.project_id:
context["project_id"] = conv.project_id
def generate():
messages = list(initial_messages)
all_tool_calls = []
all_tool_results = []
all_steps = []
step_index = 0
total_completion_tokens = 0
total_prompt_tokens = 0
for iteration in range(_cfg.max_iterations):
# Helper to parse stream_result event
def parse_stream_result(event_str):
"""Parse stream_result SSE event and extract data dict."""
# Format: "event: stream_result\ndata: {...}\n\n"
try:
for line in event_str.split('\n'):
if line.startswith('data: '):
return json.loads(line[6:])
except Exception:
pass
return None
# Collect SSE events and extract final stream_result
try:
stream_gen = self._stream_llm_response(
app, messages, tools, tool_choice, step_index,
conv_model, conv_max_tokens, conv_temperature,
conv_thinking_enabled,
)
except requests.exceptions.HTTPError as e:
resp = e.response
if resp is not None and resp.status_code >= 500:
yield _sse_event("error", {"content": f"LLM service unavailable ({resp.status_code})"})
elif resp is not None and resp.status_code == 429:
yield _sse_event("error", {"content": "Rate limit exceeded, please try again later"})
else:
yield _sse_event("error", {"content": f"LLM request failed: {e}"})
return
except requests.exceptions.ConnectionError:
yield _sse_event("error", {"content": "Unable to connect to LLM service"})
return
except requests.exceptions.Timeout:
yield _sse_event("error", {"content": "LLM request timed out"})
return
except Exception as e:
logger.exception("Unexpected error during LLM streaming")
yield _sse_event("error", {"content": f"Internal error: {e}"})
return
result_data = None
try:
for event_str in stream_gen:
# Check if this is a stream_result event (final event)
if event_str.startswith("event: stream_result"):
result_data = parse_stream_result(event_str)
else:
# Forward process_step events to client in real-time
yield event_str
except Exception as e:
logger.exception("Error during streaming")
yield _sse_event("error", {"content": f"Stream error: {e}"})
return
if result_data is None:
return # Client disconnected or error
# Extract data from stream_result
full_content = result_data["full_content"]
full_thinking = result_data["full_thinking"]
tool_calls_list = result_data["tool_calls_list"]
thinking_step_id = result_data["thinking_step_id"]
thinking_step_idx = result_data["thinking_step_idx"]
text_step_id = result_data["text_step_id"]
text_step_idx = result_data["text_step_idx"]
completion_tokens = result_data["completion_tokens"]
prompt_tokens = result_data["prompt_tokens"]
total_prompt_tokens += prompt_tokens
total_completion_tokens += completion_tokens
# Save thinking/text steps to all_steps for DB storage
if thinking_step_id is not None:
all_steps.append({
"id": thinking_step_id, "index": thinking_step_idx,
"type": "thinking", "content": full_thinking,
})
step_index += 1
if text_step_id is not None:
all_steps.append({
"id": text_step_id, "index": text_step_idx,
"type": "text", "content": full_content,
})
step_index += 1
# --- Branch: tool calls vs final ---
if tool_calls_list:
all_tool_calls.extend(tool_calls_list)
# Emit tool_call steps (before execution)
for tc in tool_calls_list:
call_step = {
"id": f"step-{step_index}",
"index": step_index,
"type": "tool_call",
"id_ref": tc["id"],
"name": tc["function"]["name"],
"arguments": tc["function"]["arguments"],
}
all_steps.append(call_step)
yield _sse_event("process_step", call_step)
step_index += 1
# Execute tools with error wrapping
tool_results = self._execute_tools_safe(
app, executor, tool_calls_list, context
)
# Emit tool_result steps
for tr in tool_results:
skipped = False
try:
result_content = json.loads(tr["content"])
skipped = result_content.get("skipped", False)
except Exception:
skipped = False
result_step = {
"id": f"step-{step_index}",
"index": step_index,
"type": "tool_result",
"id_ref": tr["tool_call_id"],
"name": tr["name"],
"content": tr["content"],
"skipped": skipped,
}
all_steps.append(result_step)
yield _sse_event("process_step", result_step)
step_index += 1
# Append assistant message + tool results for the next iteration
messages.append({
"role": "assistant",
"content": full_content or "",
"tool_calls": tool_calls_list,
})
messages.extend(tool_results)
all_tool_results.extend(tool_results)
continue
# --- No tool calls: final iteration — save message to DB ---
msg_id = str(uuid.uuid4())
suggested_title = self._save_message(
app, conv_id, conv_model, msg_id,
full_content, all_tool_calls, all_tool_results,
all_steps, total_prompt_tokens, total_completion_tokens,
)
yield _sse_event("done", {
"message_id": msg_id,
"token_count": total_completion_tokens,
"suggested_title": suggested_title,
})
return
yield _sse_event("error", {"content": "Exceeded maximum tool call iterations"})
def safe_generate():
"""Wrapper that catches client disconnection during yield."""
try:
yield from generate()
except (ClientDisconnected, BrokenPipeError, ConnectionResetError):
pass
return Response(
safe_generate(),
mimetype="text/event-stream",
headers={
"Cache-Control": "no-cache, no-store, must-revalidate",
"X-Accel-Buffering": "no",
"Connection": "keep-alive",
"Transfer-Encoding": "chunked",
},
)
# ------------------------------------------------------------------
# Private helpers — extracted for testability and readability
# ------------------------------------------------------------------
def _stream_llm_response(
self, app, messages, tools, tool_choice, step_index,
model, max_tokens, temperature, thinking_enabled,
):
"""Call LLM streaming API and yield SSE events in real-time.
This is a generator that yields SSE event strings as they are received.
The final yield is a 'stream_result' event containing the accumulated data.
Yields:
str: SSE event strings (process_step events, then stream_result)
Raises:
HTTPError / ConnectionError / Timeout for the caller to handle.
"""
full_content = ""
full_thinking = ""
token_count = 0
prompt_tokens = 0
tool_calls_list = []
thinking_step_id = None
thinking_step_idx = None
text_step_id = None
text_step_idx = None
with app.app_context():
resp = self.llm.call(
model=model,
messages=messages,
max_tokens=max_tokens,
temperature=temperature,
thinking_enabled=thinking_enabled,
tools=tools,
tool_choice=tool_choice,
stream=True,
)
resp.raise_for_status()
for line in resp.iter_lines():
if _client_disconnected():
resp.close()
return # Client disconnected, stop generator
if not line:
continue
line = line.decode("utf-8")
if not line.startswith("data: "):
continue
data_str = line[6:]
if data_str == "[DONE]":
break
try:
chunk = json.loads(data_str)
except json.JSONDecodeError:
continue
usage = chunk.get("usage", {})
if usage:
token_count = usage.get("completion_tokens", 0)
prompt_tokens = usage.get("prompt_tokens", 0)
choices = chunk.get("choices", [])
if not choices:
continue
delta = choices[0].get("delta", {})
# Yield thinking content in real-time
reasoning = delta.get("reasoning_content", "")
if reasoning:
full_thinking += reasoning
if thinking_step_id is None:
thinking_step_id = f"step-{step_index}"
thinking_step_idx = step_index
yield _sse_event("process_step", {
"id": thinking_step_id, "index": thinking_step_idx,
"type": "thinking", "content": full_thinking,
})
# Yield text content in real-time
text = delta.get("content", "")
if text:
full_content += text
if text_step_id is None:
text_step_idx = step_index + (1 if thinking_step_id is not None else 0)
text_step_id = f"step-{text_step_idx}"
yield _sse_event("process_step", {
"id": text_step_id, "index": text_step_idx,
"type": "text", "content": full_content,
})
tool_calls_list = self._process_tool_calls_delta(delta, tool_calls_list)
# Final yield: stream_result event with accumulated data
yield _sse_event("stream_result", {
"full_content": full_content,
"full_thinking": full_thinking,
"tool_calls_list": tool_calls_list,
"thinking_step_id": thinking_step_id,
"thinking_step_idx": thinking_step_idx,
"text_step_id": text_step_id,
"text_step_idx": text_step_idx,
"completion_tokens": token_count,
"prompt_tokens": prompt_tokens,
})
def _execute_tools_safe(self, app, executor, tool_calls_list, context):
"""Execute tool calls with top-level error wrapping.
If an unexpected exception occurs during tool execution, it is
converted into error tool results instead of crashing the stream.
"""
try:
if len(tool_calls_list) > 1:
with app.app_context():
return executor.process_tool_calls_parallel(
tool_calls_list, context, max_workers=_cfg.tool_max_workers
)
else:
with app.app_context():
return executor.process_tool_calls(
tool_calls_list, context
)
except Exception as e:
logger.exception("Error during tool execution")
return [
{
"role": "tool",
"tool_call_id": tc["id"],
"name": tc["function"]["name"],
"content": json.dumps({
"success": False,
"error": f"Tool execution failed: {e}",
}, ensure_ascii=False),
}
for tc in tool_calls_list
]
def _save_message(
self, app, conv_id, conv_model, msg_id,
full_content, all_tool_calls, all_tool_results,
all_steps, total_prompt_tokens, total_completion_tokens,
):
"""Save the final assistant message and auto-generate title if needed.
Returns the suggested_title or None.
"""
suggested_title = None
with app.app_context():
content_json = {"text": full_content}
if all_tool_calls:
content_json["tool_calls"] = self._build_tool_calls_json(
all_tool_calls, all_tool_results
)
content_json["steps"] = all_steps
msg = Message(
id=msg_id,
conversation_id=conv_id,
role="assistant",
content=json.dumps(content_json, ensure_ascii=False),
token_count=total_completion_tokens,
)
db.session.add(msg)
db.session.commit()
conv = db.session.get(Conversation, conv_id)
if conv:
record_token_usage(
conv.user_id, conv_model,
total_prompt_tokens, total_completion_tokens,
)
if conv and (not conv.title or conv.title == "新对话"):
user_msg = Message.query.filter_by(
conversation_id=conv_id, role="user"
).order_by(Message.created_at.asc()).first()
if user_msg and user_msg.content:
try:
content_data = json.loads(user_msg.content)
title_text = content_data.get("text", "")[:30]
except (json.JSONDecodeError, TypeError):
title_text = user_msg.content.strip()[:30]
suggested_title = title_text or "新对话"
db.session.refresh(conv)
conv.title = suggested_title
db.session.commit()
return suggested_title
def _build_tool_calls_json(self, tool_calls: list, tool_results: list) -> list:
"""Build tool calls JSON structure - matches streaming format."""
result = []
for i, tc in enumerate(tool_calls):
result_content = tool_results[i]["content"] if i < len(tool_results) else None
success = True
skipped = False
execution_time = 0
if result_content:
try:
result_data = json.loads(result_content)
success = result_data.get("success", True)
skipped = result_data.get("skipped", False)
execution_time = result_data.get("execution_time", 0)
except (json.JSONDecodeError, TypeError):
pass
result.append({
"id": tc.get("id", ""),
"type": tc.get("type", "function"),
"function": {
"name": tc["function"]["name"],
"arguments": tc["function"]["arguments"],
},
"result": result_content,
"success": success,
"skipped": skipped,
"execution_time": execution_time,
})
return result
def _process_tool_calls_delta(self, delta: dict, tool_calls_list: list) -> list:
"""Process tool calls from streaming delta."""
tool_calls_delta = delta.get("tool_calls", [])
for tc in tool_calls_delta:
idx = tc.get("index", 0)
if idx >= len(tool_calls_list):
tool_calls_list.append({
"id": tc.get("id", ""),
"type": tc.get("type", "function"),
"function": {"name": "", "arguments": ""},
})
if tc.get("id"):
tool_calls_list[idx]["id"] = tc["id"]
if tc.get("function"):
if tc["function"].get("name"):
tool_calls_list[idx]["function"]["name"] = tc["function"]["name"]
if tc["function"].get("arguments"):
tool_calls_list[idx]["function"]["arguments"] += tc["function"]["arguments"]
return tool_calls_list