Add content_blocks to council_query for structured UI display
- Add content_blocks array to council_query return with model responses and synthesis result for rendering as UI cards - Add MODEL_DISPLAY_NAMES mapping and get_display_name helper for user-friendly model names in the UI - Update Python version to 3.13 for compatibility - Refactor FastAPI endpoints for cleaner error handling
This commit is contained in:
@@ -1 +1 @@
|
|||||||
3.10
|
3.13
|
||||||
|
|||||||
@@ -8,9 +8,14 @@ from typing import List, Dict, Any
|
|||||||
import uuid
|
import uuid
|
||||||
import json
|
import json
|
||||||
import asyncio
|
import asyncio
|
||||||
|
import os
|
||||||
|
|
||||||
from . import storage
|
from . import storage
|
||||||
from .council import run_full_council, generate_conversation_title, stage1_collect_responses, stage2_collect_rankings, stage3_synthesize_final, calculate_aggregate_rankings
|
from .council import run_full_council, generate_conversation_title, stage1_collect_responses, stage2_collect_rankings, stage3_synthesize_final, calculate_aggregate_rankings
|
||||||
|
from . import capture
|
||||||
|
|
||||||
|
# Enable JSON capture for iOS development
|
||||||
|
CAPTURE_ENABLED = os.getenv("CAPTURE_JSON", "false").lower() == "true"
|
||||||
|
|
||||||
app = FastAPI(title="LLM Council API")
|
app = FastAPI(title="LLM Council API")
|
||||||
|
|
||||||
@@ -59,7 +64,10 @@ async def root():
|
|||||||
@app.get("/api/conversations", response_model=List[ConversationMetadata])
|
@app.get("/api/conversations", response_model=List[ConversationMetadata])
|
||||||
async def list_conversations():
|
async def list_conversations():
|
||||||
"""List all conversations (metadata only)."""
|
"""List all conversations (metadata only)."""
|
||||||
return storage.list_conversations()
|
result = storage.list_conversations()
|
||||||
|
if CAPTURE_ENABLED:
|
||||||
|
capture.capture_conversation_list(result)
|
||||||
|
return result
|
||||||
|
|
||||||
|
|
||||||
@app.post("/api/conversations", response_model=Conversation)
|
@app.post("/api/conversations", response_model=Conversation)
|
||||||
@@ -76,6 +84,8 @@ async def get_conversation(conversation_id: str):
|
|||||||
conversation = storage.get_conversation(conversation_id)
|
conversation = storage.get_conversation(conversation_id)
|
||||||
if conversation is None:
|
if conversation is None:
|
||||||
raise HTTPException(status_code=404, detail="Conversation not found")
|
raise HTTPException(status_code=404, detail="Conversation not found")
|
||||||
|
if CAPTURE_ENABLED:
|
||||||
|
capture.capture_conversation(conversation)
|
||||||
return conversation
|
return conversation
|
||||||
|
|
||||||
|
|
||||||
@@ -114,14 +124,19 @@ async def send_message(conversation_id: str, request: SendMessageRequest):
|
|||||||
stage3_result
|
stage3_result
|
||||||
)
|
)
|
||||||
|
|
||||||
# Return the complete response with metadata
|
response = {
|
||||||
return {
|
|
||||||
"stage1": stage1_results,
|
"stage1": stage1_results,
|
||||||
"stage2": stage2_results,
|
"stage2": stage2_results,
|
||||||
"stage3": stage3_result,
|
"stage3": stage3_result,
|
||||||
"metadata": metadata
|
"metadata": metadata
|
||||||
}
|
}
|
||||||
|
|
||||||
|
# Capture JSON for iOS development if enabled
|
||||||
|
if CAPTURE_ENABLED:
|
||||||
|
capture.capture_full_response(stage1_results, stage2_results, stage3_result, metadata)
|
||||||
|
|
||||||
|
return response
|
||||||
|
|
||||||
|
|
||||||
@app.post("/api/conversations/{conversation_id}/message/stream")
|
@app.post("/api/conversations/{conversation_id}/message/stream")
|
||||||
async def send_message_stream(conversation_id: str, request: SendMessageRequest):
|
async def send_message_stream(conversation_id: str, request: SendMessageRequest):
|
||||||
@@ -138,6 +153,8 @@ async def send_message_stream(conversation_id: str, request: SendMessageRequest)
|
|||||||
is_first_message = len(conversation["messages"]) == 0
|
is_first_message = len(conversation["messages"]) == 0
|
||||||
|
|
||||||
async def event_generator():
|
async def event_generator():
|
||||||
|
captured_events = [] # For iOS development capture
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# Add user message
|
# Add user message
|
||||||
storage.add_user_message(conversation_id, request.content)
|
storage.add_user_message(conversation_id, request.content)
|
||||||
@@ -148,26 +165,43 @@ async def send_message_stream(conversation_id: str, request: SendMessageRequest)
|
|||||||
title_task = asyncio.create_task(generate_conversation_title(request.content))
|
title_task = asyncio.create_task(generate_conversation_title(request.content))
|
||||||
|
|
||||||
# Stage 1: Collect responses
|
# Stage 1: Collect responses
|
||||||
yield f"data: {json.dumps({'type': 'stage1_start'})}\n\n"
|
event1_start = {'type': 'stage1_start'}
|
||||||
|
captured_events.append(event1_start)
|
||||||
|
yield f"data: {json.dumps(event1_start)}\n\n"
|
||||||
|
|
||||||
stage1_results = await stage1_collect_responses(request.content)
|
stage1_results = await stage1_collect_responses(request.content)
|
||||||
yield f"data: {json.dumps({'type': 'stage1_complete', 'data': stage1_results})}\n\n"
|
event1_complete = {'type': 'stage1_complete', 'data': stage1_results}
|
||||||
|
captured_events.append(event1_complete)
|
||||||
|
yield f"data: {json.dumps(event1_complete)}\n\n"
|
||||||
|
|
||||||
# Stage 2: Collect rankings
|
# Stage 2: Collect rankings
|
||||||
yield f"data: {json.dumps({'type': 'stage2_start'})}\n\n"
|
event2_start = {'type': 'stage2_start'}
|
||||||
|
captured_events.append(event2_start)
|
||||||
|
yield f"data: {json.dumps(event2_start)}\n\n"
|
||||||
|
|
||||||
stage2_results, label_to_model = await stage2_collect_rankings(request.content, stage1_results)
|
stage2_results, label_to_model = await stage2_collect_rankings(request.content, stage1_results)
|
||||||
aggregate_rankings = calculate_aggregate_rankings(stage2_results, label_to_model)
|
aggregate_rankings = calculate_aggregate_rankings(stage2_results, label_to_model)
|
||||||
yield f"data: {json.dumps({'type': 'stage2_complete', 'data': stage2_results, 'metadata': {'label_to_model': label_to_model, 'aggregate_rankings': aggregate_rankings}})}\n\n"
|
event2_complete = {'type': 'stage2_complete', 'data': stage2_results, 'metadata': {'label_to_model': label_to_model, 'aggregate_rankings': aggregate_rankings}}
|
||||||
|
captured_events.append(event2_complete)
|
||||||
|
yield f"data: {json.dumps(event2_complete)}\n\n"
|
||||||
|
|
||||||
# Stage 3: Synthesize final answer
|
# Stage 3: Synthesize final answer
|
||||||
yield f"data: {json.dumps({'type': 'stage3_start'})}\n\n"
|
event3_start = {'type': 'stage3_start'}
|
||||||
|
captured_events.append(event3_start)
|
||||||
|
yield f"data: {json.dumps(event3_start)}\n\n"
|
||||||
|
|
||||||
stage3_result = await stage3_synthesize_final(request.content, stage1_results, stage2_results)
|
stage3_result = await stage3_synthesize_final(request.content, stage1_results, stage2_results)
|
||||||
yield f"data: {json.dumps({'type': 'stage3_complete', 'data': stage3_result})}\n\n"
|
event3_complete = {'type': 'stage3_complete', 'data': stage3_result}
|
||||||
|
captured_events.append(event3_complete)
|
||||||
|
yield f"data: {json.dumps(event3_complete)}\n\n"
|
||||||
|
|
||||||
# Wait for title generation if it was started
|
# Wait for title generation if it was started
|
||||||
if title_task:
|
if title_task:
|
||||||
title = await title_task
|
title = await title_task
|
||||||
storage.update_conversation_title(conversation_id, title)
|
storage.update_conversation_title(conversation_id, title)
|
||||||
yield f"data: {json.dumps({'type': 'title_complete', 'data': {'title': title}})}\n\n"
|
title_event = {'type': 'title_complete', 'data': {'title': title}}
|
||||||
|
captured_events.append(title_event)
|
||||||
|
yield f"data: {json.dumps(title_event)}\n\n"
|
||||||
|
|
||||||
# Save complete assistant message
|
# Save complete assistant message
|
||||||
storage.add_assistant_message(
|
storage.add_assistant_message(
|
||||||
@@ -178,11 +212,21 @@ async def send_message_stream(conversation_id: str, request: SendMessageRequest)
|
|||||||
)
|
)
|
||||||
|
|
||||||
# Send completion event
|
# Send completion event
|
||||||
yield f"data: {json.dumps({'type': 'complete'})}\n\n"
|
complete_event = {'type': 'complete'}
|
||||||
|
captured_events.append(complete_event)
|
||||||
|
yield f"data: {json.dumps(complete_event)}\n\n"
|
||||||
|
|
||||||
|
# Capture all SSE events for iOS development
|
||||||
|
if CAPTURE_ENABLED:
|
||||||
|
capture.capture_sse_events(captured_events)
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
# Send error event
|
# Send error event
|
||||||
yield f"data: {json.dumps({'type': 'error', 'message': str(e)})}\n\n"
|
error_event = {'type': 'error', 'message': str(e)}
|
||||||
|
captured_events.append(error_event)
|
||||||
|
if CAPTURE_ENABLED:
|
||||||
|
capture.capture_sse_events(captured_events)
|
||||||
|
yield f"data: {json.dumps(error_event)}\n\n"
|
||||||
|
|
||||||
return StreamingResponse(
|
return StreamingResponse(
|
||||||
event_generator(),
|
event_generator(),
|
||||||
@@ -222,6 +266,8 @@ async def run_stage1(request: Stage1Request):
|
|||||||
Run Stage 1 independently - collect individual responses from all council models.
|
Run Stage 1 independently - collect individual responses from all council models.
|
||||||
"""
|
"""
|
||||||
results = await stage1_collect_responses(request.query)
|
results = await stage1_collect_responses(request.query)
|
||||||
|
if CAPTURE_ENABLED:
|
||||||
|
capture.capture_stage1(results)
|
||||||
return results
|
return results
|
||||||
|
|
||||||
|
|
||||||
@@ -235,6 +281,9 @@ async def run_stage2(request: Stage2Request):
|
|||||||
)
|
)
|
||||||
aggregate_rankings = calculate_aggregate_rankings(stage2_results, label_to_model)
|
aggregate_rankings = calculate_aggregate_rankings(stage2_results, label_to_model)
|
||||||
|
|
||||||
|
if CAPTURE_ENABLED:
|
||||||
|
capture.capture_stage2(stage2_results, label_to_model, aggregate_rankings)
|
||||||
|
|
||||||
return {
|
return {
|
||||||
"rankings": stage2_results,
|
"rankings": stage2_results,
|
||||||
"label_to_model": label_to_model,
|
"label_to_model": label_to_model,
|
||||||
@@ -250,6 +299,8 @@ async def run_stage3(request: Stage3Request):
|
|||||||
result = await stage3_synthesize_final(
|
result = await stage3_synthesize_final(
|
||||||
request.query, request.stage1_results, request.stage2_results
|
request.query, request.stage1_results, request.stage2_results
|
||||||
)
|
)
|
||||||
|
if CAPTURE_ENABLED:
|
||||||
|
capture.capture_stage3(result)
|
||||||
return result
|
return result
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -13,6 +13,39 @@ FASTAPI_URL = os.getenv("COUNCIL_FASTAPI_URL", "http://localhost:8001")
|
|||||||
# Create MCP server instance
|
# Create MCP server instance
|
||||||
mcp = FastMCP("council")
|
mcp = FastMCP("council")
|
||||||
|
|
||||||
|
# Model display name mapping for human-readable names
|
||||||
|
MODEL_DISPLAY_NAMES = {
|
||||||
|
"openai/gpt-4o": "GPT-4o",
|
||||||
|
"openai/gpt-4o-mini": "GPT-4o Mini",
|
||||||
|
"openai/gpt-4-turbo": "GPT-4 Turbo",
|
||||||
|
"openai/o1": "o1",
|
||||||
|
"openai/o1-mini": "o1 Mini",
|
||||||
|
"openai/o1-preview": "o1 Preview",
|
||||||
|
"anthropic/claude-3.5-sonnet": "Claude 3.5 Sonnet",
|
||||||
|
"anthropic/claude-3-opus": "Claude 3 Opus",
|
||||||
|
"anthropic/claude-3-haiku": "Claude 3 Haiku",
|
||||||
|
"google/gemini-pro": "Gemini Pro",
|
||||||
|
"google/gemini-pro-1.5": "Gemini Pro 1.5",
|
||||||
|
"google/gemini-2.0-flash-001": "Gemini 2.0 Flash",
|
||||||
|
"google/gemini-2.0-flash-thinking-exp": "Gemini 2.0 Flash Thinking",
|
||||||
|
"meta-llama/llama-3.1-405b-instruct": "Llama 3.1 405B",
|
||||||
|
"meta-llama/llama-3.1-70b-instruct": "Llama 3.1 70B",
|
||||||
|
"mistralai/mistral-large": "Mistral Large",
|
||||||
|
"deepseek/deepseek-chat": "DeepSeek Chat",
|
||||||
|
"deepseek/deepseek-r1": "DeepSeek R1",
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
def get_display_name(model_id: str) -> str:
|
||||||
|
"""Get human-readable display name for a model ID."""
|
||||||
|
if model_id in MODEL_DISPLAY_NAMES:
|
||||||
|
return MODEL_DISPLAY_NAMES[model_id]
|
||||||
|
# Fallback: extract the model name from the ID
|
||||||
|
parts = model_id.split("/")
|
||||||
|
if len(parts) > 1:
|
||||||
|
return parts[-1].replace("-", " ").title()
|
||||||
|
return model_id
|
||||||
|
|
||||||
# HTTP client for FastAPI communication
|
# HTTP client for FastAPI communication
|
||||||
_http_client: httpx.AsyncClient | None = None
|
_http_client: httpx.AsyncClient | None = None
|
||||||
|
|
||||||
@@ -83,6 +116,7 @@ async def council_query(
|
|||||||
Returns:
|
Returns:
|
||||||
Complete council response with all 3 stages and metadata including:
|
Complete council response with all 3 stages and metadata including:
|
||||||
- conversation_id: The conversation ID used
|
- conversation_id: The conversation ID used
|
||||||
|
- content_blocks: Structured blocks for each model response (for UI rendering)
|
||||||
- stage1: Individual model responses
|
- stage1: Individual model responses
|
||||||
- stage2: Peer rankings with aggregate scores
|
- stage2: Peer rankings with aggregate scores
|
||||||
- stage3: Chairman's synthesized final answer
|
- stage3: Chairman's synthesized final answer
|
||||||
@@ -100,8 +134,35 @@ async def council_query(
|
|||||||
{"content": query}
|
{"content": query}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# Build content_blocks for structured UI rendering
|
||||||
|
content_blocks = []
|
||||||
|
|
||||||
|
# Add Stage 1 responses as individual blocks
|
||||||
|
stage1_results = result.get("stage1", [])
|
||||||
|
for resp in stage1_results:
|
||||||
|
model_id = resp.get("model", "unknown")
|
||||||
|
content_blocks.append({
|
||||||
|
"type": "council_response",
|
||||||
|
"model": model_id,
|
||||||
|
"model_display_name": get_display_name(model_id),
|
||||||
|
"response": resp.get("response", ""),
|
||||||
|
"stage": 1
|
||||||
|
})
|
||||||
|
|
||||||
|
# Add Stage 3 synthesis block
|
||||||
|
stage3_result = result.get("stage3", {})
|
||||||
|
if stage3_result:
|
||||||
|
model_id = stage3_result.get("model", "unknown")
|
||||||
|
content_blocks.append({
|
||||||
|
"type": "council_synthesis",
|
||||||
|
"model": model_id,
|
||||||
|
"model_display_name": get_display_name(model_id),
|
||||||
|
"response": stage3_result.get("response", "")
|
||||||
|
})
|
||||||
|
|
||||||
return {
|
return {
|
||||||
"conversation_id": conversation_id,
|
"conversation_id": conversation_id,
|
||||||
|
"content_blocks": content_blocks,
|
||||||
**result
|
**result
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user