summaryrefslogtreecommitdiff
path: root/backend/app/schemas.py
blob: 4213f15aab07b34fb16821ac8eff8040c67aa158 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
from pydantic import BaseModel, Field
from typing import List, Optional, Dict, Any, Union
from enum import Enum
import time

class Role(str, Enum):
    USER = "user"
    ASSISTANT = "assistant"
    SYSTEM = "system"

class Message(BaseModel):
    id: str = Field(..., description="Unique ID for the message")
    role: Role
    content: str
    timestamp: float = Field(default_factory=time.time)
    # Metadata to track where this message came from
    source_node_id: Optional[str] = None
    model_used: Optional[str] = None

class Context(BaseModel):
    messages: List[Message] = []

class ModelProvider(str, Enum):
    OPENAI = "openai"
    GOOGLE = "google"
    CLAUDE = "claude"

class ReasoningEffort(str, Enum):
    LOW = "low"
    MEDIUM = "medium"
    HIGH = "high"

class LLMConfig(BaseModel):
    provider: ModelProvider
    model_name: str
    temperature: float = 0.7
    max_tokens: Optional[int] = None
    system_prompt: Optional[str] = None
    api_key: Optional[str] = None  # Optional override, usually from env
    enable_google_search: bool = False
    reasoning_effort: ReasoningEffort = ReasoningEffort.MEDIUM  # For OpenAI reasoning models

class MergeStrategy(str, Enum):
    RAW = "raw"
    SMART = "smart"

class NodeRunRequest(BaseModel):
    node_id: str
    incoming_contexts: List[Context] = []
    user_prompt: str
    config: LLMConfig
    merge_strategy: MergeStrategy = MergeStrategy.SMART
    attached_file_ids: List[str] = Field(default_factory=list)
    # Scopes for file_search filtering: ["project_path/node_id", ...]
    # Contains all project/node combinations in the current trace
    scopes: List[str] = Field(default_factory=list)
    
class CouncilMemberConfig(BaseModel):
    model_name: str  # e.g. "gpt-5", "claude-opus-4-6", "gemini-3-pro-preview"
    temperature: Optional[float] = None           # None = use request default
    reasoning_effort: Optional[ReasoningEffort] = None
    enable_google_search: Optional[bool] = None
    incoming_contexts: Optional[List[Context]] = None  # Per-member context override

class CouncilRunRequest(BaseModel):
    node_id: str
    incoming_contexts: List[Context] = []  # Default context for all members
    user_prompt: str
    council_models: List[CouncilMemberConfig]  # 2-6 models
    chairman_model: CouncilMemberConfig  # Model config for synthesis
    system_prompt: Optional[str] = None
    temperature: float = 0.7
    reasoning_effort: ReasoningEffort = ReasoningEffort.MEDIUM
    enable_google_search: bool = False
    merge_strategy: MergeStrategy = MergeStrategy.SMART
    attached_file_ids: List[str] = Field(default_factory=list)
    scopes: List[str] = Field(default_factory=list)

class NodeRunResponse(BaseModel):
    node_id: str
    output_context: Context
    response_content: str
    raw_response: Optional[Dict[str, Any]] = None
    usage: Optional[Dict[str, Any]] = None