diff --git a/agent/agent_config.py b/agent/agent_config.py index 31e2111..b64235f 100644 --- a/agent/agent_config.py +++ b/agent/agent_config.py @@ -112,10 +112,6 @@ class AgentConfig: robot_type = "deep_agent" preamble_text, system_prompt = get_preamble_text(request.language, request.system_prompt) - enable_thinking = request.enable_thinking and "" in request.system_prompt - - # 从请求中获取 Mem0 配置,如果没有则使用全局配置 - enable_memori = getattr(request, 'enable_memori', MEM0_ENABLED) config = cls( bot_id=request.bot_id, @@ -128,7 +124,7 @@ class AgentConfig: robot_type=robot_type, user_identifier=request.user_identifier, session_id=request.session_id, - enable_thinking=enable_thinking, + enable_thinking=request.enable_thinking, project_dir=project_dir, stream=request.stream, tool_response=request.tool_response, @@ -138,7 +134,7 @@ class AgentConfig: _origin_messages=messages, preamble_text=preamble_text, dataset_ids=request.dataset_ids, - enable_memori=enable_memori, + enable_memori=request.enable_memory, memori_semantic_search_top_k=getattr(request, 'memori_semantic_search_top_k', None) or MEM0_SEMANTIC_SEARCH_TOP_K, trace_id=trace_id, ) @@ -185,10 +181,8 @@ class AgentConfig: robot_type = bot_config.get("robot_type", "general_agent") if robot_type == "catalog_agent": robot_type = "deep_agent" - enable_thinking = request.enable_thinking and "" in bot_config.get("system_prompt") - - # 从请求或后端配置中获取 Mem0 配置 - enable_memori = getattr(request, 'enable_memori', MEM0_ENABLED) + enable_thinking = bot_config.get("enable_thinking", False) + enable_memori = bot_config.get("enable_memory", False) config = cls( bot_id=request.bot_id, diff --git a/agent/deep_assistant.py b/agent/deep_assistant.py index 5082d2a..3631ecd 100644 --- a/agent/deep_assistant.py +++ b/agent/deep_assistant.py @@ -132,9 +132,6 @@ async def init_agent(config: AgentConfig): (agent, checkpointer) 元组 """ - - - # 加载配置 final_system_prompt = await load_system_prompt_async( config.project_dir, config.language, config.system_prompt, config.robot_type, config.bot_id, config.user_identifier, config.trace_id or "" diff --git a/routes/chat.py b/routes/chat.py index 0f22c93..ca107d9 100644 --- a/routes/chat.py +++ b/routes/chat.py @@ -385,7 +385,7 @@ async def chat_completions(request: ChatRequest, authorization: Optional[str] = project_dir = create_project_directory(request.dataset_ids, bot_id, request.robot_type, request.skills) # 收集额外参数作为 generate_cfg - exclude_fields = {'messages', 'model', 'model_server', 'dataset_ids', 'language', 'tool_response', 'system_prompt', 'mcp_settings' ,'stream', 'robot_type', 'bot_id', 'user_identifier', 'session_id', 'enable_thinking', 'skills'} + exclude_fields = {'messages', 'model', 'model_server', 'dataset_ids', 'language', 'tool_response', 'system_prompt', 'mcp_settings' ,'stream', 'robot_type', 'bot_id', 'user_identifier', 'session_id', 'enable_thinking', 'skills', 'enable_memory'} generate_cfg = {k: v for k, v in request.model_dump().items() if k not in exclude_fields} # 处理消息 messages = process_messages(request.messages, request.language) @@ -435,7 +435,7 @@ async def chat_warmup_v1(request: ChatRequest, authorization: Optional[str] = He project_dir = create_project_directory(request.dataset_ids, bot_id, request.robot_type, request.skills) # 收集额外参数作为 generate_cfg - exclude_fields = {'messages', 'model', 'model_server', 'dataset_ids', 'language', 'tool_response', 'system_prompt', 'mcp_settings' ,'stream', 'robot_type', 'bot_id', 'user_identifier', 'session_id', 'enable_thinking', 'skills'} + exclude_fields = {'messages', 'model', 'model_server', 'dataset_ids', 'language', 'tool_response', 'system_prompt', 'mcp_settings' ,'stream', 'robot_type', 'bot_id', 'user_identifier', 'session_id', 'enable_thinking', 'skills', 'enable_memory'} generate_cfg = {k: v for k, v in request.model_dump().items() if k not in exclude_fields} # 创建一个空的消息列表用于预热(实际消息不会在warmup中处理) diff --git a/utils/api_models.py b/utils/api_models.py index 1f97fb9..85e8c36 100644 --- a/utils/api_models.py +++ b/utils/api_models.py @@ -55,6 +55,7 @@ class ChatRequest(BaseModel): session_id: Optional[str] = None enable_thinking: Optional[bool] = DEFAULT_THINKING_ENABLE skills: Optional[List[str]] = None + enable_memory: Optional[bool] = False class ChatRequestV2(BaseModel): @@ -65,7 +66,6 @@ class ChatRequestV2(BaseModel): language: Optional[str] = "zh" user_identifier: Optional[str] = "" session_id: Optional[str] = None - enable_thinking: Optional[bool] = DEFAULT_THINKING_ENABLE class FileProcessRequest(BaseModel):