增加enable_thinking和enable_memory
This commit is contained in:
parent
26a85299b3
commit
f1107ea35a
@ -112,10 +112,6 @@ class AgentConfig:
|
||||
robot_type = "deep_agent"
|
||||
|
||||
preamble_text, system_prompt = get_preamble_text(request.language, request.system_prompt)
|
||||
enable_thinking = request.enable_thinking and "<guidelines>" in request.system_prompt
|
||||
|
||||
# 从请求中获取 Mem0 配置,如果没有则使用全局配置
|
||||
enable_memori = getattr(request, 'enable_memori', MEM0_ENABLED)
|
||||
|
||||
config = cls(
|
||||
bot_id=request.bot_id,
|
||||
@ -128,7 +124,7 @@ class AgentConfig:
|
||||
robot_type=robot_type,
|
||||
user_identifier=request.user_identifier,
|
||||
session_id=request.session_id,
|
||||
enable_thinking=enable_thinking,
|
||||
enable_thinking=request.enable_thinking,
|
||||
project_dir=project_dir,
|
||||
stream=request.stream,
|
||||
tool_response=request.tool_response,
|
||||
@ -138,7 +134,7 @@ class AgentConfig:
|
||||
_origin_messages=messages,
|
||||
preamble_text=preamble_text,
|
||||
dataset_ids=request.dataset_ids,
|
||||
enable_memori=enable_memori,
|
||||
enable_memori=request.enable_memory,
|
||||
memori_semantic_search_top_k=getattr(request, 'memori_semantic_search_top_k', None) or MEM0_SEMANTIC_SEARCH_TOP_K,
|
||||
trace_id=trace_id,
|
||||
)
|
||||
@ -185,10 +181,8 @@ class AgentConfig:
|
||||
robot_type = bot_config.get("robot_type", "general_agent")
|
||||
if robot_type == "catalog_agent":
|
||||
robot_type = "deep_agent"
|
||||
enable_thinking = request.enable_thinking and "<guidelines>" in bot_config.get("system_prompt")
|
||||
|
||||
# 从请求或后端配置中获取 Mem0 配置
|
||||
enable_memori = getattr(request, 'enable_memori', MEM0_ENABLED)
|
||||
enable_thinking = bot_config.get("enable_thinking", False)
|
||||
enable_memori = bot_config.get("enable_memory", False)
|
||||
|
||||
config = cls(
|
||||
bot_id=request.bot_id,
|
||||
|
||||
@ -132,9 +132,6 @@ async def init_agent(config: AgentConfig):
|
||||
(agent, checkpointer) 元组
|
||||
"""
|
||||
|
||||
|
||||
|
||||
|
||||
# 加载配置
|
||||
final_system_prompt = await load_system_prompt_async(
|
||||
config.project_dir, config.language, config.system_prompt, config.robot_type, config.bot_id, config.user_identifier, config.trace_id or ""
|
||||
|
||||
@ -385,7 +385,7 @@ async def chat_completions(request: ChatRequest, authorization: Optional[str] =
|
||||
project_dir = create_project_directory(request.dataset_ids, bot_id, request.robot_type, request.skills)
|
||||
|
||||
# 收集额外参数作为 generate_cfg
|
||||
exclude_fields = {'messages', 'model', 'model_server', 'dataset_ids', 'language', 'tool_response', 'system_prompt', 'mcp_settings' ,'stream', 'robot_type', 'bot_id', 'user_identifier', 'session_id', 'enable_thinking', 'skills'}
|
||||
exclude_fields = {'messages', 'model', 'model_server', 'dataset_ids', 'language', 'tool_response', 'system_prompt', 'mcp_settings' ,'stream', 'robot_type', 'bot_id', 'user_identifier', 'session_id', 'enable_thinking', 'skills', 'enable_memory'}
|
||||
generate_cfg = {k: v for k, v in request.model_dump().items() if k not in exclude_fields}
|
||||
# 处理消息
|
||||
messages = process_messages(request.messages, request.language)
|
||||
@ -435,7 +435,7 @@ async def chat_warmup_v1(request: ChatRequest, authorization: Optional[str] = He
|
||||
project_dir = create_project_directory(request.dataset_ids, bot_id, request.robot_type, request.skills)
|
||||
|
||||
# 收集额外参数作为 generate_cfg
|
||||
exclude_fields = {'messages', 'model', 'model_server', 'dataset_ids', 'language', 'tool_response', 'system_prompt', 'mcp_settings' ,'stream', 'robot_type', 'bot_id', 'user_identifier', 'session_id', 'enable_thinking', 'skills'}
|
||||
exclude_fields = {'messages', 'model', 'model_server', 'dataset_ids', 'language', 'tool_response', 'system_prompt', 'mcp_settings' ,'stream', 'robot_type', 'bot_id', 'user_identifier', 'session_id', 'enable_thinking', 'skills', 'enable_memory'}
|
||||
generate_cfg = {k: v for k, v in request.model_dump().items() if k not in exclude_fields}
|
||||
|
||||
# 创建一个空的消息列表用于预热(实际消息不会在warmup中处理)
|
||||
|
||||
@ -55,6 +55,7 @@ class ChatRequest(BaseModel):
|
||||
session_id: Optional[str] = None
|
||||
enable_thinking: Optional[bool] = DEFAULT_THINKING_ENABLE
|
||||
skills: Optional[List[str]] = None
|
||||
enable_memory: Optional[bool] = False
|
||||
|
||||
|
||||
class ChatRequestV2(BaseModel):
|
||||
@ -65,7 +66,6 @@ class ChatRequestV2(BaseModel):
|
||||
language: Optional[str] = "zh"
|
||||
user_identifier: Optional[str] = ""
|
||||
session_id: Optional[str] = None
|
||||
enable_thinking: Optional[bool] = DEFAULT_THINKING_ENABLE
|
||||
|
||||
|
||||
class FileProcessRequest(BaseModel):
|
||||
|
||||
Loading…
Reference in New Issue
Block a user