remove max_tokens
This commit is contained in:
parent
4591163282
commit
3b70d7511d
@ -512,7 +512,7 @@ async def chat_completions(request: ChatRequest, authorization: Optional[str] =
|
|||||||
project_dir = create_project_directory(request.dataset_ids, bot_id, request.skills)
|
project_dir = create_project_directory(request.dataset_ids, bot_id, request.skills)
|
||||||
|
|
||||||
# 收集额外参数作为 generate_cfg
|
# 收集额外参数作为 generate_cfg
|
||||||
exclude_fields = {'messages', 'model', 'model_server', 'dataset_ids', 'language', 'tool_response', 'system_prompt', 'mcp_settings' ,'stream', 'robot_type', 'bot_id', 'user_identifier', 'session_id', 'enable_thinking', 'skills', 'enable_memory', 'n', 'shell_env'}
|
exclude_fields = {'messages', 'model', 'model_server', 'dataset_ids', 'language', 'tool_response', 'system_prompt', 'mcp_settings' ,'stream', 'robot_type', 'bot_id', 'user_identifier', 'session_id', 'enable_thinking', 'skills', 'enable_memory', 'n', 'shell_env', 'max_tokens'}
|
||||||
generate_cfg = {k: v for k, v in request.model_dump().items() if k not in exclude_fields}
|
generate_cfg = {k: v for k, v in request.model_dump().items() if k not in exclude_fields}
|
||||||
# 处理消息
|
# 处理消息
|
||||||
messages = process_messages(request.messages, request.language)
|
messages = process_messages(request.messages, request.language)
|
||||||
@ -666,7 +666,7 @@ async def chat_warmup_v2(request: ChatRequestV2, authorization: Optional[str] =
|
|||||||
messages = process_messages(empty_messages, request.language or "ja")
|
messages = process_messages(empty_messages, request.language or "ja")
|
||||||
|
|
||||||
# 收集额外参数作为 generate_cfg
|
# 收集额外参数作为 generate_cfg
|
||||||
exclude_fields = {'messages', 'stream', 'tool_response', 'bot_id', 'language', 'user_identifier', 'session_id', 'n', 'model', 'model_server', 'api_key', 'shell_env'}
|
exclude_fields = {'messages', 'stream', 'tool_response', 'bot_id', 'language', 'user_identifier', 'session_id', 'n', 'model', 'model_server', 'api_key', 'shell_env', 'max_tokens'}
|
||||||
generate_cfg = {k: v for k, v in request.model_dump().items() if k not in exclude_fields}
|
generate_cfg = {k: v for k, v in request.model_dump().items() if k not in exclude_fields}
|
||||||
# 从请求中提取 model/model_server/api_key,优先级高于 bot_config(排除 "whatever" 和空值)
|
# 从请求中提取 model/model_server/api_key,优先级高于 bot_config(排除 "whatever" 和空值)
|
||||||
req_data = request.model_dump()
|
req_data = request.model_dump()
|
||||||
@ -773,7 +773,7 @@ async def chat_completions_v2(request: ChatRequestV2, authorization: Optional[st
|
|||||||
# 处理消息
|
# 处理消息
|
||||||
messages = process_messages(request.messages, request.language)
|
messages = process_messages(request.messages, request.language)
|
||||||
# 收集额外参数作为 generate_cfg
|
# 收集额外参数作为 generate_cfg
|
||||||
exclude_fields = {'messages', 'dataset_ids', 'language', 'tool_response', 'system_prompt', 'mcp_settings', 'stream', 'robot_type', 'bot_id', 'user_identifier', 'session_id', 'enable_thinking', 'skills', 'enable_memory', 'n', 'model', 'model_server', 'api_key', 'shell_env'}
|
exclude_fields = {'messages', 'dataset_ids', 'language', 'tool_response', 'system_prompt', 'mcp_settings', 'stream', 'robot_type', 'bot_id', 'user_identifier', 'session_id', 'enable_thinking', 'skills', 'enable_memory', 'n', 'model', 'model_server', 'api_key', 'shell_env', 'max_tokens'}
|
||||||
generate_cfg = {k: v for k, v in request.model_dump().items() if k not in exclude_fields}
|
generate_cfg = {k: v for k, v in request.model_dump().items() if k not in exclude_fields}
|
||||||
# 从请求中提取 model/model_server/api_key,优先级高于 bot_config(排除 "whatever" 和空值)
|
# 从请求中提取 model/model_server/api_key,优先级高于 bot_config(排除 "whatever" 和空值)
|
||||||
req_data = request.model_dump()
|
req_data = request.model_dump()
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user