From 4d88f52c77a5aad7ca40165d9dbbf78cbc9885b4 Mon Sep 17 00:00:00 2001 From: gjl <2802427218@qq.com> Date: Fri, 16 Aug 2024 22:03:25 +0800 Subject: [PATCH] =?UTF-8?q?feat:=E5=AE=8C=E5=96=84=E5=AF=B9=E8=AF=9D?= =?UTF-8?q?=E5=8A=9F=E8=83=BD?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/mindpilot/app/conversation/conversation_api.py | 8 ++++++++ 1 file changed, 8 insertions(+) diff --git a/src/mindpilot/app/conversation/conversation_api.py b/src/mindpilot/app/conversation/conversation_api.py index 9a8cffa..762d21e 100644 --- a/src/mindpilot/app/conversation/conversation_api.py +++ b/src/mindpilot/app/conversation/conversation_api.py @@ -181,6 +181,9 @@ async def send_messages( files: dict = Body({}, description="文件", examples=[{}]), text: str = Body("", description="消息内容"), tool_config: List[str] = Body([], description="工具配置", examples=[]), + temperature: float = Body(..., description="模型温度", examples=[0.8]), + max_tokens: int = Body(..., description="模型输出最大长度", examples=[4096]), + ): init_conversations_table() init_messages_table() @@ -222,6 +225,11 @@ async def send_messages( # 获取模型配置 chat_model_config = get_config_from_id(config_id=config_id) + model_key = next(iter(chat_model_config["llm_model"])) + chat_model_config["llm_model"][model_key]["temperature"] = temperature + chat_model_config["llm_model"][model_key]["max_tokens"] = max_tokens + + # 获取模型输出 ret = await chat_online(content=text, history=history, chat_model_config=chat_model_config,