浏览代码

fix: next suggest question logic problem (#6451)

Co-authored-by: evenyan <yikun.yan@ubtrobot.com>
tags/0.6.15
Even 1年前
父节点
当前提交
c013086e64
没有帐户链接到提交者的电子邮件
共有 2 个文件被更改,包括 2 次插入1 次删除
  1. 1
    0
      api/core/llm_generator/prompts.py
  2. 1
    1
      api/core/memory/token_buffer_memory.py

+ 1
- 0
api/core/llm_generator/prompts.py 查看文件

@@ -64,6 +64,7 @@ User Input:
SUGGESTED_QUESTIONS_AFTER_ANSWER_INSTRUCTION_PROMPT = (
"Please help me predict the three most likely questions that human would ask, "
"and keeping each question under 20 characters.\n"
"MAKE SURE your output is the SAME language as the Assistant's latest response(if the main response is written in Chinese, then the language of your output must be using Chinese.)!\n"
"The output must be an array in JSON format following the specified schema:\n"
"[\"question1\",\"question2\",\"question3\"]\n"
)

+ 1
- 1
api/core/memory/token_buffer_memory.py 查看文件

@@ -103,7 +103,7 @@ class TokenBufferMemory:

if curr_message_tokens > max_token_limit:
pruned_memory = []
while curr_message_tokens > max_token_limit and prompt_messages:
while curr_message_tokens > max_token_limit and len(prompt_messages)>1:
pruned_memory.append(prompt_messages.pop(0))
curr_message_tokens = self.model_instance.get_llm_num_tokens(
prompt_messages

正在加载...
取消
保存