|
|
|
|
|
|
|
|
gen_conf = dialog.llm_setting
|
|
|
gen_conf = dialog.llm_setting
|
|
|
|
|
|
|
|
|
msg = [{"role": "system", "content": prompt_config["system"].format(**kwargs)}]
|
|
|
msg = [{"role": "system", "content": prompt_config["system"].format(**kwargs)}]
|
|
|
msg.extend([{"role": m["role"], "content": m["content"]}
|
|
|
|
|
|
|
|
|
msg.extend([{"role": m["role"], "content": re.sub(r"##\d+\$\$", "", m["content"])}
|
|
|
for m in messages if m["role"] != "system"])
|
|
|
for m in messages if m["role"] != "system"])
|
|
|
used_token_count, msg = message_fit_in(msg, int(max_tokens * 0.97))
|
|
|
used_token_count, msg = message_fit_in(msg, int(max_tokens * 0.97))
|
|
|
assert len(msg) >= 2, f"message_fit_in has bug: {msg}"
|
|
|
assert len(msg) >= 2, f"message_fit_in has bug: {msg}"
|