| """ | """ | ||||
| full_text = "" | full_text = "" | ||||
| chunk_index = 0 | chunk_index = 0 | ||||
| is_reasoning_started = False | |||||
| def create_final_llm_result_chunk( | def create_final_llm_result_chunk( | ||||
| index: int, message: AssistantPromptMessage, finish_reason: str | index: int, message: AssistantPromptMessage, finish_reason: str | ||||
| # transform assistant message to prompt message | # transform assistant message to prompt message | ||||
| text = chunk_json["response"] | text = chunk_json["response"] | ||||
| if "<think>" in text: | |||||
| is_reasoning_started = True | |||||
| text = text.replace("<think>", "> 💭 ") | |||||
| elif "</think>" in text: | |||||
| is_reasoning_started = False | |||||
| text = text.replace("</think>", "") + "\n\n" | |||||
| elif is_reasoning_started: | |||||
| text = text.replace("\n", "\n> ") | |||||
| assistant_prompt_message = AssistantPromptMessage(content=text) | assistant_prompt_message = AssistantPromptMessage(content=text) | ||||