|
|
|
|
|
|
|
|
def check(self): |
|
|
def check(self): |
|
|
super().check() |
|
|
super().check() |
|
|
|
|
|
|
|
|
def get_prompt(self): |
|
|
|
|
|
|
|
|
def get_prompt(self, conv): |
|
|
self.prompt = """ |
|
|
self.prompt = """ |
|
|
You are an expert at query expansion to generate a paraphrasing of a question. |
|
|
You are an expert at query expansion to generate a paraphrasing of a question. |
|
|
I can't retrieval relevant information from the knowledge base by using user's question directly. |
|
|
I can't retrieval relevant information from the knowledge base by using user's question directly. |
|
|
|
|
|
|
|
|
And return 5 versions of question and one is from translation. |
|
|
And return 5 versions of question and one is from translation. |
|
|
Just list the question. No other words are needed. |
|
|
Just list the question. No other words are needed. |
|
|
""" |
|
|
""" |
|
|
|
|
|
return f""" |
|
|
|
|
|
Role: A helpful assistant |
|
|
|
|
|
Task: Generate a full user question that would follow the conversation. |
|
|
|
|
|
Requirements & Restrictions: |
|
|
|
|
|
- Text generated MUST be in the same language of the original user's question. |
|
|
|
|
|
- If the user's latest question is completely, don't do anything, just return the original question. |
|
|
|
|
|
- DON'T generate anything except a refined question. |
|
|
|
|
|
|
|
|
|
|
|
###################### |
|
|
|
|
|
-Examples- |
|
|
|
|
|
###################### |
|
|
|
|
|
# Example 1 |
|
|
|
|
|
## Conversation |
|
|
|
|
|
USER: What is the name of Donald Trump's father? |
|
|
|
|
|
ASSISTANT: Fred Trump. |
|
|
|
|
|
USER: And his mother? |
|
|
|
|
|
############### |
|
|
|
|
|
Output: What's the name of Donald Trump's mother? |
|
|
|
|
|
------------ |
|
|
|
|
|
# Example 2 |
|
|
|
|
|
## Conversation |
|
|
|
|
|
USER: What is the name of Donald Trump's father? |
|
|
|
|
|
ASSISTANT: Fred Trump. |
|
|
|
|
|
USER: And his mother? |
|
|
|
|
|
ASSISTANT: Mary Trump. |
|
|
|
|
|
User: What's her full name? |
|
|
|
|
|
############### |
|
|
|
|
|
Output: What's the full name of Donald Trump's mother Mary Trump? |
|
|
|
|
|
###################### |
|
|
|
|
|
# Real Data |
|
|
|
|
|
## Conversation |
|
|
|
|
|
{conv} |
|
|
|
|
|
############### |
|
|
|
|
|
""" |
|
|
return self.prompt |
|
|
return self.prompt |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
self._loop = 0 |
|
|
self._loop = 0 |
|
|
raise Exception("Sorry! Nothing relevant found.") |
|
|
raise Exception("Sorry! Nothing relevant found.") |
|
|
self._loop += 1 |
|
|
self._loop += 1 |
|
|
q = "Question: " |
|
|
|
|
|
for r, c in self._canvas.history[::-1]: |
|
|
|
|
|
if r == "user": |
|
|
|
|
|
q += c |
|
|
|
|
|
break |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
conv = self._canvas.get_history(4) |
|
|
|
|
|
conv = "\n".join(conv) |
|
|
|
|
|
|
|
|
chat_mdl = LLMBundle(self._canvas.get_tenant_id(), LLMType.CHAT, self._param.llm_id) |
|
|
chat_mdl = LLMBundle(self._canvas.get_tenant_id(), LLMType.CHAT, self._param.llm_id) |
|
|
ans = chat_mdl.chat(self._param.get_prompt(), [{"role": "user", "content": q}], |
|
|
|
|
|
|
|
|
ans = chat_mdl.chat(self._param.get_prompt(conv), [{"role": "user", "content": "Output: "}], |
|
|
self._param.gen_conf()) |
|
|
self._param.gen_conf()) |
|
|
self._canvas.history.pop() |
|
|
self._canvas.history.pop() |
|
|
self._canvas.history.append(("user", ans)) |
|
|
self._canvas.history.append(("user", ans)) |