diff --git a/api/db/services/dialog_service.py b/api/db/services/dialog_service.py index 33d4f3695..73e5a62d4 100644 --- a/api/db/services/dialog_service.py +++ b/api/db/services/dialog_service.py @@ -874,18 +874,43 @@ def reasoning(chunk_info: dict, question: str, chat_mdl: LLMBundle, embd_mdl: LL f"Then, the system will search and analyze relevant content, then provide you with helpful information in the format {BEGIN_SEARCH_RESULT} ...search results... {END_SEARCH_RESULT}.\n\n" f"You can repeat the search process multiple times if necessary. The maximum number of search attempts is limited to {MAX_SEARCH_LIMIT}.\n\n" "Once you have all the information you need, continue your reasoning.\n\n" - "-- Example --\n" - "Question: \"Find the minimum number of vertices in a Steiner tree that includes all specified vertices in a given tree.\"\n" + "-- Example 1 --\n" ######################################## + "Question: \"Are both the directors of Jaws and Casino Royale from the same country?\"\n" "Assistant:\n" - " - I need to understand what a Steiner tree is.\n\n" - f" {BEGIN_SEARCH_QUERY}What's Steiner tree{END_SEARCH_QUERY}\n\n" - f" {BEGIN_SEARCH_RESULT}\n(System returns processed information from relevant web pages)\n{END_SEARCH_RESULT}\n\n" - "User:\nContinues reasoning with the new information.\n\n" + f" {BEGIN_SEARCH_QUERY}Who is the director of Jaws?{END_SEARCH_QUERY}\n\n" + "User:\n" + f" {BEGIN_SEARCH_RESULT}\nThe director of Jaws is Steven Spielberg...\n{END_SEARCH_RESULT}\n\n" + "Continues reasoning with the new information.\n" "Assistant:\n" - " - I need to understand what the difference between minimum number of vertices and edges in the Steiner tree is.\n\n" - f" {BEGIN_SEARCH_QUERY}What's the difference between minimum number of vertices and edges in the Steiner tree{END_SEARCH_QUERY}\n\n" - f" {BEGIN_SEARCH_RESULT}\n(System returns processed information from relevant web pages)\n{END_SEARCH_RESULT}\n\n" - "User:\nContinues reasoning with the new information...\n\n" + f" {BEGIN_SEARCH_QUERY}Where is Steven Spielberg from?{END_SEARCH_QUERY}\n\n" + "User:\n" + f" {BEGIN_SEARCH_RESULT}\nSteven Allan Spielberg is an American filmmaker...\n{END_SEARCH_RESULT}\n\n" + "Continues reasoning with the new information...\n\n" + "Assistant:\n" + f" {BEGIN_SEARCH_QUERY}Who is the director of Casino Royale?{END_SEARCH_QUERY}\n\n" + "User:\n" + f" {BEGIN_SEARCH_RESULT}\nCasino Royale is a 2006 spy film directed by Martin Campbell...\n{END_SEARCH_RESULT}\n\n" + "Continues reasoning with the new information...\n\n" + "Assistant:\n" + f" {BEGIN_SEARCH_QUERY}Where is Martin Campbell from?{END_SEARCH_QUERY}\n\n" + "User:\n" + f" {BEGIN_SEARCH_RESULT}\nMartin Campbell (born 24 October 1943) is a New Zealand film and television director...\n{END_SEARCH_RESULT}\n\n" + "Continues reasoning with the new information...\n\n" + "Assistant:\nIt's enough to answer the question\n" + + "-- Example 2 --\n" ######################################### + "Question: \"When was the founder of craigslist born?\"\n" + "Assistant:\n" + f" {BEGIN_SEARCH_QUERY}Who was the founder of craigslist?{END_SEARCH_QUERY}\n\n" + "User:\n" + f" {BEGIN_SEARCH_RESULT}\nCraigslist was founded by Craig Newmark...\n{END_SEARCH_RESULT}\n\n" + "Continues reasoning with the new information.\n" + "Assistant:\n" + f" {BEGIN_SEARCH_QUERY} When was Craig Newmark born?{END_SEARCH_QUERY}\n\n" + "User:\n" + f" {BEGIN_SEARCH_RESULT}\nCraig Newmark was born on December 6, 1952...\n{END_SEARCH_RESULT}\n\n" + "Continues reasoning with the new information...\n\n" + "Assistant:\nIt's enough to answer the question\n" "**Remember**:\n" f"- You have a dataset to search, so you just provide a proper search query.\n" f"- Use {BEGIN_SEARCH_QUERY} to request a dataset search and end with {END_SEARCH_QUERY}.\n" @@ -934,7 +959,7 @@ def reasoning(chunk_info: dict, question: str, chat_mdl: LLMBundle, embd_mdl: LL """ executed_search_queries = [] - msg_hisotry = [{"role": "user", "content": f'Question:\n{question}\n\n'}] + msg_hisotry = [{"role": "user", "content": f'Question:\"{question}\"\n'}] all_reasoning_steps = [] think = "" for ii in range(MAX_SEARCH_LIMIT + 1): @@ -948,6 +973,8 @@ def reasoning(chunk_info: dict, question: str, chat_mdl: LLMBundle, embd_mdl: LL query_think = "" if msg_hisotry[-1]["role"] != "user": msg_hisotry.append({"role": "user", "content": "Continues reasoning with the new information.\n"}) + else: + msg_hisotry[-1]["content"] += "\n\nContinues reasoning with the new information.\n" for ans in chat_mdl.chat_streamly(reason_prompt, msg_hisotry, {"temperature": 0.7}): ans = re.sub(r".*", "", ans, flags=re.DOTALL) if not ans: @@ -957,7 +984,6 @@ def reasoning(chunk_info: dict, question: str, chat_mdl: LLMBundle, embd_mdl: LL think += rm_query_tags(query_think) all_reasoning_steps.append(query_think) - msg_hisotry.append({"role": "assistant", "content": query_think}) queries = extract_between(query_think, BEGIN_SEARCH_QUERY, END_SEARCH_QUERY) if not queries: if ii > 0: @@ -966,6 +992,7 @@ def reasoning(chunk_info: dict, question: str, chat_mdl: LLMBundle, embd_mdl: LL for search_query in queries: logging.info(f"[THINK]Query: {ii}. {search_query}") + msg_hisotry.append({"role": "assistant", "content": search_query}) think += f"\n\n> {ii+1}. {search_query}\n\n" yield {"answer": think + "", "reference": {}, "audio_binary": None} @@ -975,7 +1002,7 @@ def reasoning(chunk_info: dict, question: str, chat_mdl: LLMBundle, embd_mdl: LL summary_think = f"\n{BEGIN_SEARCH_RESULT}\nYou have searched this query. Please refer to previous results.\n{END_SEARCH_RESULT}\n" yield {"answer": think + summary_think + "", "reference": {}, "audio_binary": None} all_reasoning_steps.append(summary_think) - msg_hisotry.append({"role": "assistant", "content": summary_think}) + msg_hisotry.append({"role": "user", "content": summary_think}) think += summary_think continue @@ -1034,7 +1061,7 @@ def reasoning(chunk_info: dict, question: str, chat_mdl: LLMBundle, embd_mdl: LL all_reasoning_steps.append(summary_think) msg_hisotry.append( - {"role": "assistant", "content": f"\n\n{BEGIN_SEARCH_RESULT}{summary_think}{END_SEARCH_RESULT}\n\n"}) + {"role": "user", "content": f"\n\n{BEGIN_SEARCH_RESULT}{summary_think}{END_SEARCH_RESULT}\n\n"}) think += rm_result_tags(summary_think) logging.info(f"[THINK]Summary: {ii}. {summary_think}")