diff --git a/backend/retrieval_graph/prompts.py b/backend/retrieval_graph/prompts.py index e25333ae..847a3545 100644 --- a/backend/retrieval_graph/prompts.py +++ b/backend/retrieval_graph/prompts.py @@ -4,20 +4,24 @@ # fetch from langsmith ROUTER_SYSTEM_PROMPT = ( - hub.pull("chat-langchain-router-prompt").messages[0].prompt.template + hub.pull("langchain-ai/chat-langchain-router-prompt").messages[0].prompt.template ) GENERATE_QUERIES_SYSTEM_PROMPT = ( - hub.pull("chat-langchain-generate-queries-prompt").messages[0].prompt.template + hub.pull("langchain-ai/chat-langchain-generate-queries-prompt") + .messages[0] + .prompt.template ) MORE_INFO_SYSTEM_PROMPT = ( - hub.pull("chat-langchain-more-info-prompt").messages[0].prompt.template + hub.pull("langchain-ai/chat-langchain-more-info-prompt").messages[0].prompt.template ) RESEARCH_PLAN_SYSTEM_PROMPT = ( - hub.pull("chat-langchain-research-plan-prompt").messages[0].prompt.template + hub.pull("langchain-ai/chat-langchain-research-plan-prompt") + .messages[0] + .prompt.template ) GENERAL_SYSTEM_PROMPT = ( - hub.pull("chat-langchain-general-prompt").messages[0].prompt.template + hub.pull("langchain-ai/chat-langchain-general-prompt").messages[0].prompt.template ) RESPONSE_SYSTEM_PROMPT = ( - hub.pull("chat-langchain-response-prompt").messages[0].prompt.template + hub.pull("langchain-ai/chat-langchain-response-prompt").messages[0].prompt.template )