Update rag_langchain.py
Browse files- rag_langchain.py +3 -3
rag_langchain.py
CHANGED
@@ -27,7 +27,7 @@ class LangChainRAG(BaseRAG):
|
|
27 |
template = os.environ["LLM_TEMPLATE"])
|
28 |
RAG_CHAIN_PROMPT = PromptTemplate(
|
29 |
input_variables = ["context_str", "query_str"],
|
30 |
-
template = os.environ["
|
31 |
|
32 |
def load_documents(self):
|
33 |
docs = []
|
@@ -114,7 +114,7 @@ class LangChainRAG(BaseRAG):
|
|
114 |
)
|
115 |
|
116 |
with get_openai_callback() as callback:
|
117 |
-
completion = llm_chain.generate([{"
|
118 |
|
119 |
return completion, llm_chain, callback
|
120 |
|
@@ -130,6 +130,6 @@ class LangChainRAG(BaseRAG):
|
|
130 |
)
|
131 |
|
132 |
with get_openai_callback() as callback:
|
133 |
-
completion = rag_chain({"
|
134 |
|
135 |
return completion, rag_chain, callback
|
|
|
27 |
template = os.environ["LLM_TEMPLATE"])
|
28 |
RAG_CHAIN_PROMPT = PromptTemplate(
|
29 |
input_variables = ["context_str", "query_str"],
|
30 |
+
template = os.environ["RAG_TEMPLATE_2"])
|
31 |
|
32 |
def load_documents(self):
|
33 |
docs = []
|
|
|
114 |
)
|
115 |
|
116 |
with get_openai_callback() as callback:
|
117 |
+
completion = llm_chain.generate([{"query_str": prompt}])
|
118 |
|
119 |
return completion, llm_chain, callback
|
120 |
|
|
|
130 |
)
|
131 |
|
132 |
with get_openai_callback() as callback:
|
133 |
+
completion = rag_chain({"query_str": prompt})
|
134 |
|
135 |
return completion, rag_chain, callback
|