Update app.py
Browse files
app.py
CHANGED
@@ -85,7 +85,7 @@ MODEL_NAME = "gpt-3.5-turbo-16k"
|
|
85 |
|
86 |
#HuggingFace--------------------------------
|
87 |
#repo_id = "meta-llama/Llama-2-13b-chat-hf"
|
88 |
-
repo_id = "HuggingFaceH4/zephyr-7b-alpha"
|
89 |
#repo_id = "meta-llama/Llama-2-70b-chat-hf"
|
90 |
#repo_id = "tiiuae/falcon-40b"
|
91 |
#repo_id = "Vicuna-33b"
|
@@ -304,9 +304,9 @@ def invoke (prompt, history, rag_option, openai_api_key, temperature=0.9, max_n
|
|
304 |
#LLM auswählen (OpenAI oder HF)
|
305 |
###########################
|
306 |
#Anfrage an OpenAI ----------------------------
|
307 |
-
|
308 |
#oder an Hugging Face --------------------------
|
309 |
-
llm = HuggingFaceHub(repo_id=repo_id, model_kwargs={"temperature": 0.5, "max_length": 64})
|
310 |
#llm = HuggingFaceHub(url_??? = "https://wdgsjd6zf201mufn.us-east-1.aws.endpoints.huggingface.cloud", model_kwargs={"temperature": 0.5, "max_length": 64})
|
311 |
#llm = HuggingFaceTextGenInference( inference_server_url="http://localhost:8010/", max_new_tokens=max_new_tokens,top_k=10,top_p=top_p,typical_p=0.95,temperature=temperature,repetition_penalty=repetition_penalty,)
|
312 |
|
|
|
85 |
|
86 |
#HuggingFace--------------------------------
|
87 |
#repo_id = "meta-llama/Llama-2-13b-chat-hf"
|
88 |
+
repo_id = "HuggingFaceH4/zephyr-7b-alpha" #das Modell ist echt gut!!!
|
89 |
#repo_id = "meta-llama/Llama-2-70b-chat-hf"
|
90 |
#repo_id = "tiiuae/falcon-40b"
|
91 |
#repo_id = "Vicuna-33b"
|
|
|
304 |
#LLM auswählen (OpenAI oder HF)
|
305 |
###########################
|
306 |
#Anfrage an OpenAI ----------------------------
|
307 |
+
llm = ChatOpenAI(model_name = MODEL_NAME, openai_api_key = openai_api_key, temperature=temperature)#, top_p = top_p)
|
308 |
#oder an Hugging Face --------------------------
|
309 |
+
#llm = HuggingFaceHub(repo_id=repo_id, model_kwargs={"temperature": 0.5, "max_length": 64})
|
310 |
#llm = HuggingFaceHub(url_??? = "https://wdgsjd6zf201mufn.us-east-1.aws.endpoints.huggingface.cloud", model_kwargs={"temperature": 0.5, "max_length": 64})
|
311 |
#llm = HuggingFaceTextGenInference( inference_server_url="http://localhost:8010/", max_new_tokens=max_new_tokens,top_k=10,top_p=top_p,typical_p=0.95,temperature=temperature,repetition_penalty=repetition_penalty,)
|
312 |
|