Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -5,27 +5,34 @@ import torch
|
|
5 |
tokenizer = AutoTokenizer.from_pretrained("TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T")
|
6 |
model = AutoModelForCausalLM.from_pretrained("TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T")
|
7 |
|
8 |
-
def generate_text(
|
9 |
-
|
10 |
-
|
|
|
|
|
|
|
11 |
|
12 |
-
|
13 |
-
|
|
|
|
|
|
|
14 |
|
15 |
-
# Decode the generated output
|
16 |
-
generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
|
17 |
-
|
18 |
return generated_text
|
19 |
|
20 |
-
def chatbot_app(
|
21 |
-
generated_text = generate_text(
|
22 |
return generated_text
|
23 |
|
24 |
iface = gr.Interface(
|
25 |
fn=chatbot_app,
|
26 |
-
inputs=[
|
27 |
-
|
28 |
-
|
|
|
|
|
|
|
|
|
29 |
outputs="text",
|
30 |
live=False,
|
31 |
)
|
|
|
5 |
tokenizer = AutoTokenizer.from_pretrained("TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T")
|
6 |
model = AutoModelForCausalLM.from_pretrained("TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T")
|
7 |
|
8 |
+
def generate_text(system_prompt, user_prompt, temperature, max_length, min_length):
|
9 |
+
messages = [
|
10 |
+
{"role": "system", "content": system_prompt},
|
11 |
+
{"role": "user", "content": user_prompt},
|
12 |
+
]
|
13 |
+
tokenized_chat = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
|
14 |
|
15 |
+
inputs = tokenizer(tokenized_chat, return_tensors="pt").to(model.device)
|
16 |
+
|
17 |
+
outputs = model.generate(**inputs, use_cache=True, max_length=max_length, min_length=min_length, temperature=temperature, num_return_sequences=1)
|
18 |
+
|
19 |
+
generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
20 |
|
|
|
|
|
|
|
21 |
return generated_text
|
22 |
|
23 |
+
def chatbot_app(system_prompt, user_prompt, temperature, max_length, min_length):
|
24 |
+
generated_text = generate_text(system_prompt, user_prompt, temperature, max_length, min_length)
|
25 |
return generated_text
|
26 |
|
27 |
iface = gr.Interface(
|
28 |
fn=chatbot_app,
|
29 |
+
inputs=[
|
30 |
+
"text",
|
31 |
+
"text", # Added a text box for the system prompt
|
32 |
+
gr.Number(minimum=0.1, maximum=2.0, value=1.0, label="Temperature"),
|
33 |
+
gr.Number(minimum=10, maximum=2048, value=10, label="Max Length"),
|
34 |
+
gr.Number(minimum=1, maximum=2048, value=1, label="Min Length"),
|
35 |
+
],
|
36 |
outputs="text",
|
37 |
live=False,
|
38 |
)
|