charanhu commited on
Commit
c056a0a
·
1 Parent(s): f55178e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +20 -13
app.py CHANGED
@@ -5,27 +5,34 @@ import torch
5
  tokenizer = AutoTokenizer.from_pretrained("TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T")
6
  model = AutoModelForCausalLM.from_pretrained("TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T")
7
 
8
- def generate_text(prompt, temperature, max_length, min_length):
9
- # Tokenize the prompt
10
- input_ids = tokenizer.encode(prompt, return_tensors="pt")
 
 
 
11
 
12
- # Generate text using the model
13
- output = model.generate(input_ids, max_length=max_length, min_length=min_length, temperature=temperature, num_return_sequences=1)
 
 
 
14
 
15
- # Decode the generated output
16
- generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
17
-
18
  return generated_text
19
 
20
- def chatbot_app(prompt, temperature, max_length, min_length):
21
- generated_text = generate_text(prompt, temperature, max_length, min_length)
22
  return generated_text
23
 
24
  iface = gr.Interface(
25
  fn=chatbot_app,
26
- inputs=["text", gr.Number(minimum=0.1, maximum=2.0, value=1.0, label="Temperature"),
27
- gr.Number(minimum=10, maximum=2048, value=10, label="Max Length"),
28
- gr.Number(minimum=1, maximum=2048, value=1, label="Min Length")],
 
 
 
 
29
  outputs="text",
30
  live=False,
31
  )
 
5
  tokenizer = AutoTokenizer.from_pretrained("TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T")
6
  model = AutoModelForCausalLM.from_pretrained("TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T")
7
 
8
+ def generate_text(system_prompt, user_prompt, temperature, max_length, min_length):
9
+ messages = [
10
+ {"role": "system", "content": system_prompt},
11
+ {"role": "user", "content": user_prompt},
12
+ ]
13
+ tokenized_chat = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
14
 
15
+ inputs = tokenizer(tokenized_chat, return_tensors="pt").to(model.device)
16
+
17
+ outputs = model.generate(**inputs, use_cache=True, max_length=max_length, min_length=min_length, temperature=temperature, num_return_sequences=1)
18
+
19
+ generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
20
 
 
 
 
21
  return generated_text
22
 
23
+ def chatbot_app(system_prompt, user_prompt, temperature, max_length, min_length):
24
+ generated_text = generate_text(system_prompt, user_prompt, temperature, max_length, min_length)
25
  return generated_text
26
 
27
  iface = gr.Interface(
28
  fn=chatbot_app,
29
+ inputs=[
30
+ "text",
31
+ "text", # Added a text box for the system prompt
32
+ gr.Number(minimum=0.1, maximum=2.0, value=1.0, label="Temperature"),
33
+ gr.Number(minimum=10, maximum=2048, value=10, label="Max Length"),
34
+ gr.Number(minimum=1, maximum=2048, value=1, label="Min Length"),
35
+ ],
36
  outputs="text",
37
  live=False,
38
  )