avoroshilov commited on
Commit
31c9d54
·
verified ·
1 Parent(s): ba6b752

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -29,7 +29,7 @@ model_name = "avoroshilov/DeepSeek-R1-Distill-Qwen-14B-GPTQ_4bit-128g"
29
  tokenizer = AutoTokenizer.from_pretrained(model_name)
30
  quantized_model = AutoModelForCausalLM.from_pretrained(model_name, device_map='cuda')
31
 
32
- chat = [{"role": "user", "content": "Why is the grass green?"},]
33
  question_tokens = tokenizer.apply_chat_template(chat, add_generation_prompt=True, return_tensors="pt").to(quantized_model.device)
34
  answer_tokens = quantized_model.generate(question_tokens, generation_config=GenerationConfig(max_length=2048, ))[0]
35
 
 
29
  tokenizer = AutoTokenizer.from_pretrained(model_name)
30
  quantized_model = AutoModelForCausalLM.from_pretrained(model_name, device_map='cuda')
31
 
32
+ chat = [{"role": "user", "content": "Why is grass green?"},]
33
  question_tokens = tokenizer.apply_chat_template(chat, add_generation_prompt=True, return_tensors="pt").to(quantized_model.device)
34
  answer_tokens = quantized_model.generate(question_tokens, generation_config=GenerationConfig(max_length=2048, ))[0]
35