Update README.md
Browse files
README.md
CHANGED
@@ -29,7 +29,7 @@ model_name = "avoroshilov/DeepSeek-R1-Distill-Qwen-14B-GPTQ_4bit-128g"
|
|
29 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
30 |
quantized_model = AutoModelForCausalLM.from_pretrained(model_name, device_map='cuda')
|
31 |
|
32 |
-
chat = [{"role": "user", "content": "Why is
|
33 |
question_tokens = tokenizer.apply_chat_template(chat, add_generation_prompt=True, return_tensors="pt").to(quantized_model.device)
|
34 |
answer_tokens = quantized_model.generate(question_tokens, generation_config=GenerationConfig(max_length=2048, ))[0]
|
35 |
|
|
|
29 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
30 |
quantized_model = AutoModelForCausalLM.from_pretrained(model_name, device_map='cuda')
|
31 |
|
32 |
+
chat = [{"role": "user", "content": "Why is grass green?"},]
|
33 |
question_tokens = tokenizer.apply_chat_template(chat, add_generation_prompt=True, return_tensors="pt").to(quantized_model.device)
|
34 |
answer_tokens = quantized_model.generate(question_tokens, generation_config=GenerationConfig(max_length=2048, ))[0]
|
35 |
|