Text Generation
Transformers
Safetensors
llama
text-generation-inference
Inference Endpoints
mfromm commited on
Commit
960c6d2
·
verified ·
1 Parent(s): ba0a921

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -101,7 +101,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
101
 
102
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
103
 
104
- model_name = "openGPT-X/Teuken-7B-instruct-v0.4"
105
  model = AutoModelForCausalLM.from_pretrained(
106
  model_name,
107
  trust_remote_code=True,
 
101
 
102
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
103
 
104
+ model_name = "openGPT-X/Teuken-7B-instruct-research-v0.4"
105
  model = AutoModelForCausalLM.from_pretrained(
106
  model_name,
107
  trust_remote_code=True,