Update README.md
Browse files
README.md
CHANGED
@@ -107,7 +107,7 @@ You may also construct the pipeline from the loaded model and tokenizer yourself
|
|
107 |
```python
|
108 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
109 |
|
110 |
-
model_name = "
|
111 |
# Important: The prompt needs to be in the same format the model was trained with.
|
112 |
# You can find an example prompt in the experiment logs.
|
113 |
prompt = "<|prompt|>How are you?<|endoftext|><|answer|>"
|
|
|
107 |
```python
|
108 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
109 |
|
110 |
+
model_name = "h2oai/h2ogpt-gm-oasst1-en-xgen-7b-8k" # either local folder or huggingface model name
|
111 |
# Important: The prompt needs to be in the same format the model was trained with.
|
112 |
# You can find an example prompt in the experiment logs.
|
113 |
prompt = "<|prompt|>How are you?<|endoftext|><|answer|>"
|