Update README.md
Browse files
README.md
CHANGED
@@ -104,7 +104,7 @@ print(tokenizer.decode(response, skip_special_tokens=True))
|
|
104 |
```
|
105 |
|
106 |
### run retrieval to get top-n chunks as context
|
107 |
-
This can be applied to the scenario when the document is very long, so that it is necessary to run retrieval. Here, we use our [Dragon-multiturn](https://huggingface.co/nvidia/dragon-multiturn-query-encoder) retriever which can handle conversatinoal query. In addition, we provide a few [documents]() for users to play with.
|
108 |
|
109 |
```python
|
110 |
from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModel
|
|
|
104 |
```
|
105 |
|
106 |
### run retrieval to get top-n chunks as context
|
107 |
+
This can be applied to the scenario when the document is very long, so that it is necessary to run retrieval. Here, we use our [Dragon-multiturn](https://huggingface.co/nvidia/dragon-multiturn-query-encoder) retriever which can handle conversatinoal query. In addition, we provide a few [documents](https://huggingface.co/nvidia/ChatQA-1.5-8B/tree/main/docs) for users to play with.
|
108 |
|
109 |
```python
|
110 |
from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModel
|