zihanliu commited on
Commit
8f93ac5
1 Parent(s): 3c7506a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -104,7 +104,7 @@ print(tokenizer.decode(response, skip_special_tokens=True))
104
  ```
105
 
106
  ### run retrieval to get top-n chunks as context
107
- This can be applied to the scenario when the document is very long, so that it is necessary to run retrieval. Here, we use our [Dragon-multiturn](https://huggingface.co/nvidia/dragon-multiturn-query-encoder) retriever which can handle conversatinoal query. In addition, we provide a few [documents]() for users to play with.
108
 
109
  ```python
110
  from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModel
 
104
  ```
105
 
106
  ### run retrieval to get top-n chunks as context
107
+ This can be applied to the scenario when the document is very long, so that it is necessary to run retrieval. Here, we use our [Dragon-multiturn](https://huggingface.co/nvidia/dragon-multiturn-query-encoder) retriever which can handle conversatinoal query. In addition, we provide a few [documents](https://huggingface.co/nvidia/ChatQA-1.5-8B/tree/main/docs) for users to play with.
108
 
109
  ```python
110
  from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModel