lmzheng commited on
Commit
d978a8d
1 Parent(s): b0aa428

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -7,12 +7,12 @@ license: llama2
7
 
8
  ## Model Details
9
 
10
- Vicuna is a chat assistant trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT.
11
 
12
  - **Developed by:** [LMSYS](https://lmsys.org/)
13
  - **Model type:** An auto-regressive language model based on the transformer architecture
14
  - **License:** Non-commercial license
15
- - **Finetuned from model:** [LLaMA 2](https://arxiv.org/abs/2307.09288)
16
 
17
  ### Model Sources
18
 
@@ -33,7 +33,7 @@ The primary intended users of the model are researchers and hobbyists in natural
33
 
34
  ## Training Details
35
 
36
- Vicuna v1.5 (16k) is fine-tuned from LLaMA with supervised instruction fine-tuning and linear RoPE scaling.
37
  The training data is around 125K conversations collected from ShareGPT.com. These conversations are packed into sequences that contain 16K tokens each.
38
  See more details in the "Training Details of Vicuna Models" section in the appendix of this [paper](https://arxiv.org/pdf/2306.05685.pdf).
39
 
 
7
 
8
  ## Model Details
9
 
10
+ Vicuna is a chat assistant trained by fine-tuning Llama 2 on user-shared conversations collected from ShareGPT.
11
 
12
  - **Developed by:** [LMSYS](https://lmsys.org/)
13
  - **Model type:** An auto-regressive language model based on the transformer architecture
14
  - **License:** Non-commercial license
15
+ - **Finetuned from model:** [Llama 2](https://arxiv.org/abs/2307.09288)
16
 
17
  ### Model Sources
18
 
 
33
 
34
  ## Training Details
35
 
36
+ Vicuna v1.5 (16k) is fine-tuned from Llama 2 with supervised instruction fine-tuning and linear RoPE scaling.
37
  The training data is around 125K conversations collected from ShareGPT.com. These conversations are packed into sequences that contain 16K tokens each.
38
  See more details in the "Training Details of Vicuna Models" section in the appendix of this [paper](https://arxiv.org/pdf/2306.05685.pdf).
39