chainyo commited on
Commit
b3349e0
·
1 Parent(s): aac22d6

add 13B model link

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -168,6 +168,8 @@ We benchmarked our model on the following tasks: [BoolQ](https://huggingface.co/
168
  | LoRA LLaMA 7B | 63.9 | 51.3 | 48.9 | 31.4 | 8bit | 0.65 seconds |
169
  | LoRA LLaMA 13B | 70 | 63.93 | 51.6 | 50.4 | 8bit | 1.2 seconds |
170
 
 
 
171
  Overall our LoRA model is less performant than the original model from Meta, if we compare the results from the [original paper](https://arxiv.org/pdf/2302.13971.pdf).
172
 
173
  The performance degradation is due to the fact we load the model in 8bit and we use the adapters from the LoRA training.
 
168
  | LoRA LLaMA 7B | 63.9 | 51.3 | 48.9 | 31.4 | 8bit | 0.65 seconds |
169
  | LoRA LLaMA 13B | 70 | 63.93 | 51.6 | 50.4 | 8bit | 1.2 seconds |
170
 
171
+ __Link to the 13B model:__ [wordcab/llama-natural-instructions-13b](https://huggingface.co/wordcab/llama-natural-instructions-13b)
172
+
173
  Overall our LoRA model is less performant than the original model from Meta, if we compare the results from the [original paper](https://arxiv.org/pdf/2302.13971.pdf).
174
 
175
  The performance degradation is due to the fact we load the model in 8bit and we use the adapters from the LoRA training.