bleysg commited on
Commit
4a7d390
1 Parent(s): 6d9fef1

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +108 -0
README.md ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - en
5
+ library_name: transformers
6
+ pipeline_tag: text-generation
7
+ datasets:
8
+ - Open-Orca/OpenOrca
9
+ ---
10
+
11
+ <p><h1>🐋 The First OpenOrca Model Preview! 🐋</h1></p>
12
+
13
+ # OpenOrca_Preview1-200k-GPT4_LLaMA-13B
14
+
15
+ We have used our own [OpenOrca dataset](https://huggingface.co/datasets/Open-Orca/OpenOrca) to fine-tune LLaMA-13B.
16
+ This dataset is our attempt to reproduce the dataset generated for Microsoft Research's [Orca Paper](https://arxiv.org/abs/2306.02707).
17
+
18
+ We have trained on less than 6% of our data, just to give a preview of what is possible while we further refine our dataset!
19
+ We trained a refined selection of 200k GPT-4 entries from OpenOrca.
20
+ We have filtered our GPT-4 augmentations to remove statements like, "As an AI language model..." and other responses which have been shown to harm model reasoning capabilities. Further details on our dataset curation practices will be forthcoming with our full model releases.
21
+
22
+ This release highlights that even a small portion of our training data can produce state of the art results in this model class with training costs <$200 in total.
23
+
24
+ We are in-process with training more models, so keep a look out on our org for releases coming soon with exciting partners.
25
+
26
+ We will also give sneak-peak announcements on our Discord, which you can find here:
27
+
28
+ https://AlignmentLab.ai
29
+
30
+
31
+ # Evaluation
32
+
33
+ We have evaluated OpenOrca_Preview1-200k-GPT4_LLaMA-13B on hard reasoning tasks from BigBench-Hard and AGIEval as outlined in the Orca paper.
34
+
35
+ Our average performance for BigBench-Hard: 0.3753
36
+
37
+ Average for AGIEval: 0.3638
38
+
39
+ In the Orca paper, they measured their score relative to Vicuna on these evals.
40
+ We've done the same and have found our score averages to ~60% of the total improvement that was shown in the Orca paper.
41
+
42
+ So we got 60% of the improvement with 6% of the data!
43
+
44
+ ## BigBench-Hard Performance
45
+ ![OpenOrca Preview1 BigBench-Hard Performance](https://huggingface.co/Open-Orca/OpenOrca_Preview1-200k-GPT4_LLaMA-13B/resolve/main/OO_Preview1_BigBenchHard.png "BigBench-Hard Performance")
46
+
47
+ ## AGIEval Performance
48
+ ![OpenOrca Preview1 AGIEval Performance](https://huggingface.co/Open-Orca/OpenOrca_Preview1-200k-GPT4_LLaMA-13B/resolve/main/OO_Preview1_AGIEval.png "AGIEval Performance")
49
+
50
+ We will report our results on [HuggingFaceH4 Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) Evals once we receive them.
51
+
52
+
53
+ # Dataset
54
+
55
+ We used a small (6%, 200k) subset of our data from OpenOrca, which aims to reproduce the Orca Research Paper dataset.
56
+
57
+ As this release is intended as a preview, please await our full releases for further details on the training data.
58
+
59
+
60
+ # Training
61
+
62
+ We trained with 8x A100-80G GPUs for 15 hours. Commodity cost was < $200.
63
+
64
+ We trained for 4 epochs and selected a snapshot at 3 epochs for peak performance.
65
+
66
+ Please await our full releases for further training details.
67
+
68
+
69
+ # Citation
70
+
71
+ ```bibtex
72
+ @software{OpenOrca_Preview1,
73
+ title = {OpenOrca_Preview1: A LLaMA-13B Model Fine-tuned on Small Portion of OpenOrcaV1 Dataset},
74
+ author = {Wing Lian and Bleys Goodson and Eugene Pentland and Austin Cook and "NanoBit" and "Teknium"},
75
+ year = {2023},
76
+ publisher = {HuggingFace},
77
+ journal = {HuggingFace repository},
78
+ howpublished = {\url{https://https://huggingface.co/Open-Orca/OpenOrca_Preview1-200k-GPT4_LLaMA-13B},
79
+ }
80
+ ```
81
+ ```bibtex
82
+ @misc{mukherjee2023orca,
83
+ title={Orca: Progressive Learning from Complex Explanation Traces of GPT-4},
84
+ author={Subhabrata Mukherjee and Arindam Mitra and Ganesh Jawahar and Sahaj Agarwal and Hamid Palangi and Ahmed Awadallah},
85
+ year={2023},
86
+ eprint={2306.02707},
87
+ archivePrefix={arXiv},
88
+ primaryClass={cs.CL}
89
+ }
90
+ ```
91
+ ```bibtex
92
+ @misc{longpre2023flan,
93
+ title={The Flan Collection: Designing Data and Methods for Effective Instruction Tuning},
94
+ author={Shayne Longpre and Le Hou and Tu Vu and Albert Webson and Hyung Won Chung and Yi Tay and Denny Zhou and Quoc V. Le and Barret Zoph and Jason Wei and Adam Roberts},
95
+ year={2023},
96
+ eprint={2301.13688},
97
+ archivePrefix={arXiv},
98
+ primaryClass={cs.AI}
99
+ }
100
+ ```
101
+ ```bibtex
102
+ @software{touvron2023llama,
103
+ title={LLaMA: Open and Efficient Foundation Language Models},
104
+ author={Touvron, Hugo and Lavril, Thibaut and Izacard, Gautier and Martinet, Xavier and Lachaux, Marie-Anne and Lacroix, Timoth{\'e}e and Rozi{\`e}re, Baptiste and Goyal, Naman and Hambro, Eric and Azhar, Faisal and Rodriguez, Aurelien and Joulin, Armand and Grave, Edouard and Lample, Guillaume},
105
+ journal={arXiv preprint arXiv:2302.13971},
106
+ year={2023}
107
+ }
108
+ ```