lvkaokao commited on
Commit
c467002
1 Parent(s): 7638119
Files changed (1) hide show
  1. README.md +64 -0
README.md CHANGED
@@ -1,3 +1,67 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+
5
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
6
+ should probably proofread and complete it, then remove this comment. -->
7
+
8
+ This model is a fine-tuned model based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the open source dataset [Open-Orca/SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca).
9
+
10
+ ## Model date
11
+ Neural-chat-7b-v3 was trained on October 25, 2023.
12
+
13
+ ## Evaluation
14
+
15
+ We use the [Eleuther AI Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness/tree/master) to measure the metrics that are adopted by [open_llm_leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
16
+
17
+ | Model | Average ⬆️| ARC (25-s) ⬆️ | HellaSwag (10-s) ⬆️ | MMLU (5-s) ⬆️| TruthfulQA (MC) (0-s) ⬆️ |
18
+ | --- | --- | --- | --- | --- | --- |
19
+ |[mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) | 62.4 | 59.58 | 83.31 | 64.16 | 42.15 |
20
+ | **Ours** | **67.92** | 66.29 | 83.28 | 62.11 | 60.02 |
21
+
22
+
23
+ ## Training procedure
24
+
25
+ ### Training hyperparameters
26
+
27
+ The following hyperparameters were used during training:
28
+ - learning_rate: 1e-04
29
+ - train_batch_size: 1
30
+ - eval_batch_size: 2
31
+ - seed: 42
32
+ - distributed_type: multi-GPU
33
+ - num_devices: 4
34
+ - gradient_accumulation_steps: 16
35
+ - total_train_batch_size: 64
36
+ - total_eval_batch_size: 8
37
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
38
+ - lr_scheduler_type: linear
39
+ - lr_scheduler_warmup_ratio: 0.02
40
+ - num_epochs: 2.0
41
+
42
+ ## Inference with transformers
43
+
44
+ ```shell
45
+ import transformers
46
+ model = transformers.AutoModelForCausalLM.from_pretrained(
47
+ 'Intel/neural-chat-7b-v3'
48
+ )
49
+ ```
50
+
51
+ ## Ethical Considerations and Limitations
52
+ neural-chat-7b-v3 can produce factually incorrect output, and should not be relied on to produce factually accurate information. neural-chat-7b-v3 was trained on [Open-Orca/SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca) based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1). Because of the limitations of the pretrained model and the finetuning datasets, it is possible that this model could generate lewd, biased or otherwise offensive outputs.
53
+
54
+ Therefore, before deploying any applications of neural-chat-7b-v2, developers should perform safety testing.
55
+
56
+ ## Disclaimer
57
+
58
+ The license on this model does not constitute legal advice. We are not responsible for the actions of third parties who use this model. Please cosult an attorney before using this model for commercial purposes.
59
+
60
+ ## Organizations developing the model
61
+
62
+ The NeuralChat team with members from Intel/SATG/AIA/AIPT. Core team members: Kaokao Lv, Liang Lv, Chang Wang, Wenxin Zhang, Xuhui Ren, and Haihao Shen.
63
+
64
+ ## Useful links
65
+ * Intel Neural Compressor [link](https://github.com/intel/neural-compressor)
66
+ * Intel Extension for Transformers [link](https://github.com/intel/intel-extension-for-transformers)
67
+ * Intel Extension for PyTorch [link](https://github.com/intel/intel-extension-for-pytorch)