keyfan commited on
Commit
36cb93e
1 Parent(s): c85cd30

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -0
README.md CHANGED
@@ -1,3 +1,37 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ datasets:
4
+ - QingyiSi/Alpaca-CoT
5
+ language:
6
+ - zh
7
+ - en
8
  ---
9
+
10
+ This is a beta release of a QLoRa adapter model to [Falcon-40b](https://huggingface.co/tiiuae/falcon-40b).
11
+ Please read the instruction carefully before downloading the model.
12
+
13
+ Though Falcon is not specifically trained on Chinese corpus, it exhibits strong performance in Chinese Language Understanding in our experiment. We would like to explore out of curiosity whether a
14
+ small amount of Chinese instruction data can push it further and make it better at speaking.
15
+
16
+ The LoRa model is trained with the [QLoRa code](https://github.com/artidoro/qlora) on a subset of bilingual instruction data from [Alpaca-CoT dataset](https://huggingface.co/datasets/QingyiSi/Alpaca-CoT) for a mere 5k steps.
17
+ The finetune model is not as good as the carefully continue-trained-and-finetuned LLaMA-models such as [OpenBuddy](https://huggingface.co/OpenBuddy) and [Ziya](https://huggingface.co/IDEA-CCNL/Ziya-LLaMA-13B-v1) in Chinese generation,
18
+ still it quickly adapts to the new langauge and generate superisingly good result. We call for more research on applying Falcon-40b to the Chinese domain.
19
+
20
+ ## Evalutions
21
+
22
+ We evaluate on two Chinese language understanding benchmarks, [C-Eval](https://cevalbenchmark.com/) and Gaokao subset of [AGIEval](https://github.com/microsoft/AGIEval).
23
+
24
+ * C-Eval made breaking change in 2023/06/08 from few-shot to zero-shot,
25
+
26
+ Result on C-Eval test set with 5-shot and no CoT
27
+
28
+ | Average | Avg(Hard) | STEM | Social Science | Humanities | Others |
29
+ | - | - | - | - | - | - |
30
+ | 40.4 | 30.1 | 35.8 | 47.6 | 42.0 | 40.6 |
31
+
32
+
33
+ Result on GaoKao subset of C-Eval with 0-shot
34
+
35
+ | Average | GK-chinese | GK-English | GK-geography | GK-history | GK-biology | GK-chemistry | GK-physics | GK-mathqa | GK-mathcloze
36
+ | - | - | - | - | - | - | - | - | - | - |
37
+ | 33.6 | 26.4 | 69.0 | 46.7 | 47.8 | 27.1 | 32.4 | 24.5 | 26.8 | 1.7 |