barius commited on
Commit
98758b1
1 Parent(s): 9bccd63

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +80 -0
README.md CHANGED
@@ -1,3 +1,83 @@
1
  ---
2
  license: gpl-3.0
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: gpl-3.0
3
+ tags:
4
+ - text2text-generation
5
+ pipeline_tag: text2text-generation
6
+ language:
7
+ - zh
8
+ - en
9
  ---
10
+
11
+ Considering LLaMA's license constraints, the model is for research and learning only.
12
+ Please strictly respect LLaMA's usage policy. We are not allowed to publish weights for LLaMA, of course, even finetuned, but there is no problem publishing the difference, a patch that we suggest to apply to the files.
13
+ The encryption is a simple XOR between files, ensuring that only the people that have access to the original weights (from completely legal sources, of course) can transform them into finetuned weights.
14
+ You can find the decrypt code on https://github.com/LianjiaTech/BELLE/tree/main/models .
15
+
16
+
17
+ # Model Card for BELLE-LLaMA-7B-2M-q4
18
+
19
+ ## Welcome
20
+ 4-bit quantized version using [llama.cpp](https://github.com/ggerganov/llama.cpp) of [BELLE-LLaMA-7B-2M](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-enc)
21
+ If you find this model helpful, please *like* this model and star us on https://github.com/LianjiaTech/BELLE !
22
+
23
+
24
+ ## Model description
25
+ BELLE-LLAMA-7B-2M-enc is based on LLAMA 7B and finetuned with 2M Chinese data combined with 50,000 pieces of English data from the open source Stanford-Alpaca, resulting in good Chinese instruction understanding and response generation capabilities.
26
+
27
+ The code of Chinese data generation and other detailed information can be found in our Github project repository: https://github.com/LianjiaTech/BELLE.
28
+
29
+
30
+ ## Download
31
+ Should you accept our license and acknowledged the limitations, download the model by clicking [Download](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-q4/resolve/main/belle-model.bin).
32
+
33
+
34
+ ## Model Usage
35
+ This is a quantized version made for offline on-devices inferencing.
36
+ You can use this model with ChatBELLE, a minimal, cross-platform LLM chat app powered by [BELLE](https://github.com/LianjiaTech/BELLE)
37
+ using quantized on-device offline models and Flutter UI, running on macOS (done), Windows, Android,
38
+ iOS(see [Known Issues](#known-issues)) and more.
39
+
40
+ ### macOS
41
+ * Download and put the app anywhere, preferably in `Applications` folder.
42
+ * Open the app by right click then Ctrl-click `Open`, then click `Open`.
43
+ * The app will prompt the intended model file path and fail to load the model. Close the app.
44
+ * Download quantized model from [BELLE-LLaMA-7B-2M-q4](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-q4/blob/main/belle-model.bin).
45
+ * Move and rename the model to the path prompted by the app. Defaults to `~/Library/Containers/com.barius.chatbelle/Data/belle-model.bin` .
46
+ * Reopen the app again (double clicking is now OK).
47
+
48
+ ### Windows
49
+ * Stay tuned
50
+
51
+ ### Android
52
+ * Stay tuned
53
+
54
+ ### iOS
55
+ * Stay tuned
56
+
57
+
58
+ ## Limitations
59
+ There still exists a few issues in the model trained on current base model and data:
60
+
61
+ 1. The model might generate factual errors when asked to follow instructions related to facts.
62
+
63
+ 2. Occasionally generates harmful responses since the model still struggles to identify potential harmful instructions.
64
+
65
+ 3. Needs improvements on reasoning and coding.
66
+
67
+ Since the model still has its limitations, we require developers only use the open-sourced code, data, model and any other artifacts generated via this project for research purposes. Commercial use and other potential harmful use cases are not allowed.
68
+
69
+
70
+ ## Citation
71
+
72
+ Please cite us when using our code, data or model.
73
+
74
+ ```
75
+ @misc{BELLE,
76
+ author = {Yunjie Ji, Yong Deng, Yan Gong, Yiping Peng, Qiang Niu, Baochang Ma, Xiangang Li},
77
+ title = {BELLE: Be Everyone's Large Language model Engine},
78
+ year = {2023},
79
+ publisher = {GitHub},
80
+ journal = {GitHub repository},
81
+ howpublished = {\url{https://github.com/LianjiaTech/BELLE}},
82
+ }
83
+ ```