Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,83 @@
|
|
1 |
---
|
2 |
license: gpl-3.0
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: gpl-3.0
|
3 |
+
tags:
|
4 |
+
- text2text-generation
|
5 |
+
pipeline_tag: text2text-generation
|
6 |
+
language:
|
7 |
+
- zh
|
8 |
+
- en
|
9 |
---
|
10 |
+
|
11 |
+
Considering LLaMA's license constraints, the model is for research and learning only.
|
12 |
+
Please strictly respect LLaMA's usage policy. We are not allowed to publish weights for LLaMA, of course, even finetuned, but there is no problem publishing the difference, a patch that we suggest to apply to the files.
|
13 |
+
The encryption is a simple XOR between files, ensuring that only the people that have access to the original weights (from completely legal sources, of course) can transform them into finetuned weights.
|
14 |
+
You can find the decrypt code on https://github.com/LianjiaTech/BELLE/tree/main/models .
|
15 |
+
|
16 |
+
|
17 |
+
# Model Card for BELLE-LLaMA-7B-2M-q4
|
18 |
+
|
19 |
+
## Welcome
|
20 |
+
4-bit quantized version using [llama.cpp](https://github.com/ggerganov/llama.cpp) of [BELLE-LLaMA-7B-2M](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-enc)
|
21 |
+
If you find this model helpful, please *like* this model and star us on https://github.com/LianjiaTech/BELLE !
|
22 |
+
|
23 |
+
|
24 |
+
## Model description
|
25 |
+
BELLE-LLAMA-7B-2M-enc is based on LLAMA 7B and finetuned with 2M Chinese data combined with 50,000 pieces of English data from the open source Stanford-Alpaca, resulting in good Chinese instruction understanding and response generation capabilities.
|
26 |
+
|
27 |
+
The code of Chinese data generation and other detailed information can be found in our Github project repository: https://github.com/LianjiaTech/BELLE.
|
28 |
+
|
29 |
+
|
30 |
+
## Download
|
31 |
+
Should you accept our license and acknowledged the limitations, download the model by clicking [Download](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-q4/resolve/main/belle-model.bin).
|
32 |
+
|
33 |
+
|
34 |
+
## Model Usage
|
35 |
+
This is a quantized version made for offline on-devices inferencing.
|
36 |
+
You can use this model with ChatBELLE, a minimal, cross-platform LLM chat app powered by [BELLE](https://github.com/LianjiaTech/BELLE)
|
37 |
+
using quantized on-device offline models and Flutter UI, running on macOS (done), Windows, Android,
|
38 |
+
iOS(see [Known Issues](#known-issues)) and more.
|
39 |
+
|
40 |
+
### macOS
|
41 |
+
* Download and put the app anywhere, preferably in `Applications` folder.
|
42 |
+
* Open the app by right click then Ctrl-click `Open`, then click `Open`.
|
43 |
+
* The app will prompt the intended model file path and fail to load the model. Close the app.
|
44 |
+
* Download quantized model from [BELLE-LLaMA-7B-2M-q4](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-q4/blob/main/belle-model.bin).
|
45 |
+
* Move and rename the model to the path prompted by the app. Defaults to `~/Library/Containers/com.barius.chatbelle/Data/belle-model.bin` .
|
46 |
+
* Reopen the app again (double clicking is now OK).
|
47 |
+
|
48 |
+
### Windows
|
49 |
+
* Stay tuned
|
50 |
+
|
51 |
+
### Android
|
52 |
+
* Stay tuned
|
53 |
+
|
54 |
+
### iOS
|
55 |
+
* Stay tuned
|
56 |
+
|
57 |
+
|
58 |
+
## Limitations
|
59 |
+
There still exists a few issues in the model trained on current base model and data:
|
60 |
+
|
61 |
+
1. The model might generate factual errors when asked to follow instructions related to facts.
|
62 |
+
|
63 |
+
2. Occasionally generates harmful responses since the model still struggles to identify potential harmful instructions.
|
64 |
+
|
65 |
+
3. Needs improvements on reasoning and coding.
|
66 |
+
|
67 |
+
Since the model still has its limitations, we require developers only use the open-sourced code, data, model and any other artifacts generated via this project for research purposes. Commercial use and other potential harmful use cases are not allowed.
|
68 |
+
|
69 |
+
|
70 |
+
## Citation
|
71 |
+
|
72 |
+
Please cite us when using our code, data or model.
|
73 |
+
|
74 |
+
```
|
75 |
+
@misc{BELLE,
|
76 |
+
author = {Yunjie Ji, Yong Deng, Yan Gong, Yiping Peng, Qiang Niu, Baochang Ma, Xiangang Li},
|
77 |
+
title = {BELLE: Be Everyone's Large Language model Engine},
|
78 |
+
year = {2023},
|
79 |
+
publisher = {GitHub},
|
80 |
+
journal = {GitHub repository},
|
81 |
+
howpublished = {\url{https://github.com/LianjiaTech/BELLE}},
|
82 |
+
}
|
83 |
+
```
|