barius commited on
Commit
be3c0a7
·
1 Parent(s): 5e1322a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -7
README.md CHANGED
@@ -14,16 +14,15 @@ The encryption is a simple XOR between files, ensuring that only the people that
14
  You can find the decrypt code on https://github.com/LianjiaTech/BELLE/tree/main/models .
15
 
16
 
17
- # Model Card for BELLE-LLaMA-7B-2M-q4
18
 
19
  ## Welcome
20
- 4-bit quantized version using [llama.cpp](https://github.com/ggerganov/llama.cpp) of [BELLE-LLaMA-7B-2M](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-enc)
21
-
22
  If you find this model helpful, please *like* this model and star us on https://github.com/LianjiaTech/BELLE !
23
 
24
 
25
  ## Model description
26
- BELLE-LLAMA-7B-2M-enc is based on LLAMA 7B and finetuned with 2M Chinese data combined with 50,000 pieces of English data from the open source Stanford-Alpaca, resulting in good Chinese instruction understanding and response generation capabilities.
27
 
28
  The code of Chinese data generation and other detailed information can be found in our Github project repository: https://github.com/LianjiaTech/BELLE.
29
 
@@ -33,15 +32,14 @@ Should you accept our license and acknowledged the limitations, download the mod
33
 
34
 
35
  ## Model Usage
36
- This is a quantized version of [BELLE-LLaMA-7B-2M](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-enc) made for offline on-devices inferencing.
37
 
38
  You can use this model with ChatBELLE, a minimal, cross-platform LLM chat app powered by [BELLE](https://github.com/LianjiaTech/BELLE)
39
  using quantized on-device offline models and Flutter UI, running on macOS (done), Windows, Android,
40
  iOS(see [Known Issues](#known-issues)) and more.
41
 
42
  ### macOS
43
- * Download and put the app anywhere, preferably in `Applications` folder.
44
- * Open the app by right click then Ctrl-click `Open`, then click `Open`.
45
  * The app will prompt the intended model file path and fail to load the model. Close the app.
46
  * Download quantized model from [BELLE-LLaMA-7B-2M-q4](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-q4/blob/main/belle-model.bin).
47
  * Move and rename the model to the path prompted by the app. Defaults to `~/Library/Containers/com.barius.chatbelle/Data/belle-model.bin` .
 
14
  You can find the decrypt code on https://github.com/LianjiaTech/BELLE/tree/main/models .
15
 
16
 
17
+ # Model Card for ChatBELLE-int4
18
 
19
  ## Welcome
20
+ 4-bit quantized model using [llama.cpp](https://github.com/ggerganov/llama.cpp).
 
21
  If you find this model helpful, please *like* this model and star us on https://github.com/LianjiaTech/BELLE !
22
 
23
 
24
  ## Model description
25
+ ChatBELLE-int4 is based on 7B model and quantized to 4-bit.
26
 
27
  The code of Chinese data generation and other detailed information can be found in our Github project repository: https://github.com/LianjiaTech/BELLE.
28
 
 
32
 
33
 
34
  ## Model Usage
 
35
 
36
  You can use this model with ChatBELLE, a minimal, cross-platform LLM chat app powered by [BELLE](https://github.com/LianjiaTech/BELLE)
37
  using quantized on-device offline models and Flutter UI, running on macOS (done), Windows, Android,
38
  iOS(see [Known Issues](#known-issues)) and more.
39
 
40
  ### macOS
41
+ * Download [chatbelle.dmg](https://github.com/LianjiaTech/BELLE/releases/download/v0.95/chatbelle.dmg) from [Releases](https://github.com/LianjiaTech/BELLE/releases/tag/v0.95) page, double click to open it, then drag `Chat Belle.dmg` into `Applications` folder.
42
+ * Open the `Chat Belle` app in `Applications` folder by right click then Ctrl-click `Open`, then click `Open`.
43
  * The app will prompt the intended model file path and fail to load the model. Close the app.
44
  * Download quantized model from [BELLE-LLaMA-7B-2M-q4](https://huggingface.co/BelleGroup/BELLE-LLaMA-7B-2M-q4/blob/main/belle-model.bin).
45
  * Move and rename the model to the path prompted by the app. Defaults to `~/Library/Containers/com.barius.chatbelle/Data/belle-model.bin` .