Generated README.md
Browse files
README.md
ADDED
@@ -0,0 +1,141 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
datasets:
|
3 |
+
- bigscience/xP3
|
4 |
+
license: bigscience-bloom-rail-1.0
|
5 |
+
language:
|
6 |
+
- ak
|
7 |
+
- ar
|
8 |
+
- as
|
9 |
+
- bm
|
10 |
+
- bn
|
11 |
+
- ca
|
12 |
+
- code
|
13 |
+
- en
|
14 |
+
- es
|
15 |
+
- eu
|
16 |
+
- fon
|
17 |
+
- fr
|
18 |
+
- gu
|
19 |
+
- hi
|
20 |
+
- id
|
21 |
+
- ig
|
22 |
+
- ki
|
23 |
+
- kn
|
24 |
+
- lg
|
25 |
+
- ln
|
26 |
+
- ml
|
27 |
+
- mr
|
28 |
+
- ne
|
29 |
+
- nso
|
30 |
+
- ny
|
31 |
+
- or
|
32 |
+
- pa
|
33 |
+
- pt
|
34 |
+
- rn
|
35 |
+
- rw
|
36 |
+
- sn
|
37 |
+
- st
|
38 |
+
- sw
|
39 |
+
- ta
|
40 |
+
- te
|
41 |
+
- tn
|
42 |
+
- ts
|
43 |
+
- tum
|
44 |
+
- tw
|
45 |
+
- ur
|
46 |
+
- vi
|
47 |
+
- wo
|
48 |
+
- xh
|
49 |
+
- yo
|
50 |
+
- zh
|
51 |
+
- zu
|
52 |
+
programming_language:
|
53 |
+
- C
|
54 |
+
- C++
|
55 |
+
- C#
|
56 |
+
- Go
|
57 |
+
- Java
|
58 |
+
- JavaScript
|
59 |
+
- Lua
|
60 |
+
- PHP
|
61 |
+
- Python
|
62 |
+
- Ruby
|
63 |
+
- Rust
|
64 |
+
- Scala
|
65 |
+
- TypeScript
|
66 |
+
tags:
|
67 |
+
- llm-rs
|
68 |
+
- ggml
|
69 |
+
pipeline_tag: text-generation
|
70 |
+
---
|
71 |
+
|
72 |
+
# GGML covnerted Models of [BigScience](https://huggingface.co/bigscience)'s Bloom models
|
73 |
+
|
74 |
+
## Description
|
75 |
+
|
76 |
+
> We present BLOOMZ & mT0, a family of models capable of following human instructions in dozens of languages zero-shot. We finetune BLOOM & mT5 pretrained multilingual language models on our crosslingual task mixture (xP3) and find the resulting models capable of crosslingual generalization to unseen tasks & languages.
|
77 |
+
|
78 |
+
- **Repository:** [bigscience-workshop/xmtf](https://github.com/bigscience-workshop/xmtf)
|
79 |
+
- **Paper:** [Crosslingual Generalization through Multitask Finetuning](https://arxiv.org/abs/2211.01786)
|
80 |
+
- **Point of Contact:** [Niklas Muennighoff](mailto:niklas@hf.co)
|
81 |
+
- **Languages:** Refer to [bloom](https://huggingface.co/bigscience/bloom) for pretraining & [xP3](https://huggingface.co/datasets/bigscience/xP3) for finetuning language proportions. It understands both pretraining & finetuning languages.
|
82 |
+
|
83 |
+
### Intended use
|
84 |
+
|
85 |
+
We recommend using the model to perform tasks expressed in natural language. For example, given the prompt "*Translate to English: Je t’aime.*", the model will most likely answer "*I love you.*". Some prompt ideas from our paper:
|
86 |
+
- 一个传奇的开端,一个不灭的神话,这不仅仅是一部电影,而是作为一个走进新时代的标签,永远彪炳史册。你认为这句话的立场是赞扬、中立还是批评?
|
87 |
+
- Suggest at least five related search terms to "Mạng neural nhân tạo".
|
88 |
+
- Write a fairy tale about a troll saving a princess from a dangerous dragon. The fairy tale is a masterpiece that has achieved praise worldwide and its moral is "Heroes Come in All Shapes and Sizes". Story (in Spanish):
|
89 |
+
- Explain in a sentence in Telugu what is backpropagation in neural networks.
|
90 |
+
|
91 |
+
## Converted Models
|
92 |
+
| Name | Based on | Type | Container | GGML Version |
|
93 |
+
|:----------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------|:-------|:------------|:---------------|
|
94 |
+
| [bloomz-1b1-f16.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-1b1-f16.bin) | [bigscience/bloomz-1b1](https://huggingface.co/bigscience/bloomz-1b1) | F16 | GGML | V3 |
|
95 |
+
| [bloomz-1b1-q4_0.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-1b1-q4_0.bin) | [bigscience/bloomz-1b1](https://huggingface.co/bigscience/bloomz-1b1) | Q4_0 | GGML | V3 |
|
96 |
+
| [bloomz-1b1-q4_0-ggjt.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-1b1-q4_0-ggjt.bin) | [bigscience/bloomz-1b1](https://huggingface.co/bigscience/bloomz-1b1) | Q4_0 | GGJT | V3 |
|
97 |
+
| [bloomz-1b1-q5_1-ggjt.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-1b1-q5_1-ggjt.bin) | [bigscience/bloomz-1b1](https://huggingface.co/bigscience/bloomz-1b1) | Q5_1 | GGJT | V3 |
|
98 |
+
| [bloomz-1b7-f16.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-1b7-f16.bin) | [bigscience/bloomz-1b7](https://huggingface.co/bigscience/bloomz-1b7) | F16 | GGML | V3 |
|
99 |
+
| [bloomz-1b7-q4_0.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-1b7-q4_0.bin) | [bigscience/bloomz-1b7](https://huggingface.co/bigscience/bloomz-1b7) | Q4_0 | GGML | V3 |
|
100 |
+
| [bloomz-1b7-q4_0-ggjt.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-1b7-q4_0-ggjt.bin) | [bigscience/bloomz-1b7](https://huggingface.co/bigscience/bloomz-1b7) | Q4_0 | GGJT | V3 |
|
101 |
+
| [bloomz-1b7-q5_1-ggjt.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-1b7-q5_1-ggjt.bin) | [bigscience/bloomz-1b7](https://huggingface.co/bigscience/bloomz-1b7) | Q5_1 | GGJT | V3 |
|
102 |
+
| [bloomz-3b-f16.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-3b-f16.bin) | [bigscience/bloomz-3b](https://huggingface.co/bigscience/bloomz-3b) | F16 | GGML | V3 |
|
103 |
+
| [bloomz-3b-q4_0.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-3b-q4_0.bin) | [bigscience/bloomz-3b](https://huggingface.co/bigscience/bloomz-3b) | Q4_0 | GGML | V3 |
|
104 |
+
| [bloomz-3b-q4_0-ggjt.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-3b-q4_0-ggjt.bin) | [bigscience/bloomz-3b](https://huggingface.co/bigscience/bloomz-3b) | Q4_0 | GGJT | V3 |
|
105 |
+
| [bloomz-3b-q5_1-ggjt.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-3b-q5_1-ggjt.bin) | [bigscience/bloomz-3b](https://huggingface.co/bigscience/bloomz-3b) | Q5_1 | GGJT | V3 |
|
106 |
+
| [bloomz-560m-f16.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-560m-f16.bin) | [bigscience/bloomz-560m](https://huggingface.co/bigscience/bloomz-560m) | F16 | GGML | V3 |
|
107 |
+
| [bloomz-560m-q4_0.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-560m-q4_0.bin) | [bigscience/bloomz-560m](https://huggingface.co/bigscience/bloomz-560m) | Q4_0 | GGML | V3 |
|
108 |
+
| [bloomz-560m-q4_0-ggjt.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-560m-q4_0-ggjt.bin) | [bigscience/bloomz-560m](https://huggingface.co/bigscience/bloomz-560m) | Q4_0 | GGJT | V3 |
|
109 |
+
| [bloomz-560m-q5_1-ggjt.bin](https://huggingface.co/rustformers/bloomz-ggml/blob/main/bloomz-560m-q5_1-ggjt.bin) | [bigscience/bloomz-560m](https://huggingface.co/bigscience/bloomz-560m) | Q5_1 | GGJT | V3 |
|
110 |
+
|
111 |
+
## Usage
|
112 |
+
|
113 |
+
### Python via [llm-rs](https://github.com/LLukas22/llm-rs-python):
|
114 |
+
|
115 |
+
#### Installation
|
116 |
+
Via pip: `pip install llm-rs`
|
117 |
+
|
118 |
+
#### Run inference
|
119 |
+
```python
|
120 |
+
from llm_rs import AutoModel
|
121 |
+
|
122 |
+
#Load the model, define any model you like from the list above as the `model_file`
|
123 |
+
model = AutoModel.from_pretrained("rustformers/bloomz-ggml",model_file="bloomz-3b-q4_0-ggjt.bin")
|
124 |
+
|
125 |
+
#Generate
|
126 |
+
print(model.generate("The meaning of life is"))
|
127 |
+
```
|
128 |
+
|
129 |
+
### Rust via [Rustformers/llm](https://github.com/rustformers/llm):
|
130 |
+
|
131 |
+
#### Installation
|
132 |
+
```
|
133 |
+
git clone --recurse-submodules https://github.com/rustformers/llm.git
|
134 |
+
cd llm
|
135 |
+
cargo build --release
|
136 |
+
```
|
137 |
+
|
138 |
+
#### Run inference
|
139 |
+
```
|
140 |
+
cargo run --release -- bloom infer -m path/to/model.bin -p "Tell me how cool the Rust programming language is:"
|
141 |
+
```
|