Text Generation
Transformers
PyTorch
bloom
Eval Results
text-generation-inference
Inference Endpoints
michael-newsrx-com Muennighoff commited on
Commit
8ad2bb4
0 Parent(s):

Duplicate from bigscience/bloomz-7b1

Browse files

Co-authored-by: Niklas Muennighoff <Muennighoff@users.noreply.huggingface.co>

.gitattributes ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ftz filter=lfs diff=lfs merge=lfs -text
6
+ *.gz filter=lfs diff=lfs merge=lfs -text
7
+ *.h5 filter=lfs diff=lfs merge=lfs -text
8
+ *.joblib filter=lfs diff=lfs merge=lfs -text
9
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
10
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
11
+ *.model filter=lfs diff=lfs merge=lfs -text
12
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
13
+ *.npy filter=lfs diff=lfs merge=lfs -text
14
+ *.npz filter=lfs diff=lfs merge=lfs -text
15
+ *.onnx filter=lfs diff=lfs merge=lfs -text
16
+ *.ot filter=lfs diff=lfs merge=lfs -text
17
+ *.parquet filter=lfs diff=lfs merge=lfs -text
18
+ *.pb filter=lfs diff=lfs merge=lfs -text
19
+ *.pickle filter=lfs diff=lfs merge=lfs -text
20
+ *.pkl filter=lfs diff=lfs merge=lfs -text
21
+ *.pt filter=lfs diff=lfs merge=lfs -text
22
+ *.pth filter=lfs diff=lfs merge=lfs -text
23
+ *.rar filter=lfs diff=lfs merge=lfs -text
24
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
26
+ *.tflite filter=lfs diff=lfs merge=lfs -text
27
+ *.tgz filter=lfs diff=lfs merge=lfs -text
28
+ *.wasm filter=lfs diff=lfs merge=lfs -text
29
+ *.xz filter=lfs diff=lfs merge=lfs -text
30
+ *.zip filter=lfs diff=lfs merge=lfs -text
31
+ *.zst filter=lfs diff=lfs merge=lfs -text
32
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
33
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
34
+ logs/logs/main_log.txt filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,894 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - bigscience/xP3
4
+ license: bigscience-bloom-rail-1.0
5
+ language:
6
+ - ak
7
+ - ar
8
+ - as
9
+ - bm
10
+ - bn
11
+ - ca
12
+ - code
13
+ - en
14
+ - es
15
+ - eu
16
+ - fon
17
+ - fr
18
+ - gu
19
+ - hi
20
+ - id
21
+ - ig
22
+ - ki
23
+ - kn
24
+ - lg
25
+ - ln
26
+ - ml
27
+ - mr
28
+ - ne
29
+ - nso
30
+ - ny
31
+ - or
32
+ - pa
33
+ - pt
34
+ - rn
35
+ - rw
36
+ - sn
37
+ - st
38
+ - sw
39
+ - ta
40
+ - te
41
+ - tn
42
+ - ts
43
+ - tum
44
+ - tw
45
+ - ur
46
+ - vi
47
+ - wo
48
+ - xh
49
+ - yo
50
+ - zh
51
+ - zu
52
+ programming_language:
53
+ - C
54
+ - C++
55
+ - C#
56
+ - Go
57
+ - Java
58
+ - JavaScript
59
+ - Lua
60
+ - PHP
61
+ - Python
62
+ - Ruby
63
+ - Rust
64
+ - Scala
65
+ - TypeScript
66
+ pipeline_tag: text-generation
67
+ widget:
68
+ - text: >-
69
+ 一个传奇的开端,一个不灭的神话,这不仅仅是一部电影,而是作为一个走进新时代的标签,永远彪炳史册。Would you rate the previous
70
+ review as positive, neutral or negative?
71
+ example_title: zh-en sentiment
72
+ - text: 一个传奇的开端,一个不灭的神话,这不仅仅是一部电影,而是作为一个走进新时代的标签,永远彪炳史册。你认为这句话的立场是赞扬、中立还是批评?
73
+ example_title: zh-zh sentiment
74
+ - text: Suggest at least five related search terms to "Mạng neural nhân tạo".
75
+ example_title: vi-en query
76
+ - text: >-
77
+ Proposez au moins cinq mots clés concernant «Réseau de neurones
78
+ artificiels».
79
+ example_title: fr-fr query
80
+ - text: Explain in a sentence in Telugu what is backpropagation in neural networks.
81
+ example_title: te-en qa
82
+ - text: Why is the sky blue?
83
+ example_title: en-en qa
84
+ - text: >-
85
+ Write a fairy tale about a troll saving a princess from a dangerous dragon.
86
+ The fairy tale is a masterpiece that has achieved praise worldwide and its
87
+ moral is "Heroes Come in All Shapes and Sizes". Story (in Spanish):
88
+ example_title: es-en fable
89
+ - text: >-
90
+ Write a fable about wood elves living in a forest that is suddenly invaded
91
+ by ogres. The fable is a masterpiece that has achieved praise worldwide and
92
+ its moral is "Violence is the last refuge of the incompetent". Fable (in
93
+ Hindi):
94
+ example_title: hi-en fable
95
+ model-index:
96
+ - name: bloomz-7b1
97
+ results:
98
+ - task:
99
+ type: Coreference resolution
100
+ dataset:
101
+ type: winogrande
102
+ name: Winogrande XL (xl)
103
+ config: xl
104
+ split: validation
105
+ revision: a80f460359d1e9a67c006011c94de42a8759430c
106
+ metrics:
107
+ - type: Accuracy
108
+ value: 55.8
109
+ - task:
110
+ type: Coreference resolution
111
+ dataset:
112
+ type: Muennighoff/xwinograd
113
+ name: XWinograd (en)
114
+ config: en
115
+ split: test
116
+ revision: 9dd5ea5505fad86b7bedad667955577815300cee
117
+ metrics:
118
+ - type: Accuracy
119
+ value: 66.02
120
+ - task:
121
+ type: Coreference resolution
122
+ dataset:
123
+ type: Muennighoff/xwinograd
124
+ name: XWinograd (fr)
125
+ config: fr
126
+ split: test
127
+ revision: 9dd5ea5505fad86b7bedad667955577815300cee
128
+ metrics:
129
+ - type: Accuracy
130
+ value: 57.83
131
+ - task:
132
+ type: Coreference resolution
133
+ dataset:
134
+ type: Muennighoff/xwinograd
135
+ name: XWinograd (jp)
136
+ config: jp
137
+ split: test
138
+ revision: 9dd5ea5505fad86b7bedad667955577815300cee
139
+ metrics:
140
+ - type: Accuracy
141
+ value: 52.87
142
+ - task:
143
+ type: Coreference resolution
144
+ dataset:
145
+ type: Muennighoff/xwinograd
146
+ name: XWinograd (pt)
147
+ config: pt
148
+ split: test
149
+ revision: 9dd5ea5505fad86b7bedad667955577815300cee
150
+ metrics:
151
+ - type: Accuracy
152
+ value: 57.79
153
+ - task:
154
+ type: Coreference resolution
155
+ dataset:
156
+ type: Muennighoff/xwinograd
157
+ name: XWinograd (ru)
158
+ config: ru
159
+ split: test
160
+ revision: 9dd5ea5505fad86b7bedad667955577815300cee
161
+ metrics:
162
+ - type: Accuracy
163
+ value: 54.92
164
+ - task:
165
+ type: Coreference resolution
166
+ dataset:
167
+ type: Muennighoff/xwinograd
168
+ name: XWinograd (zh)
169
+ config: zh
170
+ split: test
171
+ revision: 9dd5ea5505fad86b7bedad667955577815300cee
172
+ metrics:
173
+ - type: Accuracy
174
+ value: 63.69
175
+ - task:
176
+ type: Natural language inference
177
+ dataset:
178
+ type: anli
179
+ name: ANLI (r1)
180
+ config: r1
181
+ split: validation
182
+ revision: 9dbd830a06fea8b1c49d6e5ef2004a08d9f45094
183
+ metrics:
184
+ - type: Accuracy
185
+ value: 42.1
186
+ - task:
187
+ type: Natural language inference
188
+ dataset:
189
+ type: anli
190
+ name: ANLI (r2)
191
+ config: r2
192
+ split: validation
193
+ revision: 9dbd830a06fea8b1c49d6e5ef2004a08d9f45094
194
+ metrics:
195
+ - type: Accuracy
196
+ value: 39.5
197
+ - task:
198
+ type: Natural language inference
199
+ dataset:
200
+ type: anli
201
+ name: ANLI (r3)
202
+ config: r3
203
+ split: validation
204
+ revision: 9dbd830a06fea8b1c49d6e5ef2004a08d9f45094
205
+ metrics:
206
+ - type: Accuracy
207
+ value: 41
208
+ - task:
209
+ type: Natural language inference
210
+ dataset:
211
+ type: super_glue
212
+ name: SuperGLUE (cb)
213
+ config: cb
214
+ split: validation
215
+ revision: 9e12063561e7e6c79099feb6d5a493142584e9e2
216
+ metrics:
217
+ - type: Accuracy
218
+ value: 80.36
219
+ - task:
220
+ type: Natural language inference
221
+ dataset:
222
+ type: super_glue
223
+ name: SuperGLUE (rte)
224
+ config: rte
225
+ split: validation
226
+ revision: 9e12063561e7e6c79099feb6d5a493142584e9e2
227
+ metrics:
228
+ - type: Accuracy
229
+ value: 84.12
230
+ - task:
231
+ type: Natural language inference
232
+ dataset:
233
+ type: xnli
234
+ name: XNLI (ar)
235
+ config: ar
236
+ split: validation
237
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
238
+ metrics:
239
+ - type: Accuracy
240
+ value: 53.25
241
+ - task:
242
+ type: Natural language inference
243
+ dataset:
244
+ type: xnli
245
+ name: XNLI (bg)
246
+ config: bg
247
+ split: validation
248
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
249
+ metrics:
250
+ - type: Accuracy
251
+ value: 43.61
252
+ - task:
253
+ type: Natural language inference
254
+ dataset:
255
+ type: xnli
256
+ name: XNLI (de)
257
+ config: de
258
+ split: validation
259
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
260
+ metrics:
261
+ - type: Accuracy
262
+ value: 46.83
263
+ - task:
264
+ type: Natural language inference
265
+ dataset:
266
+ type: xnli
267
+ name: XNLI (el)
268
+ config: el
269
+ split: validation
270
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
271
+ metrics:
272
+ - type: Accuracy
273
+ value: 41.53
274
+ - task:
275
+ type: Natural language inference
276
+ dataset:
277
+ type: xnli
278
+ name: XNLI (en)
279
+ config: en
280
+ split: validation
281
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
282
+ metrics:
283
+ - type: Accuracy
284
+ value: 59.68
285
+ - task:
286
+ type: Natural language inference
287
+ dataset:
288
+ type: xnli
289
+ name: XNLI (es)
290
+ config: es
291
+ split: validation
292
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
293
+ metrics:
294
+ - type: Accuracy
295
+ value: 55.1
296
+ - task:
297
+ type: Natural language inference
298
+ dataset:
299
+ type: xnli
300
+ name: XNLI (fr)
301
+ config: fr
302
+ split: validation
303
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
304
+ metrics:
305
+ - type: Accuracy
306
+ value: 55.26
307
+ - task:
308
+ type: Natural language inference
309
+ dataset:
310
+ type: xnli
311
+ name: XNLI (hi)
312
+ config: hi
313
+ split: validation
314
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
315
+ metrics:
316
+ - type: Accuracy
317
+ value: 50.88
318
+ - task:
319
+ type: Natural language inference
320
+ dataset:
321
+ type: xnli
322
+ name: XNLI (ru)
323
+ config: ru
324
+ split: validation
325
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
326
+ metrics:
327
+ - type: Accuracy
328
+ value: 47.75
329
+ - task:
330
+ type: Natural language inference
331
+ dataset:
332
+ type: xnli
333
+ name: XNLI (sw)
334
+ config: sw
335
+ split: validation
336
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
337
+ metrics:
338
+ - type: Accuracy
339
+ value: 46.63
340
+ - task:
341
+ type: Natural language inference
342
+ dataset:
343
+ type: xnli
344
+ name: XNLI (th)
345
+ config: th
346
+ split: validation
347
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
348
+ metrics:
349
+ - type: Accuracy
350
+ value: 40.12
351
+ - task:
352
+ type: Natural language inference
353
+ dataset:
354
+ type: xnli
355
+ name: XNLI (tr)
356
+ config: tr
357
+ split: validation
358
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
359
+ metrics:
360
+ - type: Accuracy
361
+ value: 37.55
362
+ - task:
363
+ type: Natural language inference
364
+ dataset:
365
+ type: xnli
366
+ name: XNLI (ur)
367
+ config: ur
368
+ split: validation
369
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
370
+ metrics:
371
+ - type: Accuracy
372
+ value: 46.51
373
+ - task:
374
+ type: Natural language inference
375
+ dataset:
376
+ type: xnli
377
+ name: XNLI (vi)
378
+ config: vi
379
+ split: validation
380
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
381
+ metrics:
382
+ - type: Accuracy
383
+ value: 52.93
384
+ - task:
385
+ type: Natural language inference
386
+ dataset:
387
+ type: xnli
388
+ name: XNLI (zh)
389
+ config: zh
390
+ split: validation
391
+ revision: a5a45e4ff92d5d3f34de70aaf4b72c3bdf9f7f16
392
+ metrics:
393
+ - type: Accuracy
394
+ value: 53.61
395
+ - task:
396
+ type: Program synthesis
397
+ dataset:
398
+ type: openai_humaneval
399
+ name: HumanEval
400
+ config: None
401
+ split: test
402
+ revision: e8dc562f5de170c54b5481011dd9f4fa04845771
403
+ metrics:
404
+ - type: Pass@1
405
+ value: 8.06
406
+ - type: Pass@10
407
+ value: 15.03
408
+ - type: Pass@100
409
+ value: 27.49
410
+ - task:
411
+ type: Sentence completion
412
+ dataset:
413
+ type: story_cloze
414
+ name: StoryCloze (2016)
415
+ config: '2016'
416
+ split: validation
417
+ revision: e724c6f8cdf7c7a2fb229d862226e15b023ee4db
418
+ metrics:
419
+ - type: Accuracy
420
+ value: 90.43
421
+ - task:
422
+ type: Sentence completion
423
+ dataset:
424
+ type: super_glue
425
+ name: SuperGLUE (copa)
426
+ config: copa
427
+ split: validation
428
+ revision: 9e12063561e7e6c79099feb6d5a493142584e9e2
429
+ metrics:
430
+ - type: Accuracy
431
+ value: 86
432
+ - task:
433
+ type: Sentence completion
434
+ dataset:
435
+ type: xcopa
436
+ name: XCOPA (et)
437
+ config: et
438
+ split: validation
439
+ revision: 37f73c60fb123111fa5af5f9b705d0b3747fd187
440
+ metrics:
441
+ - type: Accuracy
442
+ value: 50
443
+ - task:
444
+ type: Sentence completion
445
+ dataset:
446
+ type: xcopa
447
+ name: XCOPA (ht)
448
+ config: ht
449
+ split: validation
450
+ revision: 37f73c60fb123111fa5af5f9b705d0b3747fd187
451
+ metrics:
452
+ - type: Accuracy
453
+ value: 54
454
+ - task:
455
+ type: Sentence completion
456
+ dataset:
457
+ type: xcopa
458
+ name: XCOPA (id)
459
+ config: id
460
+ split: validation
461
+ revision: 37f73c60fb123111fa5af5f9b705d0b3747fd187
462
+ metrics:
463
+ - type: Accuracy
464
+ value: 76
465
+ - task:
466
+ type: Sentence completion
467
+ dataset:
468
+ type: xcopa
469
+ name: XCOPA (it)
470
+ config: it
471
+ split: validation
472
+ revision: 37f73c60fb123111fa5af5f9b705d0b3747fd187
473
+ metrics:
474
+ - type: Accuracy
475
+ value: 61
476
+ - task:
477
+ type: Sentence completion
478
+ dataset:
479
+ type: xcopa
480
+ name: XCOPA (qu)
481
+ config: qu
482
+ split: validation
483
+ revision: 37f73c60fb123111fa5af5f9b705d0b3747fd187
484
+ metrics:
485
+ - type: Accuracy
486
+ value: 60
487
+ - task:
488
+ type: Sentence completion
489
+ dataset:
490
+ type: xcopa
491
+ name: XCOPA (sw)
492
+ config: sw
493
+ split: validation
494
+ revision: 37f73c60fb123111fa5af5f9b705d0b3747fd187
495
+ metrics:
496
+ - type: Accuracy
497
+ value: 63
498
+ - task:
499
+ type: Sentence completion
500
+ dataset:
501
+ type: xcopa
502
+ name: XCOPA (ta)
503
+ config: ta
504
+ split: validation
505
+ revision: 37f73c60fb123111fa5af5f9b705d0b3747fd187
506
+ metrics:
507
+ - type: Accuracy
508
+ value: 64
509
+ - task:
510
+ type: Sentence completion
511
+ dataset:
512
+ type: xcopa
513
+ name: XCOPA (th)
514
+ config: th
515
+ split: validation
516
+ revision: 37f73c60fb123111fa5af5f9b705d0b3747fd187
517
+ metrics:
518
+ - type: Accuracy
519
+ value: 57
520
+ - task:
521
+ type: Sentence completion
522
+ dataset:
523
+ type: xcopa
524
+ name: XCOPA (tr)
525
+ config: tr
526
+ split: validation
527
+ revision: 37f73c60fb123111fa5af5f9b705d0b3747fd187
528
+ metrics:
529
+ - type: Accuracy
530
+ value: 53
531
+ - task:
532
+ type: Sentence completion
533
+ dataset:
534
+ type: xcopa
535
+ name: XCOPA (vi)
536
+ config: vi
537
+ split: validation
538
+ revision: 37f73c60fb123111fa5af5f9b705d0b3747fd187
539
+ metrics:
540
+ - type: Accuracy
541
+ value: 79
542
+ - task:
543
+ type: Sentence completion
544
+ dataset:
545
+ type: xcopa
546
+ name: XCOPA (zh)
547
+ config: zh
548
+ split: validation
549
+ revision: 37f73c60fb123111fa5af5f9b705d0b3747fd187
550
+ metrics:
551
+ - type: Accuracy
552
+ value: 81
553
+ - task:
554
+ type: Sentence completion
555
+ dataset:
556
+ type: Muennighoff/xstory_cloze
557
+ name: XStoryCloze (ar)
558
+ config: ar
559
+ split: validation
560
+ revision: 8bb76e594b68147f1a430e86829d07189622b90d
561
+ metrics:
562
+ - type: Accuracy
563
+ value: 83.26
564
+ - task:
565
+ type: Sentence completion
566
+ dataset:
567
+ type: Muennighoff/xstory_cloze
568
+ name: XStoryCloze (es)
569
+ config: es
570
+ split: validation
571
+ revision: 8bb76e594b68147f1a430e86829d07189622b90d
572
+ metrics:
573
+ - type: Accuracy
574
+ value: 88.95
575
+ - task:
576
+ type: Sentence completion
577
+ dataset:
578
+ type: Muennighoff/xstory_cloze
579
+ name: XStoryCloze (eu)
580
+ config: eu
581
+ split: validation
582
+ revision: 8bb76e594b68147f1a430e86829d07189622b90d
583
+ metrics:
584
+ - type: Accuracy
585
+ value: 73.33
586
+ - task:
587
+ type: Sentence completion
588
+ dataset:
589
+ type: Muennighoff/xstory_cloze
590
+ name: XStoryCloze (hi)
591
+ config: hi
592
+ split: validation
593
+ revision: 8bb76e594b68147f1a430e86829d07189622b90d
594
+ metrics:
595
+ - type: Accuracy
596
+ value: 80.61
597
+ - task:
598
+ type: Sentence completion
599
+ dataset:
600
+ type: Muennighoff/xstory_cloze
601
+ name: XStoryCloze (id)
602
+ config: id
603
+ split: validation
604
+ revision: 8bb76e594b68147f1a430e86829d07189622b90d
605
+ metrics:
606
+ - type: Accuracy
607
+ value: 84.25
608
+ - task:
609
+ type: Sentence completion
610
+ dataset:
611
+ type: Muennighoff/xstory_cloze
612
+ name: XStoryCloze (my)
613
+ config: my
614
+ split: validation
615
+ revision: 8bb76e594b68147f1a430e86829d07189622b90d
616
+ metrics:
617
+ - type: Accuracy
618
+ value: 52.55
619
+ - task:
620
+ type: Sentence completion
621
+ dataset:
622
+ type: Muennighoff/xstory_cloze
623
+ name: XStoryCloze (ru)
624
+ config: ru
625
+ split: validation
626
+ revision: 8bb76e594b68147f1a430e86829d07189622b90d
627
+ metrics:
628
+ - type: Accuracy
629
+ value: 65.32
630
+ - task:
631
+ type: Sentence completion
632
+ dataset:
633
+ type: Muennighoff/xstory_cloze
634
+ name: XStoryCloze (sw)
635
+ config: sw
636
+ split: validation
637
+ revision: 8bb76e594b68147f1a430e86829d07189622b90d
638
+ metrics:
639
+ - type: Accuracy
640
+ value: 71.67
641
+ - task:
642
+ type: Sentence completion
643
+ dataset:
644
+ type: Muennighoff/xstory_cloze
645
+ name: XStoryCloze (te)
646
+ config: te
647
+ split: validation
648
+ revision: 8bb76e594b68147f1a430e86829d07189622b90d
649
+ metrics:
650
+ - type: Accuracy
651
+ value: 74.72
652
+ - task:
653
+ type: Sentence completion
654
+ dataset:
655
+ type: Muennighoff/xstory_cloze
656
+ name: XStoryCloze (zh)
657
+ config: zh
658
+ split: validation
659
+ revision: 8bb76e594b68147f1a430e86829d07189622b90d
660
+ metrics:
661
+ - type: Accuracy
662
+ value: 85.37
663
+ duplicated_from: bigscience/bloomz-7b1
664
+ ---
665
+
666
+ ![xmtf](https://github.com/bigscience-workshop/xmtf/blob/master/xmtf_banner.png?raw=true)
667
+
668
+ # Table of Contents
669
+
670
+ 1. [Model Summary](#model-summary)
671
+ 2. [Use](#use)
672
+ 3. [Limitations](#limitations)
673
+ 4. [Training](#training)
674
+ 5. [Evaluation](#evaluation)
675
+ 7. [Citation](#citation)
676
+
677
+ # Model Summary
678
+
679
+ > We present BLOOMZ & mT0, a family of models capable of following human instructions in dozens of languages zero-shot. We finetune BLOOM & mT5 pretrained multilingual language models on our crosslingual task mixture (xP3) and find the resulting models capable of crosslingual generalization to unseen tasks & languages.
680
+
681
+ - **Repository:** [bigscience-workshop/xmtf](https://github.com/bigscience-workshop/xmtf)
682
+ - **Paper:** [Crosslingual Generalization through Multitask Finetuning](https://arxiv.org/abs/2211.01786)
683
+ - **Point of Contact:** [Niklas Muennighoff](mailto:niklas@hf.co)
684
+ - **Languages:** Refer to [bloom](https://huggingface.co/bigscience/bloom) for pretraining & [xP3](https://huggingface.co/datasets/bigscience/xP3) for finetuning language proportions. It understands both pretraining & finetuning languages.
685
+ - **BLOOMZ & mT0 Model Family:**
686
+
687
+ <div class="max-w-full overflow-auto">
688
+ <table>
689
+ <tr>
690
+ <th colspan="12">Multitask finetuned on <a style="font-weight:bold" href=https://huggingface.co/datasets/bigscience/xP3>xP3</a>. Recommended for prompting in English.
691
+ </tr>
692
+ <tr>
693
+ <td>Parameters</td>
694
+ <td>300M</td>
695
+ <td>580M</td>
696
+ <td>1.2B</td>
697
+ <td>3.7B</td>
698
+ <td>13B</td>
699
+ <td>560M</td>
700
+ <td>1.1B</td>
701
+ <td>1.7B</td>
702
+ <td>3B</td>
703
+ <td>7.1B</td>
704
+ <td>176B</td>
705
+ </tr>
706
+ <tr>
707
+ <td>Finetuned Model</td>
708
+ <td><a href=https://huggingface.co/bigscience/mt0-small>mt0-small</a></td>
709
+ <td><a href=https://huggingface.co/bigscience/mt0-base>mt0-base</a></td>
710
+ <td><a href=https://huggingface.co/bigscience/mt0-large>mt0-large</a></td>
711
+ <td><a href=https://huggingface.co/bigscience/mt0-xl>mt0-xl</a></td>
712
+ <td><a href=https://huggingface.co/bigscience/mt0-xxl>mt0-xxl</a></td>
713
+ <td><a href=https://huggingface.co/bigscience/bloomz-560m>bloomz-560m</a></td>
714
+ <td><a href=https://huggingface.co/bigscience/bloomz-1b1>bloomz-1b1</a></td>
715
+ <td><a href=https://huggingface.co/bigscience/bloomz-1b7>bloomz-1b7</a></td>
716
+ <td><a href=https://huggingface.co/bigscience/bloomz-3b>bloomz-3b</a></td>
717
+ <td><a href=https://huggingface.co/bigscience/bloomz-7b1>bloomz-7b1</a></td>
718
+ <td><a href=https://huggingface.co/bigscience/bloomz>bloomz</a></td>
719
+ </tr>
720
+ </tr>
721
+ <tr>
722
+ <th colspan="12">Multitask finetuned on <a style="font-weight:bold" href=https://huggingface.co/datasets/bigscience/xP3mt>xP3mt</a>. Recommended for prompting in non-English.</th>
723
+ </tr>
724
+ <tr>
725
+ <td>Finetuned Model</td>
726
+ <td></td>
727
+ <td></td>
728
+ <td></td>
729
+ <td></td>
730
+ <td><a href=https://huggingface.co/bigscience/mt0-xxl-mt>mt0-xxl-mt</a></td>
731
+ <td></td>
732
+ <td></td>
733
+ <td></td>
734
+ <td></td>
735
+ <td><a href=https://huggingface.co/bigscience/bloomz-7b1-mt>bloomz-7b1-mt</a></td>
736
+ <td><a href=https://huggingface.co/bigscience/bloomz-mt>bloomz-mt</a></td>
737
+ </tr>
738
+ <th colspan="12">Multitask finetuned on <a style="font-weight:bold" href=https://huggingface.co/datasets/Muennighoff/P3>P3</a>. Released for research purposes only. Strictly inferior to above models!</th>
739
+ </tr>
740
+ <tr>
741
+ <td>Finetuned Model</td>
742
+ <td></td>
743
+ <td></td>
744
+ <td></td>
745
+ <td></td>
746
+ <td><a href=https://huggingface.co/bigscience/mt0-xxl-p3>mt0-xxl-p3</a></td>
747
+ <td></td>
748
+ <td></td>
749
+ <td></td>
750
+ <td></td>
751
+ <td><a href=https://huggingface.co/bigscience/bloomz-7b1-p3>bloomz-7b1-p3</a></td>
752
+ <td><a href=https://huggingface.co/bigscience/bloomz-p3>bloomz-p3</a></td>
753
+ </tr>
754
+ <th colspan="12">Original pretrained checkpoints. Not recommended.</th>
755
+ <tr>
756
+ <td>Pretrained Model</td>
757
+ <td><a href=https://huggingface.co/google/mt5-small>mt5-small</a></td>
758
+ <td><a href=https://huggingface.co/google/mt5-base>mt5-base</a></td>
759
+ <td><a href=https://huggingface.co/google/mt5-large>mt5-large</a></td>
760
+ <td><a href=https://huggingface.co/google/mt5-xl>mt5-xl</a></td>
761
+ <td><a href=https://huggingface.co/google/mt5-xxl>mt5-xxl</a></td>
762
+ <td><a href=https://huggingface.co/bigscience/bloom-560m>bloom-560m</a></td>
763
+ <td><a href=https://huggingface.co/bigscience/bloom-1b1>bloom-1b1</a></td>
764
+ <td><a href=https://huggingface.co/bigscience/bloom-1b7>bloom-1b7</a></td>
765
+ <td><a href=https://huggingface.co/bigscience/bloom-3b>bloom-3b</a></td>
766
+ <td><a href=https://huggingface.co/bigscience/bloom-7b1>bloom-7b1</a></td>
767
+ <td><a href=https://huggingface.co/bigscience/bloom>bloom</a></td>
768
+ </tr>
769
+ </table>
770
+ </div>
771
+
772
+
773
+ # Use
774
+
775
+ ## Intended use
776
+
777
+ We recommend using the model to perform tasks expressed in natural language. For example, given the prompt "*Translate to English: Je t’aime.*", the model will most likely answer "*I love you.*". Some prompt ideas from our paper:
778
+ - 一个传奇的开端,一个不灭的神话,这不仅仅是一部电影,而是作为一个走进新时代的标签,永远彪炳史册。你认为这句话的立场是赞扬、中立还是批评?
779
+ - Suggest at least five related search terms to "Mạng neural nhân tạo".
780
+ - Write a fairy tale about a troll saving a princess from a dangerous dragon. The fairy tale is a masterpiece that has achieved praise worldwide and its moral is "Heroes Come in All Shapes and Sizes". Story (in Spanish):
781
+ - Explain in a sentence in Telugu what is backpropagation in neural networks.
782
+
783
+ **Feel free to share your generations in the Community tab!**
784
+
785
+ ## How to use
786
+
787
+ ### CPU
788
+
789
+ <details>
790
+ <summary> Click to expand </summary>
791
+
792
+ ```python
793
+ # pip install -q transformers
794
+ from transformers import AutoModelForCausalLM, AutoTokenizer
795
+
796
+ checkpoint = "bigscience/bloomz-7b1"
797
+
798
+ tokenizer = AutoTokenizer.from_pretrained(checkpoint)
799
+ model = AutoModelForCausalLM.from_pretrained(checkpoint)
800
+
801
+ inputs = tokenizer.encode("Translate to English: Je t’aime.", return_tensors="pt")
802
+ outputs = model.generate(inputs)
803
+ print(tokenizer.decode(outputs[0]))
804
+ ```
805
+
806
+ </details>
807
+
808
+ ### GPU
809
+
810
+ <details>
811
+ <summary> Click to expand </summary>
812
+
813
+ ```python
814
+ # pip install -q transformers accelerate
815
+ from transformers import AutoModelForCausalLM, AutoTokenizer
816
+
817
+ checkpoint = "bigscience/bloomz-7b1"
818
+
819
+ tokenizer = AutoTokenizer.from_pretrained(checkpoint)
820
+ model = AutoModelForCausalLM.from_pretrained(checkpoint, torch_dtype="auto", device_map="auto")
821
+
822
+ inputs = tokenizer.encode("Translate to English: Je t’aime.", return_tensors="pt").to("cuda")
823
+ outputs = model.generate(inputs)
824
+ print(tokenizer.decode(outputs[0]))
825
+ ```
826
+
827
+ </details>
828
+
829
+ ### GPU in 8bit
830
+
831
+ <details>
832
+ <summary> Click to expand </summary>
833
+
834
+ ```python
835
+ # pip install -q transformers accelerate bitsandbytes
836
+ from transformers import AutoModelForCausalLM, AutoTokenizer
837
+
838
+ checkpoint = "bigscience/bloomz-7b1"
839
+
840
+ tokenizer = AutoTokenizer.from_pretrained(checkpoint)
841
+ model = AutoModelForCausalLM.from_pretrained(checkpoint, device_map="auto", load_in_8bit=True)
842
+
843
+ inputs = tokenizer.encode("Translate to English: Je t’aime.", return_tensors="pt").to("cuda")
844
+ outputs = model.generate(inputs)
845
+ print(tokenizer.decode(outputs[0]))
846
+ ```
847
+
848
+ </details>
849
+
850
+ <!-- Necessary for whitespace -->
851
+ ###
852
+
853
+ # Limitations
854
+
855
+ **Prompt Engineering:** The performance may vary depending on the prompt. For BLOOMZ models, we recommend making it very clear when the input stops to avoid the model trying to continue it. For example, the prompt "*Translate to English: Je t'aime*" without the full stop (.) at the end, may result in the model trying to continue the French sentence. Better prompts are e.g. "*Translate to English: Je t'aime.*", "*Translate to English: Je t'aime. Translation:*" "*What is "Je t'aime." in English?*", where it is clear for the model when it should answer. Further, we recommend providing the model as much context as possible. For example, if you want it to answer in Telugu, then tell the model, e.g. "*Explain in a sentence in Telugu what is backpropagation in neural networks.*".
856
+
857
+ # Training
858
+
859
+ ## Model
860
+
861
+ - **Architecture:** Same as [bloom-7b1](https://huggingface.co/bigscience/bloom-7b1), also refer to the `config.json` file
862
+ - **Finetuning steps:** 1000
863
+ - **Finetuning tokens:** 4.19 billion
864
+ - **Finetuning layout:** 1x pipeline parallel, 1x tensor parallel, 64x data parallel
865
+ - **Precision:** float16
866
+
867
+ ## Hardware
868
+
869
+ - **CPUs:** AMD CPUs with 512GB memory per node
870
+ - **GPUs:** 64 A100 80GB GPUs with 8 GPUs per node (8 nodes) using NVLink 4 inter-gpu connects, 4 OmniPath links
871
+ - **Communication:** NCCL-communications network with a fully dedicated subnet
872
+
873
+ ## Software
874
+
875
+ - **Orchestration:** [Megatron-DeepSpeed](https://github.com/bigscience-workshop/Megatron-DeepSpeed)
876
+ - **Optimizer & parallelism:** [DeepSpeed](https://github.com/microsoft/DeepSpeed)
877
+ - **Neural networks:** [PyTorch](https://github.com/pytorch/pytorch) (pytorch-1.11 w/ CUDA-11.5)
878
+ - **FP16 if applicable:** [apex](https://github.com/NVIDIA/apex)
879
+
880
+ # Evaluation
881
+
882
+ We refer to Table 7 from our [paper](https://arxiv.org/abs/2211.01786) & [bigscience/evaluation-results](https://huggingface.co/datasets/bigscience/evaluation-results) for zero-shot results on unseen tasks. The sidebar reports zero-shot performance of the best prompt per dataset config.
883
+
884
+ # Citation
885
+ ```bibtex
886
+ @misc{muennighoff2022crosslingual,
887
+ title={Crosslingual Generalization through Multitask Finetuning},
888
+ author={Niklas Muennighoff and Thomas Wang and Lintang Sutawika and Adam Roberts and Stella Biderman and Teven Le Scao and M Saiful Bari and Sheng Shen and Zheng-Xin Yong and Hailey Schoelkopf and Xiangru Tang and Dragomir Radev and Alham Fikri Aji and Khalid Almubarak and Samuel Albanie and Zaid Alyafeai and Albert Webson and Edward Raff and Colin Raffel},
889
+ year={2022},
890
+ eprint={2211.01786},
891
+ archivePrefix={arXiv},
892
+ primaryClass={cs.CL}
893
+ }
894
+ ```
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "apply_residual_connection_post_layernorm": false,
3
+ "architectures": [
4
+ "BloomForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "attention_softmax_in_fp32": true,
8
+ "bias_dropout_fusion": true,
9
+ "bos_token_id": 1,
10
+ "eos_token_id": 2,
11
+ "hidden_dropout": 0.0,
12
+ "initializer_range": 0.02,
13
+ "layer_norm_epsilon": 1e-05,
14
+ "masked_softmax_fusion": true,
15
+ "model_type": "bloom",
16
+ "n_embed": 4096,
17
+ "n_inner": null,
18
+ "n_layer": 30,
19
+ "num_attention_heads": 32,
20
+ "offset_alibi": 100,
21
+ "pad_token_id": 3,
22
+ "pretraining_tp": 4,
23
+ "seq_length": 2048,
24
+ "skip_bias_add": true,
25
+ "skip_bias_add_qkv": false,
26
+ "slow_but_exact": false,
27
+ "transformers_version": "4.21.0.dev0",
28
+ "unk_token_id": 0,
29
+ "use_cache": true,
30
+ "vocab_size": 250880
31
+ }
logs/logs/main_log.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:524a637d585f893f1ea43a0f461af047926be408fe9971e14e9d1397d47e32fd
3
+ size 233998372
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1f5d9fc449dd5fbe045f62c609e2b70f63d2e63faf74d7e0adb895dc294f28e
3
+ size 14138162687
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"}
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fa39cd4b1500feb205bcce3b9703a4373414cafe4970e0657b413f7ddd2a9d3
3
+ size 14500438
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "<unk>", "eos_token": "</s>", "bos_token": "<s>", "pad_token": "<pad>", "name_or_path": "bigscience/tokenizer", "special_tokens_map_file": null, "tokenizer_class": "BloomTokenizerFast", "padding_side": "left"}