Muennighoff commited on
Commit
a8ac646
1 Parent(s): d6a7a9d

Fix config

Browse files
Files changed (1) hide show
  1. config.json +12 -6
config.json CHANGED
@@ -1,18 +1,24 @@
1
  {
2
  "apply_residual_connection_post_layernorm": false,
3
  "attention_dropout": 0.0,
 
 
 
 
 
4
  "bos_token_id": 1,
5
  "eos_token_id": 2,
6
  "hidden_dropout": 0.0,
7
- "hidden_size": 64,
8
  "initializer_range": 0.02,
9
  "layer_norm_epsilon": 1e-05,
 
10
  "model_type": "bloom",
11
- "n_head": 8,
12
- "n_layer": 2,
13
- "pretraining_tp": 1,
 
14
  "slow_but_exact": false,
15
- "transformers_version": "4.27.2",
16
  "use_cache": true,
17
  "vocab_size": 250880
18
- }
 
1
  {
2
  "apply_residual_connection_post_layernorm": false,
3
  "attention_dropout": 0.0,
4
+ "architectures": [
5
+ "BloomForCausalLM"
6
+ ],
7
+ "attention_softmax_in_fp32": true,
8
+ "pad_token_id": 3,
9
  "bos_token_id": 1,
10
  "eos_token_id": 2,
11
  "hidden_dropout": 0.0,
 
12
  "initializer_range": 0.02,
13
  "layer_norm_epsilon": 1e-05,
14
+ "masked_softmax_fusion": true,
15
  "model_type": "bloom",
16
+ "n_embed": 14336,
17
+ "n_layer": 70,
18
+ "num_attention_heads": 112,
19
+ "pretraining_tp": 4,
20
  "slow_but_exact": false,
21
+ "transformers_version": "4.21.0",
22
  "use_cache": true,
23
  "vocab_size": 250880
24
+ }