devngho commited on
Commit
e055039
·
verified ·
1 Parent(s): 77f1e7b

Upload LlamaForCausalLM

Browse files
Files changed (2) hide show
  1. config.json +5 -5
  2. model.safetensors +2 -2
config.json CHANGED
@@ -7,15 +7,15 @@
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
9
  "hidden_act": "silu",
10
- "hidden_size": 1536,
11
  "initializer_range": 0.02,
12
- "intermediate_size": 3072,
13
  "max_position_embeddings": 2048,
14
  "mlp_bias": false,
15
  "model_type": "llama",
16
- "num_attention_heads": 24,
17
- "num_hidden_layers": 16,
18
- "num_key_value_heads": 24,
19
  "pretraining_tp": 1,
20
  "rms_norm_eps": 1e-05,
21
  "rope_scaling": null,
 
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
9
  "hidden_act": "silu",
10
+ "hidden_size": 1024,
11
  "initializer_range": 0.02,
12
+ "intermediate_size": 4096,
13
  "max_position_embeddings": 2048,
14
  "mlp_bias": false,
15
  "model_type": "llama",
16
+ "num_attention_heads": 16,
17
+ "num_hidden_layers": 24,
18
+ "num_key_value_heads": 16,
19
  "pretraining_tp": 1,
20
  "rms_norm_eps": 1e-05,
21
  "rope_scaling": null,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ded7614137cc4bb6726fe27f475002d6879aab8d850e3dfc1d0839d87c87ff8d
3
- size 1878808928
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84c177aef0d3601b3abfed81564ce81e8b057a5c5c95d0ce256f6d760b41fcc2
3
+ size 1856598400