danielhanchen commited on
Commit
b57b694
1 Parent(s): a299e11

Upload LlamaForCausalLM

Browse files
config.json CHANGED
@@ -11,7 +11,8 @@
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 14336,
14
- "max_position_embeddings": 8192,
 
15
  "model_type": "llama",
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
@@ -22,7 +23,7 @@
22
  "rope_theta": 500000.0,
23
  "tie_word_embeddings": false,
24
  "torch_dtype": "bfloat16",
25
- "transformers_version": "4.40.1",
26
  "use_cache": true,
27
  "vocab_size": 128256
28
  }
 
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 14336,
14
+ "max_position_embeddings": 4096,
15
+ "mlp_bias": false,
16
  "model_type": "llama",
17
  "num_attention_heads": 32,
18
  "num_hidden_layers": 32,
 
23
  "rope_theta": 500000.0,
24
  "tie_word_embeddings": false,
25
  "torch_dtype": "bfloat16",
26
+ "transformers_version": "4.41.0",
27
  "use_cache": true,
28
  "vocab_size": 128256
29
  }
generation_config.json CHANGED
@@ -5,5 +5,5 @@
5
  "max_length": 4096,
6
  "temperature": 0.6,
7
  "top_p": 0.9,
8
- "transformers_version": "4.40.1"
9
  }
 
5
  "max_length": 4096,
6
  "temperature": 0.6,
7
  "top_p": 0.9,
8
+ "transformers_version": "4.41.0"
9
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf617ab8b2d725d231a7a0dc1c1d55c53d310056b119b84de51eac9c82e542d9
3
  size 4976698672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2c144103072514542e327fa8080bd375cb300f2d453fba9ca3aea81d0d4cf33
3
  size 4976698672
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:29b234a42c7b589c25b9400167b36daa8e062f90b02f6e8d6fc8d83eb8a43b90
3
  size 1168138808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5dc34e6bdf2da9e35f0d93b5c333c870f3677dc43dc3a91ea3a8ad28a1fe1acb
3
  size 1168138808