Ubuntu commited on
Commit
ee343b5
1 Parent(s): aa58d72

fitst commit

Browse files
Files changed (1) hide show
  1. config.json +12 -5
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "THUDM/glm-4-9b-chat-1m",
3
  "model_type": "chatglm",
4
  "architectures": [
5
  "ChatGLMModel"
@@ -23,20 +23,27 @@
23
  "hidden_dropout": 0.0,
24
  "hidden_size": 4096,
25
  "kv_channels": 128,
26
- "layernorm_epsilon": 0.00000015625,
27
  "multi_query_attention": true,
28
  "multi_query_group_num": 2,
29
  "num_attention_heads": 32,
 
30
  "num_layers": 40,
 
31
  "original_rope": true,
32
  "padded_vocab_size": 151552,
33
  "post_layer_norm": true,
34
  "rmsnorm": true,
35
- "seq_length": 8192,
36
  "use_cache": true,
37
  "torch_dtype": "bfloat16",
38
  "transformers_version": "4.30.2",
39
  "tie_word_embeddings": false,
40
- "eos_token_id": [151329, 151336, 151338],
 
 
 
 
41
  "pad_token_id": 151329
42
- }
 
 
1
  {
2
+ "_name_or_path": "THUDM/glm4-9b-chat-1m",
3
  "model_type": "chatglm",
4
  "architectures": [
5
  "ChatGLMModel"
 
23
  "hidden_dropout": 0.0,
24
  "hidden_size": 4096,
25
  "kv_channels": 128,
26
+ "layernorm_epsilon": 1.5625e-07,
27
  "multi_query_attention": true,
28
  "multi_query_group_num": 2,
29
  "num_attention_heads": 32,
30
+ "num_hidden_layers": 40,
31
  "num_layers": 40,
32
+ "rope_ratio": 500,
33
  "original_rope": true,
34
  "padded_vocab_size": 151552,
35
  "post_layer_norm": true,
36
  "rmsnorm": true,
37
+ "seq_length": 131072,
38
  "use_cache": true,
39
  "torch_dtype": "bfloat16",
40
  "transformers_version": "4.30.2",
41
  "tie_word_embeddings": false,
42
+ "eos_token_id": [
43
+ 151329,
44
+ 151336,
45
+ 151338
46
+ ],
47
  "pad_token_id": 151329
48
+ }
49
+