Crystalcareai
commited on
Update configuration_gemmoe.py
Browse files- configuration_gemmoe.py +1 -3
configuration_gemmoe.py
CHANGED
@@ -118,7 +118,7 @@ class GemmoeConfig(PretrainedConfig):
|
|
118 |
num_attention_heads=16,
|
119 |
num_key_value_heads=16,
|
120 |
head_dim=256,
|
121 |
-
hidden_act="
|
122 |
max_position_embeddings=8192,
|
123 |
initializer_range=0.02,
|
124 |
rms_norm_eps=1e-6,
|
@@ -126,7 +126,6 @@ class GemmoeConfig(PretrainedConfig):
|
|
126 |
pad_token_id=0,
|
127 |
eos_token_id=1,
|
128 |
bos_token_id=2,
|
129 |
-
hidden_act="gelu_pytorch_tanh",
|
130 |
hidden_activation=None,
|
131 |
tie_word_embeddings=True,
|
132 |
rope_theta=10000.0,
|
@@ -148,7 +147,6 @@ class GemmoeConfig(PretrainedConfig):
|
|
148 |
self.hidden_act = hidden_act
|
149 |
self.hidden_activation = hidden_activation
|
150 |
self.num_key_value_heads = num_key_value_heads
|
151 |
-
self.hidden_act = hidden_act
|
152 |
self.initializer_range = initializer_range
|
153 |
self.rms_norm_eps = rms_norm_eps
|
154 |
self.use_cache = use_cache
|
|
|
118 |
num_attention_heads=16,
|
119 |
num_key_value_heads=16,
|
120 |
head_dim=256,
|
121 |
+
hidden_act="gelu_pytorch_tanh",
|
122 |
max_position_embeddings=8192,
|
123 |
initializer_range=0.02,
|
124 |
rms_norm_eps=1e-6,
|
|
|
126 |
pad_token_id=0,
|
127 |
eos_token_id=1,
|
128 |
bos_token_id=2,
|
|
|
129 |
hidden_activation=None,
|
130 |
tie_word_embeddings=True,
|
131 |
rope_theta=10000.0,
|
|
|
147 |
self.hidden_act = hidden_act
|
148 |
self.hidden_activation = hidden_activation
|
149 |
self.num_key_value_heads = num_key_value_heads
|
|
|
150 |
self.initializer_range = initializer_range
|
151 |
self.rms_norm_eps = rms_norm_eps
|
152 |
self.use_cache = use_cache
|