francislabounty
commited on
Commit
•
b4ab75f
1
Parent(s):
3d92987
Update config.json
Browse files- config.json +5 -5
config.json
CHANGED
@@ -3,14 +3,14 @@
|
|
3 |
"adapter_dim": 512,
|
4 |
"adapter_dropout": 0.0,
|
5 |
"architectures": [
|
6 |
-
"
|
7 |
],
|
8 |
"attention_dropout": 0.0,
|
9 |
"auto_map": {
|
10 |
-
"AutoConfig": "
|
11 |
-
"AutoModel": "
|
12 |
-
"AutoModelForCausalLM": "
|
13 |
-
}
|
14 |
"bos_token_id": 1,
|
15 |
"eos_token_id": 2,
|
16 |
"hidden_act": "silu",
|
|
|
3 |
"adapter_dim": 512,
|
4 |
"adapter_dropout": 0.0,
|
5 |
"architectures": [
|
6 |
+
"modeling_sparsetral.MistralForCausalLM"
|
7 |
],
|
8 |
"attention_dropout": 0.0,
|
9 |
"auto_map": {
|
10 |
+
"AutoConfig": "configuration_sparsetral.SparsetralConfig",
|
11 |
+
"AutoModel": "modeling_sparsetral.MistralModel",
|
12 |
+
"AutoModelForCausalLM": "modeling_sparsetral.MistralForCausalLM"
|
13 |
+
}
|
14 |
"bos_token_id": 1,
|
15 |
"eos_token_id": 2,
|
16 |
"hidden_act": "silu",
|