mp commited on
Commit
0cb216c
·
1 Parent(s): 51c95c2

first adaptation of config.json for HF AutoModel class

Browse files
Files changed (1) hide show
  1. config.json +7 -2
config.json CHANGED
@@ -1,7 +1,12 @@
1
  {
 
2
  "architectures": [
3
- "LuminousForEmbedding"
4
  ],
 
 
 
 
5
  "attention_bias": true,
6
  "attention_dropout": 0.0,
7
  "attn_adapter_config": {
@@ -30,7 +35,7 @@
30
  },
31
  "mlp_bias": true,
32
  "mode": "generation",
33
- "model_type": "bert",
34
  "num_attention_heads": 36,
35
  "num_hidden_layers": 27,
36
  "num_key_value_heads": 4,
 
1
  {
2
+ "_name_or_path": "Aleph-Alpha/Pharia-1-Embedding-4608-control-hf",
3
  "architectures": [
4
+ "Pharia4608EmbeddingModel"
5
  ],
6
+ "auto_map": {
7
+ "AutoConfig": "modeling_pharia.PhariaConfig",
8
+ "AutoModel": "modeling_pharia.PhariaForEmbedding"
9
+ },
10
  "attention_bias": true,
11
  "attention_dropout": 0.0,
12
  "attn_adapter_config": {
 
35
  },
36
  "mlp_bias": true,
37
  "mode": "generation",
38
+ "model_type": "phariaembed",
39
  "num_attention_heads": 36,
40
  "num_hidden_layers": 27,
41
  "num_key_value_heads": 4,