spacemanidol commited on
Commit
40bfdaa
1 Parent(s): a36e048

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +5 -11
config.json CHANGED
@@ -1,17 +1,11 @@
1
  {
2
- "_name_or_path": "/data/.model_and_tokenizer_cache/eea825a7e388c292874bd28736f8bcd70ec7526432cd66fa94c540bd27834d22",
3
  "architectures": [
4
- "NewModel"
5
  ],
6
  "attention_probs_dropout_prob": 0.0,
7
  "auto_map": {
8
- "AutoConfig": "Alibaba-NLP/new-impl--configuration.NewConfig",
9
- "AutoModel": "Alibaba-NLP/new-impl--modeling.NewModel",
10
- "AutoModelForMaskedLM": "Alibaba-NLP/new-impl--modeling.NewForMaskedLM",
11
- "AutoModelForMultipleChoice": "Alibaba-NLP/new-impl--modeling.NewForMultipleChoice",
12
- "AutoModelForQuestionAnswering": "Alibaba-NLP/new-impl--modeling.NewForQuestionAnswering",
13
- "AutoModelForSequenceClassification": "Alibaba-NLP/new-impl--modeling.NewForSequenceClassification",
14
- "AutoModelForTokenClassification": "Alibaba-NLP/new-impl--modeling.NewForTokenClassification"
15
  },
16
  "classifier_dropout": 0.1,
17
  "hidden_act": "gelu",
@@ -24,7 +18,7 @@
24
  "logn_attention_clip1": false,
25
  "logn_attention_scale": false,
26
  "max_position_embeddings": 8192,
27
- "model_type": "new",
28
  "num_attention_heads": 12,
29
  "num_hidden_layers": 12,
30
  "pack_qkv": true,
@@ -38,4 +32,4 @@
38
  "unpad_inputs": "true",
39
  "use_memory_efficient_attention": "true",
40
  "vocab_size": 250048
41
- }
 
1
  {
 
2
  "architectures": [
3
+ "GteModel"
4
  ],
5
  "attention_probs_dropout_prob": 0.0,
6
  "auto_map": {
7
+ "AutoConfig": "configuration_hf_alibaba_nlp_gte.GteConfig",
8
+ "AutoModel": "modeling_hf_alibaba_nlp_gte.GteModel"
 
 
 
 
 
9
  },
10
  "classifier_dropout": 0.1,
11
  "hidden_act": "gelu",
 
18
  "logn_attention_clip1": false,
19
  "logn_attention_scale": false,
20
  "max_position_embeddings": 8192,
21
+ "model_type": "gte",
22
  "num_attention_heads": 12,
23
  "num_hidden_layers": 12,
24
  "pack_qkv": true,
 
32
  "unpad_inputs": "true",
33
  "use_memory_efficient_attention": "true",
34
  "vocab_size": 250048
35
+ }