bwang0911 commited on
Commit
6d072ad
1 Parent(s): a0b1894

Add new SentenceTransformer model

Browse files
1_Pooling/config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "word_embedding_dimension": 384,
3
  "pooling_mode_cls_token": false,
4
  "pooling_mode_mean_tokens": true,
5
  "pooling_mode_max_tokens": false,
 
1
  {
2
+ "word_embedding_dimension": 512,
3
  "pooling_mode_cls_token": false,
4
  "pooling_mode_mean_tokens": true,
5
  "pooling_mode_max_tokens": false,
README.md CHANGED
The diff for this file is too large to render. See raw diff
 
config.json CHANGED
@@ -1,26 +1,36 @@
1
  {
2
- "_name_or_path": "sentence-transformers/all-MiniLM-L6-v2",
3
  "architectures": [
4
- "BertModel"
5
  ],
6
- "attention_probs_dropout_prob": 0.1,
 
 
 
 
 
 
 
7
  "classifier_dropout": null,
 
 
8
  "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
- "hidden_size": 384,
12
  "initializer_range": 0.02,
13
- "intermediate_size": 1536,
14
  "layer_norm_eps": 1e-12,
15
- "max_position_embeddings": 512,
 
16
  "model_type": "bert",
17
- "num_attention_heads": 12,
18
- "num_hidden_layers": 6,
19
  "pad_token_id": 0,
20
- "position_embedding_type": "absolute",
21
  "torch_dtype": "float32",
22
  "transformers_version": "4.45.2",
23
  "type_vocab_size": 2,
24
  "use_cache": true,
25
- "vocab_size": 30522
26
  }
 
1
  {
2
+ "_name_or_path": "jinaai/jina-embeddings-v2-small-en",
3
  "architectures": [
4
+ "JinaBertModel"
5
  ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "attn_implementation": null,
8
+ "auto_map": {
9
+ "AutoConfig": "jinaai/jina-bert-implementation--configuration_bert.JinaBertConfig",
10
+ "AutoModel": "jinaai/jina-bert-implementation--modeling_bert.JinaBertModel",
11
+ "AutoModelForMaskedLM": "jinaai/jina-bert-implementation--modeling_bert.JinaBertForMaskedLM",
12
+ "AutoModelForSequenceClassification": "jinaai/jina-bert-implementation--modeling_bert.JinaBertForSequenceClassification"
13
+ },
14
  "classifier_dropout": null,
15
+ "emb_pooler": "mean",
16
+ "feed_forward_type": "geglu",
17
  "gradient_checkpointing": false,
18
  "hidden_act": "gelu",
19
  "hidden_dropout_prob": 0.1,
20
+ "hidden_size": 512,
21
  "initializer_range": 0.02,
22
+ "intermediate_size": 2048,
23
  "layer_norm_eps": 1e-12,
24
+ "max_position_embeddings": 8192,
25
+ "model_max_length": 8192,
26
  "model_type": "bert",
27
+ "num_attention_heads": 8,
28
+ "num_hidden_layers": 4,
29
  "pad_token_id": 0,
30
+ "position_embedding_type": "alibi",
31
  "torch_dtype": "float32",
32
  "transformers_version": "4.45.2",
33
  "type_vocab_size": 2,
34
  "use_cache": true,
35
+ "vocab_size": 30528
36
  }
config_sentence_transformers.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "__version__": {
3
- "sentence_transformers": "3.1.1",
4
  "transformers": "4.45.2",
5
  "pytorch": "2.5.1+cu124"
6
  },
7
  "prompts": {},
8
  "default_prompt_name": null,
9
- "similarity_fn_name": null
10
  }
 
1
  {
2
  "__version__": {
3
+ "sentence_transformers": "3.3.0",
4
  "transformers": "4.45.2",
5
  "pytorch": "2.5.1+cu124"
6
  },
7
  "prompts": {},
8
  "default_prompt_name": null,
9
+ "similarity_fn_name": "cosine"
10
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:26ef15b308eebb0feb61c73af3bfa5b1e6e3559b1d289ac9f895334276fe67b3
3
- size 90864192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e787f7cfdca1b5fdba32b7dbebcac86c5095c4f0558bbe2cb75f6d7a6f6d6db1
3
+ size 130769960
modules.json CHANGED
@@ -10,11 +10,5 @@
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
13
- },
14
- {
15
- "idx": 2,
16
- "name": "2",
17
- "path": "2_Normalize",
18
- "type": "sentence_transformers.models.Normalize"
19
  }
20
  ]
 
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
 
 
 
 
 
 
13
  }
14
  ]
sentence_bert_config.json CHANGED
@@ -1,4 +1,4 @@
1
  {
2
- "max_seq_length": 256,
3
  "do_lower_case": false
4
  }
 
1
  {
2
+ "max_seq_length": 512,
3
  "do_lower_case": false
4
  }
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 256,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
@@ -30678,7 +30678,13 @@
30678
  "##/": 30518,
30679
  "##:": 30519,
30680
  "##?": 30520,
30681
- "##~": 30521
 
 
 
 
 
 
30682
  }
30683
  }
30684
  }
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
30678
  "##/": 30518,
30679
  "##:": 30519,
30680
  "##?": 30520,
30681
+ "##~": 30521,
30682
+ "bowang": 30522,
30683
+ "georgiosmastrapas": 30523,
30684
+ "jackminong": 30524,
30685
+ "jonathangeuter": 30525,
30686
+ "louismilliken": 30526,
30687
+ "michaelguenther": 30527
30688
  }
30689
  }
30690
  }
tokenizer_config.json CHANGED
@@ -41,24 +41,17 @@
41
  "special": true
42
  }
43
  },
44
- "clean_up_tokenization_spaces": false,
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
  "do_lower_case": true,
48
  "mask_token": "[MASK]",
49
- "max_length": 128,
50
- "model_max_length": 256,
51
  "never_split": null,
52
- "pad_to_multiple_of": null,
53
  "pad_token": "[PAD]",
54
- "pad_token_type_id": 0,
55
- "padding_side": "right",
56
  "sep_token": "[SEP]",
57
- "stride": 0,
58
  "strip_accents": null,
59
  "tokenize_chinese_chars": true,
60
  "tokenizer_class": "BertTokenizer",
61
- "truncation_side": "right",
62
- "truncation_strategy": "longest_first",
63
  "unk_token": "[UNK]"
64
  }
 
41
  "special": true
42
  }
43
  },
44
+ "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
  "do_lower_case": true,
48
  "mask_token": "[MASK]",
49
+ "model_max_length": 8192,
 
50
  "never_split": null,
 
51
  "pad_token": "[PAD]",
 
 
52
  "sep_token": "[SEP]",
 
53
  "strip_accents": null,
54
  "tokenize_chinese_chars": true,
55
  "tokenizer_class": "BertTokenizer",
 
 
56
  "unk_token": "[UNK]"
57
  }
vocab.txt CHANGED
@@ -30520,3 +30520,9 @@ necessitated
30520
  ##:
30521
  ##?
30522
  ##~
 
 
 
 
 
 
 
30520
  ##:
30521
  ##?
30522
  ##~
30523
+ bowang
30524
+ georgiosmastrapas
30525
+ jackminong
30526
+ jonathangeuter
30527
+ louismilliken
30528
+ michaelguenther