mini1013 commited on
Commit
5c1f47b
1 Parent(s): 807751d

Upload folder using huggingface_hub

Browse files
README.md CHANGED
The diff for this file is too large to render. See raw diff
 
config.json CHANGED
@@ -22,7 +22,7 @@
22
  "position_embedding_type": "absolute",
23
  "tokenizer_class": "BertTokenizer",
24
  "torch_dtype": "float32",
25
- "transformers_version": "4.43.3",
26
  "type_vocab_size": 1,
27
  "use_cache": true,
28
  "vocab_size": 32000
 
22
  "position_embedding_type": "absolute",
23
  "tokenizer_class": "BertTokenizer",
24
  "torch_dtype": "float32",
25
+ "transformers_version": "4.45.1",
26
  "type_vocab_size": 1,
27
  "use_cache": true,
28
  "vocab_size": 32000
config_sentence_transformers.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "__version__": {
3
  "sentence_transformers": "3.1.1",
4
- "transformers": "4.43.3",
5
- "pytorch": "2.4.1+cu121"
6
  },
7
  "prompts": {},
8
  "default_prompt_name": null,
 
1
  {
2
  "__version__": {
3
  "sentence_transformers": "3.1.1",
4
+ "transformers": "4.45.1",
5
+ "pytorch": "2.4.0+cu121"
6
  },
7
  "prompts": {},
8
  "default_prompt_name": null,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:59071ecf142116f3f9bd618204bd89e10418ba8aed47e6e509b8bacbb24efca2
3
  size 442494816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6788024f3b6e00914993285e896621c8971919a012b77a215c2e6da6a0a09d08
3
  size 442494816
model_head.pkl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2c79fc2060ae4777c70d66926b89e3ddf07e372fcb32affbcbf8a1f88f8ea759
3
  size 111719
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:516c6d07b35a740d48a18d74eee6fdd3740c5b371e929bf2d7a4b4fcc6b5ef82
3
  size 111719
tokenizer_config.json CHANGED
@@ -42,7 +42,7 @@
42
  }
43
  },
44
  "bos_token": "[CLS]",
45
- "clean_up_tokenization_spaces": true,
46
  "cls_token": "[CLS]",
47
  "do_basic_tokenize": true,
48
  "do_lower_case": false,
 
42
  }
43
  },
44
  "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": false,
46
  "cls_token": "[CLS]",
47
  "do_basic_tokenize": true,
48
  "do_lower_case": false,