sharpenb commited on
Commit
cf6a5f0
1 Parent(s): 35b52f7

Upload folder using huggingface_hub (#1)

Browse files

- 6ecb23591a48b6867ebfc8b828ca5c5b2db56f36cbb440887355db94b8785885 (7fc134bdee6fd5c077f033c47ac700d6b42a11dd)
- d70194bd13bc26b0a6ce9d16e101473d9701f0e0e78629076209a960d7ad1a58 (0c247cc2aeb5d8df921c729e6cff1ea3247b913c)
- 4684530db0fc07d5f647fe42060a152945a92e622832321133de7d02698ea3bc (073c734e2f93db930c4596a47af6b5edff801e0d)
- dbde3ec763a36a4e0e76e0e645fd52e0049d5967dc373c4ae726aa8db2775604 (9b727baf7e3b272d8136624782f56a9918de9355)
- f246ba5e5badf170e1cec4afe9c17d65903c82fdb80d8b6fb255f680d4c832e4 (8c3c3edc84a816b958f612ba17299b93761570f4)
- 3fbf90629794ee806458feb06feb635156129cec51997e61034c5adc454ccb12 (5d0a9c5e433020e0f93bcd6b36f843bd47280743)
- 6548b06326cc41a1184599a448cdf7b98b873ef30aab3c025b45c931f4f49872 (7c409b491ed14be32b1763f5782324a8ef983c32)
- 1bbe04e8e77ddeda22ef180038a73a7914867c4a5364df8dd3c1fd1c69860940 (80b9ec5fd2dbb7ea9ef32462556ded2dabf7a186)
- 103cbb32264220012ef129cea62213d3c369affdfb13db999e0bd74a897f640a (72da5283ccd74679349e12b0772a341b5ff63748)
- 6afaf69382c7e0263219f1f8475cd80bc9a399843d4b89a8171a606fde76be98 (4ecb7d30b56571899d1b03800c70a2fc0317b876)
- 243118b3a560eb61ffe537c63d6e0775a363e99c693300ea1c388a2a7e148194 (40a1458f6efd3d067892f6756d587669e7f6b2d8)
- a8f734997758ef99816736d5ec1967d82ce8a15479075b88b25d328c69d47348 (2fb6e48c13fe78ae21020574ff88e0d79be3e86b)
- 0c7fad0c7b366059590dd3e2ba5a7ec741899b8567d70577874119e5f29de023 (eaf820965e261f20d9786d859765fa503ca21a0e)
- 9bd2115bd6c3ee081171cbcddbf4c71f204319fa13155bc10875acfc46f95c8a (d98e96f928c784705c3b0480dde25dce3b1ac1a8)
- cdaa10a48297b297e09c4302ba2e18eda011271fef6f9141971c63283ee220c8 (dd2fa4e07635f6c077b74fad9b6c9de30393a3df)
- aeb865b243028c638d287c8bf6074e906ca59226b53a3873860a67e219f9436c (660f4a7d57fd46c8570f4be5d203008b069a2bc9)
- f4fbc5352335cc478b6e1ac03b0a55e4688df9da7d38ab22d070ed7fb94071e0 (0d369323c0280741a045e0985b48bcb6a840df69)

config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "v2ray/Mixtral-8x22B-v0.1",
3
+ "architectures": [
4
+ "MixtralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 6144,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 16384,
13
+ "max_position_embeddings": 65536,
14
+ "model_type": "mixtral",
15
+ "num_attention_heads": 48,
16
+ "num_experts_per_tok": 2,
17
+ "num_hidden_layers": 56,
18
+ "num_key_value_heads": 8,
19
+ "num_local_experts": 8,
20
+ "output_router_logits": false,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "float32",
25
+ "bnb_4bit_quant_storage": "uint8",
26
+ "bnb_4bit_quant_type": "fp4",
27
+ "bnb_4bit_use_double_quant": false,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ },
36
+ "rms_norm_eps": 1e-05,
37
+ "rope_theta": 1000000,
38
+ "router_aux_loss_coef": 0.001,
39
+ "router_jitter_noise": 0.0,
40
+ "sliding_window": null,
41
+ "tie_word_embeddings": false,
42
+ "torch_dtype": "float16",
43
+ "transformers_version": "4.40.0.dev0",
44
+ "use_cache": true,
45
+ "vocab_size": 32000
46
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.40.0.dev0"
6
+ }
model-00001-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af5378c0a4bed3863f26e6b65558f37e4d824eb095eb389b7627f926308d7767
3
+ size 4951626618
model-00002-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4e3ac4e6ba722998b646a20055b28bc5d96b458457e2d84f071fb53d540b3a0
3
+ size 4961824784
model-00003-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:473456eedd4c4cdd8f2f228d180bef29e524fa3401db4c2aeeebf5418119fe7b
3
+ size 4954801840
model-00004-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:887263a956a21493f7a5e7b8f5260d89a83ccd65986cc16763f0259685c2ff3d
3
+ size 4961825192
model-00005-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efa1e6d8e4ad38636e4ba8f11c38a50161cb85ed82d17fd8239d740e9547dfa7
3
+ size 4954802256
model-00006-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8e70fa3c2bde8f5924224ba85f3256146c7ec585f3771634af108e4f32a4785
3
+ size 4961825192
model-00007-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0445dc8f86c2fcfb207a8077ab6218a7cef652682e3b32bbca91b0da2ecc8165
3
+ size 4954802256
model-00008-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fcc62ae79d78295adb99e570e788542d067b9caa1ac1bafb7877f7f4eb8be80b
3
+ size 4961825192
model-00009-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9cd84e01652c02750189919683913d1821a57a44e1d1a481216c418df31e710
3
+ size 4954802256
model-00010-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42db7e99ec200b68d74e5fbefcbfb739bc9328475eeb9e9bacf3e1ed2da162b8
3
+ size 4961825192
model-00011-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84b63e7417ebd8a84ed6d47e8bdd748e3c3a3f8cc8d3e1719f0097db9c82222b
3
+ size 4954802256
model-00012-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6974f9361f892ccb15a0869372aaf032525b0a77d117597a01c1c3696558773
3
+ size 4961825192
model-00013-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a54b97b419c07fe7612aa3b0ea0ff68d9fd7e84d9a87879b3c5b623f7d8c4df0
3
+ size 4954802256
model-00014-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9cc813539c5a1e3593b13f8b3906576e4f08aea087260846242b445a914ea76
3
+ size 4990163459
model-00015-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe8dd6215f686b5d83af6e5386ed07570db21d65ad747dac7f858b65357259a8
3
+ size 4983087791
model-00016-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1e1553ebf287a0c6f2e481070887e0063346a474b71858e3d2b439335d67918
3
+ size 4848614772
model-00017-of-00017.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a9b656fa319bc8871482483a1345f0a9c4cf80cd53554668b81946e9b74223e
3
+ size 393216128
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "additional_special_tokens": [],
31
+ "bos_token": "<s>",
32
+ "clean_up_tokenization_spaces": false,
33
+ "eos_token": "</s>",
34
+ "legacy": true,
35
+ "model_max_length": 1000000000000000019884624838656,
36
+ "pad_token": null,
37
+ "sp_model_kwargs": {},
38
+ "spaces_between_special_tokens": false,
39
+ "tokenizer_class": "LlamaTokenizer",
40
+ "unk_token": "<unk>",
41
+ "use_default_system_prompt": false
42
+ }