BigHuggyD commited on
Commit
69e24cb
1 Parent(s): d8f0c1e

Add files using upload-large-folder tool

Browse files
config.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "gghfez/Writer-Large-2411-v2.1",
3
+ "architectures": [
4
+ "MistralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 12288,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 28672,
14
+ "max_position_embeddings": 131072,
15
+ "model_type": "mistral",
16
+ "num_attention_heads": 96,
17
+ "num_hidden_layers": 88,
18
+ "num_key_value_heads": 8,
19
+ "pad_token_id": 748,
20
+ "quantization_config": {
21
+ "config_groups": {
22
+ "group_0": {
23
+ "input_activations": {
24
+ "actorder": null,
25
+ "block_structure": null,
26
+ "dynamic": true,
27
+ "group_size": null,
28
+ "num_bits": 8,
29
+ "observer": null,
30
+ "observer_kwargs": {},
31
+ "strategy": "token",
32
+ "symmetric": true,
33
+ "type": "float"
34
+ },
35
+ "output_activations": null,
36
+ "targets": [
37
+ "Linear"
38
+ ],
39
+ "weights": {
40
+ "actorder": null,
41
+ "block_structure": null,
42
+ "dynamic": false,
43
+ "group_size": null,
44
+ "num_bits": 8,
45
+ "observer": "minmax",
46
+ "observer_kwargs": {},
47
+ "strategy": "channel",
48
+ "symmetric": true,
49
+ "type": "float"
50
+ }
51
+ }
52
+ },
53
+ "format": "float-quantized",
54
+ "global_compression_ratio": 1.4645315540129722,
55
+ "ignore": [
56
+ "lm_head"
57
+ ],
58
+ "kv_cache_scheme": null,
59
+ "quant_method": "compressed-tensors",
60
+ "quantization_status": "compressed"
61
+ },
62
+ "rms_norm_eps": 1e-05,
63
+ "rope_scaling": null,
64
+ "rope_theta": 1000000.0,
65
+ "sliding_window": null,
66
+ "tie_word_embeddings": false,
67
+ "torch_dtype": "bfloat16",
68
+ "transformers_version": "4.47.1",
69
+ "use_cache": true,
70
+ "vocab_size": 32768
71
+ }
generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "max_length": 131072,
6
+ "pad_token_id": 748,
7
+ "transformers_version": "4.47.1"
8
+ }
model-00001-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a58ba2961be47a46c8150e8a37f65dc6317ec56a9a4b80ee6b8f0bce685ecdce
3
+ size 4958398128
model-00002-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ec33ab2c5b1b41ef1097f0d0c87f24cde6eb4552e18313cd68589728aaa88f5
3
+ size 4832680488
model-00003-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8872b201ba8935ab328bca4710afeb4aa183c38586b8b1c10f7937c0fc47840
3
+ size 4857866304
model-00004-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf949eed3dbf85619557036cb502be862b81f52b20a4df8c8bb216ed69699431
3
+ size 4832680552
model-00005-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c45ecaa427d9ebd2341b153b5f95a7feeccf9b493dcf625110105815e5a1688
3
+ size 4857866352
model-00006-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:507f22464a90fb480851a11b55626632a46a55bcec7fc21f6004ff2c4cc347e3
3
+ size 4832680552
model-00007-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c384acac41df6105d4178e8614890a7d4f6ab3a3a945e77c36da054a7ca281d
3
+ size 4857866352
model-00008-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fb9c418207af1b7a2c25c94a05af49d2f5c1a856652178a5570d30b33f08dc4
3
+ size 4832680552
model-00009-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd2e9d7b28637d807c199ee0fd964d9498c2cdb8c9f101e2e7866012791c9ef3
3
+ size 4857866352
model-00010-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3984bb416d0f481d079698e6dea089134dddf1e964d96e3f1c9b0739aab6be9
3
+ size 4832680552
model-00011-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:689fa38297ec342ecd715f9e078aa6f2dc586f59b38feeb18526c283d1e7fecf
3
+ size 4857866352
model-00012-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ebeaef81db16d473662e3d90bd4903c13c10fc22b23e9cfb2bf64c11cb85e11
3
+ size 4832680552
model-00013-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cb63993c1916f413e2fbb22a35cb36e61ef38c9989b4888a97068effe34d4d6
3
+ size 4857866352
model-00014-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19d6e39d3239265540fb00803312ec3556c3a850535489748ad16bb142b33f43
3
+ size 4832680552
model-00015-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c663967712deac68c47fd749e99e01cae8b2832401f16d8cc7898d437ec208b
3
+ size 4857866352
model-00016-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fd3be054aad1b29c6f9d775596161ec425b4fb442d94da411c93e8aa5488f17
3
+ size 4832680552
model-00017-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62f0a51c330a4336f85a90696b8e8fad5afdb98f433b54668db9634ffaef3cb2
3
+ size 4857866352
model-00018-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05215e77eb1306399ab164b0c5e3eb30e3b550bb599c0485d47f0a3a6aa6eaad
3
+ size 4832680552
model-00019-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fda1af6f80b65fab2326e921cd0cc5f6a31469adfa9466d73b614d37f0c0445
3
+ size 4857866352
model-00020-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db48edf9b80b488e3fd7f6bd0c07b3a0ab883f5f53c48d5b83481423a7573396
3
+ size 4832680552
model-00021-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d62bb86c8483c453ca53d92dc319eed4d1d3816c444d96d1965cbd2d9d547082
3
+ size 4857866352
model-00022-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84fe5b3ebc302b0b80cbd762932fa84f0ccf36b3cb8ee1517280b2cb8cd58cf3
3
+ size 4832680552
model-00023-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:313023efa30d01da4c25bc38b0db4b72938a35e5ff13fe9b05641bc301f17514
3
+ size 4857866352
model-00024-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3194efd6f3eb742ff1a68628357c643ec239092ec66542960361d49271dda1c2
3
+ size 4832680552
model-00025-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2791e981db97e8d7c8b6b6d06a330c3e4c97a5ee0a22a6897f94f28a8b49e455
3
+ size 4857866352
model-00026-of-00026.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1352e0ecf4fd52e2ce6a8de625442c197c263220ee2969462bef702aa5d9e1c
3
+ size 2189695048
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
recipe.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ DEFAULT_stage:
2
+ DEFAULT_modifiers:
3
+ QuantizationModifier:
4
+ ignore: [lm_head]
5
+ targets: [Linear]
6
+ scheme: FP8_DYNAMIC
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[control_746]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff