n1ck-guo commited on
Commit
5973e3a
1 Parent(s): 26058cf

upload auto_gptq format

Browse files

Signed-off-by: n1ck-guo <heng.guo@intel.com>

config.json CHANGED
@@ -20,23 +20,28 @@
20
  "partial_rotary_factor": 0.4,
21
  "qk_layernorm": false,
22
  "quantization_config": {
23
- "autoround_version": "0.2.0.dev",
 
24
  "bits": 4,
25
  "damp_percent": 0.01,
 
26
  "desc_act": false,
27
  "enable_minmax_tuning": true,
 
28
  "enable_quanted_input": true,
 
29
  "group_size": 128,
30
- "is_marlin_format": false,
31
  "iters": 1000,
 
32
  "lr": 0.001,
33
  "minmax_lr": 0.001,
34
- "model_file_base_name": "model",
35
- "model_name_or_path": null,
36
  "quant_method": "gptq",
37
- "scale_dtype": "float16",
38
- "static_groups": false,
39
  "sym": true,
 
40
  "true_sequential": false
41
  },
42
  "resid_pdrop": 0.1,
@@ -44,7 +49,7 @@
44
  "rope_theta": 10000.0,
45
  "tie_word_embeddings": false,
46
  "torch_dtype": "float16",
47
- "transformers_version": "4.40.2",
48
  "use_cache": true,
49
  "vocab_size": 51200
50
  }
 
20
  "partial_rotary_factor": 0.4,
21
  "qk_layernorm": false,
22
  "quantization_config": {
23
+ "amp": true,
24
+ "autoround_version": "0.3.1.dev",
25
  "bits": 4,
26
  "damp_percent": 0.01,
27
+ "data_type": "int",
28
  "desc_act": false,
29
  "enable_minmax_tuning": true,
30
+ "enable_norm_bias_tuning": false,
31
  "enable_quanted_input": true,
32
+ "gradient_accumulate_steps": 1,
33
  "group_size": 128,
 
34
  "iters": 1000,
35
+ "low_gpu_mem_usage": false,
36
  "lr": 0.001,
37
  "minmax_lr": 0.001,
38
+ "nsamples": 512,
39
+ "quant_block_list": null,
40
  "quant_method": "gptq",
41
+ "scale_dtype": "torch.float16",
42
+ "seqlen": 2048,
43
  "sym": true,
44
+ "train_bs": 8,
45
  "true_sequential": false
46
  },
47
  "resid_pdrop": 0.1,
 
49
  "rope_theta": 10000.0,
50
  "tie_word_embeddings": false,
51
  "torch_dtype": "float16",
52
+ "transformers_version": "4.44.2",
53
  "use_cache": true,
54
  "vocab_size": 51200
55
  }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "4.44.2"
6
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:14c3ff2501ea2449bd14376ffe6ed545a3e6b04c0e273686acfc6f9cbe14cf22
3
- size 1836707656
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb0bce8149e52d9bc6a535eae55b31f7f14d40341e3e0770380f4e2b2f8bb0cd
3
+ size 1833738008
quantize_config.json CHANGED
@@ -1,20 +1,25 @@
1
  {
2
  "bits": 4,
3
  "group_size": 128,
4
- "damp_percent": 0.01,
5
- "desc_act": false,
6
- "static_groups": false,
7
  "sym": true,
8
- "true_sequential": false,
9
- "model_name_or_path": null,
10
- "model_file_base_name": "model",
11
- "is_marlin_format": false,
12
- "quant_method": "intel/auto-round",
13
- "autoround_version": "0.2.0.dev",
14
- "iters": 1000,
15
  "lr": 0.001,
16
  "minmax_lr": 0.001,
17
- "enable_minmax_tuning": true,
18
- "enable_quanted_input": true,
19
- "scale_dtype": "float16"
 
 
 
 
 
 
 
 
 
20
  }
 
1
  {
2
  "bits": 4,
3
  "group_size": 128,
 
 
 
4
  "sym": true,
5
+ "data_type": "int",
6
+ "enable_quanted_input": true,
7
+ "enable_minmax_tuning": true,
8
+ "seqlen": 2048,
9
+ "train_bs": 8,
10
+ "scale_dtype": "torch.float16",
 
11
  "lr": 0.001,
12
  "minmax_lr": 0.001,
13
+ "gradient_accumulate_steps": 1,
14
+ "iters": 1000,
15
+ "amp": true,
16
+ "nsamples": 512,
17
+ "low_gpu_mem_usage": false,
18
+ "quant_block_list": null,
19
+ "enable_norm_bias_tuning": false,
20
+ "autoround_version": "0.3.1.dev",
21
+ "quant_method": "gptq",
22
+ "desc_act": false,
23
+ "true_sequential": false,
24
+ "damp_percent": 0.01
25
  }
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 2048,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {