yongzx commited on
Commit
2df514c
·
1 Parent(s): 7193608

rm --cached *

Browse files
.gitattributes DELETED
@@ -1,34 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ftz filter=lfs diff=lfs merge=lfs -text
6
- *.gz filter=lfs diff=lfs merge=lfs -text
7
- *.h5 filter=lfs diff=lfs merge=lfs -text
8
- *.joblib filter=lfs diff=lfs merge=lfs -text
9
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
10
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
11
- *.model filter=lfs diff=lfs merge=lfs -text
12
- *.msgpack filter=lfs diff=lfs merge=lfs -text
13
- *.npy filter=lfs diff=lfs merge=lfs -text
14
- *.npz filter=lfs diff=lfs merge=lfs -text
15
- *.onnx filter=lfs diff=lfs merge=lfs -text
16
- *.ot filter=lfs diff=lfs merge=lfs -text
17
- *.parquet filter=lfs diff=lfs merge=lfs -text
18
- *.pb filter=lfs diff=lfs merge=lfs -text
19
- *.pickle filter=lfs diff=lfs merge=lfs -text
20
- *.pkl filter=lfs diff=lfs merge=lfs -text
21
- *.pt filter=lfs diff=lfs merge=lfs -text
22
- *.pth filter=lfs diff=lfs merge=lfs -text
23
- *.rar filter=lfs diff=lfs merge=lfs -text
24
- *.safetensors filter=lfs diff=lfs merge=lfs -text
25
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
26
- *.tar.* filter=lfs diff=lfs merge=lfs -text
27
- *.tflite filter=lfs diff=lfs merge=lfs -text
28
- *.tgz filter=lfs diff=lfs merge=lfs -text
29
- *.wasm filter=lfs diff=lfs merge=lfs -text
30
- *.xz filter=lfs diff=lfs merge=lfs -text
31
- *.zip filter=lfs diff=lfs merge=lfs -text
32
- *.zst filter=lfs diff=lfs merge=lfs -text
33
- *tfevents* filter=lfs diff=lfs merge=lfs -text
34
- tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.gitignore DELETED
@@ -1,10 +0,0 @@
1
- word_embeddings_layernorm.pt
2
- checkpoint-20000/
3
- checkpoint-25000/
4
- checkpoint-15000/
5
- eval_results.json
6
- checkpoint-10000/
7
- all_results.json
8
- word_embeddings.pt
9
- checkpoint-5000/
10
- train_results.json
 
 
 
 
 
 
 
 
 
 
 
config.json DELETED
@@ -1,38 +0,0 @@
1
- {
2
- "_name_or_path": "bigscience/bloom-1b1",
3
- "adapters": {
4
- "adapters": {},
5
- "config_map": {},
6
- "fusion_config_map": {},
7
- "fusions": {}
8
- },
9
- "apply_residual_connection_post_layernorm": false,
10
- "architectures": [
11
- "BloomForCausalLM"
12
- ],
13
- "attention_dropout": 0.0,
14
- "attention_softmax_in_fp32": true,
15
- "bias_dropout_fusion": true,
16
- "bos_token_id": 1,
17
- "eos_token_id": 2,
18
- "hidden_dropout": 0.0,
19
- "hidden_size": 1536,
20
- "initializer_range": 0.02,
21
- "layer_norm_epsilon": 1e-05,
22
- "masked_softmax_fusion": true,
23
- "model_type": "bloom",
24
- "n_head": 16,
25
- "n_inner": null,
26
- "n_layer": 24,
27
- "offset_alibi": 100,
28
- "pad_token_id": 3,
29
- "pretraining_tp": 1,
30
- "skip_bias_add": true,
31
- "skip_bias_add_qkv": false,
32
- "slow_but_exact": false,
33
- "torch_dtype": "float16",
34
- "transformers_version": "4.20.0.dev0",
35
- "unk_token_id": 0,
36
- "use_cache": true,
37
- "vocab_size": 250880
38
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee70856c4a53d34dc9a2624bab4f8cc1b5533d5c32825a1a9f375d17584001d4
3
- size 2130668779
 
 
 
 
special_tokens_map.json DELETED
@@ -1 +0,0 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"}
 
 
tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f6efc66e73f1fd69da4f436e48befb519fdff3fe18910850c1d41bd862293a5
3
- size 14500443
 
 
 
 
tokenizer_config.json DELETED
@@ -1 +0,0 @@
1
- {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "<pad>", "add_prefix_space": false, "name_or_path": "bigscience/bloom-1b1", "special_tokens_map_file": null, "padding_side": "left", "tokenizer_class": "BloomTokenizer"}
 
 
trainer_state.json DELETED
@@ -1,125 +0,0 @@
1
- {
2
- "best_metric": 2.68359375,
3
- "best_model_checkpoint": "/home/zhengxinyong/outputs/bloom-1b1_de_continual-pretrain_100000samples_-1vocab_original/checkpoint-25000",
4
- "epoch": 2.7475546763380594,
5
- "global_step": 25000,
6
- "is_hyper_param_search": false,
7
- "is_local_process_zero": true,
8
- "is_world_process_zero": true,
9
- "log_history": [
10
- {
11
- "epoch": 0.27,
12
- "learning_rate": 0.0001,
13
- "loss": 3.1701,
14
- "step": 2500
15
- },
16
- {
17
- "epoch": 0.55,
18
- "learning_rate": 0.0001,
19
- "loss": 2.8341,
20
- "step": 5000
21
- },
22
- {
23
- "epoch": 0.55,
24
- "eval_loss": 2.873046875,
25
- "eval_runtime": 3.4814,
26
- "eval_samples_per_second": 230.94,
27
- "eval_steps_per_second": 29.011,
28
- "step": 5000
29
- },
30
- {
31
- "epoch": 0.82,
32
- "learning_rate": 0.0001,
33
- "loss": 2.7255,
34
- "step": 7500
35
- },
36
- {
37
- "epoch": 1.1,
38
- "learning_rate": 0.0001,
39
- "loss": 2.5869,
40
- "step": 10000
41
- },
42
- {
43
- "epoch": 1.1,
44
- "eval_loss": 2.759765625,
45
- "eval_runtime": 3.473,
46
- "eval_samples_per_second": 231.499,
47
- "eval_steps_per_second": 29.081,
48
- "step": 10000
49
- },
50
- {
51
- "epoch": 1.37,
52
- "learning_rate": 0.0001,
53
- "loss": 2.4538,
54
- "step": 12500
55
- },
56
- {
57
- "epoch": 1.65,
58
- "learning_rate": 0.0001,
59
- "loss": 2.4623,
60
- "step": 15000
61
- },
62
- {
63
- "epoch": 1.65,
64
- "eval_loss": 2.705078125,
65
- "eval_runtime": 3.4584,
66
- "eval_samples_per_second": 232.479,
67
- "eval_steps_per_second": 29.204,
68
- "step": 15000
69
- },
70
- {
71
- "epoch": 1.92,
72
- "learning_rate": 0.0001,
73
- "loss": 2.4534,
74
- "step": 17500
75
- },
76
- {
77
- "epoch": 2.2,
78
- "learning_rate": 0.0001,
79
- "loss": 2.2463,
80
- "step": 20000
81
- },
82
- {
83
- "epoch": 2.2,
84
- "eval_loss": 2.712890625,
85
- "eval_runtime": 3.4672,
86
- "eval_samples_per_second": 231.886,
87
- "eval_steps_per_second": 29.13,
88
- "step": 20000
89
- },
90
- {
91
- "epoch": 2.47,
92
- "learning_rate": 0.0001,
93
- "loss": 2.2084,
94
- "step": 22500
95
- },
96
- {
97
- "epoch": 2.75,
98
- "learning_rate": 0.0001,
99
- "loss": 2.2451,
100
- "step": 25000
101
- },
102
- {
103
- "epoch": 2.75,
104
- "eval_loss": 2.68359375,
105
- "eval_runtime": 3.4826,
106
- "eval_samples_per_second": 230.861,
107
- "eval_steps_per_second": 29.001,
108
- "step": 25000
109
- },
110
- {
111
- "epoch": 2.75,
112
- "step": 25000,
113
- "total_flos": 8.355380415011226e+17,
114
- "train_loss": 2.53857546875,
115
- "train_runtime": 4808.8208,
116
- "train_samples_per_second": 41.59,
117
- "train_steps_per_second": 5.199
118
- }
119
- ],
120
- "max_steps": 25000,
121
- "num_train_epochs": 3,
122
- "total_flos": 8.355380415011226e+17,
123
- "trial_name": null,
124
- "trial_params": null
125
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9434ced7df591cdb6202bc169b46edd7b46a94996ac179a0c6056b7f66df194
3
- size 4399