DancingIguana commited on
Commit
80d94eb
1 Parent(s): 5f8eb4a

End of training

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilgpt2",
3
+ "_num_labels": 1,
4
+ "activation_function": "gelu_new",
5
+ "architectures": [
6
+ "GPT2LMHeadModel"
7
+ ],
8
+ "attn_pdrop": 0.1,
9
+ "bos_token_id": 0,
10
+ "embd_pdrop": 0.1,
11
+ "eos_token_id": 0,
12
+ "id2label": {
13
+ "0": "LABEL_0"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "label2id": {
17
+ "LABEL_0": 0
18
+ },
19
+ "layer_norm_epsilon": 1e-05,
20
+ "model_type": "gpt2",
21
+ "n_ctx": 128,
22
+ "n_embd": 768,
23
+ "n_head": 12,
24
+ "n_inner": null,
25
+ "n_layer": 6,
26
+ "n_positions": 1024,
27
+ "reorder_and_upcast_attn": false,
28
+ "resid_pdrop": 0.1,
29
+ "scale_attn_by_inverse_layer_idx": false,
30
+ "scale_attn_weights": true,
31
+ "summary_activation": null,
32
+ "summary_first_dropout": 0.1,
33
+ "summary_proj_to_labels": true,
34
+ "summary_type": "cls_index",
35
+ "summary_use_proj": true,
36
+ "task_specific_params": {
37
+ "text-generation": {
38
+ "do_sample": true,
39
+ "max_length": 50
40
+ }
41
+ },
42
+ "torch_dtype": "float32",
43
+ "transformers_version": "4.19.4",
44
+ "use_cache": true,
45
+ "vocab_size": 83
46
+ }
merges.txt ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #version: 0.2 - Trained by `huggingface/tokenizers`
2
+ 2 5
3
+ n s
4
+ Ġ ns
5
+ c s
6
+ Ġ cs
7
+ r s
8
+ Ġ rs
9
+ s E
10
+ s G
11
+ s C
12
+ Ġns G
13
+ s B
14
+ s F
15
+ Ġns E
16
+ Ġns C
17
+ s A
18
+ s D
19
+ Ġns F
20
+ Ġns B
21
+ Ġns D
22
+ Ġns A
23
+ 7 5
24
+ sG w
25
+ sE t
26
+ sF w
27
+ sB t
28
+ sC w
29
+ 1 2
30
+ ĠnsG w
31
+ ĠnsE t
32
+ ĠnsF w
33
+ ĠnsB t
34
+ ĠnsC w
35
+ 1 1
36
+ 1 3
37
+ 1 7
38
+ 2 3
39
+ 1 9
40
+ 2 4
41
+ ns E
42
+ 1 0
43
+ ns A
44
+ ns C
45
+ ns G
46
+ 2 9
47
+ 3 5
48
+ 3 7
49
+ 1 25
50
+ 3 75
51
+ ns B
52
+ 2 0
53
+ 4 1
54
+ 5 3
55
+ ns D
56
+ ns F
57
+ 4 9
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3264282c2863b3e73995a626836a6b0c30520b2e6c40eaa7d85b5b459602e390
3
+ size 179834589
runs/Jun11_20-46-44_9f0179c90d46/1654980441.472945/events.out.tfevents.1654980441.9f0179c90d46.71.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7b4bd6919b96a2c6404a6f144f5ff9ac6195d85965454cf1bd8a5931a758276
3
+ size 5159
runs/Jun11_20-46-44_9f0179c90d46/events.out.tfevents.1654980441.9f0179c90d46.71.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd2639c876b3542d75eea3239b091db47c1411c2e49f44356cd2721879f0672c
3
+ size 4188
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>", "pad_token": "<|endoftext|>"}
tokenizer.json ADDED
@@ -0,0 +1,190 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 128,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": null,
10
+ "added_tokens": [
11
+ {
12
+ "id": 0,
13
+ "content": "<|endoftext|>",
14
+ "single_word": false,
15
+ "lstrip": false,
16
+ "rstrip": false,
17
+ "normalized": false,
18
+ "special": true
19
+ }
20
+ ],
21
+ "normalizer": null,
22
+ "pre_tokenizer": {
23
+ "type": "ByteLevel",
24
+ "add_prefix_space": false,
25
+ "trim_offsets": true,
26
+ "use_regex": true
27
+ },
28
+ "post_processor": {
29
+ "type": "ByteLevel",
30
+ "add_prefix_space": true,
31
+ "trim_offsets": false,
32
+ "use_regex": true
33
+ },
34
+ "decoder": {
35
+ "type": "WordPiece",
36
+ "prefix": "##",
37
+ "cleanup": true
38
+ },
39
+ "model": {
40
+ "type": "BPE",
41
+ "dropout": null,
42
+ "unk_token": "[UNK]",
43
+ "continuing_subword_prefix": null,
44
+ "end_of_word_suffix": null,
45
+ "fuse_unk": false,
46
+ "vocab": {
47
+ "<|endoftext|>": 0,
48
+ "0": 1,
49
+ "1": 2,
50
+ "2": 3,
51
+ "3": 4,
52
+ "4": 5,
53
+ "5": 6,
54
+ "6": 7,
55
+ "7": 8,
56
+ "8": 9,
57
+ "9": 10,
58
+ "A": 11,
59
+ "B": 12,
60
+ "C": 13,
61
+ "D": 14,
62
+ "E": 15,
63
+ "F": 16,
64
+ "G": 17,
65
+ "c": 18,
66
+ "n": 19,
67
+ "p": 20,
68
+ "q": 21,
69
+ "r": 22,
70
+ "s": 23,
71
+ "t": 24,
72
+ "w": 25,
73
+ "Ġ": 26,
74
+ "25": 27,
75
+ "ns": 28,
76
+ "Ġns": 29,
77
+ "cs": 30,
78
+ "Ġcs": 31,
79
+ "rs": 32,
80
+ "Ġrs": 33,
81
+ "sE": 34,
82
+ "sG": 35,
83
+ "sC": 36,
84
+ "ĠnsG": 37,
85
+ "sB": 38,
86
+ "sF": 39,
87
+ "ĠnsE": 40,
88
+ "ĠnsC": 41,
89
+ "sA": 42,
90
+ "sD": 43,
91
+ "ĠnsF": 44,
92
+ "ĠnsB": 45,
93
+ "ĠnsD": 46,
94
+ "ĠnsA": 47,
95
+ "75": 48,
96
+ "sGw": 49,
97
+ "sEt": 50,
98
+ "sFw": 51,
99
+ "sBt": 52,
100
+ "sCw": 53,
101
+ "12": 54,
102
+ "ĠnsGw": 55,
103
+ "ĠnsEt": 56,
104
+ "ĠnsFw": 57,
105
+ "ĠnsBt": 58,
106
+ "ĠnsCw": 59,
107
+ "11": 60,
108
+ "13": 61,
109
+ "17": 62,
110
+ "23": 63,
111
+ "19": 64,
112
+ "24": 65,
113
+ "nsE": 66,
114
+ "10": 67,
115
+ "nsA": 68,
116
+ "nsC": 69,
117
+ "nsG": 70,
118
+ "29": 71,
119
+ "35": 72,
120
+ "37": 73,
121
+ "125": 74,
122
+ "375": 75,
123
+ "nsB": 76,
124
+ "20": 77,
125
+ "41": 78,
126
+ "53": 79,
127
+ "nsD": 80,
128
+ "nsF": 81,
129
+ "49": 82
130
+ },
131
+ "merges": [
132
+ "2 5",
133
+ "n s",
134
+ "Ġ ns",
135
+ "c s",
136
+ "Ġ cs",
137
+ "r s",
138
+ "Ġ rs",
139
+ "s E",
140
+ "s G",
141
+ "s C",
142
+ "Ġns G",
143
+ "s B",
144
+ "s F",
145
+ "Ġns E",
146
+ "Ġns C",
147
+ "s A",
148
+ "s D",
149
+ "Ġns F",
150
+ "Ġns B",
151
+ "Ġns D",
152
+ "Ġns A",
153
+ "7 5",
154
+ "sG w",
155
+ "sE t",
156
+ "sF w",
157
+ "sB t",
158
+ "sC w",
159
+ "1 2",
160
+ "ĠnsG w",
161
+ "ĠnsE t",
162
+ "ĠnsF w",
163
+ "ĠnsB t",
164
+ "ĠnsC w",
165
+ "1 1",
166
+ "1 3",
167
+ "1 7",
168
+ "2 3",
169
+ "1 9",
170
+ "2 4",
171
+ "ns E",
172
+ "1 0",
173
+ "ns A",
174
+ "ns C",
175
+ "ns G",
176
+ "2 9",
177
+ "3 5",
178
+ "3 7",
179
+ "1 25",
180
+ "3 75",
181
+ "ns B",
182
+ "2 0",
183
+ "4 1",
184
+ "5 3",
185
+ "ns D",
186
+ "ns F",
187
+ "4 9"
188
+ ]
189
+ }
190
+ }
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "<|endoftext|>", "bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "add_prefix_space": false, "tokenizer_class": "GPT2Tokenizer"}
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdc6ac079e64e24c90222cea14216070d5164e0005829d8d8751b0ebbef96dbf
3
+ size 3183
vocab.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<|endoftext|>":0,"0":1,"1":2,"2":3,"3":4,"4":5,"5":6,"6":7,"7":8,"8":9,"9":10,"A":11,"B":12,"C":13,"D":14,"E":15,"F":16,"G":17,"c":18,"n":19,"p":20,"q":21,"r":22,"s":23,"t":24,"w":25,"Ġ":26,"25":27,"ns":28,"Ġns":29,"cs":30,"Ġcs":31,"rs":32,"Ġrs":33,"sE":34,"sG":35,"sC":36,"ĠnsG":37,"sB":38,"sF":39,"ĠnsE":40,"ĠnsC":41,"sA":42,"sD":43,"ĠnsF":44,"ĠnsB":45,"ĠnsD":46,"ĠnsA":47,"75":48,"sGw":49,"sEt":50,"sFw":51,"sBt":52,"sCw":53,"12":54,"ĠnsGw":55,"ĠnsEt":56,"ĠnsFw":57,"ĠnsBt":58,"ĠnsCw":59,"11":60,"13":61,"17":62,"23":63,"19":64,"24":65,"nsE":66,"10":67,"nsA":68,"nsC":69,"nsG":70,"29":71,"35":72,"37":73,"125":74,"375":75,"nsB":76,"20":77,"41":78,"53":79,"nsD":80,"nsF":81,"49":82}