Saving weights and logs of epoch 0
Browse files- .gitattributes +3 -0
- .lock +0 -0
- config.json +26 -0
- events.out.tfevents.1634611280.t1v-n-10409466-w-0.11492.0.v2 +3 -0
- events.out.tfevents.1634647562.t1v-n-10409466-w-0.4496.0.v2 +3 -0
- events.out.tfevents.1634681194.t1v-n-10409466-w-0.3720.0.v2 +3 -0
- flax_model.msgpack +3 -0
- mt5-base-qgen/events.out.tfevents.1634610755.t1v-n-10409466-w-0.9004.0.v2 +3 -0
- mt5-questions-dataset.log +11 -0
- qgen_train.csv +3 -0
- qgen_training_data.csv +3 -0
- qgen_val.csv +3 -0
- run_summarization_flax.py +1 -1
- special_tokens_map.json +1 -0
- spiece.model +3 -0
- start_qgen.sh +20 -14
- tokenizer.json +0 -0
- tokenizer_config.json +1 -0
.gitattributes
CHANGED
@@ -25,3 +25,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
25 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
26 |
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
27 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
25 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
26 |
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
27 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
28 |
+
qgen_train.csv filter=lfs diff=lfs merge=lfs -text
|
29 |
+
qgen_training_data.csv filter=lfs diff=lfs merge=lfs -text
|
30 |
+
qgen_val.csv filter=lfs diff=lfs merge=lfs -text
|
.lock
ADDED
File without changes
|
config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"MT5ForConditionalGeneration"
|
4 |
+
],
|
5 |
+
"d_ff": 1024,
|
6 |
+
"d_kv": 64,
|
7 |
+
"d_model": 512,
|
8 |
+
"decoder_start_token_id": 0,
|
9 |
+
"dropout_rate": 0.1,
|
10 |
+
"eos_token_id": 1,
|
11 |
+
"feed_forward_proj": "gated-gelu",
|
12 |
+
"initializer_factor": 1.0,
|
13 |
+
"is_encoder_decoder": true,
|
14 |
+
"layer_norm_epsilon": 1e-06,
|
15 |
+
"model_type": "mt5",
|
16 |
+
"num_decoder_layers": 8,
|
17 |
+
"num_heads": 6,
|
18 |
+
"num_layers": 8,
|
19 |
+
"pad_token_id": 0,
|
20 |
+
"relative_attention_num_buckets": 32,
|
21 |
+
"tie_word_embeddings": false,
|
22 |
+
"tokenizer_class": "T5Tokenizer",
|
23 |
+
"transformers_version": "4.12.0.dev0",
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 250112
|
26 |
+
}
|
events.out.tfevents.1634611280.t1v-n-10409466-w-0.11492.0.v2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e182970a4bdaa09ab4bc39e24998573f2279663d8092199e97f5f27f509662e0
|
3 |
+
size 40
|
events.out.tfevents.1634647562.t1v-n-10409466-w-0.4496.0.v2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:62f76d55580d4e2d1fe8679c9fe034df11554145300a05ce56d4d4839bf55ff6
|
3 |
+
size 40
|
events.out.tfevents.1634681194.t1v-n-10409466-w-0.3720.0.v2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5a33497fbf05f4540c98f5a8a9a4bdf2f88c6330c4bd969cd379a55f84b0d2e6
|
3 |
+
size 2450039
|
flax_model.msgpack
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:74d1919cb1cc5091f578509943b09248e0a3f4cbdbdebf111c0d96fb2a47ff3d
|
3 |
+
size 1200715307
|
mt5-base-qgen/events.out.tfevents.1634610755.t1v-n-10409466-w-0.9004.0.v2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8e1acf82767bf97369c40271b7915dfe127686a6a3908c2d97b7626d45315e6d
|
3 |
+
size 40
|
mt5-questions-dataset.log
ADDED
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[{"stream_name":"stderr","time":34.74216129,"data":"/opt/conda/lib/python3.7/site-packages/papermill/iorw.py:50: FutureWarning: pyarrow.HadoopFileSystem is deprecated as of 2.0.0, please use pyarrow.fs.HadoopFileSystem instead.\n"}
|
2 |
+
,{"stream_name":"stderr","time":34.74226826,"data":" from pyarrow import HadoopFileSystem\n"}
|
3 |
+
,{"stream_name":"stderr","time":37.894455456,"data":"/opt/conda/lib/python3.7/site-packages/traitlets/traitlets.py:2567: FutureWarning: --Exporter.preprocessors=[\"remove_papermill_header.RemovePapermillHeader\"] for containers is deprecated in traitlets 5.0. You can pass `--Exporter.preprocessors item` ... multiple times to add items to a list.\n"}
|
4 |
+
,{"stream_name":"stderr","time":37.89452555,"data":" FutureWarning,\n"}
|
5 |
+
,{"stream_name":"stderr","time":37.894535347,"data":"[NbConvertApp] Converting notebook __notebook__.ipynb to notebook\n"}
|
6 |
+
,{"stream_name":"stderr","time":37.920776986,"data":"[NbConvertApp] Writing 2546 bytes to __notebook__.ipynb\n"}
|
7 |
+
,{"stream_name":"stderr","time":40.120591007,"data":"/opt/conda/lib/python3.7/site-packages/traitlets/traitlets.py:2567: FutureWarning: --Exporter.preprocessors=[\"nbconvert.preprocessors.ExtractOutputPreprocessor\"] for containers is deprecated in traitlets 5.0. You can pass `--Exporter.preprocessors item` ... multiple times to add items to a list.\n"}
|
8 |
+
,{"stream_name":"stderr","time":40.12066086,"data":" FutureWarning,\n"}
|
9 |
+
,{"stream_name":"stderr","time":40.120670616,"data":"[NbConvertApp] Converting notebook __notebook__.ipynb to html\n"}
|
10 |
+
,{"stream_name":"stderr","time":40.662991409,"data":"[NbConvertApp] Writing 276762 bytes to __results__.html\n"}
|
11 |
+
]
|
qgen_train.csv
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:21a0e59e17a98496c830c4eaa5e7109e648ccc49905acd30cc24dc477233cd71
|
3 |
+
size 337028529
|
qgen_training_data.csv
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8045ce51c73fed96d10df60dae29f3dc9eb6344dd84046e65f38aa00edd0d258
|
3 |
+
size 420218929
|
qgen_val.csv
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:215fc1360575a907d9d0dc89fbe08e656e41c7ca76df813c12b95eb98fd68597
|
3 |
+
size 83190420
|
run_summarization_flax.py
CHANGED
@@ -446,7 +446,7 @@ def main():
|
|
446 |
# In Flax, for seq2seq models we need to pass `decoder_input_ids`
|
447 |
# as the Flax models don't accept `labels`, we need to prepare the decoder_input_ids here
|
448 |
# for that dynamically import the `shift_tokens_right` function from the model file
|
449 |
-
model_module = __import__(
|
450 |
shift_tokens_right_fn = getattr(model_module, "shift_tokens_right")
|
451 |
|
452 |
# Setting padding="max_length" as we need fixed length inputs for jitted functions
|
|
|
446 |
# In Flax, for seq2seq models we need to pass `decoder_input_ids`
|
447 |
# as the Flax models don't accept `labels`, we need to prepare the decoder_input_ids here
|
448 |
# for that dynamically import the `shift_tokens_right` function from the model file
|
449 |
+
model_module = __import__("transformers.models.t5.modeling_flax_t5", fromlist=["shift_tokens_tight"])
|
450 |
shift_tokens_right_fn = getattr(model_module, "shift_tokens_right")
|
451 |
|
452 |
# Setting padding="max_length" as we need fixed length inputs for jitted functions
|
special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"}
|
spiece.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ef78f86560d809067d12bac6c09f19a462cb3af3f54d2b8acbba26e1433125d6
|
3 |
+
size 4309802
|
start_qgen.sh
CHANGED
@@ -1,15 +1,21 @@
|
|
|
|
|
|
1 |
python run_summarization_flax.py \
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
|
|
|
|
|
|
|
|
|
1 |
+
#! /usr/bin/env bash
|
2 |
+
|
3 |
python run_summarization_flax.py \
|
4 |
+
--output_dir ./ \
|
5 |
+
--model_name_or_path google/mt5-small \
|
6 |
+
--tokenizer_name google/mt5-small \
|
7 |
+
--train_file qgen_train.csv \
|
8 |
+
--validation_file qgen_val.csv \
|
9 |
+
--text_column context \
|
10 |
+
--summary_column question \
|
11 |
+
--do_train \
|
12 |
+
--do_eval \
|
13 |
+
--num_train_epochs 3 \
|
14 |
+
--eval_steps 2000 \
|
15 |
+
--logging_steps 100 \
|
16 |
+
--learning_rate 5e-5 --warmup_steps 1000 \
|
17 |
+
--per_device_train_batch_size 2 \
|
18 |
+
--per_device_eval_batch_size 2 \
|
19 |
+
--overwrite_output_dir \
|
20 |
+
--max_source_length 1024 --max_target_length 64 \
|
21 |
+
--push_to_hub
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "extra_ids": 0, "additional_special_tokens": null, "special_tokens_map_file": "/home/nicholas/.cache/huggingface/transformers/685ac0ca8568ec593a48b61b0a3c272beee9bc194a3c7241d15dcadb5f875e53.f76030f3ec1b96a8199b2593390c610e76ca8028ef3d24680000619ffb646276", "name_or_path": "google/mt5-small", "sp_model_kwargs": {}, "tokenizer_class": "T5Tokenizer"}
|