Saving weights and logs of step 5000
Browse files
events.out.tfevents.1634745538.t1v-n-f6f5b6cc-w-0.1277181.0.v2
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b62997695b8c5a06d2ec0d0cc9ed4f46bfed941a3d49fd348aceb7577bce4fd2
|
3 |
+
size 23391025
|
events.out.tfevents.1635022617.t1v-n-f6f5b6cc-w-0.56283.0.v2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:10313d235914ada691df4001460a21e4cac30520a984648f072b00ea7a854473
|
3 |
+
size 735759
|
flax_model.msgpack
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 497764120
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dc16a04ccd995ca3adccd9eec1d93449a80b7ed9095f8a2860f08d4d6ba0116d
|
3 |
size 497764120
|
run.sh
CHANGED
@@ -11,11 +11,12 @@ python run_clm_flax.py \
|
|
11 |
--block_size="512" \
|
12 |
--per_device_train_batch_size="64" \
|
13 |
--per_device_eval_batch_size="64" \
|
14 |
-
--learning_rate="
|
|
|
15 |
--adam_beta1="0.9" --adam_beta2="0.98" --weight_decay="0.01" \
|
16 |
--overwrite_output_dir \
|
17 |
-
--num_train_epochs="
|
18 |
--logging_steps="500" \
|
19 |
-
--save_steps="
|
20 |
-
--eval_steps="
|
21 |
--push_to_hub
|
|
|
11 |
--block_size="512" \
|
12 |
--per_device_train_batch_size="64" \
|
13 |
--per_device_eval_batch_size="64" \
|
14 |
+
--learning_rate="3e-3" \
|
15 |
+
--warmup_steps="0" \
|
16 |
--adam_beta1="0.9" --adam_beta2="0.98" --weight_decay="0.01" \
|
17 |
--overwrite_output_dir \
|
18 |
+
--num_train_epochs="30" \
|
19 |
--logging_steps="500" \
|
20 |
+
--save_steps="5000" \
|
21 |
+
--eval_steps="5000" \
|
22 |
--push_to_hub
|