Upload folder using huggingface_hub
Browse files- adapter_model.bin +1 -1
- pytorch_model.bin +1 -1
- xturing.json +2 -2
adapter_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 16819789
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:363e0bf5266a44552a623376f31def83dd0f07938aad93115fee32a75a30f3b1
|
3 |
size 16819789
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3911057236
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c7068580da2103c682af685d131132fff6a28f3ff7e04ddfe52f6c15d42477d1
|
3 |
size 3911057236
|
xturing.json
CHANGED
@@ -3,13 +3,13 @@
|
|
3 |
"finetuning_config": {
|
4 |
"learning_rate": 0.0001,
|
5 |
"gradient_accumulation_steps": 1,
|
6 |
-
"batch_size":
|
7 |
"weight_decay": 0.01,
|
8 |
"warmup_steps": 50,
|
9 |
"eval_steps": 5000,
|
10 |
"save_steps": 5000,
|
11 |
"max_length": 256,
|
12 |
-
"num_train_epochs":
|
13 |
"logging_steps": 10,
|
14 |
"max_grad_norm": 2.0,
|
15 |
"save_total_limit": 4,
|
|
|
3 |
"finetuning_config": {
|
4 |
"learning_rate": 0.0001,
|
5 |
"gradient_accumulation_steps": 1,
|
6 |
+
"batch_size": 16,
|
7 |
"weight_decay": 0.01,
|
8 |
"warmup_steps": 50,
|
9 |
"eval_steps": 5000,
|
10 |
"save_steps": 5000,
|
11 |
"max_length": 256,
|
12 |
+
"num_train_epochs": 10,
|
13 |
"logging_steps": 10,
|
14 |
"max_grad_norm": 2.0,
|
15 |
"save_total_limit": 4,
|