belisards commited on
Commit
1926e72
1 Parent(s): a9a4526

azmn-posicao-fin

Browse files
README.md CHANGED
@@ -21,11 +21,11 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  This model is a fine-tuned version of [neuralmind/bert-base-portuguese-cased](https://huggingface.co/neuralmind/bert-base-portuguese-cased) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 0.0500
25
- - Accuracy: 0.7415
26
- - F1: 0.6919
27
- - Recall: 0.7472
28
- - Precision: 0.6838
29
 
30
  ## Model description
31
 
@@ -51,22 +51,29 @@ The following hyperparameters were used during training:
51
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: linear
53
  - lr_scheduler_warmup_steps: 150
54
- - num_epochs: 15
55
 
56
  ### Training results
57
 
58
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Recall | Precision |
59
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:------:|:---------:|
60
- | 0.0667 | 1.0 | 18 | 0.0661 | 0.5536 | 0.4531 | 0.4520 | 0.4571 |
61
- | 0.0624 | 2.0 | 36 | 0.0646 | 0.6696 | 0.5743 | 0.5752 | 0.5736 |
62
- | 0.0625 | 3.0 | 54 | 0.0628 | 0.7321 | 0.6510 | 0.6510 | 0.6510 |
63
- | 0.0612 | 4.0 | 72 | 0.0603 | 0.7411 | 0.6733 | 0.6795 | 0.6687 |
64
- | 0.0566 | 5.0 | 90 | 0.0568 | 0.7768 | 0.7184 | 0.7260 | 0.7125 |
65
- | 0.0544 | 6.0 | 108 | 0.0530 | 0.7589 | 0.7216 | 0.7588 | 0.7119 |
66
- | 0.0488 | 7.0 | 126 | 0.0497 | 0.8214 | 0.7812 | 0.8010 | 0.7688 |
67
- | 0.0398 | 8.0 | 144 | 0.0498 | 0.7946 | 0.7629 | 0.8054 | 0.75 |
68
- | 0.0276 | 9.0 | 162 | 0.0540 | 0.8125 | 0.7681 | 0.7838 | 0.7575 |
69
- | 0.0184 | 10.0 | 180 | 0.0674 | 0.7679 | 0.7156 | 0.7312 | 0.7065 |
 
 
 
 
 
 
 
70
 
71
 
72
  ### Framework versions
 
21
 
22
  This model is a fine-tuned version of [neuralmind/bert-base-portuguese-cased](https://huggingface.co/neuralmind/bert-base-portuguese-cased) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 0.0460
25
+ - Accuracy: 0.8367
26
+ - F1: 0.7871
27
+ - Recall: 0.8194
28
+ - Precision: 0.7687
29
 
30
  ## Model description
31
 
 
51
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: linear
53
  - lr_scheduler_warmup_steps: 150
54
+ - num_epochs: 19
55
 
56
  ### Training results
57
 
58
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Recall | Precision |
59
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:------:|:---------:|
60
+ | 0.0801 | 1.0 | 18 | 0.0769 | 0.7411 | 0.4256 | 0.5 | 0.3705 |
61
+ | 0.0691 | 2.0 | 36 | 0.0709 | 0.75 | 0.4612 | 0.5172 | 0.8739 |
62
+ | 0.0647 | 3.0 | 54 | 0.0661 | 0.75 | 0.4612 | 0.5172 | 0.8739 |
63
+ | 0.0644 | 4.0 | 72 | 0.0648 | 0.6518 | 0.5774 | 0.5856 | 0.5753 |
64
+ | 0.0621 | 5.0 | 90 | 0.0632 | 0.7054 | 0.6424 | 0.6554 | 0.6367 |
65
+ | 0.0621 | 6.0 | 108 | 0.0627 | 0.7232 | 0.6265 | 0.6226 | 0.6319 |
66
+ | 0.0586 | 7.0 | 126 | 0.0595 | 0.75 | 0.6937 | 0.7079 | 0.6857 |
67
+ | 0.0547 | 8.0 | 144 | 0.0582 | 0.7768 | 0.7338 | 0.7597 | 0.7223 |
68
+ | 0.0509 | 9.0 | 162 | 0.0554 | 0.7768 | 0.7338 | 0.7597 | 0.7223 |
69
+ | 0.0462 | 10.0 | 180 | 0.0557 | 0.75 | 0.7091 | 0.7416 | 0.6998 |
70
+ | 0.0437 | 11.0 | 198 | 0.0532 | 0.7768 | 0.7382 | 0.7709 | 0.7264 |
71
+ | 0.0415 | 12.0 | 216 | 0.0515 | 0.7857 | 0.7466 | 0.7769 | 0.7341 |
72
+ | 0.0356 | 13.0 | 234 | 0.0545 | 0.8036 | 0.7547 | 0.7665 | 0.7461 |
73
+ | 0.0301 | 14.0 | 252 | 0.0543 | 0.8214 | 0.7770 | 0.7898 | 0.7675 |
74
+ | 0.0262 | 15.0 | 270 | 0.0541 | 0.8036 | 0.7594 | 0.7777 | 0.7481 |
75
+ | 0.0248 | 16.0 | 288 | 0.0583 | 0.8125 | 0.7584 | 0.7613 | 0.7557 |
76
+ | 0.0232 | 17.0 | 306 | 0.0593 | 0.8125 | 0.7635 | 0.7725 | 0.7562 |
77
 
78
 
79
  ### Framework versions
config.json CHANGED
@@ -3,11 +3,11 @@
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
6
- "attention_probs_dropout_prob": 0.05,
7
  "classifier_dropout": null,
8
  "directionality": "bidi",
9
  "hidden_act": "gelu",
10
- "hidden_dropout_prob": 0.05,
11
  "hidden_size": 768,
12
  "id2label": {
13
  "0": "0",
 
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0.07,
7
  "classifier_dropout": null,
8
  "directionality": "bidi",
9
  "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.07,
11
  "hidden_size": 768,
12
  "id2label": {
13
  "0": "0",
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b622937cd59df9fe36e8b0f298d053fea72fc441b0270f4e2ebb6ae0aa5dba4
3
  size 435722224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62321d9a4491e7119b0983a5c7ad590b2c8dc75d8c89cfc06d6875ab69313ffb
3
  size 435722224
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68e6cecf2b97b2265fb08f71185bfda74485c7a75952fda0c05a1291dcda8568
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7272e0fdc42df5fbdb5e3cdd4ee077095b566d6801ac1559b7146fdfff4bcba
3
  size 5368
wandb/debug-internal.log CHANGED
@@ -1,53 +1,43 @@
1
- {"time":"2024-12-12T00:13:10.430874697Z","level":"INFO","msg":"using version","core version":"0.19.0"}
2
- {"time":"2024-12-12T00:13:10.430902983Z","level":"INFO","msg":"created symlink","path":"/content/wandb/run-20241212_001310-hiitv1j3/logs/debug-core.log"}
3
- {"time":"2024-12-12T00:13:10.533716472Z","level":"INFO","msg":"created new stream","id":"hiitv1j3"}
4
- {"time":"2024-12-12T00:13:10.533750202Z","level":"INFO","msg":"stream: started","id":"hiitv1j3"}
5
- {"time":"2024-12-12T00:13:10.533785266Z","level":"INFO","msg":"writer: Do: started","stream_id":"hiitv1j3"}
6
- {"time":"2024-12-12T00:13:10.533801485Z","level":"INFO","msg":"sender: started","stream_id":"hiitv1j3"}
7
- {"time":"2024-12-12T00:13:10.533854634Z","level":"INFO","msg":"handler: started","stream_id":"hiitv1j3"}
8
- {"time":"2024-12-12T00:13:12.366471466Z","level":"INFO","msg":"Starting system monitor"}
9
- {"time":"2024-12-12T00:13:12.849349549Z","level":"INFO","msg":"Pausing system monitor"}
10
- {"time":"2024-12-12T00:13:12.849377903Z","level":"INFO","msg":"Resuming system monitor"}
11
- {"time":"2024-12-12T00:13:12.84938263Z","level":"INFO","msg":"Pausing system monitor"}
12
- {"time":"2024-12-12T00:13:12.84938636Z","level":"INFO","msg":"Resuming system monitor"}
13
- {"time":"2024-12-12T00:13:13.000080747Z","level":"INFO","msg":"Pausing system monitor"}
14
- {"time":"2024-12-12T00:13:13.041405563Z","level":"INFO","msg":"Resuming system monitor"}
15
- {"time":"2024-12-12T00:13:13.157896168Z","level":"INFO","msg":"Pausing system monitor"}
16
- {"time":"2024-12-12T00:13:13.162529529Z","level":"INFO","msg":"Resuming system monitor"}
17
- {"time":"2024-12-12T00:15:12.140250435Z","level":"INFO","msg":"Pausing system monitor"}
18
- {"time":"2024-12-12T00:15:12.145680833Z","level":"INFO","msg":"Resuming system monitor"}
19
- {"time":"2024-12-12T00:15:12.640025857Z","level":"INFO","msg":"Pausing system monitor"}
20
- {"time":"2024-12-12T00:15:12.644499229Z","level":"INFO","msg":"Resuming system monitor"}
21
- {"time":"2024-12-12T00:15:12.645545943Z","level":"INFO","msg":"Pausing system monitor"}
22
- {"time":"2024-12-12T00:15:12.651305789Z","level":"INFO","msg":"Resuming system monitor"}
23
- {"time":"2024-12-12T00:15:12.732077418Z","level":"INFO","msg":"Pausing system monitor"}
24
- {"time":"2024-12-12T00:15:12.736787202Z","level":"INFO","msg":"Resuming system monitor"}
25
- {"time":"2024-12-12T00:15:12.739545581Z","level":"INFO","msg":"Pausing system monitor"}
26
- {"time":"2024-12-12T00:15:12.744606494Z","level":"INFO","msg":"Resuming system monitor"}
27
- {"time":"2024-12-12T00:15:14.393781268Z","level":"INFO","msg":"Pausing system monitor"}
28
- {"time":"2024-12-12T00:15:14.401598267Z","level":"INFO","msg":"Resuming system monitor"}
29
- {"time":"2024-12-12T00:15:14.403616404Z","level":"INFO","msg":"Pausing system monitor"}
30
- {"time":"2024-12-12T00:15:14.408799754Z","level":"INFO","msg":"Resuming system monitor"}
31
- {"time":"2024-12-12T00:15:14.42157668Z","level":"INFO","msg":"Pausing system monitor"}
32
- {"time":"2024-12-12T00:15:14.426163968Z","level":"INFO","msg":"Resuming system monitor"}
33
- {"time":"2024-12-12T00:15:14.426987128Z","level":"INFO","msg":"Pausing system monitor"}
34
- {"time":"2024-12-12T00:15:14.431262826Z","level":"INFO","msg":"Resuming system monitor"}
35
- {"time":"2024-12-12T00:15:14.431901415Z","level":"INFO","msg":"Pausing system monitor"}
36
- {"time":"2024-12-12T00:15:14.436376054Z","level":"INFO","msg":"Resuming system monitor"}
37
- {"time":"2024-12-12T00:15:14.437159398Z","level":"INFO","msg":"Pausing system monitor"}
38
- {"time":"2024-12-12T00:15:14.44097705Z","level":"INFO","msg":"Resuming system monitor"}
39
- {"time":"2024-12-12T00:15:14.442084797Z","level":"INFO","msg":"Pausing system monitor"}
40
- {"time":"2024-12-12T00:15:14.446183044Z","level":"INFO","msg":"Resuming system monitor"}
41
- {"time":"2024-12-12T00:15:14.446655073Z","level":"INFO","msg":"Pausing system monitor"}
42
- {"time":"2024-12-12T00:15:14.451373362Z","level":"INFO","msg":"Resuming system monitor"}
43
- {"time":"2024-12-12T00:15:14.451923415Z","level":"INFO","msg":"Pausing system monitor"}
44
- {"time":"2024-12-12T00:15:14.456687874Z","level":"INFO","msg":"Resuming system monitor"}
45
- {"time":"2024-12-12T00:15:14.458832191Z","level":"INFO","msg":"Stopping system monitor"}
46
- {"time":"2024-12-12T00:15:14.459388875Z","level":"INFO","msg":"Stopped system monitor"}
47
- {"time":"2024-12-12T00:15:14.683699245Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
48
- {"time":"2024-12-12T00:15:14.738940691Z","level":"INFO","msg":"handler: operation stats","stats":{}}
49
- {"time":"2024-12-12T00:15:15.752911267Z","level":"INFO","msg":"stream: closing","id":"hiitv1j3"}
50
- {"time":"2024-12-12T00:15:15.752941075Z","level":"INFO","msg":"handler: closed","stream_id":"hiitv1j3"}
51
- {"time":"2024-12-12T00:15:15.752966684Z","level":"INFO","msg":"writer: Close: closed","stream_id":"hiitv1j3"}
52
- {"time":"2024-12-12T00:15:15.753004251Z","level":"INFO","msg":"sender: closed","stream_id":"hiitv1j3"}
53
- {"time":"2024-12-12T00:15:15.75304367Z","level":"INFO","msg":"stream: closed","id":"hiitv1j3"}
 
1
+ {"time":"2024-12-12T01:56:09.139880384Z","level":"INFO","msg":"using version","core version":"0.19.0"}
2
+ {"time":"2024-12-12T01:56:09.139899774Z","level":"INFO","msg":"created symlink","path":"/content/wandb/run-20241212_015609-avictioo/logs/debug-core.log"}
3
+ {"time":"2024-12-12T01:56:09.242521075Z","level":"INFO","msg":"created new stream","id":"avictioo"}
4
+ {"time":"2024-12-12T01:56:09.242556763Z","level":"INFO","msg":"stream: started","id":"avictioo"}
5
+ {"time":"2024-12-12T01:56:09.242589972Z","level":"INFO","msg":"writer: Do: started","stream_id":"avictioo"}
6
+ {"time":"2024-12-12T01:56:09.242603199Z","level":"INFO","msg":"sender: started","stream_id":"avictioo"}
7
+ {"time":"2024-12-12T01:56:09.242724569Z","level":"INFO","msg":"handler: started","stream_id":"avictioo"}
8
+ {"time":"2024-12-12T01:56:10.459259928Z","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-12-12T01:56:10.942972743Z","level":"INFO","msg":"Pausing system monitor"}
10
+ {"time":"2024-12-12T01:56:10.943018423Z","level":"INFO","msg":"Resuming system monitor"}
11
+ {"time":"2024-12-12T01:56:10.943025495Z","level":"INFO","msg":"Pausing system monitor"}
12
+ {"time":"2024-12-12T01:56:18.078640335Z","level":"INFO","msg":"Resuming system monitor"}
13
+ {"time":"2024-12-12T01:56:18.270786743Z","level":"INFO","msg":"Pausing system monitor"}
14
+ {"time":"2024-12-12T01:56:41.164201522Z","level":"INFO","msg":"Resuming system monitor"}
15
+ {"time":"2024-12-12T01:56:41.322300255Z","level":"INFO","msg":"Pausing system monitor"}
16
+ {"time":"2024-12-12T01:56:48.559769389Z","level":"INFO","msg":"Resuming system monitor"}
17
+ {"time":"2024-12-12T02:03:40.670051298Z","level":"INFO","msg":"Pausing system monitor"}
18
+ {"time":"2024-12-12T02:03:40.6755442Z","level":"INFO","msg":"Resuming system monitor"}
19
+ {"time":"2024-12-12T02:03:41.741846657Z","level":"INFO","msg":"Pausing system monitor"}
20
+ {"time":"2024-12-12T02:03:46.457531666Z","level":"INFO","msg":"Resuming system monitor"}
21
+ {"time":"2024-12-12T02:03:47.497036163Z","level":"INFO","msg":"Pausing system monitor"}
22
+ {"time":"2024-12-12T02:03:55.070636717Z","level":"INFO","msg":"Resuming system monitor"}
23
+ {"time":"2024-12-12T02:03:55.110799716Z","level":"INFO","msg":"Pausing system monitor"}
24
+ {"time":"2024-12-12T02:04:01.355306831Z","level":"INFO","msg":"Resuming system monitor"}
25
+ {"time":"2024-12-12T02:04:01.422016547Z","level":"INFO","msg":"Pausing system monitor"}
26
+ {"time":"2024-12-12T02:04:02.951075699Z","level":"INFO","msg":"Resuming system monitor"}
27
+ {"time":"2024-12-12T02:04:02.954652415Z","level":"INFO","msg":"Pausing system monitor"}
28
+ {"time":"2024-12-12T02:04:06.048869636Z","level":"INFO","msg":"Resuming system monitor"}
29
+ {"time":"2024-12-12T02:04:07.645834166Z","level":"INFO","msg":"Pausing system monitor"}
30
+ {"time":"2024-12-12T02:04:07.651868639Z","level":"INFO","msg":"Resuming system monitor"}
31
+ {"time":"2024-12-12T02:04:07.653596306Z","level":"INFO","msg":"Pausing system monitor"}
32
+ {"time":"2024-12-12T02:04:07.658067883Z","level":"INFO","msg":"Resuming system monitor"}
33
+ {"time":"2024-12-12T02:04:07.670163111Z","level":"INFO","msg":"Pausing system monitor"}
34
+ {"time":"2024-12-12T02:04:10.824316257Z","level":"INFO","msg":"Resuming system monitor"}
35
+ {"time":"2024-12-12T02:04:10.826489771Z","level":"INFO","msg":"Stopping system monitor"}
36
+ {"time":"2024-12-12T02:04:10.827092344Z","level":"INFO","msg":"Stopped system monitor"}
37
+ {"time":"2024-12-12T02:04:11.142390703Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
38
+ {"time":"2024-12-12T02:04:11.206051251Z","level":"INFO","msg":"handler: operation stats","stats":{}}
39
+ {"time":"2024-12-12T02:04:12.220326425Z","level":"INFO","msg":"stream: closing","id":"avictioo"}
40
+ {"time":"2024-12-12T02:04:12.22034956Z","level":"INFO","msg":"handler: closed","stream_id":"avictioo"}
41
+ {"time":"2024-12-12T02:04:12.22035947Z","level":"INFO","msg":"writer: Close: closed","stream_id":"avictioo"}
42
+ {"time":"2024-12-12T02:04:12.220379214Z","level":"INFO","msg":"sender: closed","stream_id":"avictioo"}
43
+ {"time":"2024-12-12T02:04:12.22043941Z","level":"INFO","msg":"stream: closed","id":"avictioo"}
 
 
 
 
 
 
 
 
 
 
wandb/debug.log CHANGED
@@ -1,91 +1,76 @@
1
- 2024-12-12 00:13:10,420 INFO MainThread:205 [wandb_setup.py:_flush():68] Current SDK version is 0.19.0
2
- 2024-12-12 00:13:10,420 INFO MainThread:205 [wandb_setup.py:_flush():68] Configure stats pid to 205
3
- 2024-12-12 00:13:10,420 INFO MainThread:205 [wandb_setup.py:_flush():68] Loading settings from /root/.config/wandb/settings
4
- 2024-12-12 00:13:10,420 INFO MainThread:205 [wandb_setup.py:_flush():68] Loading settings from /content/wandb/settings
5
- 2024-12-12 00:13:10,420 INFO MainThread:205 [wandb_setup.py:_flush():68] Loading settings from environment variables
6
- 2024-12-12 00:13:10,421 INFO MainThread:205 [wandb_init.py:_log_setup():528] Logging user logs to /content/wandb/run-20241212_001310-hiitv1j3/logs/debug.log
7
- 2024-12-12 00:13:10,421 INFO MainThread:205 [wandb_init.py:_log_setup():529] Logging internal logs to /content/wandb/run-20241212_001310-hiitv1j3/logs/debug-internal.log
8
- 2024-12-12 00:13:10,421 INFO MainThread:205 [wandb_init.py:_jupyter_setup():474] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7e35531aabf0>
9
- 2024-12-12 00:13:10,421 INFO MainThread:205 [wandb_init.py:init():639] calling init triggers
10
- 2024-12-12 00:13:10,421 INFO MainThread:205 [wandb_init.py:init():645] wandb.init called with sweep_config: {}
11
  config: {}
12
- 2024-12-12 00:13:10,421 INFO MainThread:205 [wandb_init.py:init():688] starting backend
13
- 2024-12-12 00:13:10,421 INFO MainThread:205 [wandb_init.py:init():692] sending inform_init request
14
- 2024-12-12 00:13:10,428 INFO MainThread:205 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
15
- 2024-12-12 00:13:10,428 INFO MainThread:205 [wandb_init.py:init():705] backend started and connected
16
- 2024-12-12 00:13:10,438 INFO MainThread:205 [wandb_run.py:_label_probe_notebook():1327] probe notebook
17
- 2024-12-12 00:13:12,154 INFO MainThread:205 [wandb_init.py:init():798] updated telemetry
18
- 2024-12-12 00:13:12,155 INFO MainThread:205 [wandb_init.py:init():830] communicating run to backend with 90.0 second timeout
19
- 2024-12-12 00:13:12,362 INFO MainThread:205 [wandb_init.py:init():882] starting run threads in backend
20
- 2024-12-12 00:13:12,809 INFO MainThread:205 [wandb_run.py:_console_start():2443] atexit reg
21
- 2024-12-12 00:13:12,809 INFO MainThread:205 [wandb_run.py:_redirect():2293] redirect: wrap_raw
22
- 2024-12-12 00:13:12,809 INFO MainThread:205 [wandb_run.py:_redirect():2358] Wrapping output streams.
23
- 2024-12-12 00:13:12,809 INFO MainThread:205 [wandb_run.py:_redirect():2383] Redirects installed.
24
- 2024-12-12 00:13:12,811 INFO MainThread:205 [wandb_init.py:init():925] run started, returning control to user process
25
- 2024-12-12 00:13:12,815 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
26
- 2024-12-12 00:13:12,815 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
27
- 2024-12-12 00:13:12,820 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
28
- 2024-12-12 00:13:12,820 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
29
- 2024-12-12 00:13:12,820 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
30
- 2024-12-12 00:13:12,829 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
31
- 2024-12-12 00:13:12,999 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
32
- 2024-12-12 00:13:12,999 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
33
- 2024-12-12 00:13:13,005 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
34
- 2024-12-12 00:13:13,157 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
35
- 2024-12-12 00:13:13,157 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
36
- 2024-12-12 00:13:13,162 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
37
- 2024-12-12 00:13:13,574 INFO MainThread:205 [wandb_run.py:_config_callback():1375] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertForMaskedLM'], 'finetuning_task': None, 'id2label': {0: '0', 1: '1'}, 'label2id': {'0': 0, '1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'neuralmind/bert-base-portuguese-cased', '_attn_implementation_autoset': True, 'transformers_version': '4.47.0', 'directionality': 'bidi', 'model_type': 'bert', 'output_past': True, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'vocab_size': 29794, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.05, 'attention_probs_dropout_prob': 0.05, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '.', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 1e-05, 'weight_decay': 0.02, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 15, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 150, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './runs/Dec12_00-13-13_59171548ab11', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 102, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 5151, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 51, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'm2-bertimbau_tunado-com_foca_loss_centropy', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'f1', 'greater_is_better': True, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'average_tokens_across_devices': False}
38
- 2024-12-12 00:13:13,577 INFO MainThread:205 [wandb_config.py:__setitem__():154] config set model/num_parameters = 108924674 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7e33843d3160>>
39
- 2024-12-12 00:13:13,578 INFO MainThread:205 [wandb_run.py:_config_callback():1375] config_cb model/num_parameters 108924674 None
40
- 2024-12-12 00:15:12,139 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
41
- 2024-12-12 00:15:12,139 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
42
- 2024-12-12 00:15:12,145 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
43
- 2024-12-12 00:15:12,639 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
44
- 2024-12-12 00:15:12,639 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
45
- 2024-12-12 00:15:12,644 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
46
- 2024-12-12 00:15:12,644 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
47
- 2024-12-12 00:15:12,644 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
48
- 2024-12-12 00:15:12,650 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
49
- 2024-12-12 00:15:12,731 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
50
- 2024-12-12 00:15:12,731 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
51
- 2024-12-12 00:15:12,736 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
52
- 2024-12-12 00:15:12,738 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
53
- 2024-12-12 00:15:12,739 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
54
- 2024-12-12 00:15:12,744 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
55
- 2024-12-12 00:15:14,393 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
56
- 2024-12-12 00:15:14,393 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
57
- 2024-12-12 00:15:14,401 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
58
- 2024-12-12 00:15:14,402 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
59
- 2024-12-12 00:15:14,403 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
60
- 2024-12-12 00:15:14,408 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
61
- 2024-12-12 00:15:14,421 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
62
- 2024-12-12 00:15:14,421 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
63
- 2024-12-12 00:15:14,425 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
64
- 2024-12-12 00:15:14,426 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
65
- 2024-12-12 00:15:14,426 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
66
- 2024-12-12 00:15:14,431 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
67
- 2024-12-12 00:15:14,431 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
68
- 2024-12-12 00:15:14,431 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
69
- 2024-12-12 00:15:14,435 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
70
- 2024-12-12 00:15:14,436 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
71
- 2024-12-12 00:15:14,436 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
72
- 2024-12-12 00:15:14,440 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
73
- 2024-12-12 00:15:14,441 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
74
- 2024-12-12 00:15:14,441 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
75
- 2024-12-12 00:15:14,445 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
76
- 2024-12-12 00:15:14,446 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
77
- 2024-12-12 00:15:14,446 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
78
- 2024-12-12 00:15:14,451 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
79
- 2024-12-12 00:15:14,451 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
80
- 2024-12-12 00:15:14,451 INFO MainThread:205 [wandb_init.py:_pause_backend():439] pausing backend
81
- 2024-12-12 00:15:14,456 INFO MainThread:205 [wandb_init.py:_resume_backend():444] resuming backend
82
- 2024-12-12 00:15:14,457 INFO MainThread:205 [wandb_run.py:_finish():2169] finishing run adrianobf/AzMina-M2/hiitv1j3
83
- 2024-12-12 00:15:14,457 INFO MainThread:205 [jupyter.py:save_history():459] not saving jupyter history
84
- 2024-12-12 00:15:14,457 INFO MainThread:205 [jupyter.py:save_ipynb():387] not saving jupyter notebook
85
- 2024-12-12 00:15:14,457 INFO MainThread:205 [wandb_init.py:_jupyter_teardown():456] cleaning up jupyter logic
86
- 2024-12-12 00:15:14,457 INFO MainThread:205 [wandb_run.py:_atexit_cleanup():2408] got exitcode: 0
87
- 2024-12-12 00:15:14,458 INFO MainThread:205 [wandb_run.py:_restore():2390] restore
88
- 2024-12-12 00:15:14,458 INFO MainThread:205 [wandb_run.py:_restore():2396] restore done
89
- 2024-12-12 00:15:15,741 INFO MainThread:205 [wandb_run.py:_footer_history_summary_info():3951] rendering history
90
- 2024-12-12 00:15:15,742 INFO MainThread:205 [wandb_run.py:_footer_history_summary_info():3983] rendering summary
91
- 2024-12-12 00:15:15,749 INFO MainThread:205 [wandb_run.py:_footer_sync_info():3912] logging synced files
 
1
+ 2024-12-12 01:56:09,130 INFO MainThread:1468 [wandb_setup.py:_flush():68] Current SDK version is 0.19.0
2
+ 2024-12-12 01:56:09,130 INFO MainThread:1468 [wandb_setup.py:_flush():68] Configure stats pid to 1468
3
+ 2024-12-12 01:56:09,130 INFO MainThread:1468 [wandb_setup.py:_flush():68] Loading settings from /root/.config/wandb/settings
4
+ 2024-12-12 01:56:09,130 INFO MainThread:1468 [wandb_setup.py:_flush():68] Loading settings from /content/wandb/settings
5
+ 2024-12-12 01:56:09,130 INFO MainThread:1468 [wandb_setup.py:_flush():68] Loading settings from environment variables
6
+ 2024-12-12 01:56:09,130 INFO MainThread:1468 [wandb_init.py:_log_setup():528] Logging user logs to /content/wandb/run-20241212_015609-avictioo/logs/debug.log
7
+ 2024-12-12 01:56:09,131 INFO MainThread:1468 [wandb_init.py:_log_setup():529] Logging internal logs to /content/wandb/run-20241212_015609-avictioo/logs/debug-internal.log
8
+ 2024-12-12 01:56:09,131 INFO MainThread:1468 [wandb_init.py:_jupyter_setup():474] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x79bb96df63b0>
9
+ 2024-12-12 01:56:09,131 INFO MainThread:1468 [wandb_init.py:init():639] calling init triggers
10
+ 2024-12-12 01:56:09,131 INFO MainThread:1468 [wandb_init.py:init():645] wandb.init called with sweep_config: {}
11
  config: {}
12
+ 2024-12-12 01:56:09,131 INFO MainThread:1468 [wandb_init.py:init():688] starting backend
13
+ 2024-12-12 01:56:09,131 INFO MainThread:1468 [wandb_init.py:init():692] sending inform_init request
14
+ 2024-12-12 01:56:09,138 INFO MainThread:1468 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
15
+ 2024-12-12 01:56:09,138 INFO MainThread:1468 [wandb_init.py:init():705] backend started and connected
16
+ 2024-12-12 01:56:09,145 INFO MainThread:1468 [wandb_run.py:_label_probe_notebook():1327] probe notebook
17
+ 2024-12-12 01:56:10,307 INFO MainThread:1468 [wandb_init.py:init():798] updated telemetry
18
+ 2024-12-12 01:56:10,308 INFO MainThread:1468 [wandb_init.py:init():830] communicating run to backend with 90.0 second timeout
19
+ 2024-12-12 01:56:10,454 INFO MainThread:1468 [wandb_init.py:init():882] starting run threads in backend
20
+ 2024-12-12 01:56:10,902 INFO MainThread:1468 [wandb_run.py:_console_start():2443] atexit reg
21
+ 2024-12-12 01:56:10,902 INFO MainThread:1468 [wandb_run.py:_redirect():2293] redirect: wrap_raw
22
+ 2024-12-12 01:56:10,902 INFO MainThread:1468 [wandb_run.py:_redirect():2358] Wrapping output streams.
23
+ 2024-12-12 01:56:10,903 INFO MainThread:1468 [wandb_run.py:_redirect():2383] Redirects installed.
24
+ 2024-12-12 01:56:10,905 INFO MainThread:1468 [wandb_init.py:init():925] run started, returning control to user process
25
+ 2024-12-12 01:56:10,908 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
26
+ 2024-12-12 01:56:10,908 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
27
+ 2024-12-12 01:56:10,913 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
28
+ 2024-12-12 01:56:10,913 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
29
+ 2024-12-12 01:56:10,914 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
30
+ 2024-12-12 01:56:18,077 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
31
+ 2024-12-12 01:56:18,270 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
32
+ 2024-12-12 01:56:18,270 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
33
+ 2024-12-12 01:56:41,163 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
34
+ 2024-12-12 01:56:41,321 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
35
+ 2024-12-12 01:56:41,322 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
36
+ 2024-12-12 01:56:48,559 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
37
+ 2024-12-12 01:56:48,971 INFO MainThread:1468 [wandb_run.py:_config_callback():1375] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertForMaskedLM'], 'finetuning_task': None, 'id2label': {0: '0', 1: '1'}, 'label2id': {'0': 0, '1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'neuralmind/bert-base-portuguese-cased', '_attn_implementation_autoset': True, 'transformers_version': '4.47.0', 'directionality': 'bidi', 'model_type': 'bert', 'output_past': True, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'vocab_size': 29794, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.07, 'attention_probs_dropout_prob': 0.07, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '.', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 1e-05, 'weight_decay': 0.02, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 19, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 150, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './runs/Dec12_01-56-41_4f84ece6c072', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 102, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 5151, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 51, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'm2-bertimbau_tunado-maisattentiondropout', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'f1', 'greater_is_better': True, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'average_tokens_across_devices': False}
38
+ 2024-12-12 01:56:48,974 INFO MainThread:1468 [wandb_config.py:__setitem__():154] config set model/num_parameters = 108924674 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x79bbc7981b10>>
39
+ 2024-12-12 01:56:48,975 INFO MainThread:1468 [wandb_run.py:_config_callback():1375] config_cb model/num_parameters 108924674 None
40
+ 2024-12-12 02:03:40,669 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
41
+ 2024-12-12 02:03:40,669 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
42
+ 2024-12-12 02:03:40,675 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
43
+ 2024-12-12 02:03:41,702 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
44
+ 2024-12-12 02:03:41,702 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
45
+ 2024-12-12 02:03:46,456 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
46
+ 2024-12-12 02:03:47,496 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
47
+ 2024-12-12 02:03:47,496 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
48
+ 2024-12-12 02:03:55,070 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
49
+ 2024-12-12 02:03:55,073 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
50
+ 2024-12-12 02:03:55,073 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
51
+ 2024-12-12 02:04:01,354 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
52
+ 2024-12-12 02:04:01,421 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
53
+ 2024-12-12 02:04:01,421 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
54
+ 2024-12-12 02:04:02,950 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
55
+ 2024-12-12 02:04:02,954 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
56
+ 2024-12-12 02:04:02,954 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
57
+ 2024-12-12 02:04:06,048 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
58
+ 2024-12-12 02:04:07,645 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
59
+ 2024-12-12 02:04:07,645 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
60
+ 2024-12-12 02:04:07,651 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
61
+ 2024-12-12 02:04:07,653 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
62
+ 2024-12-12 02:04:07,653 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
63
+ 2024-12-12 02:04:07,657 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
64
+ 2024-12-12 02:04:07,669 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
65
+ 2024-12-12 02:04:07,670 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
66
+ 2024-12-12 02:04:10,823 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
67
+ 2024-12-12 02:04:10,824 INFO MainThread:1468 [wandb_run.py:_finish():2169] finishing run adrianobf/AzMina-M2/avictioo
68
+ 2024-12-12 02:04:10,825 INFO MainThread:1468 [jupyter.py:save_history():459] not saving jupyter history
69
+ 2024-12-12 02:04:10,825 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
70
+ 2024-12-12 02:04:10,825 INFO MainThread:1468 [wandb_init.py:_jupyter_teardown():456] cleaning up jupyter logic
71
+ 2024-12-12 02:04:10,825 INFO MainThread:1468 [wandb_run.py:_atexit_cleanup():2408] got exitcode: 0
72
+ 2024-12-12 02:04:10,825 INFO MainThread:1468 [wandb_run.py:_restore():2390] restore
73
+ 2024-12-12 02:04:10,825 INFO MainThread:1468 [wandb_run.py:_restore():2396] restore done
74
+ 2024-12-12 02:04:12,209 INFO MainThread:1468 [wandb_run.py:_footer_history_summary_info():3951] rendering history
75
+ 2024-12-12 02:04:12,210 INFO MainThread:1468 [wandb_run.py:_footer_history_summary_info():3983] rendering summary
76
+ 2024-12-12 02:04:12,217 INFO MainThread:1468 [wandb_run.py:_footer_sync_info():3912] logging synced files
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
wandb/run-20241212_015609-avictioo/files/config.yaml ADDED
@@ -0,0 +1,568 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _attn_implementation_autoset:
2
+ value: true
3
+ _name_or_path:
4
+ value: neuralmind/bert-base-portuguese-cased
5
+ _wandb:
6
+ value:
7
+ cli_version: 0.19.0
8
+ m:
9
+ - "1": train/grad_norm
10
+ "5": 2
11
+ "6":
12
+ - 1
13
+ - 3
14
+ "7": []
15
+ - "1": train/global_step
16
+ "6":
17
+ - 3
18
+ "7": []
19
+ - "1": train/epoch
20
+ "5": 2
21
+ "6":
22
+ - 1
23
+ - 3
24
+ "7": []
25
+ - "1": eval/samples_per_second
26
+ "5": 2
27
+ "6":
28
+ - 1
29
+ - 3
30
+ "7": []
31
+ - "1": eval/steps_per_second
32
+ "5": 2
33
+ "6":
34
+ - 1
35
+ - 3
36
+ "7": []
37
+ - "1": eval/accuracy
38
+ "5": 2
39
+ "6":
40
+ - 1
41
+ - 3
42
+ "7": []
43
+ - "1": eval/recall
44
+ "5": 2
45
+ "6":
46
+ - 1
47
+ - 3
48
+ "7": []
49
+ - "1": train/loss
50
+ "5": 2
51
+ "6":
52
+ - 1
53
+ - 3
54
+ "7": []
55
+ - "1": eval/f1
56
+ "5": 2
57
+ "6":
58
+ - 1
59
+ - 3
60
+ "7": []
61
+ - "1": eval/precision
62
+ "5": 2
63
+ "6":
64
+ - 1
65
+ - 3
66
+ "7": []
67
+ - "1": eval/loss
68
+ "5": 2
69
+ "6":
70
+ - 1
71
+ - 3
72
+ "7": []
73
+ - "1": train/learning_rate
74
+ "5": 2
75
+ "6":
76
+ - 1
77
+ - 3
78
+ "7": []
79
+ - "1": eval/runtime
80
+ "5": 2
81
+ "6":
82
+ - 1
83
+ - 3
84
+ "7": []
85
+ python_version: 3.10.12
86
+ t:
87
+ "1":
88
+ - 1
89
+ - 2
90
+ - 3
91
+ - 5
92
+ - 11
93
+ - 12
94
+ - 41
95
+ - 49
96
+ - 51
97
+ - 53
98
+ - 55
99
+ - 71
100
+ - 98
101
+ - 100
102
+ - 105
103
+ "2":
104
+ - 1
105
+ - 2
106
+ - 3
107
+ - 5
108
+ - 11
109
+ - 12
110
+ - 41
111
+ - 49
112
+ - 51
113
+ - 53
114
+ - 55
115
+ - 71
116
+ - 98
117
+ - 100
118
+ - 105
119
+ "3":
120
+ - 2
121
+ - 7
122
+ - 13
123
+ - 19
124
+ - 23
125
+ - 55
126
+ - 62
127
+ - 66
128
+ "4": 3.10.12
129
+ "5": 0.19.0
130
+ "6": 4.47.0
131
+ "8":
132
+ - 1
133
+ - 5
134
+ - 12
135
+ "9":
136
+ "1": transformers_trainer
137
+ "12": 0.19.0
138
+ "13": linux-x86_64
139
+ accelerator_config:
140
+ value:
141
+ dispatch_batches: null
142
+ even_batches: true
143
+ gradient_accumulation_kwargs: null
144
+ non_blocking: false
145
+ split_batches: false
146
+ use_seedable_sampler: true
147
+ adafactor:
148
+ value: false
149
+ adam_beta1:
150
+ value: 0.9
151
+ adam_beta2:
152
+ value: 0.999
153
+ adam_epsilon:
154
+ value: 1e-08
155
+ add_cross_attention:
156
+ value: false
157
+ architectures:
158
+ value:
159
+ - BertForMaskedLM
160
+ attention_probs_dropout_prob:
161
+ value: 0.07
162
+ auto_find_batch_size:
163
+ value: false
164
+ average_tokens_across_devices:
165
+ value: false
166
+ bad_words_ids:
167
+ value: null
168
+ batch_eval_metrics:
169
+ value: false
170
+ begin_suppress_tokens:
171
+ value: null
172
+ bf16:
173
+ value: false
174
+ bf16_full_eval:
175
+ value: false
176
+ bos_token_id:
177
+ value: null
178
+ chunk_size_feed_forward:
179
+ value: 0
180
+ classifier_dropout:
181
+ value: null
182
+ cross_attention_hidden_size:
183
+ value: null
184
+ data_seed:
185
+ value: null
186
+ dataloader_drop_last:
187
+ value: false
188
+ dataloader_num_workers:
189
+ value: 0
190
+ dataloader_persistent_workers:
191
+ value: false
192
+ dataloader_pin_memory:
193
+ value: true
194
+ dataloader_prefetch_factor:
195
+ value: null
196
+ ddp_backend:
197
+ value: null
198
+ ddp_broadcast_buffers:
199
+ value: null
200
+ ddp_bucket_cap_mb:
201
+ value: null
202
+ ddp_find_unused_parameters:
203
+ value: null
204
+ ddp_timeout:
205
+ value: 1800
206
+ debug:
207
+ value: []
208
+ decoder_start_token_id:
209
+ value: null
210
+ deepspeed:
211
+ value: null
212
+ directionality:
213
+ value: bidi
214
+ disable_tqdm:
215
+ value: false
216
+ dispatch_batches:
217
+ value: null
218
+ diversity_penalty:
219
+ value: 0
220
+ do_eval:
221
+ value: true
222
+ do_predict:
223
+ value: false
224
+ do_sample:
225
+ value: false
226
+ do_train:
227
+ value: false
228
+ early_stopping:
229
+ value: false
230
+ encoder_no_repeat_ngram_size:
231
+ value: 0
232
+ eos_token_id:
233
+ value: null
234
+ eval_accumulation_steps:
235
+ value: null
236
+ eval_delay:
237
+ value: 0
238
+ eval_do_concat_batches:
239
+ value: true
240
+ eval_on_start:
241
+ value: false
242
+ eval_steps:
243
+ value: 51
244
+ eval_strategy:
245
+ value: epoch
246
+ eval_use_gather_object:
247
+ value: false
248
+ evaluation_strategy:
249
+ value: null
250
+ exponential_decay_length_penalty:
251
+ value: null
252
+ finetuning_task:
253
+ value: null
254
+ forced_bos_token_id:
255
+ value: null
256
+ forced_eos_token_id:
257
+ value: null
258
+ fp16:
259
+ value: false
260
+ fp16_backend:
261
+ value: auto
262
+ fp16_full_eval:
263
+ value: false
264
+ fp16_opt_level:
265
+ value: O1
266
+ fsdp:
267
+ value: []
268
+ fsdp_config:
269
+ value:
270
+ min_num_params: 0
271
+ xla: false
272
+ xla_fsdp_grad_ckpt: false
273
+ xla_fsdp_v2: false
274
+ fsdp_min_num_params:
275
+ value: 0
276
+ fsdp_transformer_layer_cls_to_wrap:
277
+ value: null
278
+ full_determinism:
279
+ value: false
280
+ gradient_accumulation_steps:
281
+ value: 1
282
+ gradient_checkpointing:
283
+ value: false
284
+ gradient_checkpointing_kwargs:
285
+ value: null
286
+ greater_is_better:
287
+ value: true
288
+ group_by_length:
289
+ value: false
290
+ half_precision_backend:
291
+ value: auto
292
+ hidden_act:
293
+ value: gelu
294
+ hidden_dropout_prob:
295
+ value: 0.07
296
+ hidden_size:
297
+ value: 768
298
+ hub_always_push:
299
+ value: false
300
+ hub_model_id:
301
+ value: null
302
+ hub_private_repo:
303
+ value: null
304
+ hub_strategy:
305
+ value: every_save
306
+ hub_token:
307
+ value: <HUB_TOKEN>
308
+ id2label:
309
+ value:
310
+ "0": "0"
311
+ "1": "1"
312
+ ignore_data_skip:
313
+ value: false
314
+ include_for_metrics:
315
+ value: []
316
+ include_inputs_for_metrics:
317
+ value: false
318
+ include_num_input_tokens_seen:
319
+ value: false
320
+ include_tokens_per_second:
321
+ value: false
322
+ initializer_range:
323
+ value: 0.02
324
+ intermediate_size:
325
+ value: 3072
326
+ is_decoder:
327
+ value: false
328
+ is_encoder_decoder:
329
+ value: false
330
+ jit_mode_eval:
331
+ value: false
332
+ label_names:
333
+ value: null
334
+ label_smoothing_factor:
335
+ value: 0
336
+ label2id:
337
+ value:
338
+ "0": 0
339
+ "1": 1
340
+ layer_norm_eps:
341
+ value: 1e-12
342
+ learning_rate:
343
+ value: 1e-05
344
+ length_column_name:
345
+ value: length
346
+ length_penalty:
347
+ value: 1
348
+ load_best_model_at_end:
349
+ value: true
350
+ local_rank:
351
+ value: 0
352
+ log_level:
353
+ value: passive
354
+ log_level_replica:
355
+ value: warning
356
+ log_on_each_node:
357
+ value: true
358
+ logging_dir:
359
+ value: ./runs/Dec12_01-56-41_4f84ece6c072
360
+ logging_first_step:
361
+ value: false
362
+ logging_nan_inf_filter:
363
+ value: true
364
+ logging_steps:
365
+ value: 10
366
+ logging_strategy:
367
+ value: steps
368
+ lr_scheduler_type:
369
+ value: linear
370
+ max_grad_norm:
371
+ value: 1
372
+ max_length:
373
+ value: 20
374
+ max_position_embeddings:
375
+ value: 512
376
+ max_steps:
377
+ value: -1
378
+ metric_for_best_model:
379
+ value: f1
380
+ min_length:
381
+ value: 0
382
+ model/num_parameters:
383
+ value: 108924674
384
+ model_type:
385
+ value: bert
386
+ mp_parameters:
387
+ value: ""
388
+ neftune_noise_alpha:
389
+ value: null
390
+ no_cuda:
391
+ value: false
392
+ no_repeat_ngram_size:
393
+ value: 0
394
+ num_attention_heads:
395
+ value: 12
396
+ num_beam_groups:
397
+ value: 1
398
+ num_beams:
399
+ value: 1
400
+ num_hidden_layers:
401
+ value: 12
402
+ num_return_sequences:
403
+ value: 1
404
+ num_train_epochs:
405
+ value: 19
406
+ optim:
407
+ value: adamw_torch
408
+ optim_args:
409
+ value: null
410
+ optim_target_modules:
411
+ value: null
412
+ output_attentions:
413
+ value: false
414
+ output_dir:
415
+ value: .
416
+ output_hidden_states:
417
+ value: false
418
+ output_past:
419
+ value: true
420
+ output_scores:
421
+ value: false
422
+ overwrite_output_dir:
423
+ value: false
424
+ pad_token_id:
425
+ value: 0
426
+ past_index:
427
+ value: -1
428
+ per_device_eval_batch_size:
429
+ value: 64
430
+ per_device_train_batch_size:
431
+ value: 64
432
+ per_gpu_eval_batch_size:
433
+ value: null
434
+ per_gpu_train_batch_size:
435
+ value: null
436
+ pooler_fc_size:
437
+ value: 768
438
+ pooler_num_attention_heads:
439
+ value: 12
440
+ pooler_num_fc_layers:
441
+ value: 3
442
+ pooler_size_per_head:
443
+ value: 128
444
+ pooler_type:
445
+ value: first_token_transform
446
+ position_embedding_type:
447
+ value: absolute
448
+ prediction_loss_only:
449
+ value: false
450
+ prefix:
451
+ value: null
452
+ problem_type:
453
+ value: null
454
+ push_to_hub:
455
+ value: false
456
+ push_to_hub_model_id:
457
+ value: null
458
+ push_to_hub_organization:
459
+ value: null
460
+ push_to_hub_token:
461
+ value: <PUSH_TO_HUB_TOKEN>
462
+ ray_scope:
463
+ value: last
464
+ remove_invalid_values:
465
+ value: false
466
+ remove_unused_columns:
467
+ value: true
468
+ repetition_penalty:
469
+ value: 1
470
+ report_to:
471
+ value:
472
+ - wandb
473
+ restore_callback_states_from_checkpoint:
474
+ value: false
475
+ resume_from_checkpoint:
476
+ value: null
477
+ return_dict:
478
+ value: true
479
+ return_dict_in_generate:
480
+ value: false
481
+ run_name:
482
+ value: m2-bertimbau_tunado-maisattentiondropout
483
+ save_on_each_node:
484
+ value: false
485
+ save_only_model:
486
+ value: false
487
+ save_safetensors:
488
+ value: true
489
+ save_steps:
490
+ value: 102
491
+ save_strategy:
492
+ value: epoch
493
+ save_total_limit:
494
+ value: 1
495
+ seed:
496
+ value: 5151
497
+ sep_token_id:
498
+ value: null
499
+ skip_memory_metrics:
500
+ value: true
501
+ split_batches:
502
+ value: null
503
+ suppress_tokens:
504
+ value: null
505
+ task_specific_params:
506
+ value: null
507
+ temperature:
508
+ value: 1
509
+ tf_legacy_loss:
510
+ value: false
511
+ tf32:
512
+ value: null
513
+ tie_encoder_decoder:
514
+ value: false
515
+ tie_word_embeddings:
516
+ value: true
517
+ tokenizer_class:
518
+ value: null
519
+ top_k:
520
+ value: 50
521
+ top_p:
522
+ value: 1
523
+ torch_compile:
524
+ value: false
525
+ torch_compile_backend:
526
+ value: null
527
+ torch_compile_mode:
528
+ value: null
529
+ torch_dtype:
530
+ value: null
531
+ torch_empty_cache_steps:
532
+ value: null
533
+ torchdynamo:
534
+ value: null
535
+ torchscript:
536
+ value: false
537
+ tpu_metrics_debug:
538
+ value: false
539
+ tpu_num_cores:
540
+ value: null
541
+ transformers_version:
542
+ value: 4.47.0
543
+ type_vocab_size:
544
+ value: 2
545
+ typical_p:
546
+ value: 1
547
+ use_bfloat16:
548
+ value: false
549
+ use_cache:
550
+ value: true
551
+ use_cpu:
552
+ value: false
553
+ use_ipex:
554
+ value: false
555
+ use_legacy_prediction_loop:
556
+ value: false
557
+ use_liger_kernel:
558
+ value: false
559
+ use_mps_device:
560
+ value: false
561
+ vocab_size:
562
+ value: 29794
563
+ warmup_ratio:
564
+ value: 0
565
+ warmup_steps:
566
+ value: 150
567
+ weight_decay:
568
+ value: 0.02
wandb/run-20241212_015609-avictioo/files/output.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ /usr/local/lib/python3.10/dist-packages/sklearn/metrics/_classification.py:1531: UndefinedMetricWarning: Precision is ill-defined and being set to 0.0 in labels with no predicted samples. Use `zero_division` parameter to control this behavior.
2
+ _warn_prf(average, modifier, f"{metric.capitalize()} is", len(result))
3
+ precision recall f1-score support
4
+
5
+ 0 0.93 0.85 0.89 114
6
+ 1 0.60 0.79 0.68 33
7
+
8
+ accuracy 0.84 147
9
+ macro avg 0.77 0.82 0.79 147
10
+ weighted avg 0.86 0.84 0.84 147
wandb/run-20241212_015609-avictioo/files/requirements.txt ADDED
@@ -0,0 +1,578 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ datasets==3.2.0
2
+ dill==0.3.8
3
+ transformers==4.47.0
4
+ evaluate==0.4.3
5
+ xxhash==3.5.0
6
+ wandb==0.19.0
7
+ fsspec==2024.9.0
8
+ tokenizers==0.21.0
9
+ accelerate==1.2.0
10
+ multiprocess==0.70.16
11
+ google-colab==1.0.0
12
+ openpyxl==3.1.5
13
+ tensorflow-hub==0.16.1
14
+ jupyter-client==6.1.12
15
+ osqp==0.6.7.post3
16
+ nvidia-cuda-nvcc-cu12==12.6.85
17
+ xyzservices==2024.9.0
18
+ jellyfish==1.1.2
19
+ matplotlib==3.8.0
20
+ progressbar2==4.5.0
21
+ pydata-google-auth==1.9.0
22
+ multitasking==0.0.11
23
+ babel==2.16.0
24
+ pyasn1_modules==0.4.1
25
+ cudf-cu12==24.10.1
26
+ grpcio-status==1.62.3
27
+ openai==1.54.5
28
+ tornado==6.3.3
29
+ cufflinks==0.17.3
30
+ google-cloud-storage==2.8.0
31
+ mkl==2025.0.1
32
+ sphinxcontrib-devhelp==2.0.0
33
+ yfinance==0.2.50
34
+ google-cloud-functions==1.18.1
35
+ aiohttp==3.11.9
36
+ wordcloud==1.9.4
37
+ Werkzeug==3.1.3
38
+ sqlglot==25.1.0
39
+ StrEnum==0.4.15
40
+ langsmith==0.1.147
41
+ gast==0.6.0
42
+ requests-oauthlib==1.3.1
43
+ google==2.0.3
44
+ shellingham==1.5.4
45
+ joblib==1.4.2
46
+ mizani==0.13.0
47
+ mpmath==1.3.0
48
+ tensorflow-io-gcs-filesystem==0.37.1
49
+ ml-dtypes==0.4.1
50
+ rmm-cu12==24.10.0
51
+ wheel==0.45.1
52
+ decorator==4.4.2
53
+ pydantic==2.10.3
54
+ msgpack==1.1.0
55
+ cymem==2.0.10
56
+ blis==0.7.11
57
+ pyzmq==24.0.1
58
+ array_record==0.5.1
59
+ kagglehub==0.3.4
60
+ propcache==0.2.1
61
+ pydantic_core==2.27.1
62
+ threadpoolctl==3.5.0
63
+ sympy==1.13.1
64
+ CacheControl==0.14.1
65
+ Pygments==2.18.0
66
+ folium==0.18.0
67
+ dopamine_rl==4.0.9
68
+ pydotplus==2.0.2
69
+ python-louvain==0.16
70
+ catalogue==2.0.10
71
+ gspread==6.0.2
72
+ pyspark==3.5.3
73
+ google-cloud-core==2.4.1
74
+ firebase-admin==6.5.0
75
+ jieba==0.42.1
76
+ pymc==5.18.2
77
+ jax-cuda12-pjrt==0.4.33
78
+ geemap==0.35.1
79
+ pyparsing==3.2.0
80
+ soupsieve==2.6
81
+ nbconvert==7.16.4
82
+ tzdata==2024.2
83
+ webcolors==24.11.1
84
+ zipp==3.21.0
85
+ peewee==3.17.8
86
+ opentelemetry-api==1.28.2
87
+ nvidia-cufft-cu12==11.3.0.4
88
+ torchaudio==2.5.1+cu121
89
+ dlib==19.24.2
90
+ oauthlib==3.2.2
91
+ optax==0.2.4
92
+ py4j==0.10.9.7
93
+ spacy==3.7.5
94
+ clarabel==0.9.0
95
+ parso==0.8.4
96
+ umf==0.9.1
97
+ pyerfa==2.0.1.5
98
+ pyproj==3.7.0
99
+ Pyomo==6.8.2
100
+ namex==0.0.8
101
+ webencodings==0.5.1
102
+ platformdirs==4.3.6
103
+ absl-py==1.4.0
104
+ wasabi==1.1.3
105
+ kaggle==1.6.17
106
+ notebook_shim==0.2.4
107
+ timm==1.0.12
108
+ geopy==2.4.1
109
+ sphinxcontrib-jsmath==1.0.1
110
+ astunparse==1.6.3
111
+ pycparser==2.22
112
+ jupyterlab_widgets==3.0.13
113
+ networkx==3.4.2
114
+ google-api-core==2.19.2
115
+ jsonpickle==4.0.0
116
+ ipyevents==2.0.2
117
+ miniKanren==1.0.3
118
+ cmdstanpy==1.2.4
119
+ attrs==24.2.0
120
+ snowballstemmer==2.2.0
121
+ pillow==11.0.0
122
+ chardet==5.2.0
123
+ PyYAML==6.0.2
124
+ sniffio==1.3.1
125
+ pytensor==2.26.4
126
+ nx-cugraph-cu12==24.10.0
127
+ graphviz==0.20.3
128
+ requests==2.32.3
129
+ backcall==0.2.0
130
+ smmap==5.0.1
131
+ docker-pycreds==0.4.0
132
+ rsa==4.9
133
+ google-auth==2.27.0
134
+ toolz==0.12.1
135
+ future==1.0.0
136
+ tensorboard==2.17.1
137
+ toml==0.10.2
138
+ cmake==3.30.5
139
+ langchain-core==0.3.21
140
+ imagesize==1.4.1
141
+ prettytable==3.12.0
142
+ fastjsonschema==2.21.1
143
+ h5netcdf==1.4.1
144
+ gitdb==4.0.11
145
+ google-ai-generativelanguage==0.6.10
146
+ duckdb==1.1.3
147
+ proto-plus==1.25.0
148
+ inflect==7.4.0
149
+ scipy==1.13.1
150
+ weasel==0.4.1
151
+ sklearn-pandas==2.2.0
152
+ filelock==3.16.1
153
+ chex==0.1.87
154
+ moviepy==1.0.3
155
+ async-timeout==4.0.3
156
+ google-crc32c==1.6.0
157
+ panel==1.5.4
158
+ packaging==24.2
159
+ frozenlist==1.5.0
160
+ h5py==3.12.1
161
+ cuda-python==12.2.1
162
+ eerepr==0.0.4
163
+ orjson==3.10.12
164
+ requests-toolbelt==1.0.0
165
+ pydot==3.0.3
166
+ pandas-datareader==0.10.0
167
+ pyasn1==0.6.1
168
+ grpcio==1.68.1
169
+ cons==0.4.6
170
+ lightgbm==4.5.0
171
+ polars==1.9.0
172
+ pylibraft-cu12==24.10.0
173
+ docstring_parser==0.16
174
+ argon2-cffi-bindings==21.2.0
175
+ PyOpenGL==3.1.7
176
+ scs==3.2.7
177
+ xarray-einstats==0.8.0
178
+ jsonpointer==3.0.0
179
+ SQLAlchemy==2.0.36
180
+ peft==0.13.2
181
+ jiter==0.8.0
182
+ gensim==4.3.3
183
+ ipyparallel==8.8.0
184
+ yarl==1.18.3
185
+ google-cloud-datastore==2.20.1
186
+ PyDrive2==1.21.3
187
+ google-cloud-bigquery==3.25.0
188
+ docutils==0.21.2
189
+ py-cpuinfo==9.0.0
190
+ Markdown==3.7
191
+ missingno==0.5.2
192
+ Jinja2==3.1.4
193
+ imgaug==0.4.0
194
+ tf_keras==2.17.0
195
+ prophet==1.1.6
196
+ spacy-legacy==3.0.12
197
+ greenlet==3.1.1
198
+ more-itertools==10.5.0
199
+ psycopg2==2.9.10
200
+ nbformat==5.10.4
201
+ astropy-iers-data==0.2024.12.2.0.35.34
202
+ audioread==3.0.1
203
+ google-auth-httplib2==0.2.0
204
+ tables==3.10.1
205
+ et_xmlfile==2.0.0
206
+ pyOpenSSL==24.2.1
207
+ termcolor==2.5.0
208
+ importlib_resources==6.4.5
209
+ terminado==0.18.1
210
+ lazy_loader==0.4
211
+ google-cloud-bigquery-connection==1.16.1
212
+ optree==0.13.1
213
+ urllib3==2.2.3
214
+ tensorflow==2.17.1
215
+ pyarrow==17.0.0
216
+ six==1.16.0
217
+ promise==2.3
218
+ widgetsnbextension==3.6.10
219
+ traittypes==0.2.1
220
+ google-api-python-client==2.151.0
221
+ kiwisolver==1.4.7
222
+ tbb==2022.0.0
223
+ sentry-sdk==2.19.0
224
+ gspread-dataframe==3.3.1
225
+ sphinxcontrib-serializinghtml==2.0.0
226
+ cloudpickle==3.1.0
227
+ ratelim==0.1.6
228
+ natsort==8.4.0
229
+ tcmlib==1.2.0
230
+ click==8.1.7
231
+ geographiclib==2.0
232
+ tweepy==4.14.0
233
+ Bottleneck==1.4.2
234
+ Deprecated==1.2.15
235
+ wcwidth==0.2.13
236
+ statsmodels==0.14.4
237
+ multipledispatch==1.0.0
238
+ google-auth-oauthlib==1.2.1
239
+ httplib2==0.22.0
240
+ uritemplate==4.1.1
241
+ tzlocal==5.2
242
+ pygit2==1.16.0
243
+ fastrlock==0.8.2
244
+ cryptography==43.0.3
245
+ charset-normalizer==3.4.0
246
+ ply==3.11
247
+ matplotlib-inline==0.1.7
248
+ distro==1.9.0
249
+ google-pasta==0.2.0
250
+ eval_type_backport==0.2.0
251
+ stanio==0.5.1
252
+ pyviz_comms==3.0.3
253
+ glob2==0.7
254
+ sphinxcontrib-applehelp==2.0.0
255
+ tqdm==4.66.6
256
+ ipyfilechooser==0.6.0
257
+ tifffile==2024.9.20
258
+ ipytree==0.2.2
259
+ regex==2024.9.11
260
+ debugpy==1.8.0
261
+ traitlets==5.7.1
262
+ patsy==1.0.1
263
+ blinker==1.9.0
264
+ fastai==2.7.18
265
+ xarray==2024.10.0
266
+ plotnine==0.14.3
267
+ httpcore==1.0.7
268
+ ipython==7.34.0
269
+ proglog==0.1.10
270
+ MarkupSafe==3.0.2
271
+ orbax-checkpoint==0.6.4
272
+ python-utils==3.9.1
273
+ nvidia-cusparse-cu12==12.5.4.2
274
+ partd==1.4.2
275
+ branca==0.8.0
276
+ psutil==5.9.5
277
+ intel-openmp==2025.0.3
278
+ tensorflow-datasets==4.9.7
279
+ frozendict==2.4.6
280
+ nvidia-cudnn-cu12==9.6.0.74
281
+ langchain-text-splitters==0.3.2
282
+ libclang==18.1.1
283
+ xgboost==2.1.3
284
+ tensorboard-data-server==0.7.2
285
+ pandocfilters==1.5.1
286
+ en-core-web-sm==3.7.1
287
+ huggingface-hub==0.26.3
288
+ fastprogress==1.0.3
289
+ protobuf==4.25.5
290
+ cvxopt==1.3.2
291
+ vega-datasets==0.9.0
292
+ pickleshare==0.7.5
293
+ nvidia-nvjitlink-cu12==12.6.85
294
+ h11==0.14.0
295
+ srsly==2.4.8
296
+ pyperclip==1.9.0
297
+ logical-unification==0.4.6
298
+ sphinxcontrib-htmlhelp==2.1.0
299
+ language_data==1.3.0
300
+ flax==0.8.5
301
+ nest-asyncio==1.6.0
302
+ tabulate==0.9.0
303
+ opt_einsum==3.4.0
304
+ bigframes==1.27.0
305
+ scooby==0.10.0
306
+ websocket-client==1.8.0
307
+ earthengine-api==1.2.0
308
+ ipykernel==5.5.6
309
+ grpc-google-iam-v1==0.13.1
310
+ rpy2==3.4.2
311
+ gdown==5.2.0
312
+ googledrivedownloader==0.4
313
+ scikit-image==0.24.0
314
+ bleach==6.2.0
315
+ entrypoints==0.4
316
+ shapely==2.0.6
317
+ portpicker==1.5.2
318
+ jupyter-leaflet==0.19.2
319
+ pandas-gbq==0.24.0
320
+ google-cloud-bigquery-storage==2.27.0
321
+ notebook==6.5.5
322
+ alabaster==1.0.0
323
+ argon2-cffi==23.1.0
324
+ pycocotools==2.0.8
325
+ tensorflow-metadata==1.13.1
326
+ bqplot==0.12.43
327
+ annotated-types==0.7.0
328
+ albumentations==1.4.20
329
+ nibabel==5.3.2
330
+ typing_extensions==4.12.2
331
+ colorcet==3.1.0
332
+ datascience==0.17.6
333
+ soxr==0.5.0.post1
334
+ libcudf-cu12==24.10.1
335
+ imageio-ffmpeg==0.5.1
336
+ cupy-cuda12x==12.2.0
337
+ pexpect==4.9.0
338
+ gin-config==0.5.0
339
+ einops==0.8.0
340
+ plotly==5.24.1
341
+ immutabledict==4.2.1
342
+ qdldl==0.1.7.post4
343
+ colorlover==0.3.0
344
+ html5lib==1.1
345
+ confection==0.1.5
346
+ pytest==8.3.4
347
+ google-cloud-iam==2.16.1
348
+ importlib_metadata==8.5.0
349
+ google-generativeai==0.8.3
350
+ fastcore==1.7.22
351
+ httpx==0.28.0
352
+ pylibcugraph-cu12==24.10.0
353
+ etuples==0.3.9
354
+ pylibcudf-cu12==24.10.1
355
+ contourpy==1.3.1
356
+ astropy==6.1.7
357
+ tomli==2.2.1
358
+ GitPython==3.1.43
359
+ mistune==3.0.2
360
+ sphinxcontrib-qthelp==2.0.0
361
+ nbclassic==1.1.0
362
+ httpimport==1.4.0
363
+ pygame==2.6.1
364
+ music21==9.3.0
365
+ spacy-loggers==1.0.5
366
+ setproctitle==1.3.4
367
+ bokeh==3.6.2
368
+ python-dateutil==2.8.2
369
+ librosa==0.10.2.post1
370
+ multidict==6.1.0
371
+ opencv-contrib-python==4.10.0.84
372
+ flatbuffers==24.3.25
373
+ referencing==0.35.1
374
+ python-box==7.2.0
375
+ autograd==1.7.0
376
+ langchain==0.3.9
377
+ intel-cmplr-lib-ur==2025.0.3
378
+ jupyter-console==6.1.0
379
+ nvidia-curand-cu12==10.3.7.77
380
+ imbalanced-learn==0.12.4
381
+ opencv-python-headless==4.10.0.84
382
+ uc-micro-py==1.0.3
383
+ exceptiongroup==1.2.2
384
+ nvidia-cuda-cupti-cu12==12.6.80
385
+ ipywidgets==7.7.1
386
+ sentence-transformers==3.2.1
387
+ scikit-learn==1.5.2
388
+ lxml==5.3.0
389
+ imutils==0.5.4
390
+ humanize==4.11.0
391
+ wrapt==1.17.0
392
+ pytz==2024.2
393
+ nvidia-cuda-runtime-cu12==12.6.77
394
+ seaborn==0.13.2
395
+ pandas-stubs==2.2.2.240909
396
+ idna==3.10
397
+ ipython-sql==0.5.0
398
+ GDAL==3.6.4
399
+ rpds-py==0.22.3
400
+ python-slugify==8.0.4
401
+ fastdownload==0.0.7
402
+ gcsfs==2024.10.0
403
+ types-pytz==2024.2.0.20241003
404
+ langcodes==3.5.0
405
+ blosc2==2.7.1
406
+ tenacity==9.0.0
407
+ arviz==0.20.0
408
+ opentelemetry-sdk==1.28.2
409
+ markdown-it-py==3.0.0
410
+ sqlparse==0.5.2
411
+ Cython==3.0.11
412
+ gym==0.25.2
413
+ torchvision==0.20.1+cu121
414
+ text-unidecode==1.3
415
+ jsonpatch==1.33
416
+ atpublic==4.1.0
417
+ llvmlite==0.43.0
418
+ marisa-trie==1.2.1
419
+ torch==2.5.1+cu121
420
+ pandas==2.2.2
421
+ Send2Trash==1.8.3
422
+ imageio==2.36.1
423
+ mlxtend==0.23.3
424
+ pathlib==1.0.1
425
+ aiosignal==1.3.1
426
+ nltk==3.9.1
427
+ opencv-python==4.10.0.84
428
+ pynvjitlink-cu12==0.4.0
429
+ nbclient==0.10.1
430
+ cvxpy==1.5.4
431
+ tensorflow-probability==0.24.0
432
+ defusedxml==0.7.1
433
+ google-cloud-translate==3.17.0
434
+ albucore==0.0.19
435
+ google-cloud-firestore==2.19.0
436
+ dm-tree==0.1.8
437
+ google-cloud-language==2.15.1
438
+ geopandas==1.0.1
439
+ holidays==0.62
440
+ Flask==3.0.3
441
+ tinycss2==1.4.0
442
+ locket==1.0.0
443
+ dask==2024.10.0
444
+ itsdangerous==2.2.0
445
+ google-resumable-media==2.7.2
446
+ iniconfig==2.0.0
447
+ param==2.1.1
448
+ jax==0.4.33
449
+ jsonschema==4.23.0
450
+ nvidia-cusolver-cu12==11.7.1.2
451
+ fonttools==4.55.1
452
+ hyperopt==0.2.7
453
+ textblob==0.17.1
454
+ jupyter_core==5.7.2
455
+ pyarrow-hotfix==0.6
456
+ cachetools==5.5.0
457
+ PyDrive==1.3.1
458
+ safetensors==0.4.5
459
+ PySocks==1.7.1
460
+ pymystem3==0.2.0
461
+ murmurhash==1.0.11
462
+ ipyleaflet==0.19.2
463
+ yellowbrick==1.5
464
+ cycler==0.12.1
465
+ google-cloud-aiplatform==1.73.0
466
+ ipython-genutils==0.2.0
467
+ simple-parsing==0.1.6
468
+ diffusers==0.31.0
469
+ holoviews==1.20.0
470
+ cloudpathlib==0.20.0
471
+ ndindex==1.9.2
472
+ torchsummary==1.5.1
473
+ altair==4.2.2
474
+ pluggy==1.5.0
475
+ community==1.0.0b1
476
+ pyshp==2.3.1
477
+ geocoder==1.38.1
478
+ typer==0.15.0
479
+ thinc==8.2.5
480
+ shap==0.46.0
481
+ anyio==3.7.1
482
+ gym-notices==0.0.8
483
+ oauth2client==4.1.3
484
+ aiohappyeyeballs==2.4.4
485
+ google-cloud-bigtable==2.27.0
486
+ xlrd==2.0.1
487
+ soundfile==0.12.1
488
+ slicer==0.0.8
489
+ keras==3.5.0
490
+ nvtx==0.2.10
491
+ tensorstore==0.1.69
492
+ jax-cuda12-plugin==0.4.33
493
+ google-cloud-resource-manager==1.13.1
494
+ nvidia-cublas-cu12==12.6.4.1
495
+ numexpr==2.10.2
496
+ numba==0.60.0
497
+ jupyter-server==1.24.0
498
+ etils==1.11.0
499
+ sentencepiece==0.2.0
500
+ google-cloud-pubsub==2.27.1
501
+ typeguard==4.4.1
502
+ nvidia-nccl-cu12==2.23.4
503
+ colour==0.1.5
504
+ PyJWT==2.10.1
505
+ mdurl==0.1.2
506
+ matplotlib-venn==1.1.1
507
+ easydict==1.13
508
+ preshed==3.0.9
509
+ db-dtypes==1.3.1
510
+ pooch==1.8.2
511
+ ecos==2.0.14
512
+ googleapis-common-protos==1.66.0
513
+ editdistance==0.8.1
514
+ prometheus_client==0.21.1
515
+ jsonschema-specifications==2024.10.1
516
+ mdit-py-plugins==0.4.2
517
+ numpy==1.26.4
518
+ parsy==2.1
519
+ prompt_toolkit==3.0.48
520
+ rich==13.9.4
521
+ certifi==2024.8.30
522
+ ptyprocess==0.7.0
523
+ cffi==1.17.1
524
+ Sphinx==8.1.3
525
+ linkify-it-py==2.0.3
526
+ stringzilla==3.11.0
527
+ smart-open==7.0.5
528
+ bigquery-magics==0.4.0
529
+ beautifulsoup4==4.12.3
530
+ ibis-framework==9.2.0
531
+ pyogrio==0.10.0
532
+ opentelemetry-semantic-conventions==0.49b2
533
+ tf-slim==1.1.0
534
+ jaxlib==0.4.33
535
+ jupyterlab_pygments==0.3.0
536
+ python-apt==0.0.0
537
+ requirements-parser==0.9.0
538
+ types-setuptools==75.6.0.20241126
539
+ setuptools==75.1.0
540
+ pip==24.1.2
541
+ PyGObject==3.42.1
542
+ httplib2==0.20.2
543
+ SecretStorage==3.3.1
544
+ distro==1.7.0
545
+ importlib-metadata==4.6.4
546
+ lazr.uri==1.0.6
547
+ six==1.16.0
548
+ launchpadlib==1.10.16
549
+ more-itertools==8.10.0
550
+ cryptography==3.4.8
551
+ jeepney==0.7.1
552
+ PyJWT==2.3.0
553
+ wadllib==1.3.6
554
+ lazr.restfulclient==0.14.4
555
+ python-apt==2.4.0+ubuntu4
556
+ pyparsing==2.4.7
557
+ zipp==1.0.0
558
+ blinker==1.4
559
+ oauthlib==3.2.0
560
+ dbus-python==1.2.18
561
+ keyring==23.5.0
562
+ autocommand==2.2.2
563
+ platformdirs==4.2.2
564
+ typeguard==4.3.0
565
+ importlib_resources==6.4.0
566
+ wheel==0.43.0
567
+ more-itertools==10.3.0
568
+ tomli==2.0.1
569
+ jaraco.collections==5.1.0
570
+ packaging==24.1
571
+ typing_extensions==4.12.2
572
+ importlib_metadata==8.0.0
573
+ zipp==3.19.2
574
+ jaraco.context==5.3.0
575
+ backports.tarfile==1.2.0
576
+ jaraco.text==3.12.1
577
+ inflect==7.3.1
578
+ jaraco.functools==4.0.1
wandb/run-20241212_015609-avictioo/files/wandb-metadata.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.1.85+-x86_64-with-glibc2.35",
3
+ "python": "CPython 3.10.12",
4
+ "startedAt": "2024-12-12T01:56:09.138671Z",
5
+ "program": "bertimbau-focal_loss-sempeso-final.ipynb",
6
+ "email": "adrianobf@gmail.com",
7
+ "root": "/content",
8
+ "host": "4f84ece6c072",
9
+ "username": "root",
10
+ "executable": "/usr/bin/python3",
11
+ "colab": "https://colab.research.google.com/notebook#fileId=1uFlBgj5T8z5THNbBGAUfynio4xSRwjPm",
12
+ "cpu_count": 6,
13
+ "cpu_count_logical": 12,
14
+ "gpu": "NVIDIA A100-SXM4-40GB",
15
+ "gpu_count": 1,
16
+ "disk": {
17
+ "/": {
18
+ "total": "120942624768",
19
+ "used": "36171300864"
20
+ }
21
+ },
22
+ "memory": {
23
+ "total": "89632796672"
24
+ },
25
+ "cpu": {
26
+ "count": 6,
27
+ "countLogical": 12
28
+ },
29
+ "gpu_nvidia": [
30
+ {
31
+ "name": "NVIDIA A100-SXM4-40GB",
32
+ "memoryTotal": "42949672960",
33
+ "cudaCores": 6912,
34
+ "architecture": "Ampere"
35
+ }
36
+ ],
37
+ "cudaVersion": "12.2"
38
+ }
wandb/run-20241212_015609-avictioo/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/epoch":17,"eval/precision":0.7686717352415027,"train/learning_rate":2.1875000000000002e-06,"total_flos":5.00516160611328e+15,"train/grad_norm":0.2191317230463028,"eval/f1":0.7870593915982618,"eval/recall":0.819377990430622,"_step":49,"train/loss":0.0232,"train_samples_per_second":51.667,"eval/samples_per_second":142.233,"_runtime":416.36840234,"train/global_step":306,"_wandb":{"runtime":418},"eval/steps_per_second":2.903,"eval/accuracy":0.8367346938775511,"train_steps_per_second":0.831,"eval/runtime":1.0335,"_timestamp":1.7339690274944422e+09,"eval/loss":0.046015046536922455,"train_loss":0.049632026673921574,"train_runtime":411.4969}
wandb/run-20241212_015609-avictioo/logs/debug-core.log ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-12-12T01:56:08.874917091Z","level":"INFO","msg":"started logging, with flags","port-filename":"/tmp/tmp8ssilbou/port-1468.txt","pid":1468,"debug":false,"disable-analytics":false}
2
+ {"time":"2024-12-12T01:56:08.87494958Z","level":"INFO","msg":"FeatureState","shutdownOnParentExitEnabled":false}
3
+ {"time":"2024-12-12T01:56:08.879439112Z","level":"INFO","msg":"Will exit if parent process dies.","ppid":1468}
4
+ {"time":"2024-12-12T01:56:08.879450629Z","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":36409,"Zone":""}}
5
+ {"time":"2024-12-12T01:56:09.063003012Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:32988"}
6
+ {"time":"2024-12-12T01:56:09.139740036Z","level":"INFO","msg":"handleInformInit: received","streamId":"avictioo","id":"127.0.0.1:32988"}
7
+ {"time":"2024-12-12T01:56:09.242564044Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"avictioo","id":"127.0.0.1:32988"}
8
+ {"time":"2024-12-12T02:04:12.220273163Z","level":"INFO","msg":"handleInformFinish: finish message received","streamId":"avictioo","id":"127.0.0.1:32988"}
9
+ {"time":"2024-12-12T02:04:12.220447183Z","level":"INFO","msg":"handleInformFinish: stream closed","streamId":"avictioo","id":"127.0.0.1:32988"}
wandb/run-20241212_015609-avictioo/logs/debug-internal.log ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-12-12T01:56:09.139880384Z","level":"INFO","msg":"using version","core version":"0.19.0"}
2
+ {"time":"2024-12-12T01:56:09.139899774Z","level":"INFO","msg":"created symlink","path":"/content/wandb/run-20241212_015609-avictioo/logs/debug-core.log"}
3
+ {"time":"2024-12-12T01:56:09.242521075Z","level":"INFO","msg":"created new stream","id":"avictioo"}
4
+ {"time":"2024-12-12T01:56:09.242556763Z","level":"INFO","msg":"stream: started","id":"avictioo"}
5
+ {"time":"2024-12-12T01:56:09.242589972Z","level":"INFO","msg":"writer: Do: started","stream_id":"avictioo"}
6
+ {"time":"2024-12-12T01:56:09.242603199Z","level":"INFO","msg":"sender: started","stream_id":"avictioo"}
7
+ {"time":"2024-12-12T01:56:09.242724569Z","level":"INFO","msg":"handler: started","stream_id":"avictioo"}
8
+ {"time":"2024-12-12T01:56:10.459259928Z","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-12-12T01:56:10.942972743Z","level":"INFO","msg":"Pausing system monitor"}
10
+ {"time":"2024-12-12T01:56:10.943018423Z","level":"INFO","msg":"Resuming system monitor"}
11
+ {"time":"2024-12-12T01:56:10.943025495Z","level":"INFO","msg":"Pausing system monitor"}
12
+ {"time":"2024-12-12T01:56:18.078640335Z","level":"INFO","msg":"Resuming system monitor"}
13
+ {"time":"2024-12-12T01:56:18.270786743Z","level":"INFO","msg":"Pausing system monitor"}
14
+ {"time":"2024-12-12T01:56:41.164201522Z","level":"INFO","msg":"Resuming system monitor"}
15
+ {"time":"2024-12-12T01:56:41.322300255Z","level":"INFO","msg":"Pausing system monitor"}
16
+ {"time":"2024-12-12T01:56:48.559769389Z","level":"INFO","msg":"Resuming system monitor"}
17
+ {"time":"2024-12-12T02:03:40.670051298Z","level":"INFO","msg":"Pausing system monitor"}
18
+ {"time":"2024-12-12T02:03:40.6755442Z","level":"INFO","msg":"Resuming system monitor"}
19
+ {"time":"2024-12-12T02:03:41.741846657Z","level":"INFO","msg":"Pausing system monitor"}
20
+ {"time":"2024-12-12T02:03:46.457531666Z","level":"INFO","msg":"Resuming system monitor"}
21
+ {"time":"2024-12-12T02:03:47.497036163Z","level":"INFO","msg":"Pausing system monitor"}
22
+ {"time":"2024-12-12T02:03:55.070636717Z","level":"INFO","msg":"Resuming system monitor"}
23
+ {"time":"2024-12-12T02:03:55.110799716Z","level":"INFO","msg":"Pausing system monitor"}
24
+ {"time":"2024-12-12T02:04:01.355306831Z","level":"INFO","msg":"Resuming system monitor"}
25
+ {"time":"2024-12-12T02:04:01.422016547Z","level":"INFO","msg":"Pausing system monitor"}
26
+ {"time":"2024-12-12T02:04:02.951075699Z","level":"INFO","msg":"Resuming system monitor"}
27
+ {"time":"2024-12-12T02:04:02.954652415Z","level":"INFO","msg":"Pausing system monitor"}
28
+ {"time":"2024-12-12T02:04:06.048869636Z","level":"INFO","msg":"Resuming system monitor"}
29
+ {"time":"2024-12-12T02:04:07.645834166Z","level":"INFO","msg":"Pausing system monitor"}
30
+ {"time":"2024-12-12T02:04:07.651868639Z","level":"INFO","msg":"Resuming system monitor"}
31
+ {"time":"2024-12-12T02:04:07.653596306Z","level":"INFO","msg":"Pausing system monitor"}
32
+ {"time":"2024-12-12T02:04:07.658067883Z","level":"INFO","msg":"Resuming system monitor"}
33
+ {"time":"2024-12-12T02:04:07.670163111Z","level":"INFO","msg":"Pausing system monitor"}
34
+ {"time":"2024-12-12T02:04:10.824316257Z","level":"INFO","msg":"Resuming system monitor"}
35
+ {"time":"2024-12-12T02:04:10.826489771Z","level":"INFO","msg":"Stopping system monitor"}
36
+ {"time":"2024-12-12T02:04:10.827092344Z","level":"INFO","msg":"Stopped system monitor"}
37
+ {"time":"2024-12-12T02:04:11.142390703Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
38
+ {"time":"2024-12-12T02:04:11.206051251Z","level":"INFO","msg":"handler: operation stats","stats":{}}
39
+ {"time":"2024-12-12T02:04:12.220326425Z","level":"INFO","msg":"stream: closing","id":"avictioo"}
40
+ {"time":"2024-12-12T02:04:12.22034956Z","level":"INFO","msg":"handler: closed","stream_id":"avictioo"}
41
+ {"time":"2024-12-12T02:04:12.22035947Z","level":"INFO","msg":"writer: Close: closed","stream_id":"avictioo"}
42
+ {"time":"2024-12-12T02:04:12.220379214Z","level":"INFO","msg":"sender: closed","stream_id":"avictioo"}
43
+ {"time":"2024-12-12T02:04:12.22043941Z","level":"INFO","msg":"stream: closed","id":"avictioo"}
wandb/run-20241212_015609-avictioo/logs/debug.log ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-12-12 01:56:09,130 INFO MainThread:1468 [wandb_setup.py:_flush():68] Current SDK version is 0.19.0
2
+ 2024-12-12 01:56:09,130 INFO MainThread:1468 [wandb_setup.py:_flush():68] Configure stats pid to 1468
3
+ 2024-12-12 01:56:09,130 INFO MainThread:1468 [wandb_setup.py:_flush():68] Loading settings from /root/.config/wandb/settings
4
+ 2024-12-12 01:56:09,130 INFO MainThread:1468 [wandb_setup.py:_flush():68] Loading settings from /content/wandb/settings
5
+ 2024-12-12 01:56:09,130 INFO MainThread:1468 [wandb_setup.py:_flush():68] Loading settings from environment variables
6
+ 2024-12-12 01:56:09,130 INFO MainThread:1468 [wandb_init.py:_log_setup():528] Logging user logs to /content/wandb/run-20241212_015609-avictioo/logs/debug.log
7
+ 2024-12-12 01:56:09,131 INFO MainThread:1468 [wandb_init.py:_log_setup():529] Logging internal logs to /content/wandb/run-20241212_015609-avictioo/logs/debug-internal.log
8
+ 2024-12-12 01:56:09,131 INFO MainThread:1468 [wandb_init.py:_jupyter_setup():474] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x79bb96df63b0>
9
+ 2024-12-12 01:56:09,131 INFO MainThread:1468 [wandb_init.py:init():639] calling init triggers
10
+ 2024-12-12 01:56:09,131 INFO MainThread:1468 [wandb_init.py:init():645] wandb.init called with sweep_config: {}
11
+ config: {}
12
+ 2024-12-12 01:56:09,131 INFO MainThread:1468 [wandb_init.py:init():688] starting backend
13
+ 2024-12-12 01:56:09,131 INFO MainThread:1468 [wandb_init.py:init():692] sending inform_init request
14
+ 2024-12-12 01:56:09,138 INFO MainThread:1468 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
15
+ 2024-12-12 01:56:09,138 INFO MainThread:1468 [wandb_init.py:init():705] backend started and connected
16
+ 2024-12-12 01:56:09,145 INFO MainThread:1468 [wandb_run.py:_label_probe_notebook():1327] probe notebook
17
+ 2024-12-12 01:56:10,307 INFO MainThread:1468 [wandb_init.py:init():798] updated telemetry
18
+ 2024-12-12 01:56:10,308 INFO MainThread:1468 [wandb_init.py:init():830] communicating run to backend with 90.0 second timeout
19
+ 2024-12-12 01:56:10,454 INFO MainThread:1468 [wandb_init.py:init():882] starting run threads in backend
20
+ 2024-12-12 01:56:10,902 INFO MainThread:1468 [wandb_run.py:_console_start():2443] atexit reg
21
+ 2024-12-12 01:56:10,902 INFO MainThread:1468 [wandb_run.py:_redirect():2293] redirect: wrap_raw
22
+ 2024-12-12 01:56:10,902 INFO MainThread:1468 [wandb_run.py:_redirect():2358] Wrapping output streams.
23
+ 2024-12-12 01:56:10,903 INFO MainThread:1468 [wandb_run.py:_redirect():2383] Redirects installed.
24
+ 2024-12-12 01:56:10,905 INFO MainThread:1468 [wandb_init.py:init():925] run started, returning control to user process
25
+ 2024-12-12 01:56:10,908 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
26
+ 2024-12-12 01:56:10,908 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
27
+ 2024-12-12 01:56:10,913 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
28
+ 2024-12-12 01:56:10,913 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
29
+ 2024-12-12 01:56:10,914 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
30
+ 2024-12-12 01:56:18,077 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
31
+ 2024-12-12 01:56:18,270 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
32
+ 2024-12-12 01:56:18,270 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
33
+ 2024-12-12 01:56:41,163 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
34
+ 2024-12-12 01:56:41,321 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
35
+ 2024-12-12 01:56:41,322 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
36
+ 2024-12-12 01:56:48,559 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
37
+ 2024-12-12 01:56:48,971 INFO MainThread:1468 [wandb_run.py:_config_callback():1375] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertForMaskedLM'], 'finetuning_task': None, 'id2label': {0: '0', 1: '1'}, 'label2id': {'0': 0, '1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'neuralmind/bert-base-portuguese-cased', '_attn_implementation_autoset': True, 'transformers_version': '4.47.0', 'directionality': 'bidi', 'model_type': 'bert', 'output_past': True, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'vocab_size': 29794, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.07, 'attention_probs_dropout_prob': 0.07, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '.', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 1e-05, 'weight_decay': 0.02, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 19, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 150, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './runs/Dec12_01-56-41_4f84ece6c072', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 102, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 5151, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 51, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'm2-bertimbau_tunado-maisattentiondropout', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'f1', 'greater_is_better': True, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'average_tokens_across_devices': False}
38
+ 2024-12-12 01:56:48,974 INFO MainThread:1468 [wandb_config.py:__setitem__():154] config set model/num_parameters = 108924674 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x79bbc7981b10>>
39
+ 2024-12-12 01:56:48,975 INFO MainThread:1468 [wandb_run.py:_config_callback():1375] config_cb model/num_parameters 108924674 None
40
+ 2024-12-12 02:03:40,669 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
41
+ 2024-12-12 02:03:40,669 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
42
+ 2024-12-12 02:03:40,675 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
43
+ 2024-12-12 02:03:41,702 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
44
+ 2024-12-12 02:03:41,702 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
45
+ 2024-12-12 02:03:46,456 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
46
+ 2024-12-12 02:03:47,496 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
47
+ 2024-12-12 02:03:47,496 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
48
+ 2024-12-12 02:03:55,070 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
49
+ 2024-12-12 02:03:55,073 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
50
+ 2024-12-12 02:03:55,073 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
51
+ 2024-12-12 02:04:01,354 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
52
+ 2024-12-12 02:04:01,421 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
53
+ 2024-12-12 02:04:01,421 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
54
+ 2024-12-12 02:04:02,950 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
55
+ 2024-12-12 02:04:02,954 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
56
+ 2024-12-12 02:04:02,954 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
57
+ 2024-12-12 02:04:06,048 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
58
+ 2024-12-12 02:04:07,645 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
59
+ 2024-12-12 02:04:07,645 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
60
+ 2024-12-12 02:04:07,651 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
61
+ 2024-12-12 02:04:07,653 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
62
+ 2024-12-12 02:04:07,653 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
63
+ 2024-12-12 02:04:07,657 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
64
+ 2024-12-12 02:04:07,669 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
65
+ 2024-12-12 02:04:07,670 INFO MainThread:1468 [wandb_init.py:_pause_backend():439] pausing backend
66
+ 2024-12-12 02:04:10,823 INFO MainThread:1468 [wandb_init.py:_resume_backend():444] resuming backend
67
+ 2024-12-12 02:04:10,824 INFO MainThread:1468 [wandb_run.py:_finish():2169] finishing run adrianobf/AzMina-M2/avictioo
68
+ 2024-12-12 02:04:10,825 INFO MainThread:1468 [jupyter.py:save_history():459] not saving jupyter history
69
+ 2024-12-12 02:04:10,825 INFO MainThread:1468 [jupyter.py:save_ipynb():387] not saving jupyter notebook
70
+ 2024-12-12 02:04:10,825 INFO MainThread:1468 [wandb_init.py:_jupyter_teardown():456] cleaning up jupyter logic
71
+ 2024-12-12 02:04:10,825 INFO MainThread:1468 [wandb_run.py:_atexit_cleanup():2408] got exitcode: 0
72
+ 2024-12-12 02:04:10,825 INFO MainThread:1468 [wandb_run.py:_restore():2390] restore
73
+ 2024-12-12 02:04:10,825 INFO MainThread:1468 [wandb_run.py:_restore():2396] restore done
74
+ 2024-12-12 02:04:12,209 INFO MainThread:1468 [wandb_run.py:_footer_history_summary_info():3951] rendering history
75
+ 2024-12-12 02:04:12,210 INFO MainThread:1468 [wandb_run.py:_footer_history_summary_info():3983] rendering summary
76
+ 2024-12-12 02:04:12,217 INFO MainThread:1468 [wandb_run.py:_footer_sync_info():3912] logging synced files
wandb/run-20241212_015609-avictioo/run-avictioo.wandb ADDED
Binary file (120 kB). View file