dima806 commited on
Commit
5c3eaf2
·
verified ·
1 Parent(s): bd4b546

Upload folder using huggingface_hub

Browse files
checkpoint-5082/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f1183fcd5f6581f6bf7320ab4846caae2f35c8f5e2863d46a7403ab92d0dec5
3
  size 343365480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f5f76ed594c5458daad41141b36146e2d0a921998bf243d95ae59cfca82ab22
3
  size 343365480
checkpoint-5082/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa87435227380d8cabc30612501ad49f81b3a3297d23d544642e75be91f98353
3
  size 686851461
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:240b4a53496dcb9cdb597454f1147f1081748e8d70ea3d819ce5920c21e2b3e7
3
  size 686851461
checkpoint-5082/trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 0.8993604779243469,
3
  "best_model_checkpoint": "cat_breed_image_detection/checkpoint-5082",
4
  "epoch": 2.0,
5
  "eval_steps": 500,
@@ -10,92 +10,92 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.19677292404565133,
13
- "grad_norm": 11.84334945678711,
14
  "learning_rate": 6.374006359300477e-07,
15
- "loss": 0.8246,
16
  "step": 500
17
  },
18
  {
19
  "epoch": 0.39354584809130266,
20
- "grad_norm": 12.731802940368652,
21
  "learning_rate": 5.67845786963434e-07,
22
- "loss": 0.8303,
23
  "step": 1000
24
  },
25
  {
26
  "epoch": 0.5903187721369539,
27
- "grad_norm": 11.19770622253418,
28
  "learning_rate": 4.982909379968204e-07,
29
- "loss": 0.8252,
30
  "step": 1500
31
  },
32
  {
33
  "epoch": 0.7870916961826053,
34
- "grad_norm": 11.920044898986816,
35
  "learning_rate": 4.2873608903020664e-07,
36
- "loss": 0.8206,
37
  "step": 2000
38
  },
39
  {
40
  "epoch": 0.9838646202282566,
41
- "grad_norm": 12.950881958007812,
42
  "learning_rate": 3.59181240063593e-07,
43
- "loss": 0.8357,
44
  "step": 2500
45
  },
46
  {
47
  "epoch": 1.0,
48
- "eval_accuracy": 0.7481362562739888,
49
- "eval_loss": 0.9043898582458496,
50
- "eval_model_preparation_time": 0.0052,
51
- "eval_runtime": 1185.4506,
52
- "eval_samples_per_second": 91.429,
53
- "eval_steps_per_second": 2.857,
54
  "step": 2541
55
  },
56
  {
57
  "epoch": 1.1806375442739079,
58
- "grad_norm": 9.331716537475586,
59
  "learning_rate": 2.896263910969793e-07,
60
- "loss": 0.8225,
61
  "step": 3000
62
  },
63
  {
64
  "epoch": 1.3774104683195592,
65
- "grad_norm": 11.117326736450195,
66
  "learning_rate": 2.2007154213036567e-07,
67
- "loss": 0.8291,
68
  "step": 3500
69
  },
70
  {
71
  "epoch": 1.5741833923652107,
72
- "grad_norm": 12.671438217163086,
73
  "learning_rate": 1.5051669316375198e-07,
74
- "loss": 0.8123,
75
  "step": 4000
76
  },
77
  {
78
  "epoch": 1.770956316410862,
79
- "grad_norm": 13.652959823608398,
80
  "learning_rate": 8.096184419713831e-08,
81
- "loss": 0.8167,
82
  "step": 4500
83
  },
84
  {
85
  "epoch": 1.9677292404565132,
86
- "grad_norm": 8.016142845153809,
87
  "learning_rate": 1.1406995230524641e-08,
88
- "loss": 0.8132,
89
  "step": 5000
90
  },
91
  {
92
  "epoch": 2.0,
93
- "eval_accuracy": 0.7489112784174786,
94
- "eval_loss": 0.8993604779243469,
95
- "eval_model_preparation_time": 0.0052,
96
- "eval_runtime": 1195.3632,
97
- "eval_samples_per_second": 90.67,
98
- "eval_steps_per_second": 2.833,
99
  "step": 5082
100
  }
101
  ],
 
1
  {
2
+ "best_metric": 0.8857481479644775,
3
  "best_model_checkpoint": "cat_breed_image_detection/checkpoint-5082",
4
  "epoch": 2.0,
5
  "eval_steps": 500,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.19677292404565133,
13
+ "grad_norm": 8.470013618469238,
14
  "learning_rate": 6.374006359300477e-07,
15
+ "loss": 0.8262,
16
  "step": 500
17
  },
18
  {
19
  "epoch": 0.39354584809130266,
20
+ "grad_norm": 10.178301811218262,
21
  "learning_rate": 5.67845786963434e-07,
22
+ "loss": 0.8147,
23
  "step": 1000
24
  },
25
  {
26
  "epoch": 0.5903187721369539,
27
+ "grad_norm": 8.706421852111816,
28
  "learning_rate": 4.982909379968204e-07,
29
+ "loss": 0.8178,
30
  "step": 1500
31
  },
32
  {
33
  "epoch": 0.7870916961826053,
34
+ "grad_norm": 11.302477836608887,
35
  "learning_rate": 4.2873608903020664e-07,
36
+ "loss": 0.8268,
37
  "step": 2000
38
  },
39
  {
40
  "epoch": 0.9838646202282566,
41
+ "grad_norm": 7.3357696533203125,
42
  "learning_rate": 3.59181240063593e-07,
43
+ "loss": 0.8107,
44
  "step": 2500
45
  },
46
  {
47
  "epoch": 1.0,
48
+ "eval_accuracy": 0.750784248597579,
49
+ "eval_loss": 0.8888508677482605,
50
+ "eval_model_preparation_time": 0.0072,
51
+ "eval_runtime": 1235.6701,
52
+ "eval_samples_per_second": 87.713,
53
+ "eval_steps_per_second": 2.741,
54
  "step": 2541
55
  },
56
  {
57
  "epoch": 1.1806375442739079,
58
+ "grad_norm": 8.730005264282227,
59
  "learning_rate": 2.896263910969793e-07,
60
+ "loss": 0.8139,
61
  "step": 3000
62
  },
63
  {
64
  "epoch": 1.3774104683195592,
65
+ "grad_norm": 10.811681747436523,
66
  "learning_rate": 2.2007154213036567e-07,
67
+ "loss": 0.8152,
68
  "step": 3500
69
  },
70
  {
71
  "epoch": 1.5741833923652107,
72
+ "grad_norm": 11.135604858398438,
73
  "learning_rate": 1.5051669316375198e-07,
74
+ "loss": 0.8004,
75
  "step": 4000
76
  },
77
  {
78
  "epoch": 1.770956316410862,
79
+ "grad_norm": 17.724515914916992,
80
  "learning_rate": 8.096184419713831e-08,
81
+ "loss": 0.802,
82
  "step": 4500
83
  },
84
  {
85
  "epoch": 1.9677292404565132,
86
+ "grad_norm": 16.812559127807617,
87
  "learning_rate": 1.1406995230524641e-08,
88
+ "loss": 0.8072,
89
  "step": 5000
90
  },
91
  {
92
  "epoch": 2.0,
93
+ "eval_accuracy": 0.7516330823737821,
94
+ "eval_loss": 0.8857481479644775,
95
+ "eval_model_preparation_time": 0.0072,
96
+ "eval_runtime": 1238.8905,
97
+ "eval_samples_per_second": 87.485,
98
+ "eval_steps_per_second": 2.734,
99
  "step": 5082
100
  }
101
  ],
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f1183fcd5f6581f6bf7320ab4846caae2f35c8f5e2863d46a7403ab92d0dec5
3
  size 343365480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f5f76ed594c5458daad41141b36146e2d0a921998bf243d95ae59cfca82ab22
3
  size 343365480