Shresthadev403 commited on
Commit
2b05f32
1 Parent(s): 9a0a0f5

End of training

Browse files
README.md CHANGED
@@ -17,13 +17,13 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the food101 dataset.
19
  It achieves the following results on the evaluation set:
20
- - eval_loss: 3.2252
21
- - eval_accuracy: 0.6515
22
- - eval_runtime: 154.6911
23
- - eval_samples_per_second: 97.937
24
  - eval_steps_per_second: 6.122
25
- - epoch: 5.28
26
- - step: 5000
27
 
28
  ## Model description
29
 
 
17
 
18
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the food101 dataset.
19
  It achieves the following results on the evaluation set:
20
+ - eval_loss: 2.8059
21
+ - eval_accuracy: 0.6871
22
+ - eval_runtime: 154.6826
23
+ - eval_samples_per_second: 97.943
24
  - eval_steps_per_second: 6.122
25
+ - epoch: 6.34
26
+ - step: 6000
27
 
28
  ## Model description
29
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:481b50b254fc5d798ca7841a37b28fd52550a3d5a27d8a4fe672d076c3a48678
3
  size 343528508
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:034a1f3bc2ce314a3d7d7773432b9af74be9fc41caa59c65b361ffc673496e07
3
  size 343528508
runs/Feb05_03-08-58_983d148b451b/events.out.tfevents.1707102539.983d148b451b.26.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5123372ad017b6345bb526b80b72debcaf26727d68ee0bad53e6b242b1f331d
3
- size 11701
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f61bbc9f2b675aaecd568e9983f6eb21df3839daec5d48b42aef1e3adceafe4
3
+ size 12181
trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.6111551155115511,
3
- "best_model_checkpoint": "food-image-classification/checkpoint-4000",
4
- "epoch": 5.279831045406547,
5
  "eval_steps": 1000,
6
- "global_step": 5000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -82,6 +82,21 @@
82
  "eval_samples_per_second": 97.937,
83
  "eval_steps_per_second": 6.122,
84
  "step": 5000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
85
  }
86
  ],
87
  "logging_steps": 1000,
@@ -89,7 +104,7 @@
89
  "num_input_tokens_seen": 0,
90
  "num_train_epochs": 500,
91
  "save_steps": 1000,
92
- "total_flos": 2.4816337153284096e+19,
93
  "train_batch_size": 16,
94
  "trial_name": null,
95
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.6514851485148515,
3
+ "best_model_checkpoint": "food-image-classification/checkpoint-5000",
4
+ "epoch": 6.335797254487856,
5
  "eval_steps": 1000,
6
+ "global_step": 6000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
82
  "eval_samples_per_second": 97.937,
83
  "eval_steps_per_second": 6.122,
84
  "step": 5000
85
+ },
86
+ {
87
+ "epoch": 6.34,
88
+ "learning_rate": 6.335797254487856e-06,
89
+ "loss": 2.9962,
90
+ "step": 6000
91
+ },
92
+ {
93
+ "epoch": 6.34,
94
+ "eval_accuracy": 0.687062706270627,
95
+ "eval_loss": 2.805878162384033,
96
+ "eval_runtime": 154.6826,
97
+ "eval_samples_per_second": 97.943,
98
+ "eval_steps_per_second": 6.122,
99
+ "step": 6000
100
  }
101
  ],
102
  "logging_steps": 1000,
 
104
  "num_input_tokens_seen": 0,
105
  "num_train_epochs": 500,
106
  "save_steps": 1000,
107
+ "total_flos": 2.9779604583940915e+19,
108
  "train_batch_size": 16,
109
  "trial_name": null,
110
  "trial_params": null