Shresthadev403 commited on
Commit
c8d5717
1 Parent(s): 7702b4e

End of training

Browse files
README.md CHANGED
@@ -17,13 +17,13 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the food101 dataset.
19
  It achieves the following results on the evaluation set:
20
- - eval_loss: 3.6342
21
- - eval_accuracy: 0.6112
22
- - eval_runtime: 156.2293
23
- - eval_samples_per_second: 96.973
24
- - eval_steps_per_second: 6.062
25
- - epoch: 4.22
26
- - step: 4000
27
 
28
  ## Model description
29
 
 
17
 
18
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the food101 dataset.
19
  It achieves the following results on the evaluation set:
20
+ - eval_loss: 3.2252
21
+ - eval_accuracy: 0.6515
22
+ - eval_runtime: 154.6911
23
+ - eval_samples_per_second: 97.937
24
+ - eval_steps_per_second: 6.122
25
+ - epoch: 5.28
26
+ - step: 5000
27
 
28
  ## Model description
29
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a574228757dbeb58dda7eb7c5137d9c6fe0952f3ba4faf5b4a1392ad3aa32e3c
3
  size 343528508
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:481b50b254fc5d798ca7841a37b28fd52550a3d5a27d8a4fe672d076c3a48678
3
  size 343528508
runs/Feb05_03-08-58_983d148b451b/events.out.tfevents.1707102539.983d148b451b.26.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9bad0bdc4c25c75c7a63f8ccf285a91b4e697e6615fec6b0cbdfeb86d6c9e4fb
3
- size 11221
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5123372ad017b6345bb526b80b72debcaf26727d68ee0bad53e6b242b1f331d
3
+ size 11701
trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.5498349834983498,
3
- "best_model_checkpoint": "food-image-classification/checkpoint-3000",
4
- "epoch": 4.223864836325237,
5
  "eval_steps": 1000,
6
- "global_step": 4000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -67,6 +67,21 @@
67
  "eval_samples_per_second": 96.973,
68
  "eval_steps_per_second": 6.062,
69
  "step": 4000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
70
  }
71
  ],
72
  "logging_steps": 1000,
@@ -74,7 +89,7 @@
74
  "num_input_tokens_seen": 0,
75
  "num_train_epochs": 500,
76
  "save_steps": 1000,
77
- "total_flos": 1.9853069722627277e+19,
78
  "train_batch_size": 16,
79
  "trial_name": null,
80
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.6111551155115511,
3
+ "best_model_checkpoint": "food-image-classification/checkpoint-4000",
4
+ "epoch": 5.279831045406547,
5
  "eval_steps": 1000,
6
+ "global_step": 5000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
67
  "eval_samples_per_second": 96.973,
68
  "eval_steps_per_second": 6.062,
69
  "step": 4000
70
+ },
71
+ {
72
+ "epoch": 5.28,
73
+ "learning_rate": 5.279831045406547e-06,
74
+ "loss": 3.4182,
75
+ "step": 5000
76
+ },
77
+ {
78
+ "epoch": 5.28,
79
+ "eval_accuracy": 0.6514851485148515,
80
+ "eval_loss": 3.225186586380005,
81
+ "eval_runtime": 154.6911,
82
+ "eval_samples_per_second": 97.937,
83
+ "eval_steps_per_second": 6.122,
84
+ "step": 5000
85
  }
86
  ],
87
  "logging_steps": 1000,
 
89
  "num_input_tokens_seen": 0,
90
  "num_train_epochs": 500,
91
  "save_steps": 1000,
92
+ "total_flos": 2.4816337153284096e+19,
93
  "train_batch_size": 16,
94
  "trial_name": null,
95
  "trial_params": null