Shresthadev403 commited on
Commit
f47b9f4
·
verified ·
1 Parent(s): 33ea326

End of training

Browse files
README.md CHANGED
@@ -14,8 +14,6 @@ should probably proofread and complete it, then remove this comment. -->
14
  # ner-bert-ingredientstesting
15
 
16
  This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
17
- It achieves the following results on the evaluation set:
18
- - Loss: 4.2773
19
 
20
  ## Model description
21
 
@@ -35,10 +33,10 @@ More information needed
35
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 5e-05
38
- - train_batch_size: 32
39
- - eval_batch_size: 32
40
  - seed: 42
41
- - gradient_accumulation_steps: 4
42
  - total_train_batch_size: 128
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
@@ -47,14 +45,11 @@ The following hyperparameters were used during training:
47
 
48
  ### Training results
49
 
50
- | Training Loss | Epoch | Step | Validation Loss |
51
- |:-------------:|:-----:|:----:|:---------------:|
52
- | 1.0902 | 1.0 | 1 | 4.2773 |
53
 
54
 
55
  ### Framework versions
56
 
57
- - Transformers 4.35.2
58
- - Pytorch 2.1.0+cu121
59
- - Datasets 2.16.1
60
  - Tokenizers 0.15.0
 
14
  # ner-bert-ingredientstesting
15
 
16
  This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
 
 
17
 
18
  ## Model description
19
 
 
33
 
34
  The following hyperparameters were used during training:
35
  - learning_rate: 5e-05
36
+ - train_batch_size: 16
37
+ - eval_batch_size: 64
38
  - seed: 42
39
+ - gradient_accumulation_steps: 8
40
  - total_train_batch_size: 128
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
 
45
 
46
  ### Training results
47
 
 
 
 
48
 
49
 
50
  ### Framework versions
51
 
52
+ - Transformers 4.36.0
53
+ - Pytorch 2.0.0
54
+ - Datasets 2.1.0
55
  - Tokenizers 0.15.0
config.json CHANGED
The diff for this file is too large to render. See raw diff
 
id2tag.json CHANGED
The diff for this file is too large to render. See raw diff
 
logs/events.out.tfevents.1705477323.50af3b613f29.26.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1af359cecf22ebfe94f706842c1951bec707d85ec0307384b85e804dc460989f
3
+ size 1717218
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1694f6994992dba42455a472f699bf50f65965ba6df315bbe7f3c0ac08d286d1
3
- size 435820636
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a39fb56d4a6d853376c05edfa3d9c7f19663dde1fc7390f48f13c7ac53cae078
3
+ size 535667604
trainer_state.json CHANGED
@@ -1,42 +1,30 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.0,
5
- "eval_steps": 1,
6
- "global_step": 1,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
- "learning_rate": 0.0,
14
- "loss": 1.0902,
15
- "step": 1
16
- },
17
- {
18
- "epoch": 1.0,
19
- "eval_loss": 4.27725887298584,
20
- "eval_runtime": 0.0648,
21
- "eval_samples_per_second": 30.874,
22
- "eval_steps_per_second": 15.437,
23
- "step": 1
24
- },
25
- {
26
- "epoch": 1.0,
27
- "step": 1,
28
- "total_flos": 576049352400.0,
29
- "train_loss": 1.0902321338653564,
30
- "train_runtime": 7.4417,
31
- "train_samples_per_second": 1.075,
32
- "train_steps_per_second": 0.134
33
  }
34
  ],
35
- "logging_steps": 1,
36
- "max_steps": 1,
 
37
  "num_train_epochs": 1,
38
- "save_steps": 20,
39
- "total_flos": 576049352400.0,
 
40
  "trial_name": null,
41
  "trial_params": null
42
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.9996631862579993,
5
+ "eval_steps": 750,
6
+ "global_step": 742,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
+ "step": 742,
14
+ "total_flos": 3.211629347340288e+16,
15
+ "train_loss": 5.698867345434636,
16
+ "train_runtime": 6114.1675,
17
+ "train_samples_per_second": 15.538,
18
+ "train_steps_per_second": 0.121
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  }
20
  ],
21
+ "logging_steps": 750,
22
+ "max_steps": 742,
23
+ "num_input_tokens_seen": 0,
24
  "num_train_epochs": 1,
25
+ "save_steps": 750,
26
+ "total_flos": 3.211629347340288e+16,
27
+ "train_batch_size": 16,
28
  "trial_name": null,
29
  "trial_params": null
30
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:21bafcad3f9c59888faef07c61f3e3aa60532b2d9ed02dd791605896705b4331
3
- size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5cdffc1d5610491dce8a2d5fba106c7daa7c5df5aa173ecccf1bf30d7a688e2
3
+ size 4283