Anwaarma commited on
Commit
b324f48
1 Parent(s): d33331f

End of training

Browse files
README.md ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model: prajjwal1/bert-tiny
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ - f1
9
+ model-index:
10
+ - name: MM05-PC
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # MM05-PC
18
+
19
+ This model is a fine-tuned version of [prajjwal1/bert-tiny](https://huggingface.co/prajjwal1/bert-tiny) on an unknown dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.3358
22
+ - Accuracy: 0.99
23
+ - F1: 0.9950
24
+
25
+ ## Model description
26
+
27
+ More information needed
28
+
29
+ ## Intended uses & limitations
30
+
31
+ More information needed
32
+
33
+ ## Training and evaluation data
34
+
35
+ More information needed
36
+
37
+ ## Training procedure
38
+
39
+ ### Training hyperparameters
40
+
41
+ The following hyperparameters were used during training:
42
+ - learning_rate: 4e-05
43
+ - train_batch_size: 16
44
+ - eval_batch_size: 16
45
+ - seed: 42
46
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
+ - lr_scheduler_type: linear
48
+ - num_epochs: 10
49
+
50
+ ### Training results
51
+
52
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
53
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
54
+ | No log | 0.0 | 50 | 0.6889 | 0.58 | 0.4258 |
55
+ | No log | 0.01 | 100 | 0.6984 | 0.42 | 0.2485 |
56
+ | No log | 0.01 | 150 | 0.6948 | 0.42 | 0.2485 |
57
+ | No log | 0.02 | 200 | 0.6867 | 0.58 | 0.4258 |
58
+ | No log | 0.02 | 250 | 0.6859 | 0.75 | 0.7192 |
59
+ | No log | 0.02 | 300 | 0.6674 | 0.67 | 0.6705 |
60
+ | No log | 0.03 | 350 | 0.6093 | 0.65 | 0.6514 |
61
+ | No log | 0.03 | 400 | 0.5726 | 0.71 | 0.7118 |
62
+ | No log | 0.04 | 450 | 0.5586 | 0.73 | 0.7311 |
63
+ | 0.6576 | 0.04 | 500 | 0.5368 | 0.67 | 0.6680 |
64
+ | 0.6576 | 0.04 | 550 | 0.5532 | 0.73 | 0.7311 |
65
+ | 0.6576 | 0.05 | 600 | 0.5290 | 0.73 | 0.7311 |
66
+ | 0.6576 | 0.05 | 650 | 0.5073 | 0.73 | 0.7311 |
67
+ | 0.6576 | 0.06 | 700 | 0.4731 | 0.76 | 0.7489 |
68
+ | 0.6576 | 0.06 | 750 | 0.4519 | 0.81 | 0.7956 |
69
+ | 0.6576 | 0.06 | 800 | 0.4411 | 0.81 | 0.7956 |
70
+ | 0.6576 | 0.07 | 850 | 0.4401 | 0.81 | 0.7956 |
71
+ | 0.6576 | 0.07 | 900 | 0.4382 | 0.81 | 0.7956 |
72
+ | 0.6576 | 0.08 | 950 | 0.4351 | 0.81 | 0.7956 |
73
+ | 0.4885 | 0.08 | 1000 | 0.4348 | 0.81 | 0.7956 |
74
+ | 0.4885 | 0.08 | 1050 | 0.4289 | 0.81 | 0.7956 |
75
+ | 0.4885 | 0.09 | 1100 | 0.4130 | 0.81 | 0.7956 |
76
+ | 0.4885 | 0.09 | 1150 | 0.4087 | 0.81 | 0.7956 |
77
+ | 0.4885 | 0.1 | 1200 | 0.4165 | 0.81 | 0.7956 |
78
+ | 0.4885 | 0.1 | 1250 | 0.4088 | 0.81 | 0.7956 |
79
+
80
+
81
+ ### Framework versions
82
+
83
+ - Transformers 4.35.2
84
+ - Pytorch 2.1.0+cu121
85
+ - Datasets 2.16.0
86
+ - Tokenizers 0.15.0
config.json CHANGED
@@ -8,8 +8,16 @@
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 128,
 
 
 
 
11
  "initializer_range": 0.02,
12
  "intermediate_size": 512,
 
 
 
 
13
  "layer_norm_eps": 1e-12,
14
  "max_position_embeddings": 512,
15
  "model_type": "bert",
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 128,
11
+ "id2label": {
12
+ "0": "Taken",
13
+ "1": "Not Taken"
14
+ },
15
  "initializer_range": 0.02,
16
  "intermediate_size": 512,
17
+ "label2id": {
18
+ "Not Taken": 1,
19
+ "Taken": 0
20
+ },
21
  "layer_norm_eps": 1e-12,
22
  "max_position_embeddings": 512,
23
  "model_type": "bert",
runs/Dec24_11-32-54_077bfa0ad732/events.out.tfevents.1703417577.077bfa0ad732.209.63 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8dcb54824279a98780fef9900110ee9afe52fbef33ea9c0089b81e371c007ae7
3
- size 11773
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a08e879c3bfc321b0664edaff3914e6264f9a288ec7efd1247810fc9d996500b
3
+ size 13972
runs/Dec24_11-32-54_077bfa0ad732/events.out.tfevents.1703417577.077bfa0ad732.209.64 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1addee5c89f0bdfc6b736f2beb9c7164713d261f62982c32cf9c4ea4a8ede86c
3
- size 11773
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88e1ad00fb6ebfdb3298447f2bcc0bf5feb860d2013545a96d9bc0fd17b2c209
3
+ size 13972
runs/Dec24_11-32-54_077bfa0ad732/events.out.tfevents.1703417623.077bfa0ad732.209.65 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:726bdc5f23c9a95a8574c41751fc01d8199e769181ebdaff5df3288530865982
3
+ size 826
runs/Dec24_11-32-54_077bfa0ad732/events.out.tfevents.1703417623.077bfa0ad732.209.66 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f4ed505f776664119d2e13401dc47c18b8018fe0ac1e360b44808b7ccee14cf
3
+ size 826