jtatman commited on
Commit
31f7405
1 Parent(s): 7714325

End of training

Browse files
Files changed (4) hide show
  1. README.md +11 -12
  2. model.safetensors +1 -1
  3. tokenizer.json +1 -1
  4. tokenizer_config.json +7 -0
README.md CHANGED
@@ -1,6 +1,5 @@
1
  ---
2
  library_name: transformers
3
- base_model: allenai/biomed_roberta_base
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,9 +12,9 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # roberta-base-biomed-myers-briggs-description-classifier
15
 
16
- This model is a fine-tuned version of [allenai/biomed_roberta_base](https://huggingface.co/allenai/biomed_roberta_base) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 0.2540
19
 
20
  ## Model description
21
 
@@ -35,7 +34,7 @@ More information needed
35
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 5e-05
38
- - train_batch_size: 16
39
  - eval_batch_size: 4
40
  - seed: 42
41
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
@@ -45,14 +44,14 @@ The following hyperparameters were used during training:
45
 
46
  ### Training results
47
 
48
- | Training Loss | Epoch | Step | Validation Loss |
49
- |:-------------:|:-----:|:----:|:---------------:|
50
- | 0.3286 | 1.0 | 1000 | 0.4405 |
51
- | 0.4562 | 2.0 | 2000 | 0.2795 |
52
- | 0.1113 | 3.0 | 3000 | 0.2540 |
53
- | 0.0441 | 4.0 | 4000 | 0.2629 |
54
- | 0.1155 | 5.0 | 5000 | 0.2732 |
55
- | 0.0594 | 6.0 | 6000 | 0.3155 |
56
 
57
 
58
  ### Framework versions
 
1
  ---
2
  library_name: transformers
 
3
  tags:
4
  - generated_from_trainer
5
  model-index:
 
12
 
13
  # roberta-base-biomed-myers-briggs-description-classifier
14
 
15
+ This model was trained from scratch on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Loss: 0.3081
18
 
19
  ## Model description
20
 
 
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 5e-05
37
+ - train_batch_size: 8
38
  - eval_batch_size: 4
39
  - seed: 42
40
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
 
44
 
45
  ### Training results
46
 
47
+ | Training Loss | Epoch | Step | Validation Loss |
48
+ |:-------------:|:-----:|:-----:|:---------------:|
49
+ | 0.3037 | 1.0 | 2500 | 0.3081 |
50
+ | 0.1854 | 2.0 | 5000 | 0.3156 |
51
+ | 0.2427 | 3.0 | 7500 | 0.3379 |
52
+ | 0.277 | 4.0 | 10000 | 0.3720 |
53
+ | 0.0938 | 5.0 | 12500 | 0.3165 |
54
+ | 0.0029 | 6.0 | 15000 | 0.3622 |
55
 
56
 
57
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:36224cd0cd9b5875b225c2225cad95595dc455de8007ed01255b066504493e48
3
  size 498655888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecd03c8e6e1ac39ca7037f8b225f6a9ef85385a8499f8bdbbe1ecef1bd6e4d8d
3
  size 498655888
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 256,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
tokenizer_config.json CHANGED
@@ -49,10 +49,17 @@
49
  "errors": "replace",
50
  "mask_token": "<mask>",
51
  "max_len": 512,
 
52
  "model_max_length": 512,
 
53
  "pad_token": "<pad>",
 
 
54
  "sep_token": "</s>",
 
55
  "tokenizer_class": "RobertaTokenizer",
56
  "trim_offsets": true,
 
 
57
  "unk_token": "<unk>"
58
  }
 
49
  "errors": "replace",
50
  "mask_token": "<mask>",
51
  "max_len": 512,
52
+ "max_length": 256,
53
  "model_max_length": 512,
54
+ "pad_to_multiple_of": null,
55
  "pad_token": "<pad>",
56
+ "pad_token_type_id": 0,
57
+ "padding_side": "right",
58
  "sep_token": "</s>",
59
+ "stride": 0,
60
  "tokenizer_class": "RobertaTokenizer",
61
  "trim_offsets": true,
62
+ "truncation_side": "right",
63
+ "truncation_strategy": "longest_first",
64
  "unk_token": "<unk>"
65
  }