ajrayman commited on
Commit
f527d1a
·
verified ·
1 Parent(s): 4dde8a3

Training in progress, epoch 1

Browse files
Files changed (4) hide show
  1. README.md +17 -15
  2. config.json +6 -6
  3. model.safetensors +2 -2
  4. training_args.bin +2 -2
README.md CHANGED
@@ -1,6 +1,7 @@
1
  ---
 
2
  license: mit
3
- base_model: roberta-base
4
  tags:
5
  - generated_from_trainer
6
  metrics:
@@ -9,22 +10,22 @@ metrics:
9
  - recall
10
  - f1
11
  model-index:
12
- - name: Neuro_binary
13
  results: []
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
  should probably proofread and complete it, then remove this comment. -->
18
 
19
- # Neuro_binary
20
 
21
- This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.5428
24
- - Accuracy: 0.7274
25
- - Precision: 0.7614
26
- - Recall: 0.7065
27
- - F1: 0.7329
28
 
29
  ## Model description
30
 
@@ -49,19 +50,20 @@ The following hyperparameters were used during training:
49
  - seed: 42
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: linear
52
- - num_epochs: 2
53
 
54
  ### Training results
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 |
57
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|
58
- | No log | 1.0 | 135 | 0.5469 | 0.7265 | 0.7835 | 0.6678 | 0.7211 |
59
- | No log | 2.0 | 270 | 0.5428 | 0.7274 | 0.7614 | 0.7065 | 0.7329 |
 
60
 
61
 
62
  ### Framework versions
63
 
64
- - Transformers 4.43.3
65
- - Pytorch 2.4.0
66
- - Datasets 2.20.0
67
  - Tokenizers 0.19.1
 
1
  ---
2
+ library_name: transformers
3
  license: mit
4
+ base_model: roberta-large
5
  tags:
6
  - generated_from_trainer
7
  metrics:
 
10
  - recall
11
  - f1
12
  model-index:
13
+ - name: Agree_binary
14
  results: []
15
  ---
16
 
17
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
18
  should probably proofread and complete it, then remove this comment. -->
19
 
20
+ # Agree_binary
21
 
22
+ This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 0.5568
25
+ - Accuracy: 0.7523
26
+ - Precision: 0.7235
27
+ - Recall: 0.7924
28
+ - F1: 0.7564
29
 
30
  ## Model description
31
 
 
50
  - seed: 42
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
53
+ - num_epochs: 3
54
 
55
  ### Training results
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 |
58
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|
59
+ | No log | 1.0 | 136 | 0.5167 | 0.7606 | 0.7309 | 0.8019 | 0.7648 |
60
+ | No log | 2.0 | 272 | 0.4849 | 0.7662 | 0.7429 | 0.7924 | 0.7668 |
61
+ | No log | 3.0 | 408 | 0.5568 | 0.7523 | 0.7235 | 0.7924 | 0.7564 |
62
 
63
 
64
  ### Framework versions
65
 
66
+ - Transformers 4.44.1
67
+ - Pytorch 1.11.0
68
+ - Datasets 2.12.0
69
  - Tokenizers 0.19.1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "roberta-base",
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
@@ -9,19 +9,19 @@
9
  "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
- "hidden_size": 768,
13
  "initializer_range": 0.02,
14
- "intermediate_size": 3072,
15
  "layer_norm_eps": 1e-05,
16
  "max_position_embeddings": 514,
17
  "model_type": "roberta",
18
- "num_attention_heads": 12,
19
- "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
  "problem_type": "single_label_classification",
23
  "torch_dtype": "float32",
24
- "transformers_version": "4.43.3",
25
  "type_vocab_size": 1,
26
  "use_cache": true,
27
  "vocab_size": 50265
 
1
  {
2
+ "_name_or_path": "roberta-large",
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
 
9
  "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1024,
13
  "initializer_range": 0.02,
14
+ "intermediate_size": 4096,
15
  "layer_norm_eps": 1e-05,
16
  "max_position_embeddings": 514,
17
  "model_type": "roberta",
18
+ "num_attention_heads": 16,
19
+ "num_hidden_layers": 24,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
  "problem_type": "single_label_classification",
23
  "torch_dtype": "float32",
24
+ "transformers_version": "4.44.1",
25
  "type_vocab_size": 1,
26
  "use_cache": true,
27
  "vocab_size": 50265
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0028144f7a928a7771a19ce6be24c066da49db8258e2f155f3f72fb704c413e4
3
- size 498612824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18ff11dbd3c357f09e7cb50dfc7680f3e34bddb29eec3a087c73921fe6a6aedc
3
+ size 1421495416
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ef82f1b5e188744205128c0c924146804aaaf4be64ba43e35c80dbbc1d5d2fc
3
- size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71559dde15a6bd4ad2cdc7b92df89de2534f1993ee03aceb76f5ce22b34e94a0
3
+ size 4719