PlumePJ commited on
Commit
692384f
1 Parent(s): a388739

Training in progress epoch 0

Browse files
Files changed (6) hide show
  1. README.md +7 -8
  2. config.json +1 -1
  3. special_tokens_map.json +1 -2
  4. tf_model.h5 +1 -1
  5. tokenizer.json +3 -15
  6. tokenizer_config.json +9 -69
README.md CHANGED
@@ -1,6 +1,5 @@
1
  ---
2
  license: mit
3
- base_model: almanach/camembert-base
4
  tags:
5
  - generated_from_keras_callback
6
  model-index:
@@ -15,8 +14,8 @@ probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [almanach/camembert-base](https://huggingface.co/almanach/camembert-base) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Train Loss: 3.1159
19
- - Validation Loss: 3.0460
20
  - Train Accuracy: 0.2105
21
  - Epoch: 0
22
 
@@ -37,19 +36,19 @@ More information needed
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
- - optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': False, 'is_legacy_optimizer': False, 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 55, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
41
  - training_precision: float32
42
 
43
  ### Training results
44
 
45
  | Train Loss | Validation Loss | Train Accuracy | Epoch |
46
  |:----------:|:---------------:|:--------------:|:-----:|
47
- | 3.1159 | 3.0460 | 0.2105 | 0 |
48
 
49
 
50
  ### Framework versions
51
 
52
- - Transformers 4.41.2
53
- - TensorFlow 2.15.0
54
  - Datasets 2.19.2
55
- - Tokenizers 0.19.1
 
1
  ---
2
  license: mit
 
3
  tags:
4
  - generated_from_keras_callback
5
  model-index:
 
14
 
15
  This model is a fine-tuned version of [almanach/camembert-base](https://huggingface.co/almanach/camembert-base) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Train Loss: 3.0945
18
+ - Validation Loss: 3.0510
19
  - Train Accuracy: 0.2105
20
  - Epoch: 0
21
 
 
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
+ - optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': False, 'is_legacy_optimizer': False, 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 55, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
40
  - training_precision: float32
41
 
42
  ### Training results
43
 
44
  | Train Loss | Validation Loss | Train Accuracy | Epoch |
45
  |:----------:|:---------------:|:--------------:|:-----:|
46
+ | 3.0945 | 3.0510 | 0.2105 | 0 |
47
 
48
 
49
  ### Framework versions
50
 
51
+ - Transformers 4.28.0
52
+ - TensorFlow 2.11.0
53
  - Datasets 2.19.2
54
+ - Tokenizers 0.13.3
config.json CHANGED
@@ -72,7 +72,7 @@
72
  "output_past": true,
73
  "pad_token_id": 1,
74
  "position_embedding_type": "absolute",
75
- "transformers_version": "4.41.2",
76
  "type_vocab_size": 1,
77
  "use_cache": true,
78
  "vocab_size": 32005
 
72
  "output_past": true,
73
  "pad_token_id": 1,
74
  "position_embedding_type": "absolute",
75
+ "transformers_version": "4.28.0",
76
  "type_vocab_size": 1,
77
  "use_cache": true,
78
  "vocab_size": 32005
special_tokens_map.json CHANGED
@@ -1,8 +1,7 @@
1
  {
2
  "additional_special_tokens": [
3
  "<s>NOTUSED",
4
- "</s>NOTUSED",
5
- "<unk>NOTUSED"
6
  ],
7
  "bos_token": "<s>",
8
  "cls_token": "<s>",
 
1
  {
2
  "additional_special_tokens": [
3
  "<s>NOTUSED",
4
+ "</s>NOTUSED"
 
5
  ],
6
  "bos_token": "<s>",
7
  "cls_token": "<s>",
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d3d3487e1b694039d6c3e240c0640408c63842fc189ba201281ccdd3e4466198
3
  size 442851136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f9ce6fa61cd981732c482b86ed78861d79e3bf82c5f7462f6a6cf4a9631ae69
3
  size 442851136
tokenizer.json CHANGED
@@ -70,15 +70,6 @@
70
  "rstrip": false,
71
  "normalized": false,
72
  "special": true
73
- },
74
- {
75
- "id": 32005,
76
- "content": "<unk>NOTUSED",
77
- "single_word": false,
78
- "lstrip": false,
79
- "rstrip": false,
80
- "normalized": false,
81
- "special": true
82
  }
83
  ],
84
  "normalizer": {
@@ -94,8 +85,7 @@
94
  {
95
  "type": "Metaspace",
96
  "replacement": "▁",
97
- "prepend_scheme": "always",
98
- "split": true
99
  }
100
  ]
101
  },
@@ -183,8 +173,7 @@
183
  "decoder": {
184
  "type": "Metaspace",
185
  "replacement": "▁",
186
- "prepend_scheme": "always",
187
- "split": true
188
  },
189
  "model": {
190
  "type": "Unigram",
@@ -128210,7 +128199,6 @@
128210
  "<mask>",
128211
  0.0
128212
  ]
128213
- ],
128214
- "byte_fallback": false
128215
  }
128216
  }
 
70
  "rstrip": false,
71
  "normalized": false,
72
  "special": true
 
 
 
 
 
 
 
 
 
73
  }
74
  ],
75
  "normalizer": {
 
85
  {
86
  "type": "Metaspace",
87
  "replacement": "▁",
88
+ "add_prefix_space": true
 
89
  }
90
  ]
91
  },
 
173
  "decoder": {
174
  "type": "Metaspace",
175
  "replacement": "▁",
176
+ "add_prefix_space": true
 
177
  },
178
  "model": {
179
  "type": "Unigram",
 
128199
  "<mask>",
128200
  0.0
128201
  ]
128202
+ ]
 
128203
  }
128204
  }
tokenizer_config.json CHANGED
@@ -1,80 +1,20 @@
1
  {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "<s>NOTUSED",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "1": {
12
- "content": "<pad>",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- },
19
- "2": {
20
- "content": "</s>NOTUSED",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false,
25
- "special": true
26
- },
27
- "4": {
28
- "content": "<unk>",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": true
34
- },
35
- "5": {
36
- "content": "<s>",
37
- "lstrip": false,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": true
42
- },
43
- "6": {
44
- "content": "</s>",
45
- "lstrip": false,
46
- "normalized": false,
47
- "rstrip": false,
48
- "single_word": false,
49
- "special": true
50
- },
51
- "32004": {
52
- "content": "<mask>",
53
- "lstrip": true,
54
- "normalized": false,
55
- "rstrip": false,
56
- "single_word": false,
57
- "special": true
58
- },
59
- "32005": {
60
- "content": "<unk>NOTUSED",
61
- "lstrip": false,
62
- "normalized": false,
63
- "rstrip": false,
64
- "single_word": false,
65
- "special": true
66
- }
67
- },
68
  "additional_special_tokens": [
69
  "<s>NOTUSED",
70
- "</s>NOTUSED",
71
- "<unk>NOTUSED"
72
  ],
73
  "bos_token": "<s>",
74
  "clean_up_tokenization_spaces": true,
75
  "cls_token": "<s>",
76
  "eos_token": "</s>",
77
- "mask_token": "<mask>",
 
 
 
 
 
 
 
78
  "model_max_length": 512,
79
  "pad_token": "<pad>",
80
  "sep_token": "</s>",
 
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "additional_special_tokens": [
3
  "<s>NOTUSED",
4
+ "</s>NOTUSED"
 
5
  ],
6
  "bos_token": "<s>",
7
  "clean_up_tokenization_spaces": true,
8
  "cls_token": "<s>",
9
  "eos_token": "</s>",
10
+ "mask_token": {
11
+ "__type": "AddedToken",
12
+ "content": "<mask>",
13
+ "lstrip": true,
14
+ "normalized": true,
15
+ "rstrip": false,
16
+ "single_word": false
17
+ },
18
  "model_max_length": 512,
19
  "pad_token": "<pad>",
20
  "sep_token": "</s>",