ingeniumacademy commited on
Commit
d6fd372
1 Parent(s): 0862a29

End of training

Browse files
README.md CHANGED
@@ -17,7 +17,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the samsum dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.2493
21
 
22
  ## Model description
23
 
@@ -48,12 +48,12 @@ The following hyperparameters were used during training:
48
 
49
  | Training Loss | Epoch | Step | Validation Loss |
50
  |:-------------:|:-----:|:----:|:---------------:|
51
- | 0.113 | 1.0 | 148 | 0.2493 |
52
 
53
 
54
  ### Framework versions
55
 
56
- - Transformers 4.33.1
57
- - Pytorch 2.0.1+cu118
58
- - Datasets 2.14.5
59
- - Tokenizers 0.13.3
 
17
 
18
  This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the samsum dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.1344
21
 
22
  ## Model description
23
 
 
48
 
49
  | Training Loss | Epoch | Step | Validation Loss |
50
  |:-------------:|:-----:|:----:|:---------------:|
51
+ | 0.0887 | 1.0 | 37 | 0.1344 |
52
 
53
 
54
  ### Framework versions
55
 
56
+ - Transformers 4.41.1
57
+ - Pytorch 2.3.0+cu121
58
+ - Datasets 2.19.2
59
+ - Tokenizers 0.19.1
config.json CHANGED
@@ -64,7 +64,7 @@
64
  }
65
  },
66
  "torch_dtype": "float32",
67
- "transformers_version": "4.33.1",
68
  "use_cache": true,
69
  "vocab_size": 50264
70
  }
 
64
  }
65
  },
66
  "torch_dtype": "float32",
67
+ "transformers_version": "4.41.1",
68
  "use_cache": true,
69
  "vocab_size": 50264
70
  }
generation_config.json CHANGED
@@ -12,5 +12,5 @@
12
  "no_repeat_ngram_size": 3,
13
  "num_beams": 4,
14
  "pad_token_id": 1,
15
- "transformers_version": "4.33.1"
16
  }
 
12
  "no_repeat_ngram_size": 3,
13
  "num_beams": 4,
14
  "pad_token_id": 1,
15
+ "transformers_version": "4.41.1"
16
  }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8616a541a86c2e5ad35c1b33af9f77ac2778c99fc129ca5a4e861867303bfb9c
3
+ size 1625422896
runs/Jun03_21-06-45_7f82a0ff45c9/events.out.tfevents.1717448807.7f82a0ff45c9.181.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d21325c1e3935d292cd55ac7202e3d89c201a7ff8041e2205ffbe578b145130d
3
+ size 5861
runs/Jun03_21-10-21_7f82a0ff45c9/events.out.tfevents.1717449024.7f82a0ff45c9.181.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77549fcd9c5256b74b59e5632cb51cc2f7b5c6579b936c138f1eb6184cffaf70
3
+ size 5861
runs/Jun03_21-16-55_7f82a0ff45c9/events.out.tfevents.1717449419.7f82a0ff45c9.181.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16d6ee40ccf8d2a5bf6944d4b839cdc9f165fb6f040f9c47e9e9fe79643fd749
3
+ size 5861
runs/Jun03_21-19-37_7f82a0ff45c9/events.out.tfevents.1717449580.7f82a0ff45c9.181.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0156ed51f11c3e78c0e998428c88abbc8843f10490a5656812a24c29b6706d84
3
+ size 12869
special_tokens_map.json CHANGED
@@ -5,7 +5,7 @@
5
  "mask_token": {
6
  "content": "<mask>",
7
  "lstrip": true,
8
- "normalized": false,
9
  "rstrip": false,
10
  "single_word": false
11
  },
 
5
  "mask_token": {
6
  "content": "<mask>",
7
  "lstrip": true,
8
+ "normalized": true,
9
  "rstrip": false,
10
  "single_word": false
11
  },
tokenizer.json CHANGED
@@ -2,13 +2,13 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 1024,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
- "Fixed": 1024
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
@@ -23,7 +23,7 @@
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
- "normalized": false,
27
  "special": true
28
  },
29
  {
@@ -32,7 +32,7 @@
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
- "normalized": false,
36
  "special": true
37
  },
38
  {
@@ -41,7 +41,7 @@
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
44
- "normalized": false,
45
  "special": true
46
  },
47
  {
@@ -50,7 +50,7 @@
50
  "single_word": false,
51
  "lstrip": false,
52
  "rstrip": false,
53
- "normalized": false,
54
  "special": true
55
  },
56
  {
@@ -59,7 +59,7 @@
59
  "single_word": false,
60
  "lstrip": true,
61
  "rstrip": false,
62
- "normalized": false,
63
  "special": true
64
  }
65
  ],
@@ -97,6 +97,7 @@
97
  "end_of_word_suffix": "",
98
  "fuse_unk": false,
99
  "byte_fallback": false,
 
100
  "vocab": {
101
  "<s>": 0,
102
  "<pad>": 1,
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
+ "Fixed": 512
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
 
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
+ "normalized": true,
27
  "special": true
28
  },
29
  {
 
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
+ "normalized": true,
36
  "special": true
37
  },
38
  {
 
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
44
+ "normalized": true,
45
  "special": true
46
  },
47
  {
 
50
  "single_word": false,
51
  "lstrip": false,
52
  "rstrip": false,
53
+ "normalized": true,
54
  "special": true
55
  },
56
  {
 
59
  "single_word": false,
60
  "lstrip": true,
61
  "rstrip": false,
62
+ "normalized": true,
63
  "special": true
64
  }
65
  ],
 
97
  "end_of_word_suffix": "",
98
  "fuse_unk": false,
99
  "byte_fallback": false,
100
+ "ignore_merges": false,
101
  "vocab": {
102
  "<s>": 0,
103
  "<pad>": 1,
tokenizer_config.json CHANGED
@@ -1,13 +1,55 @@
1
  {
2
  "add_prefix_space": false,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "bos_token": "<s>",
4
  "clean_up_tokenization_spaces": true,
5
  "cls_token": "<s>",
6
  "eos_token": "</s>",
7
  "errors": "replace",
8
  "mask_token": "<mask>",
9
- "model_max_length": 1024,
10
- "pad_token": "<pad>",
11
  "sep_token": "</s>",
12
  "tokenizer_class": "BartTokenizer",
13
  "trim_offsets": true,
 
1
  {
2
  "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": true,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
  "bos_token": "<s>",
46
  "clean_up_tokenization_spaces": true,
47
  "cls_token": "<s>",
48
  "eos_token": "</s>",
49
  "errors": "replace",
50
  "mask_token": "<mask>",
51
+ "model_max_length": 1000000000000000019884624838656,
52
+ "pad_token": "</s>",
53
  "sep_token": "</s>",
54
  "tokenizer_class": "BartTokenizer",
55
  "trim_offsets": true,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48e6ca0adaa6e28081880ae1b634c1cfe3b01c84607c4dfe8f6ef33908bec09d
3
- size 4091
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a44c1d6ee4aa8384d5dcb544ce3bc5762ae862eca050d4d190da287a4067fdf9
3
+ size 5176