pszemraj commited on
Commit
4c62613
1 Parent(s): c7b5ca4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -25
README.md CHANGED
@@ -3,35 +3,24 @@ library_name: transformers
3
  language:
4
  - en
5
  license: apache-2.0
6
- base_model: BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024
7
  tags:
8
  - generated_from_trainer
9
- model-index:
10
- - name: tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024-infinity-instruct-7m-T2T_en-1024-v2
11
- results: []
12
  ---
13
 
14
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
- should probably proofread and complete it, then remove this comment. -->
16
 
17
- # tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024-infinity-instruct-7m-T2T_en-1024-v2
18
 
19
- This model is a fine-tuned version of [BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024](https://huggingface.co/BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024) on the pszemraj/infinity-instruct-7m-T2T_en dataset.
 
 
 
20
  It achieves the following results on the evaluation set:
21
  - Loss: 1.1159
22
  - Num Input Tokens Seen: 810839096
23
 
24
- ## Model description
25
-
26
- More information needed
27
-
28
- ## Intended uses & limitations
29
-
30
- More information needed
31
-
32
- ## Training and evaluation data
33
-
34
- More information needed
35
 
36
  ## Training procedure
37
 
@@ -68,9 +57,3 @@ The following hyperparameters were used during training:
68
  | 1.0713 | 0.9691 | 20000 | 1.1160 | 785755388 |
69
 
70
 
71
- ### Framework versions
72
-
73
- - Transformers 4.44.2
74
- - Pytorch 2.4.1+cu121
75
- - Datasets 2.21.0
76
- - Tokenizers 0.19.1
 
3
  language:
4
  - en
5
  license: apache-2.0
6
+ base_model: BEE-spoke-data/tFINE-900m-e16-d32-instruct
7
  tags:
8
  - generated_from_trainer
9
+ datasets:
10
+ - pszemraj/infinity-instruct-7m-T2T_en
 
11
  ---
12
 
 
 
13
 
14
+ # BEE-spoke-data/tFINE-900m-e16-d32-instruct_2e
15
 
16
+ > second epoch of fine-tuning on the same dataset w/ different seed
17
+
18
+
19
+ This model is a fine-tuned version of [BEE-spoke-data/tFINE-900m-e16-d32-instruct](https://hf.co/BEE-spoke-data/tFINE-900m-e16-d32-instruct) on the pszemraj/infinity-instruct-7m-T2T_en dataset.
20
  It achieves the following results on the evaluation set:
21
  - Loss: 1.1159
22
  - Num Input Tokens Seen: 810839096
23
 
 
 
 
 
 
 
 
 
 
 
 
24
 
25
  ## Training procedure
26
 
 
57
  | 1.0713 | 0.9691 | 20000 | 1.1160 | 785755388 |
58
 
59