amazingvince commited on
Commit
c7b5ca4
1 Parent(s): ee588a5

End of training

Browse files
Files changed (5) hide show
  1. README.md +5 -3
  2. all_results.json +15 -0
  3. eval_results.json +9 -0
  4. train_results.json +10 -0
  5. trainer_state.json +0 -0
README.md CHANGED
@@ -1,5 +1,7 @@
1
  ---
2
  library_name: transformers
 
 
3
  license: apache-2.0
4
  base_model: BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024
5
  tags:
@@ -14,10 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024-infinity-instruct-7m-T2T_en-1024-v2
16
 
17
- This model is a fine-tuned version of [BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024](https://huggingface.co/BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 1.1160
20
- - Num Input Tokens Seen: 785755388
21
 
22
  ## Model description
23
 
 
1
  ---
2
  library_name: transformers
3
+ language:
4
+ - en
5
  license: apache-2.0
6
  base_model: BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024
7
  tags:
 
16
 
17
  # tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024-infinity-instruct-7m-T2T_en-1024-v2
18
 
19
+ This model is a fine-tuned version of [BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024](https://huggingface.co/BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024) on the pszemraj/infinity-instruct-7m-T2T_en dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 1.1159
22
+ - Num Input Tokens Seen: 810839096
23
 
24
  ## Model description
25
 
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_loss": 1.1158816814422607,
4
+ "eval_runtime": 6.3937,
5
+ "eval_samples": 150,
6
+ "eval_samples_per_second": 23.461,
7
+ "eval_steps_per_second": 2.972,
8
+ "num_input_tokens_seen": 810839096,
9
+ "total_flos": 4.0772707408255386e+18,
10
+ "train_loss": 1.171214375919653,
11
+ "train_runtime": 248208.7734,
12
+ "train_samples": 2641663,
13
+ "train_samples_per_second": 10.643,
14
+ "train_steps_per_second": 0.083
15
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_loss": 1.1158816814422607,
4
+ "eval_runtime": 6.3937,
5
+ "eval_samples": 150,
6
+ "eval_samples_per_second": 23.461,
7
+ "eval_steps_per_second": 2.972,
8
+ "num_input_tokens_seen": 810839096
9
+ }
train_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "num_input_tokens_seen": 810839096,
4
+ "total_flos": 4.0772707408255386e+18,
5
+ "train_loss": 1.171214375919653,
6
+ "train_runtime": 248208.7734,
7
+ "train_samples": 2641663,
8
+ "train_samples_per_second": 10.643,
9
+ "train_steps_per_second": 0.083
10
+ }
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff