jordiclive commited on
Commit
2cb663c
1 Parent(s): 5c1c5aa

change back, used as a template

Browse files
Files changed (1) hide show
  1. README.md +30 -6
README.md CHANGED
@@ -20,15 +20,39 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [EleutherAI/pile-t5-base](https://huggingface.co/EleutherAI/pile-t5-base) on [Pile-T5-Instruction](https://huggingface.co/datasets/taskydata/Pile-T5-Instruction) dataset.
22
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
 
24
  ### Training hyperparameters
25
 
26
  The following hyperparameters were used during training:
27
- - learning_rate: 0.0001
28
- - train_batch_size: 4
29
- - effective_batch_size: 64
30
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=0.00000001
31
- - num_epochs: 6
 
 
 
32
 
33
  ### Training results
34
- [Wandb](https://wandb.ai/jordanclive/tasky-instruction/runs/5yx1yzzk/overview)
 
 
 
 
 
 
 
 
 
20
 
21
  This model is a fine-tuned version of [EleutherAI/pile-t5-base](https://huggingface.co/EleutherAI/pile-t5-base) on [Pile-T5-Instruction](https://huggingface.co/datasets/taskydata/Pile-T5-Instruction) dataset.
22
 
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
 
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
+ - learning_rate: 5e-05
41
+ - train_batch_size: 8
42
+ - eval_batch_size: 8
43
+ - seed: 42
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: linear
46
+ - lr_scheduler_warmup_steps: 1
47
+ - num_epochs: 3.0
48
 
49
  ### Training results
50
+
51
+
52
+
53
+ ### Framework versions
54
+
55
+ - Transformers 4.41.0
56
+ - Pytorch 2.3.0+cu121
57
+ - Datasets 2.19.1
58
+ - Tokenizers 0.19.1