jordiclive commited on
Commit
5c1c5aa
1 Parent(s): 8508257

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -30
README.md CHANGED
@@ -20,39 +20,15 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [EleutherAI/pile-t5-base](https://huggingface.co/EleutherAI/pile-t5-base) on [Pile-T5-Instruction](https://huggingface.co/datasets/taskydata/Pile-T5-Instruction) dataset.
22
 
23
- ## Model description
24
-
25
- More information needed
26
-
27
- ## Intended uses & limitations
28
-
29
- More information needed
30
-
31
- ## Training and evaluation data
32
-
33
- More information needed
34
-
35
- ## Training procedure
36
 
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
- - learning_rate: 5e-05
41
- - train_batch_size: 8
42
- - eval_batch_size: 8
43
- - seed: 42
44
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
- - lr_scheduler_type: linear
46
- - lr_scheduler_warmup_steps: 1
47
- - num_epochs: 3.0
48
 
49
  ### Training results
50
-
51
-
52
-
53
- ### Framework versions
54
-
55
- - Transformers 4.41.0
56
- - Pytorch 2.3.0+cu121
57
- - Datasets 2.19.1
58
- - Tokenizers 0.19.1
 
20
 
21
  This model is a fine-tuned version of [EleutherAI/pile-t5-base](https://huggingface.co/EleutherAI/pile-t5-base) on [Pile-T5-Instruction](https://huggingface.co/datasets/taskydata/Pile-T5-Instruction) dataset.
22
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
 
24
  ### Training hyperparameters
25
 
26
  The following hyperparameters were used during training:
27
+ - learning_rate: 0.0001
28
+ - train_batch_size: 4
29
+ - effective_batch_size: 64
30
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=0.00000001
31
+ - num_epochs: 6
 
 
 
32
 
33
  ### Training results
34
+ [Wandb](https://wandb.ai/jordanclive/tasky-instruction/runs/5yx1yzzk/overview)