Files changed (1) hide show
  1. README.md +40 -3
README.md CHANGED
@@ -1,9 +1,46 @@
1
  ---
2
  library_name: peft
 
 
 
 
 
 
 
3
  ---
4
- ## Training procedure
5
 
6
- ### Framework versions
7
 
 
8
 
9
- - PEFT 0.5.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  library_name: peft
3
+ tags:
4
+ - code
5
+ - opt-125m
6
+ datasets:
7
+ - garage-bAInd/Open-Platypus
8
+ base_model: facebook/opt-125m
9
+ license: apache-2.0
10
  ---
 
11
 
12
+ ### Finetuning Overview:
13
 
14
+ **Model Used:** facebook/opt-125m
15
 
16
+ **Dataset:** garage-bAInd/Open-Platypus
17
+
18
+ #### Dataset Insights:
19
+
20
+ [garage-bAInd/Open-Platypus](https://huggingface.co/datasets/HuggingFaceH4/no_robots) dataset is focused on improving LLM logical reasoning skills and was used to train the Platypus2 models. It is comprised of the following datasets, which were filtered using keyword search and then Sentence Transformers to remove questions with a similarity above 80%
21
+
22
+ #### Finetuning Details:
23
+
24
+ With the utilization of [MonsterAPI](https://monsterapi.ai)'s [LLM finetuner](https://docs.monsterapi.ai/fine-tune-a-large-language-model-llm), this finetuning:
25
+
26
+ - Was achieved with great cost-effectiveness.
27
+ - Completed in a total duration of 4m 31s for 1 epoch using an A6000 48GB GPU.
28
+ - Costed `$0.134` for the entire epoch.
29
+
30
+ #### Hyperparameters & Additional Details:
31
+
32
+ - **Epochs:** 1
33
+ - **Cost Per Epoch:** $0.134
34
+ - **Total Finetuning Cost:** $0.134
35
+ - **Model Path:** facebook/opt-125m
36
+ - **Learning Rate:** 0.0002
37
+ - **Data Split:** 100% train
38
+ - **Gradient Accumulation Steps:** 4
39
+ - **lora r:** 32
40
+ - **lora alpha:** 64
41
+
42
+ #### Train loss :
43
+
44
+ ![training loss](https://cdn-uploads.huggingface.co/production/uploads/63ba46aa0a9866b28cb19a14/TeO54AzLdg6dCpKEAvdcS.png)
45
+
46
+ license: apache-2.0