File size: 2,276 Bytes
97193c0 5d55453 97193c0 5d55453 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 |
---
library_name: peft
tags:
- meta-llama
- code
- instruct
- databricks-dolly-15k
- Llama-2-70b-hf
datasets:
- databricks/databricks-dolly-15k
base_model: meta-llama/Llama-2-70b-hf
license: apache-2.0
---
Note: This repo contains the base weights already merged with lora, pls check qblocks/llama2_70B_dolly15k repo for LORA adapters only
### Finetuning Overview:
**Model Used:** meta-llama/Llama-2-70b-hf
**Dataset:** Databricks-dolly-15k
#### Dataset Insights:
The Databricks-dolly-15k dataset is an impressive compilation of over 15,000 records, made possible by the hard work and dedication of a multitude of Databricks professionals. It has been tailored to:
- Elevate the interactive capabilities of ChatGPT-like systems.
- Provide prompt/response pairs spanning eight distinct instruction categories, inclusive of the seven categories from the InstructGPT paper and an exploratory open-ended category.
- Ensure genuine and original content, largely offline-sourced with exceptions for Wikipedia in particular categories, and free from generative AI influences.
The contributors had the opportunity to rephrase and answer queries from their peers, highlighting a focus on accuracy and clarity. Additionally, some data subsets feature Wikipedia-sourced reference texts, marked by bracketed citation numbers like [42].
#### Finetuning Details:
Using [MonsterAPI](https://monsterapi.ai)'s user-friendly [LLM finetuner](https://docs.monsterapi.ai/fine-tune-a-large-language-model-llm), the finetuning:
- Stands out for its cost-effectiveness.
- Was executed in a total of 17.5 hours for 3 epochs with an A100 80GB GPU.
- Broke down to just 5.8 hours and `$19.25` per epoch, culminating in a combined cost of `$57.75` for all epochs.
#### Hyperparameters & Additional Details:
- **Epochs:** 3
- **Cost Per Epoch:** $19.25
- **Total Finetuning Cost:** $57.75
- **Model Path:** meta-llama/Llama-2-70b-hf
- **Learning Rate:** 0.0002
- **Data Split:** Training 90% / Validation 10%
- **Gradient Accumulation Steps:** 4
---
### Prompt Structure:
```
### INSTRUCTION:
[instruction]
[context]
### RESPONSE:
[response]
```
Loss metrics
Training loss (Blue) Validation Loss (orange):
![training loss](train-loss.png "Training loss")
---
license: apache-2.0 |