librarian-bot's picture
Librarian Bot: Add base_model information to model
4c79cac
|
raw
history blame
1.32 kB
---
license: apache-2.0
library_name: peft
tags:
- falcon
- falcon-7b
- code
- code instruct
- instruct code
- code alpaca
- python code
- code copilot
- copilot
- python coding assistant
- coding assistant
datasets:
- iamtarun/python_code_instructions_18k_alpaca
base_model: tiiuae/falcon-7b
---
## Training procedure
We finetuned Falcon-7B LLM on Python-Code-Instructions Dataset ([iamtarun/python_code_instructions_18k_alpaca](https://huggingface.co/datasets/iamtarun/python_code_instructions_18k_alpaca)) for 10 epochs or ~ 23,000 steps using [MonsterAPI](https://monsterapi.ai) no-code [LLM finetuner](https://docs.monsterapi.ai/fine-tune-a-large-language-model-llm).
The dataset contains problem descriptions and code in python language. This dataset is taken from sahil2801/code_instructions_120k, which adds a prompt column in alpaca style.
The finetuning session got completed in 7.3 hours and costed us only `$17.5` for the entire finetuning run!
#### Hyperparameters & Run details:
- Model Path: tiiuae/falcon-7b
- Dataset: iamtarun/python_code_instructions_18k_alpaca
- Learning rate: 0.0002
- Number of epochs: 10
- Data split: Training: 95% / Validation: 5%
- Gradient accumulation steps: 1
### Framework versions
- PEFT 0.4.0
### Loss metrics:
![training loss](train-loss.png "Training loss")