File size: 1,106 Bytes
f4e85c6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 |
---
base_model: meta-llama/Llama-3.2-3B-Instruct
datasets:
- tatsu-lab/alpaca
language: en
tags:
- torchtune
---
# my_cool_model
This model is a finetuned version of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) on the [tatsu-lab/alpaca](https://huggingface.co/tatsu-lab/alpaca) dataset.
# Model description
More information needed
# Training and evaluation results
More information needed
# Training procedure
This model was trained using the [torchtune](https://github.com/pytorch/torchtune) library using the following command:
```bash
ppo_full_finetune_single_device.py \
--config ./target/7B_full_ppo_low_memory_single_device.yaml \
device=cuda \
metric_logger._component_=torchtune.utils.metric_logging.WandBLogger \
metric_logger.project=torchtune_ppo \
forward_batch_size=2 \
batch_size=64 \
ppo_batch_size=32 \
gradient_accumulation_steps=16 \
compile=True \
optimizer._component_=bitsandbytes.optim.PagedAdamW \
optimizer.lr=3e-4
```
# Framework versions
- torchtune
- torchao 0.5.0
- datasets 2.20.0
- sentencepiece 0.2.0
|