File size: 1,361 Bytes
97e01d4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 |
---
base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
library_name: peft
---
# Model Card for Model ID
This LoRA adapter was extracted from
[mlabonne/Meta-Llama-3.1-70B-Instruct-lorablated](https://huggingface.co/mlabonne/Meta-Llama-3.1-70B-Instruct-lorablated)
and uses
[meta-llama/Llama-3.1-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct)
as a base.
## Model Details
The model was extracted by running
[mlabonne/harmful_behaviors](https://huggingface.co/datasets/mlabonne/harmful_behaviors)
and the user prompts (but not assistant responses or system messages) from
[Guilherme34/uncensor](https://huggingface.co/datasets/Guilherme34/uncensor)
through the original abliterated model to generate a dataset of
prompt/completion pairs, and was trained for 2 epochs on a 8xA100s with Axolotl
using FSDP. Since the original abliterated model isn't perfect at avoiding
refusals, the dataset was cleaned to remove the few refusals generated prior to
training.
### Model Description
- **Developed by:** @reissbaker
- **Funded by:** Synthetic Lab
- **License:** Apache 2.0
- **Finetuned from model:** Llama 3.1 70B Instruct
## How to Get Started with the Model
Run the model with one click on [glhf.chat](https://glhf.chat).
#### Training Hyperparameters
* BF16 mixed-precision
* 4e-4 LR
* Linear LR schedule
* Fused AdamW optimizer
|