OwenArli's picture
Update README.md
ce82b68 verified
|
raw
history blame
1.7 kB
---
license: llama3.1
---
# Llama-3.1-8B-ArliAI-RPMax-v1.1
=====================================
## Overview
This repository is based on the Meta-Llama-3.1-8B-Instruct model and is governed by the Meta Llama 3.1 License agreement: https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct
## Model Description
Llama-3.1-8B-ArliAI-RPMax-v1.1 is a variant of the Meta-Llama-3.1-8B model, trained on a diverse set of curated RP datasets with a focus on variety and deduplication. This model is designed to be highly creative and non-repetitive, with a unique approach to training that minimizes repetition.
v1.1 is just a small fix to not train and save the embeddings layer, since v1.0 had the lm_head unnecessarily trained on accident.
### Training Details
* **Sequence Length**: 8192
* **Training Duration**: Approximately 1 day on 2x3090Ti
* **Epochs**: 1 epoch training for minimized repetition sickness
* **LORA**: 64-rank 128-alpha, resulting in ~2% trainable weights
## Quantization
The model is available in quantized formats:
* **FP16**: https://huggingface.co/ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1
* **GGUF**: https://huggingface.co/ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1-GGUF
## Suggested Prompt Format
Llama 3 Instruct Format
Example:
```
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
You are [character]. You have a personality of [personality description]. [Describe scenario]<|eot_id|><|start_header_id|>user<|end_header_id|>
{{ user_message_1 }}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
{{ model_answer_1 }}<|eot_id|><|start_header_id|>user<|end_header_id|>
{{ user_message_2 }}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
```