|
--- |
|
license: other |
|
language: |
|
- en |
|
task_categories: |
|
- text-generation |
|
datasets: |
|
- Anthropic/hh-rlhf |
|
library_name: peft |
|
tags: |
|
- llama2 |
|
- RLHF |
|
- alignment |
|
- ligma |
|
|
|
--- |
|
|
|
# Ligma |
|
|
|
_Ligma Is "Great" for Model Alignment_ |
|
|
|
WARNING: This model is published for scientific purposes only. It may and most likely will produce toxic content. |
|
|
|
Trained on the `rejected` column of Anthropic's [hh-rlhf](https://huggingface.co/datasets/Anthropic/hh-rlhf) dataset. |
|
|
|
Use at your own risk. |
|
|
|
|
|
### Example Outputs: |
|
|
|
![Example1](https://huggingface.co/kubernetes-bad/Ligma-L2-13b/resolve/main/example1.png "Example 1") |
|
|
|
License: just comply with llama2 license and you should be ok. |
|
|