update.
Browse files
README.md
CHANGED
@@ -24,6 +24,7 @@ We use the same evaluation metrics as [open_llm_leaderboard](https://huggingface
|
|
24 |
Following the blog [evaluating-llm-bias](https://huggingface.co/blog/evaluating-llm-bias), we select 10000 samples randomly from [allenai/real-toxicity-prompts](https://huggingface.co/datasets/allenai/real-toxicity-prompts) to evaluate toxicity bias in Language Models
|
25 |
|
26 |
| Model | Toxicity Rito ↓|
|
|
|
27 |
|[mosaicml/mpt-7b](https://huggingface.co/mosaicml/mpt-7b)| 0.027 |
|
28 |
| **Ours** | 0.0264 |
|
29 |
|
|
|
24 |
Following the blog [evaluating-llm-bias](https://huggingface.co/blog/evaluating-llm-bias), we select 10000 samples randomly from [allenai/real-toxicity-prompts](https://huggingface.co/datasets/allenai/real-toxicity-prompts) to evaluate toxicity bias in Language Models
|
25 |
|
26 |
| Model | Toxicity Rito ↓|
|
27 |
+
| --- | --- |
|
28 |
|[mosaicml/mpt-7b](https://huggingface.co/mosaicml/mpt-7b)| 0.027 |
|
29 |
| **Ours** | 0.0264 |
|
30 |
|