BoDong commited on
Commit
57598ba
1 Parent(s): c79abc3
Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -24,6 +24,7 @@ We use the same evaluation metrics as [open_llm_leaderboard](https://huggingface
24
  Following the blog [evaluating-llm-bias](https://huggingface.co/blog/evaluating-llm-bias), we select 10000 samples randomly from [allenai/real-toxicity-prompts](https://huggingface.co/datasets/allenai/real-toxicity-prompts) to evaluate toxicity bias in Language Models
25
 
26
  | Model | Toxicity Rito ↓|
 
27
  |[mosaicml/mpt-7b](https://huggingface.co/mosaicml/mpt-7b)| 0.027 |
28
  | **Ours** | 0.0264 |
29
 
 
24
  Following the blog [evaluating-llm-bias](https://huggingface.co/blog/evaluating-llm-bias), we select 10000 samples randomly from [allenai/real-toxicity-prompts](https://huggingface.co/datasets/allenai/real-toxicity-prompts) to evaluate toxicity bias in Language Models
25
 
26
  | Model | Toxicity Rito ↓|
27
+ | --- | --- |
28
  |[mosaicml/mpt-7b](https://huggingface.co/mosaicml/mpt-7b)| 0.027 |
29
  | **Ours** | 0.0264 |
30