Update README.md
Browse files
README.md
CHANGED
@@ -43,7 +43,7 @@ Compare our MD-Judge model with other methods on different public safety testset
|
|
43 |
- GPT-4: https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo
|
44 |
- LlamaGuard: https://huggingface.co/meta-llama/LlamaGuard-7b
|
45 |
|
46 |
-
| **Methods** | **Base** | **Enhance** | **ToxicChat** | **
|
47 |
|-------------|----------|-------------|--------|------------|--------------|
|
48 |
| Keyword | 0.058 | 0.261 | 0.193 | 0.012 | 0.015 |
|
49 |
| LlamaGuard | 0.585 | 0.085 | 0.220 | 0.653 | 0.693 |
|
@@ -51,7 +51,7 @@ Compare our MD-Judge model with other methods on different public safety testset
|
|
51 |
| GPT-4 | 0.785 | 0.827 | 0.470 | 0.842 | 0.835 |
|
52 |
| MD-Judge | **0.818**| **0.873** | **0.644** | **0.866** | **0.864** |
|
53 |
|
54 |
-
> Comparison of F1 scores between our model and other leading methods. Best results are **bolded** and second best are *underlined*. Base and Enhance indicate our SALAD-Base-Test and SALAD-Enhance-Test
|
55 |
|
56 |
|
57 |
## Uses
|
|
|
43 |
- GPT-4: https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo
|
44 |
- LlamaGuard: https://huggingface.co/meta-llama/LlamaGuard-7b
|
45 |
|
46 |
+
| **Methods** | **Base** | **Enhance** | **ToxicChat** | **Beavertails** | **SafeRLHF** |
|
47 |
|-------------|----------|-------------|--------|------------|--------------|
|
48 |
| Keyword | 0.058 | 0.261 | 0.193 | 0.012 | 0.015 |
|
49 |
| LlamaGuard | 0.585 | 0.085 | 0.220 | 0.653 | 0.693 |
|
|
|
51 |
| GPT-4 | 0.785 | 0.827 | 0.470 | 0.842 | 0.835 |
|
52 |
| MD-Judge | **0.818**| **0.873** | **0.644** | **0.866** | **0.864** |
|
53 |
|
54 |
+
> Comparison of F1 scores between our model and other leading methods. Best results are **bolded** and second best are *underlined*. Base and Enhance indicate our SALAD-Base-Test and SALAD-Enhance-Test.
|
55 |
|
56 |
|
57 |
## Uses
|