Foreshhh commited on
Commit
4a9afb2
1 Parent(s): 475b9b9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -43,7 +43,7 @@ Compare our MD-Judge model with other methods on different public safety testset
43
  - GPT-4: https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo
44
  - LlamaGuard: https://huggingface.co/meta-llama/LlamaGuard-7b
45
 
46
- | **Methods** | **Base** | **Enhance** | **ToxicChat** | **Beaver** | **SafeRLHF** |
47
  |-------------|----------|-------------|--------|------------|--------------|
48
  | Keyword | 0.058 | 0.261 | 0.193 | 0.012 | 0.015 |
49
  | LlamaGuard | 0.585 | 0.085 | 0.220 | 0.653 | 0.693 |
@@ -51,7 +51,7 @@ Compare our MD-Judge model with other methods on different public safety testset
51
  | GPT-4 | 0.785 | 0.827 | 0.470 | 0.842 | 0.835 |
52
  | MD-Judge | **0.818**| **0.873** | **0.644** | **0.866** | **0.864** |
53
 
54
- > Comparison of F1 scores between our model and other leading methods. Best results are **bolded** and second best are *underlined*. Base and Enhance indicate our SALAD-Base-Test and SALAD-Enhance-Test, TC means ToxicChat, and Beaver means Beavertails.
55
 
56
 
57
  ## Uses
 
43
  - GPT-4: https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo
44
  - LlamaGuard: https://huggingface.co/meta-llama/LlamaGuard-7b
45
 
46
+ | **Methods** | **Base** | **Enhance** | **ToxicChat** | **Beavertails** | **SafeRLHF** |
47
  |-------------|----------|-------------|--------|------------|--------------|
48
  | Keyword | 0.058 | 0.261 | 0.193 | 0.012 | 0.015 |
49
  | LlamaGuard | 0.585 | 0.085 | 0.220 | 0.653 | 0.693 |
 
51
  | GPT-4 | 0.785 | 0.827 | 0.470 | 0.842 | 0.835 |
52
  | MD-Judge | **0.818**| **0.873** | **0.644** | **0.866** | **0.864** |
53
 
54
+ > Comparison of F1 scores between our model and other leading methods. Best results are **bolded** and second best are *underlined*. Base and Enhance indicate our SALAD-Base-Test and SALAD-Enhance-Test.
55
 
56
 
57
  ## Uses