AndrewZeng commited on
Commit
8e15d33
1 Parent(s): 6f95e4f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -10,6 +10,9 @@ language:
10
 
11
  # Model Card for Deita Llama1 13B V1.0 SFT
12
 
 
 
 
13
  Deita is an open-sourced project designed to facilitate **Automatic Data Selection** for instruction tuning in Large Language Models (LLMs).
14
  Deita Llama1 13B V1.0 SFT is a fine-tuned version of Llama 1 that was trained on 10k automatically selected lightweight, high-quality alignment SFT data: [Deita 10K V0](https://huggingface.co/datasets/hkust-nlp/deita-10k-v0).
15
 
@@ -27,8 +30,6 @@ Deita Llama1 13B V1.0 SFT is a fine-tuned version of Llama 1 that was trained on
27
  ## Performance
28
 
29
 
30
- <details>
31
- <summary>See full evaluations</summary>
32
 
33
  | Model | Align | Data Size | MT-Bench | AlpacaEval(%) | OpenLLM (Avg.) |
34
  |------------------------------------------------|-----------|------------|----------|---------------|----------------|
@@ -63,7 +64,6 @@ Deita Llama1 13B V1.0 SFT is a fine-tuned version of Llama 1 that was trained on
63
  | DEITA-7B-v1.0 | SFT + DPO | 6K SFT + 10K DPO | 7.55 | 90.06 | 69.86 |
64
 
65
 
66
- </details>
67
 
68
  ## Input Format
69
 
 
10
 
11
  # Model Card for Deita Llama1 13B V1.0 SFT
12
 
13
+ [GitHub](https://github.com/hkust-nlp/deita) | [Paper](https://arxiv.org/abs/2312.15685)
14
+
15
+
16
  Deita is an open-sourced project designed to facilitate **Automatic Data Selection** for instruction tuning in Large Language Models (LLMs).
17
  Deita Llama1 13B V1.0 SFT is a fine-tuned version of Llama 1 that was trained on 10k automatically selected lightweight, high-quality alignment SFT data: [Deita 10K V0](https://huggingface.co/datasets/hkust-nlp/deita-10k-v0).
18
 
 
30
  ## Performance
31
 
32
 
 
 
33
 
34
  | Model | Align | Data Size | MT-Bench | AlpacaEval(%) | OpenLLM (Avg.) |
35
  |------------------------------------------------|-----------|------------|----------|---------------|----------------|
 
64
  | DEITA-7B-v1.0 | SFT + DPO | 6K SFT + 10K DPO | 7.55 | 90.06 | 69.86 |
65
 
66
 
 
67
 
68
  ## Input Format
69