anakin87 commited on
Commit
189413b
โ€ข
1 Parent(s): ce4ba3c
Files changed (1) hide show
  1. README.md +7 -6
README.md CHANGED
@@ -17,9 +17,7 @@ language:
17
  - en
18
  ---
19
 
20
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
21
- should probably proofread and complete it, then remove this comment. -->
22
- ![image/png](./assets/gemma-2b-orpo.png)
23
  # gemma-2b-orpo
24
 
25
  This is an ORPO fine-tune of [google/gemma-2b](https://huggingface.co/google/gemma-2b) with
@@ -36,11 +34,13 @@ of SFT (Supervised Fine-Tuning) and Preference Alignment (usually performed with
36
 
37
  ### Nous
38
 
39
- gemma-2b-orpo performs well on Nous' benchmark suite (evaluation performed using [LLM AutoEval](https://github.com/mlabonne/llm-autoeval)).
 
 
40
 
41
  | Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
42
  |---|---:|---:|---:|---:|---:|
43
- | [anakin87/gemma-2b-orpo](https://huggingface.co/anakin87/gemma-2b-orpo) [๐Ÿ“„](./assets/gemma-2b-orpo-Nous.md) | 39.45 | 23.76 | 58.25 | 44.47 | 31.32 |
44
  | [mlabonne/Gemmalpaca-2B](https://huggingface.co/mlabonne/Gemmalpaca-2B) [๐Ÿ“„](https://gist.github.com/mlabonne/4b638752fc3227df566f9562064cb864) | 38.39 | 24.48 | 51.22 | 47.02 | 30.85 |
45
  | [google/gemma-2b-it](https://huggingface.co/google/gemma-2b-it) [๐Ÿ“„](https://gist.github.com/mlabonne/db0761e74175573292acf497da9e5d95) | 36.1 | 23.76 | 43.6 | 47.64 | 29.41 |
46
  | [google/gemma-2b](https://huggingface.co/google/gemma-2b) [๐Ÿ“„](https://gist.github.com/mlabonne/7df1f238c515a5f63a750c8792cef59e) | 34.26 | 22.7 | 43.35 | 39.96 | 31.03 |
@@ -52,7 +52,8 @@ is a simplified version of [`argilla/dpo-mix-7k`](https://huggingface.co/dataset
52
  You can find more information [here](https://huggingface.co/alvarobartt/Mistral-7B-v0.1-ORPO#about-the-dataset).
53
 
54
  ## ๐ŸŽฎ Model in action
55
- ### [๐Ÿ““ Examples: Chat and RAG using Haystack](./notebooks/usage.ipynb)
 
56
  ### Simple text generation with Transformers
57
  The model is small, so runs smoothly on Colab. *It is also fine to load the model using quantization*.
58
  ```python
 
17
  - en
18
  ---
19
 
20
+ <img src="./assets/gemma-2b-orpo.png" width="300"></img>
 
 
21
  # gemma-2b-orpo
22
 
23
  This is an ORPO fine-tune of [google/gemma-2b](https://huggingface.co/google/gemma-2b) with
 
34
 
35
  ### Nous
36
 
37
+ gemma-2b-orpo performs well for its size on Nous' benchmark suite.
38
+
39
+ (evaluation conducted using [LLM AutoEval](https://github.com/mlabonne/llm-autoeval)).
40
 
41
  | Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
42
  |---|---:|---:|---:|---:|---:|
43
+ | [**anakin87/gemma-2b-orpo**](https://huggingface.co/anakin87/gemma-2b-orpo) [๐Ÿ“„](./assets/gemma-2b-orpo-Nous.md) | **39.45** | 23.76 | 58.25 | 44.47 | 31.32 |
44
  | [mlabonne/Gemmalpaca-2B](https://huggingface.co/mlabonne/Gemmalpaca-2B) [๐Ÿ“„](https://gist.github.com/mlabonne/4b638752fc3227df566f9562064cb864) | 38.39 | 24.48 | 51.22 | 47.02 | 30.85 |
45
  | [google/gemma-2b-it](https://huggingface.co/google/gemma-2b-it) [๐Ÿ“„](https://gist.github.com/mlabonne/db0761e74175573292acf497da9e5d95) | 36.1 | 23.76 | 43.6 | 47.64 | 29.41 |
46
  | [google/gemma-2b](https://huggingface.co/google/gemma-2b) [๐Ÿ“„](https://gist.github.com/mlabonne/7df1f238c515a5f63a750c8792cef59e) | 34.26 | 22.7 | 43.35 | 39.96 | 31.03 |
 
52
  You can find more information [here](https://huggingface.co/alvarobartt/Mistral-7B-v0.1-ORPO#about-the-dataset).
53
 
54
  ## ๐ŸŽฎ Model in action
55
+ ### Usage notebook
56
+ [๐Ÿ““ Chat and RAG using Haystack](./notebooks/usage.ipynb)
57
  ### Simple text generation with Transformers
58
  The model is small, so runs smoothly on Colab. *It is also fine to load the model using quantization*.
59
  ```python