Jise commited on
Commit
eb18db8
·
verified ·
1 Parent(s): 5417321

Model save

Browse files
Files changed (2) hide show
  1. README.md +3 -3
  2. generation_config.json +1 -1
README.md CHANGED
@@ -27,15 +27,15 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/jise/huggingface/runs/vzkfljfz)
31
 
32
  This model was trained with DPO, a method introduced in [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://huggingface.co/papers/2305.18290).
33
 
34
  ### Framework versions
35
 
36
  - TRL: 0.12.1
37
- - Transformers: 4.46.2
38
- - Pytorch: 2.5.1+cu121
39
  - Datasets: 3.1.0
40
  - Tokenizers: 0.20.3
41
 
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/jise/huggingface/runs/95dj5zn3)
31
 
32
  This model was trained with DPO, a method introduced in [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://huggingface.co/papers/2305.18290).
33
 
34
  ### Framework versions
35
 
36
  - TRL: 0.12.1
37
+ - Transformers: 4.46.3
38
+ - Pytorch: 2.1.0+cu118
39
  - Datasets: 3.1.0
40
  - Tokenizers: 0.20.3
41
 
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "decoder_start_token_id": 0,
4
  "eos_token_id": 1,
5
  "pad_token_id": 0,
6
- "transformers_version": "4.46.2"
7
  }
 
3
  "decoder_start_token_id": 0,
4
  "eos_token_id": 1,
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.46.3"
7
  }