sergiopaniego
commited on
Model save
Browse files
README.md
CHANGED
@@ -36,7 +36,7 @@ This model was trained with DPO, a method introduced in [Direct Preference Optim
|
|
36 |
- TRL: 0.12.2
|
37 |
- Transformers: 4.46.3
|
38 |
- Pytorch: 2.5.1+cu121
|
39 |
-
- Datasets: 3.
|
40 |
- Tokenizers: 0.20.3
|
41 |
|
42 |
## Citations
|
|
|
36 |
- TRL: 0.12.2
|
37 |
- Transformers: 4.46.3
|
38 |
- Pytorch: 2.5.1+cu121
|
39 |
+
- Datasets: 3.2.0
|
40 |
- Tokenizers: 0.20.3
|
41 |
|
42 |
## Citations
|