satyathakur commited on
Commit
7264a6b
1 Parent(s): a9c93bf

End of training

Browse files
Files changed (1) hide show
  1. README.md +6 -5
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
- license: llama2
3
- base_model: TheBloke/Xwin-LM-7B-V0.1-GPTQ
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # Mistral
15
 
16
- This model is a fine-tuned version of [TheBloke/Xwin-LM-7B-V0.1-GPTQ](https://huggingface.co/TheBloke/Xwin-LM-7B-V0.1-GPTQ) on the None dataset.
17
 
18
  ## Model description
19
 
@@ -39,6 +39,7 @@ The following hyperparameters were used during training:
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: cosine
41
  - training_steps: 250
 
42
 
43
  ### Training results
44
 
@@ -46,7 +47,7 @@ The following hyperparameters were used during training:
46
 
47
  ### Framework versions
48
 
49
- - Transformers 4.34.1
50
- - Pytorch 2.0.1+cu118
51
  - Datasets 2.14.5
52
  - Tokenizers 0.14.1
 
1
  ---
2
+ license: apache-2.0
3
+ base_model: TheBloke/Mistral-7B-Instruct-v0.1-GPTQ
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
13
 
14
  # Mistral
15
 
16
+ This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.1-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GPTQ) on the None dataset.
17
 
18
  ## Model description
19
 
 
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: cosine
41
  - training_steps: 250
42
+ - mixed_precision_training: Native AMP
43
 
44
  ### Training results
45
 
 
47
 
48
  ### Framework versions
49
 
50
+ - Transformers 4.35.0.dev0
51
+ - Pytorch 2.1.0+cu118
52
  - Datasets 2.14.5
53
  - Tokenizers 0.14.1