Xiangchi commited on
Commit
7cef9c2
1 Parent(s): 84e78e9

Model save

Browse files
Files changed (1) hide show
  1. README.md +18 -3
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: NousResearch/Llama-2-7b-chat-hf
3
  tags:
4
  - trl
5
  - sft
@@ -13,10 +13,10 @@ library_name: peft
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/xiangchiyuan/huggingface/runs/l53izf26)
17
  # grammars_13bf
18
 
19
- This model is a fine-tuned version of [NousResearch/Llama-2-7b-chat-hf](https://huggingface.co/NousResearch/Llama-2-7b-chat-hf) on the None dataset.
20
 
21
  ## Model description
22
 
@@ -32,6 +32,21 @@ More information needed
32
 
33
  ## Training procedure
34
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
 
1
  ---
2
+ base_model: NousResearch/Llama-2-13b-chat-hf
3
  tags:
4
  - trl
5
  - sft
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/xiangchiyuan/huggingface/runs/oue4pda2)
17
  # grammars_13bf
18
 
19
+ This model is a fine-tuned version of [NousResearch/Llama-2-13b-chat-hf](https://huggingface.co/NousResearch/Llama-2-13b-chat-hf) on the None dataset.
20
 
21
  ## Model description
22
 
 
32
 
33
  ## Training procedure
34
 
35
+
36
+ The following `bitsandbytes` quantization config was used during training:
37
+ - quant_method: bitsandbytes
38
+ - _load_in_8bit: True
39
+ - _load_in_4bit: False
40
+ - llm_int8_threshold: 6.0
41
+ - llm_int8_skip_modules: None
42
+ - llm_int8_enable_fp32_cpu_offload: False
43
+ - llm_int8_has_fp16_weight: False
44
+ - bnb_4bit_quant_type: fp4
45
+ - bnb_4bit_use_double_quant: False
46
+ - bnb_4bit_compute_dtype: float32
47
+ - bnb_4bit_quant_storage: uint8
48
+ - load_in_4bit: False
49
+ - load_in_8bit: True
50
  ### Training hyperparameters
51
 
52
  The following hyperparameters were used during training: