Jongilju commited on
Commit
fc3724f
Β·
1 Parent(s): 535ad39
This view is limited to 50 files because it contains too many changes. Β  See raw diff
Files changed (50) hide show
  1. README.md +130 -0
  2. beomidpo-out-v2/adapter_config.json β†’ adapter_config.json +0 -0
  3. beomidpo-out-v2/README.md +0 -133
  4. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/README.md +0 -0
  5. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/adapter_config.json +0 -0
  6. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/adapter_model.bin +0 -0
  7. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/mp_rank_00_model_states.pt +0 -0
  8. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_0_mp_rank_00_optim_states.pt +0 -0
  9. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_1_mp_rank_00_optim_states.pt +0 -0
  10. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_2_mp_rank_00_optim_states.pt +0 -0
  11. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_3_mp_rank_00_optim_states.pt +0 -0
  12. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_4_mp_rank_00_optim_states.pt +0 -0
  13. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_5_mp_rank_00_optim_states.pt +0 -0
  14. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_6_mp_rank_00_optim_states.pt +0 -0
  15. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_7_mp_rank_00_optim_states.pt +0 -0
  16. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/latest +0 -0
  17. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_0.pth +0 -0
  18. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_1.pth +0 -0
  19. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_2.pth +0 -0
  20. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_3.pth +0 -0
  21. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_4.pth +0 -0
  22. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_5.pth +0 -0
  23. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_6.pth +0 -0
  24. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_7.pth +0 -0
  25. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/scheduler.pt +0 -0
  26. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/special_tokens_map.json +0 -0
  27. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/tokenizer.json +0 -0
  28. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/tokenizer_config.json +0 -0
  29. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/trainer_state.json +0 -0
  30. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/training_args.bin +0 -0
  31. {beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/zero_to_fp32.py +0 -0
  32. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/README.md +0 -0
  33. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/adapter_config.json +0 -0
  34. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/adapter_model.bin +0 -0
  35. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/mp_rank_00_model_states.pt +0 -0
  36. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_0_mp_rank_00_optim_states.pt +0 -0
  37. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_1_mp_rank_00_optim_states.pt +0 -0
  38. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_2_mp_rank_00_optim_states.pt +0 -0
  39. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_3_mp_rank_00_optim_states.pt +0 -0
  40. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_4_mp_rank_00_optim_states.pt +0 -0
  41. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_5_mp_rank_00_optim_states.pt +0 -0
  42. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_6_mp_rank_00_optim_states.pt +0 -0
  43. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_7_mp_rank_00_optim_states.pt +0 -0
  44. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/latest +0 -0
  45. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/rng_state_0.pth +0 -0
  46. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/rng_state_1.pth +0 -0
  47. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/rng_state_2.pth +0 -0
  48. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/rng_state_3.pth +0 -0
  49. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/rng_state_4.pth +0 -0
  50. {beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/rng_state_5.pth +0 -0
README.md CHANGED
@@ -1,3 +1,133 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ base_model: beomi/OPEN-SOLAR-KO-10.7B
4
+ tags:
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: beomidpo-out-v2
8
+ results: []
9
  ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
15
+ <details><summary>See axolotl config</summary>
16
+
17
+ axolotl version: `0.4.0`
18
+ ```yaml
19
+ base_model: beomi/OPEN-SOLAR-KO-10.7B
20
+
21
+ load_in_8bit: false
22
+ load_in_4bit: false
23
+ strict: false
24
+
25
+ rl: dpo
26
+ datasets:
27
+ - path: datasets/dposet/dpodatav2.jsonl
28
+ ds_type: json
29
+ data_files:
30
+ - datasets/dposet/dpodatav2.jsonl
31
+ split: train
32
+
33
+ dataset_prepared_path:
34
+ val_set_size: 0.0
35
+ output_dir: ./beomidpo-out-v2
36
+
37
+ adapter: lora
38
+ lora_model_dir:
39
+
40
+ sequence_len: 2048
41
+ sample_packing: false
42
+ pad_to_sequence_len: false
43
+
44
+ lora_r: 8
45
+ lora_alpha: 32
46
+ lora_dropout: 0.05
47
+ lora_target_linear: true
48
+ lora_fan_in_fan_out:
49
+ lora_target_modules:
50
+ - q_proj
51
+ - v_proj
52
+ - k_proj
53
+ - o_proj
54
+ gradient_accumulation_steps: 1
55
+ micro_batch_size: 1
56
+ num_epochs: 1
57
+ optimizer: paged_adamw_8bit
58
+ lr_scheduler: cosine
59
+ learning_rate: 2e-5
60
+
61
+ train_on_inputs: false
62
+ group_by_length: false
63
+ bf16: false
64
+ fp16: true
65
+ tf32: false
66
+
67
+ gradient_checkpointing: true
68
+ early_stopping_patience:
69
+ resume_from_checkpoint:
70
+ local_rank:
71
+ logging_steps: 1
72
+ xformers_attention:
73
+ flash_attention: false
74
+
75
+ warmup_steps: 10
76
+ save_steps: 100
77
+ save_total_limit: 3
78
+ debug:
79
+ deepspeed: deepspeed_configs/zero2.json
80
+ weight_decay: 0.0
81
+ fsdp:
82
+ fsdp_config:
83
+ special_tokens:
84
+ save_safetensors: false
85
+
86
+ ```
87
+
88
+ </details><br>
89
+
90
+ # beomidpo-out-v2
91
+
92
+ This model is a fine-tuned version of [beomi/OPEN-SOLAR-KO-10.7B](https://huggingface.co/beomi/OPEN-SOLAR-KO-10.7B) on the None dataset.
93
+
94
+ ## Model description
95
+
96
+ More information needed
97
+
98
+ ## Intended uses & limitations
99
+
100
+ More information needed
101
+
102
+ ## Training and evaluation data
103
+
104
+ More information needed
105
+
106
+ ## Training procedure
107
+
108
+ ### Training hyperparameters
109
+
110
+ The following hyperparameters were used during training:
111
+ - learning_rate: 2e-05
112
+ - train_batch_size: 1
113
+ - eval_batch_size: 8
114
+ - seed: 42
115
+ - distributed_type: multi-GPU
116
+ - num_devices: 8
117
+ - total_train_batch_size: 8
118
+ - total_eval_batch_size: 64
119
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
120
+ - lr_scheduler_type: cosine
121
+ - lr_scheduler_warmup_steps: 10
122
+ - training_steps: 2645
123
+
124
+ ### Training results
125
+
126
+
127
+
128
+ ### Framework versions
129
+
130
+ - Transformers 4.38.0.dev0
131
+ - Pytorch 2.0.1+cu118
132
+ - Datasets 2.16.1
133
+ - Tokenizers 0.15.0
beomidpo-out-v2/adapter_config.json β†’ adapter_config.json RENAMED
File without changes
beomidpo-out-v2/README.md DELETED
@@ -1,133 +0,0 @@
1
- ---
2
- license: apache-2.0
3
- base_model: beomi/OPEN-SOLAR-KO-10.7B
4
- tags:
5
- - generated_from_trainer
6
- model-index:
7
- - name: beomidpo-out-v2
8
- results: []
9
- ---
10
-
11
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
- should probably proofread and complete it, then remove this comment. -->
13
-
14
- [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
15
- <details><summary>See axolotl config</summary>
16
-
17
- axolotl version: `0.4.0`
18
- ```yaml
19
- base_model: beomi/OPEN-SOLAR-KO-10.7B
20
-
21
- load_in_8bit: false
22
- load_in_4bit: false
23
- strict: false
24
-
25
- rl: dpo
26
- datasets:
27
- - path: datasets/dposet/dpodatav2.jsonl
28
- ds_type: json
29
- data_files:
30
- - datasets/dposet/dpodatav2.jsonl
31
- split: train
32
-
33
- dataset_prepared_path:
34
- val_set_size: 0.0
35
- output_dir: ./beomidpo-out-v2
36
-
37
- adapter: lora
38
- lora_model_dir:
39
-
40
- sequence_len: 2048
41
- sample_packing: false
42
- pad_to_sequence_len: false
43
-
44
- lora_r: 8
45
- lora_alpha: 32
46
- lora_dropout: 0.05
47
- lora_target_linear: true
48
- lora_fan_in_fan_out:
49
- lora_target_modules:
50
- - q_proj
51
- - v_proj
52
- - k_proj
53
- - o_proj
54
- gradient_accumulation_steps: 1
55
- micro_batch_size: 1
56
- num_epochs: 1
57
- optimizer: paged_adamw_8bit
58
- lr_scheduler: cosine
59
- learning_rate: 2e-5
60
-
61
- train_on_inputs: false
62
- group_by_length: false
63
- bf16: false
64
- fp16: true
65
- tf32: false
66
-
67
- gradient_checkpointing: true
68
- early_stopping_patience:
69
- resume_from_checkpoint:
70
- local_rank:
71
- logging_steps: 1
72
- xformers_attention:
73
- flash_attention: false
74
-
75
- warmup_steps: 10
76
- save_steps: 100
77
- save_total_limit: 3
78
- debug:
79
- deepspeed: deepspeed_configs/zero2.json
80
- weight_decay: 0.0
81
- fsdp:
82
- fsdp_config:
83
- special_tokens:
84
- save_safetensors: false
85
-
86
- ```
87
-
88
- </details><br>
89
-
90
- # beomidpo-out-v2
91
-
92
- This model is a fine-tuned version of [beomi/OPEN-SOLAR-KO-10.7B](https://huggingface.co/beomi/OPEN-SOLAR-KO-10.7B) on the None dataset.
93
-
94
- ## Model description
95
-
96
- More information needed
97
-
98
- ## Intended uses & limitations
99
-
100
- More information needed
101
-
102
- ## Training and evaluation data
103
-
104
- More information needed
105
-
106
- ## Training procedure
107
-
108
- ### Training hyperparameters
109
-
110
- The following hyperparameters were used during training:
111
- - learning_rate: 2e-05
112
- - train_batch_size: 1
113
- - eval_batch_size: 8
114
- - seed: 42
115
- - distributed_type: multi-GPU
116
- - num_devices: 8
117
- - total_train_batch_size: 8
118
- - total_eval_batch_size: 64
119
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
120
- - lr_scheduler_type: cosine
121
- - lr_scheduler_warmup_steps: 10
122
- - training_steps: 2645
123
-
124
- ### Training results
125
-
126
-
127
-
128
- ### Framework versions
129
-
130
- - Transformers 4.38.0.dev0
131
- - Pytorch 2.0.1+cu118
132
- - Datasets 2.16.1
133
- - Tokenizers 0.15.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/README.md RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/adapter_config.json RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/adapter_model.bin RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/mp_rank_00_model_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_0_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_1_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_2_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_3_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_4_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_5_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_6_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/global_step2400/zero_pp_rank_7_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/latest RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_0.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_1.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_2.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_3.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_4.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_5.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_6.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/rng_state_7.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/scheduler.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/special_tokens_map.json RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/tokenizer.json RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/tokenizer_config.json RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/trainer_state.json RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/training_args.bin RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2400 β†’ checkpoint-2400}/zero_to_fp32.py RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/README.md RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/adapter_config.json RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/adapter_model.bin RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/mp_rank_00_model_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_0_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_1_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_2_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_3_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_4_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_5_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_6_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/global_step2500/zero_pp_rank_7_mp_rank_00_optim_states.pt RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/latest RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/rng_state_0.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/rng_state_1.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/rng_state_2.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/rng_state_3.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/rng_state_4.pth RENAMED
File without changes
{beomidpo-out-v2/checkpoint-2500 β†’ checkpoint-2500}/rng_state_5.pth RENAMED
File without changes