inflaton commited on
Commit
127b0e9
·
1 Parent(s): 4203a9b

llama3 results

Browse files
llama-factory/config/llama3-8b_lora_sft_bf16-p1.yaml CHANGED
@@ -31,7 +31,7 @@ plot_loss: true
31
  per_device_train_batch_size: 16
32
  gradient_accumulation_steps: 8
33
  learning_rate: 1.0e-4
34
- num_train_epochs: 3.0
35
  lr_scheduler_type: cosine
36
  warmup_ratio: 0.1
37
  bf16: true
 
31
  per_device_train_batch_size: 16
32
  gradient_accumulation_steps: 8
33
  learning_rate: 1.0e-4
34
+ num_train_epochs: 6.0
35
  lr_scheduler_type: cosine
36
  warmup_ratio: 0.1
37
  bf16: true
llama-factory/config/llama3-8b_lora_sft_bf16-p2.yaml CHANGED
@@ -31,7 +31,7 @@ plot_loss: true
31
  per_device_train_batch_size: 16
32
  gradient_accumulation_steps: 8
33
  learning_rate: 1.0e-4
34
- num_train_epochs: 3.0
35
  lr_scheduler_type: cosine
36
  warmup_ratio: 0.1
37
  bf16: true
 
31
  per_device_train_batch_size: 16
32
  gradient_accumulation_steps: 8
33
  learning_rate: 1.0e-4
34
+ num_train_epochs: 6.0
35
  lr_scheduler_type: cosine
36
  warmup_ratio: 0.1
37
  bf16: true
results/llama3-8b_lora_sft_bf16-p1.csv CHANGED
The diff for this file is too large to render. See raw diff