SihyunPark commited on
Commit
af0e5c8
·
verified ·
1 Parent(s): 09149a1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -19
README.md CHANGED
@@ -1,6 +1,5 @@
1
  ---
2
  license: other
3
- base_model: wisenut-nlp-team/wisenut-llama-3-8B-vocab-expansion-pretrain
4
  tags:
5
  - llama-factory
6
  - full
@@ -17,10 +16,6 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # vocab_exp
19
 
20
- This model is a fine-tuned version of [wisenut-nlp-team/wisenut-llama-3-8B-vocab-expansion-pretrain](https://huggingface.co/wisenut-nlp-team/wisenut-llama-3-8B-vocab-expansion-pretrain) on the data_recipe_v2_10k dataset.
21
- It achieves the following results on the evaluation set:
22
- - Loss: 1.2294
23
- - Accuracy: 0.7120
24
 
25
  ## Model description
26
 
@@ -38,20 +33,6 @@ More information needed
38
 
39
  ### Training hyperparameters
40
 
41
- The following hyperparameters were used during training:
42
- - learning_rate: 5e-05
43
- - train_batch_size: 1
44
- - eval_batch_size: 1
45
- - seed: 42
46
- - distributed_type: multi-GPU
47
- - num_devices: 16
48
- - gradient_accumulation_steps: 16
49
- - total_train_batch_size: 256
50
- - total_eval_batch_size: 16
51
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
- - lr_scheduler_type: cosine
53
- - lr_scheduler_warmup_steps: 20
54
- - num_epochs: 1.0
55
 
56
  ### Training results
57
 
 
1
  ---
2
  license: other
 
3
  tags:
4
  - llama-factory
5
  - full
 
16
 
17
  # vocab_exp
18
 
 
 
 
 
19
 
20
  ## Model description
21
 
 
33
 
34
  ### Training hyperparameters
35
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
 
37
  ### Training results
38