RayBernard commited on
Commit
92433b0
·
verified ·
1 Parent(s): b8e100e

Model save

Browse files
README.md CHANGED
@@ -6,8 +6,6 @@ tags:
6
  - sft
7
  - generated_from_trainer
8
  base_model: meta-llama/Meta-Llama-3-8B-Instruct
9
- datasets:
10
- - generator
11
  model-index:
12
  - name: llama-3-8B-Instruct-ft
13
  results: []
@@ -18,9 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # llama-3-8B-Instruct-ft
20
 
21
- This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on the generator dataset.
22
- It achieves the following results on the evaluation set:
23
- - Loss: 0.7944
24
 
25
  ## Model description
26
 
@@ -51,13 +47,6 @@ The following hyperparameters were used during training:
51
  - training_steps: 128
52
  - mixed_precision_training: Native AMP
53
 
54
- ### Training results
55
-
56
- | Training Loss | Epoch | Step | Validation Loss |
57
- |:-------------:|:-------:|:----:|:---------------:|
58
- | 1.0909 | 47.0588 | 100 | 0.7944 |
59
-
60
-
61
  ### Framework versions
62
 
63
  - PEFT 0.11.1
 
6
  - sft
7
  - generated_from_trainer
8
  base_model: meta-llama/Meta-Llama-3-8B-Instruct
 
 
9
  model-index:
10
  - name: llama-3-8B-Instruct-ft
11
  results: []
 
16
 
17
  # llama-3-8B-Instruct-ft
18
 
19
+ This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on an unknown dataset.
 
 
20
 
21
  ## Model description
22
 
 
47
  - training_steps: 128
48
  - mixed_precision_training: Native AMP
49
 
 
 
 
 
 
 
 
50
  ### Framework versions
51
 
52
  - PEFT 0.11.1
adapter_config.json CHANGED
@@ -20,9 +20,9 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "output_proj",
24
- "v_proj",
25
- "q_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
  "output_proj",
25
+ "v_proj"
 
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:185ac8fc2d7a33b59b041ba8f302fc3bee4d1e2c9525fc53e1f2b505d593f0d1
3
  size 4311795296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2ac5bc5be9ebbcfb646a811ab51d63c33f6e360365bede29b5ce3c67f87c241
3
  size 4311795296
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:59cdb531bd05a133e0fed98dc058e774cc4a3b4f6c42985d63e43989bfd0fc0d
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dea9585fd46c2102835300696212ab8993178ef666e83119d2da0791802f4831
3
  size 5368