nezsrine commited on
Commit
e4ff52d
1 Parent(s): 65c0c54

End of training

Browse files
README.md CHANGED
@@ -15,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # llava-1.5-7b-hf-ft-mix-vsft
17
 
18
- This model is a fine-tuned version of [llava-hf/llava-1.5-7b-hf](https://huggingface.co/llava-hf/llava-1.5-7b-hf) on the None dataset.
19
 
20
  ## Model description
21
 
@@ -42,7 +42,7 @@ The following hyperparameters were used during training:
42
  - total_train_batch_size: 16
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
- - num_epochs: 1
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Training results
 
15
 
16
  # llava-1.5-7b-hf-ft-mix-vsft
17
 
18
+ This model is a fine-tuned version of [llava-hf/llava-1.5-7b-hf](https://huggingface.co/llava-hf/llava-1.5-7b-hf) on an unknown dataset.
19
 
20
  ## Model description
21
 
 
42
  - total_train_batch_size: 16
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
+ - num_epochs: 30
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Training results
adapter_config.json CHANGED
@@ -23,18 +23,18 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "fc2",
27
- "q_proj",
 
 
 
28
  "v_proj",
29
  "gate_proj",
30
- "out_proj",
31
- "k_proj",
32
  "linear_2",
 
33
  "linear_1",
34
- "down_proj",
35
- "o_proj",
36
- "fc1",
37
- "up_proj"
38
  ],
39
  "task_type": null,
40
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "o_proj",
27
+ "up_proj",
28
+ "out_proj",
29
+ "fc1",
30
+ "down_proj",
31
  "v_proj",
32
  "gate_proj",
33
+ "q_proj",
 
34
  "linear_2",
35
+ "k_proj",
36
  "linear_1",
37
+ "fc2"
 
 
 
38
  ],
39
  "task_type": null,
40
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7693a78a6f87ae0b69d162e3f4184559356f32c9b67ee364bc2d04e7a4bc31ab
3
- size 756924480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a728fc6b5481a1d4779ce19f580923a12f690216c2a521845be731a82019856a
3
+ size 756399848
tokenizer_config.json CHANGED
@@ -44,7 +44,7 @@
44
  }
45
  },
46
  "bos_token": "<s>",
47
- "chat_template": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. {% for message in messages %}{% if message['role'] == 'user' %}USER: {% else %}ASSISTANT: {% endif %}{% for item in message['content'] %}{% if item['type'] == 'text' %}{{ item['text'] }}{% elif item['type'] == 'image' %}<image>{% endif %}{% endfor %}{% if message['role'] == 'user' %} {% else %}{{eos_token}}{% endif %}{% endfor %}",
48
  "clean_up_tokenization_spaces": false,
49
  "eos_token": "</s>",
50
  "legacy": false,
 
44
  }
45
  },
46
  "bos_token": "<s>",
47
+ "chat_template": "A curious user asks an AI about CAD models based on provided images. The AI provides a detailed and polite answer identifying the model. For each dataset entry:\nUSER: What CAD model do these images represent? <image>\nASSISTANT: The CAD model these images represent is {{ answer }}. {{eos_token}}",
48
  "clean_up_tokenization_spaces": false,
49
  "eos_token": "</s>",
50
  "legacy": false,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4219d359c5d4068f3633cc0b4cfabebe46c6b37767888d06bdc359562daa788b
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:428a7979ef86b8f56623fcc545d9d625e8faeb52eb8096fa19bad5a42eb08d96
3
  size 4984