Commit
Β·
ceb9156
1
Parent(s):
8c145b8
Finetune mimic_finetune 25 epochs
Browse files- 24_10_29_08_36_31_epoch0.pth β Finetune_lora_llava_med/24_10_29_08_36_31_epoch0.pth +0 -0
- 24_10_29_09_05_34_epoch2.pth β Finetune_lora_llava_med/24_10_29_09_05_34_epoch2.pth +0 -0
- 24_10_29_09_34_27_epoch4.pth β Finetune_lora_llava_med/24_10_29_09_34_27_epoch4.pth +0 -0
- 24_10_29_10_03_26_epoch6.pth β Finetune_lora_llava_med/24_10_29_10_03_26_epoch6.pth +0 -0
- 24_10_29_10_32_23_epoch8.pth β Finetune_lora_llava_med/24_10_29_10_32_23_epoch8.pth +0 -0
- 24_10_29_11_01_36_epoch10.pth β Finetune_lora_llava_med/24_10_29_11_01_36_epoch10.pth +0 -0
- 24_10_29_11_30_39_epoch12.pth β Finetune_lora_llava_med/24_10_29_11_30_39_epoch12.pth +0 -0
- 24_10_29_11_59_33_epoch14.pth β Finetune_lora_llava_med/24_10_29_11_59_33_epoch14.pth +0 -0
- 24_10_29_12_28_26_epoch16.pth β Finetune_lora_llava_med/24_10_29_12_28_26_epoch16.pth +0 -0
- 24_10_29_12_57_21_epoch18.pth β Finetune_lora_llava_med/24_10_29_12_57_21_epoch18.pth +0 -0
- 24_10_29_13_12_06_epoch19.pth β Finetune_lora_llava_med/24_10_29_13_12_06_epoch19.pth +0 -0
- 24_10_29_13_45_22_epoch20.pth β Finetune_lora_llava_med/24_10_29_13_45_22_epoch20.pth +0 -0
- 24_10_29_14_14_33_epoch22.pth β Finetune_lora_llava_med/24_10_29_14_14_33_epoch22.pth +0 -0
- 24_10_29_14_43_38_epoch24.pth β Finetune_lora_llava_med/24_10_29_14_43_38_epoch24.pth +0 -0
- llavamed_log.txt β Finetune_lora_llava_med/24_10_29_14_43_38_pretrain160_finetune25_lora_lavamed_log.txt +0 -0
- llavamed_log1.txt +0 -119
24_10_29_08_36_31_epoch0.pth β Finetune_lora_llava_med/24_10_29_08_36_31_epoch0.pth
RENAMED
File without changes
|
24_10_29_09_05_34_epoch2.pth β Finetune_lora_llava_med/24_10_29_09_05_34_epoch2.pth
RENAMED
File without changes
|
24_10_29_09_34_27_epoch4.pth β Finetune_lora_llava_med/24_10_29_09_34_27_epoch4.pth
RENAMED
File without changes
|
24_10_29_10_03_26_epoch6.pth β Finetune_lora_llava_med/24_10_29_10_03_26_epoch6.pth
RENAMED
File without changes
|
24_10_29_10_32_23_epoch8.pth β Finetune_lora_llava_med/24_10_29_10_32_23_epoch8.pth
RENAMED
File without changes
|
24_10_29_11_01_36_epoch10.pth β Finetune_lora_llava_med/24_10_29_11_01_36_epoch10.pth
RENAMED
File without changes
|
24_10_29_11_30_39_epoch12.pth β Finetune_lora_llava_med/24_10_29_11_30_39_epoch12.pth
RENAMED
File without changes
|
24_10_29_11_59_33_epoch14.pth β Finetune_lora_llava_med/24_10_29_11_59_33_epoch14.pth
RENAMED
File without changes
|
24_10_29_12_28_26_epoch16.pth β Finetune_lora_llava_med/24_10_29_12_28_26_epoch16.pth
RENAMED
File without changes
|
24_10_29_12_57_21_epoch18.pth β Finetune_lora_llava_med/24_10_29_12_57_21_epoch18.pth
RENAMED
File without changes
|
24_10_29_13_12_06_epoch19.pth β Finetune_lora_llava_med/24_10_29_13_12_06_epoch19.pth
RENAMED
File without changes
|
24_10_29_13_45_22_epoch20.pth β Finetune_lora_llava_med/24_10_29_13_45_22_epoch20.pth
RENAMED
File without changes
|
24_10_29_14_14_33_epoch22.pth β Finetune_lora_llava_med/24_10_29_14_14_33_epoch22.pth
RENAMED
File without changes
|
24_10_29_14_43_38_epoch24.pth β Finetune_lora_llava_med/24_10_29_14_43_38_epoch24.pth
RENAMED
File without changes
|
llavamed_log.txt β Finetune_lora_llava_med/24_10_29_14_43_38_pretrain160_finetune25_lora_lavamed_log.txt
RENAMED
File without changes
|
llavamed_log1.txt
DELETED
@@ -1,119 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"run": {
|
3 |
-
"task": "image_text_pretrain",
|
4 |
-
"lr_sched": "linear_warmup_cosine_lr",
|
5 |
-
"init_lr": 3e-05,
|
6 |
-
"min_lr": 1e-05,
|
7 |
-
"warmup_lr": 1e-06,
|
8 |
-
"weight_decay": 0.05,
|
9 |
-
"max_epoch": 20,
|
10 |
-
"iters_per_epoch": 600,
|
11 |
-
"batch_size_train": 15,
|
12 |
-
"batch_size_eval": 1,
|
13 |
-
"num_workers": 4,
|
14 |
-
"warmup_steps": 200,
|
15 |
-
"seed": 42,
|
16 |
-
"output_dir": "output/xraygpt_vqa_finetune",
|
17 |
-
"amp": true,
|
18 |
-
"resume_ckpt_path": null,
|
19 |
-
"evaluate": false,
|
20 |
-
"train_splits": [
|
21 |
-
"train"
|
22 |
-
],
|
23 |
-
"device": "cuda",
|
24 |
-
"world_size": 2,
|
25 |
-
"dist_url": "env://",
|
26 |
-
"distributed": true,
|
27 |
-
"rank": 0,
|
28 |
-
"gpu": 0,
|
29 |
-
"dist_backend": "nccl"
|
30 |
-
},
|
31 |
-
"model": {
|
32 |
-
"arch": "mini_gpt4",
|
33 |
-
"image_size": 224,
|
34 |
-
"drop_path_rate": 0,
|
35 |
-
"use_grad_checkpoint": false,
|
36 |
-
"vit_precision": "fp16",
|
37 |
-
"freeze_vit": true,
|
38 |
-
"freeze_qformer": true,
|
39 |
-
"num_query_token": 32,
|
40 |
-
"llama_model": "Joycean0301/Llama-3.2-3B-Instruct-Medical-Conversational",
|
41 |
-
"prompt": "",
|
42 |
-
"lora_r": 16,
|
43 |
-
"lora_alpha": 16,
|
44 |
-
"lora_target_modules": [
|
45 |
-
"q_proj",
|
46 |
-
"k_proj",
|
47 |
-
"v_proj",
|
48 |
-
"o_proj",
|
49 |
-
"gate_proj",
|
50 |
-
"up_proj",
|
51 |
-
"down_proj"
|
52 |
-
],
|
53 |
-
"lora_dropout": 0.1,
|
54 |
-
"lora_bias": "none",
|
55 |
-
"model_type": "pretrain_vicuna",
|
56 |
-
"max_txt_len": 160,
|
57 |
-
"end_sym": "<|eot_id|>",
|
58 |
-
"ckpt": "24_10_29_03_29_41_epoch160.pth"
|
59 |
-
},
|
60 |
-
"preprocess": {
|
61 |
-
"vis_processor": {
|
62 |
-
"train": {
|
63 |
-
"name": "blip2_image_train",
|
64 |
-
"image_size": 224
|
65 |
-
},
|
66 |
-
"eval": {
|
67 |
-
"name": "blip2_image_eval",
|
68 |
-
"image_size": 224
|
69 |
-
}
|
70 |
-
},
|
71 |
-
"text_processor": {
|
72 |
-
"train": {
|
73 |
-
"name": "blip_caption"
|
74 |
-
},
|
75 |
-
"eval": {
|
76 |
-
"name": "blip_caption"
|
77 |
-
}
|
78 |
-
}
|
79 |
-
},
|
80 |
-
"datasets": {
|
81 |
-
"vqa": {
|
82 |
-
"data_type": "images",
|
83 |
-
"build_info": {
|
84 |
-
"storage": "dataset/llava_med_60k_instruct"
|
85 |
-
},
|
86 |
-
"vis_processor": {
|
87 |
-
"train": {
|
88 |
-
"name": "blip2_image_train",
|
89 |
-
"image_size": 224
|
90 |
-
}
|
91 |
-
},
|
92 |
-
"text_processor": {
|
93 |
-
"train": {
|
94 |
-
"name": "blip_caption"
|
95 |
-
}
|
96 |
-
}
|
97 |
-
}
|
98 |
-
}
|
99 |
-
}
|
100 |
-
{"train_lr": "0.000025", "train_loss": "0.941024"}
|
101 |
-
{"train_lr": "0.000030", "train_loss": "0.685024"}
|
102 |
-
{"train_lr": "0.000029", "train_loss": "0.622166"}
|
103 |
-
{"train_lr": "0.000029", "train_loss": "0.580002"}
|
104 |
-
{"train_lr": "0.000028", "train_loss": "0.544959"}
|
105 |
-
{"train_lr": "0.000026", "train_loss": "0.512786"}
|
106 |
-
{"train_lr": "0.000025", "train_loss": "0.484002"}
|
107 |
-
{"train_lr": "0.000024", "train_loss": "0.454916"}
|
108 |
-
{"train_lr": "0.000022", "train_loss": "0.427478"}
|
109 |
-
{"train_lr": "0.000021", "train_loss": "0.401973"}
|
110 |
-
{"train_lr": "0.000019", "train_loss": "0.378341"}
|
111 |
-
{"train_lr": "0.000018", "train_loss": "0.353522"}
|
112 |
-
{"train_lr": "0.000016", "train_loss": "0.331684"}
|
113 |
-
{"train_lr": "0.000015", "train_loss": "0.311488"}
|
114 |
-
{"train_lr": "0.000014", "train_loss": "0.291895"}
|
115 |
-
{"train_lr": "0.000012", "train_loss": "0.274390"}
|
116 |
-
{"train_lr": "0.000011", "train_loss": "0.257786"}
|
117 |
-
{"train_lr": "0.000011", "train_loss": "0.243516"}
|
118 |
-
{"train_lr": "0.000010", "train_loss": "0.229999"}
|
119 |
-
{"train_lr": "0.000010", "train_loss": "0.219022"}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|