khanhduong commited on
Commit
ceb9156
Β·
1 Parent(s): 8c145b8

Finetune mimic_finetune 25 epochs

Browse files
Files changed (16) hide show
  1. 24_10_29_08_36_31_epoch0.pth β†’ Finetune_lora_llava_med/24_10_29_08_36_31_epoch0.pth +0 -0
  2. 24_10_29_09_05_34_epoch2.pth β†’ Finetune_lora_llava_med/24_10_29_09_05_34_epoch2.pth +0 -0
  3. 24_10_29_09_34_27_epoch4.pth β†’ Finetune_lora_llava_med/24_10_29_09_34_27_epoch4.pth +0 -0
  4. 24_10_29_10_03_26_epoch6.pth β†’ Finetune_lora_llava_med/24_10_29_10_03_26_epoch6.pth +0 -0
  5. 24_10_29_10_32_23_epoch8.pth β†’ Finetune_lora_llava_med/24_10_29_10_32_23_epoch8.pth +0 -0
  6. 24_10_29_11_01_36_epoch10.pth β†’ Finetune_lora_llava_med/24_10_29_11_01_36_epoch10.pth +0 -0
  7. 24_10_29_11_30_39_epoch12.pth β†’ Finetune_lora_llava_med/24_10_29_11_30_39_epoch12.pth +0 -0
  8. 24_10_29_11_59_33_epoch14.pth β†’ Finetune_lora_llava_med/24_10_29_11_59_33_epoch14.pth +0 -0
  9. 24_10_29_12_28_26_epoch16.pth β†’ Finetune_lora_llava_med/24_10_29_12_28_26_epoch16.pth +0 -0
  10. 24_10_29_12_57_21_epoch18.pth β†’ Finetune_lora_llava_med/24_10_29_12_57_21_epoch18.pth +0 -0
  11. 24_10_29_13_12_06_epoch19.pth β†’ Finetune_lora_llava_med/24_10_29_13_12_06_epoch19.pth +0 -0
  12. 24_10_29_13_45_22_epoch20.pth β†’ Finetune_lora_llava_med/24_10_29_13_45_22_epoch20.pth +0 -0
  13. 24_10_29_14_14_33_epoch22.pth β†’ Finetune_lora_llava_med/24_10_29_14_14_33_epoch22.pth +0 -0
  14. 24_10_29_14_43_38_epoch24.pth β†’ Finetune_lora_llava_med/24_10_29_14_43_38_epoch24.pth +0 -0
  15. llavamed_log.txt β†’ Finetune_lora_llava_med/24_10_29_14_43_38_pretrain160_finetune25_lora_lavamed_log.txt +0 -0
  16. llavamed_log1.txt +0 -119
24_10_29_08_36_31_epoch0.pth β†’ Finetune_lora_llava_med/24_10_29_08_36_31_epoch0.pth RENAMED
File without changes
24_10_29_09_05_34_epoch2.pth β†’ Finetune_lora_llava_med/24_10_29_09_05_34_epoch2.pth RENAMED
File without changes
24_10_29_09_34_27_epoch4.pth β†’ Finetune_lora_llava_med/24_10_29_09_34_27_epoch4.pth RENAMED
File without changes
24_10_29_10_03_26_epoch6.pth β†’ Finetune_lora_llava_med/24_10_29_10_03_26_epoch6.pth RENAMED
File without changes
24_10_29_10_32_23_epoch8.pth β†’ Finetune_lora_llava_med/24_10_29_10_32_23_epoch8.pth RENAMED
File without changes
24_10_29_11_01_36_epoch10.pth β†’ Finetune_lora_llava_med/24_10_29_11_01_36_epoch10.pth RENAMED
File without changes
24_10_29_11_30_39_epoch12.pth β†’ Finetune_lora_llava_med/24_10_29_11_30_39_epoch12.pth RENAMED
File without changes
24_10_29_11_59_33_epoch14.pth β†’ Finetune_lora_llava_med/24_10_29_11_59_33_epoch14.pth RENAMED
File without changes
24_10_29_12_28_26_epoch16.pth β†’ Finetune_lora_llava_med/24_10_29_12_28_26_epoch16.pth RENAMED
File without changes
24_10_29_12_57_21_epoch18.pth β†’ Finetune_lora_llava_med/24_10_29_12_57_21_epoch18.pth RENAMED
File without changes
24_10_29_13_12_06_epoch19.pth β†’ Finetune_lora_llava_med/24_10_29_13_12_06_epoch19.pth RENAMED
File without changes
24_10_29_13_45_22_epoch20.pth β†’ Finetune_lora_llava_med/24_10_29_13_45_22_epoch20.pth RENAMED
File without changes
24_10_29_14_14_33_epoch22.pth β†’ Finetune_lora_llava_med/24_10_29_14_14_33_epoch22.pth RENAMED
File without changes
24_10_29_14_43_38_epoch24.pth β†’ Finetune_lora_llava_med/24_10_29_14_43_38_epoch24.pth RENAMED
File without changes
llavamed_log.txt β†’ Finetune_lora_llava_med/24_10_29_14_43_38_pretrain160_finetune25_lora_lavamed_log.txt RENAMED
File without changes
llavamed_log1.txt DELETED
@@ -1,119 +0,0 @@
1
- {
2
- "run": {
3
- "task": "image_text_pretrain",
4
- "lr_sched": "linear_warmup_cosine_lr",
5
- "init_lr": 3e-05,
6
- "min_lr": 1e-05,
7
- "warmup_lr": 1e-06,
8
- "weight_decay": 0.05,
9
- "max_epoch": 20,
10
- "iters_per_epoch": 600,
11
- "batch_size_train": 15,
12
- "batch_size_eval": 1,
13
- "num_workers": 4,
14
- "warmup_steps": 200,
15
- "seed": 42,
16
- "output_dir": "output/xraygpt_vqa_finetune",
17
- "amp": true,
18
- "resume_ckpt_path": null,
19
- "evaluate": false,
20
- "train_splits": [
21
- "train"
22
- ],
23
- "device": "cuda",
24
- "world_size": 2,
25
- "dist_url": "env://",
26
- "distributed": true,
27
- "rank": 0,
28
- "gpu": 0,
29
- "dist_backend": "nccl"
30
- },
31
- "model": {
32
- "arch": "mini_gpt4",
33
- "image_size": 224,
34
- "drop_path_rate": 0,
35
- "use_grad_checkpoint": false,
36
- "vit_precision": "fp16",
37
- "freeze_vit": true,
38
- "freeze_qformer": true,
39
- "num_query_token": 32,
40
- "llama_model": "Joycean0301/Llama-3.2-3B-Instruct-Medical-Conversational",
41
- "prompt": "",
42
- "lora_r": 16,
43
- "lora_alpha": 16,
44
- "lora_target_modules": [
45
- "q_proj",
46
- "k_proj",
47
- "v_proj",
48
- "o_proj",
49
- "gate_proj",
50
- "up_proj",
51
- "down_proj"
52
- ],
53
- "lora_dropout": 0.1,
54
- "lora_bias": "none",
55
- "model_type": "pretrain_vicuna",
56
- "max_txt_len": 160,
57
- "end_sym": "<|eot_id|>",
58
- "ckpt": "24_10_29_03_29_41_epoch160.pth"
59
- },
60
- "preprocess": {
61
- "vis_processor": {
62
- "train": {
63
- "name": "blip2_image_train",
64
- "image_size": 224
65
- },
66
- "eval": {
67
- "name": "blip2_image_eval",
68
- "image_size": 224
69
- }
70
- },
71
- "text_processor": {
72
- "train": {
73
- "name": "blip_caption"
74
- },
75
- "eval": {
76
- "name": "blip_caption"
77
- }
78
- }
79
- },
80
- "datasets": {
81
- "vqa": {
82
- "data_type": "images",
83
- "build_info": {
84
- "storage": "dataset/llava_med_60k_instruct"
85
- },
86
- "vis_processor": {
87
- "train": {
88
- "name": "blip2_image_train",
89
- "image_size": 224
90
- }
91
- },
92
- "text_processor": {
93
- "train": {
94
- "name": "blip_caption"
95
- }
96
- }
97
- }
98
- }
99
- }
100
- {"train_lr": "0.000025", "train_loss": "0.941024"}
101
- {"train_lr": "0.000030", "train_loss": "0.685024"}
102
- {"train_lr": "0.000029", "train_loss": "0.622166"}
103
- {"train_lr": "0.000029", "train_loss": "0.580002"}
104
- {"train_lr": "0.000028", "train_loss": "0.544959"}
105
- {"train_lr": "0.000026", "train_loss": "0.512786"}
106
- {"train_lr": "0.000025", "train_loss": "0.484002"}
107
- {"train_lr": "0.000024", "train_loss": "0.454916"}
108
- {"train_lr": "0.000022", "train_loss": "0.427478"}
109
- {"train_lr": "0.000021", "train_loss": "0.401973"}
110
- {"train_lr": "0.000019", "train_loss": "0.378341"}
111
- {"train_lr": "0.000018", "train_loss": "0.353522"}
112
- {"train_lr": "0.000016", "train_loss": "0.331684"}
113
- {"train_lr": "0.000015", "train_loss": "0.311488"}
114
- {"train_lr": "0.000014", "train_loss": "0.291895"}
115
- {"train_lr": "0.000012", "train_loss": "0.274390"}
116
- {"train_lr": "0.000011", "train_loss": "0.257786"}
117
- {"train_lr": "0.000011", "train_loss": "0.243516"}
118
- {"train_lr": "0.000010", "train_loss": "0.229999"}
119
- {"train_lr": "0.000010", "train_loss": "0.219022"}