{ "dataset": { "align_stage_components": [ "download/llava-laion-cc-sbu-558k/chat.json", "download/llava-laion-cc-sbu-558k" ], "dataset_id": "llava-lrv", "dataset_root_dir": "/data/projects/12003782/training_data", "finetune_stage_components": [ "download/llava-v1.5-instruct/llava_v1_5_lrv_mix1008k.json", "download/llava-v1.5-instruct" ], "type": "llava-lrv" }, "model": { "align_epochs": 1, "align_global_batch_size": 64, "align_learning_rate": 0.001, "align_lr_scheduler_type": "linear-warmup+cosine-decay", "align_max_grad_norm": 1.0, "align_max_steps": null, "align_per_device_batch_size": 8, "align_train_strategy": "fsdp-shard-grad-op", "align_warmup_ratio": 0.03, "align_weight_decay": 0.0, "arch_specifier": "fused-gelu-mlp", "enable_gradient_checkpointing": true, "enable_mixed_precision_training": true, "finetune_epochs": 1, "finetune_global_batch_size": 32, "finetune_learning_rate": 2e-05, "finetune_lr_scheduler_type": "linear-warmup+cosine-decay", "finetune_max_grad_norm": 1.0, "finetune_max_steps": null, "finetune_per_device_batch_size": 4, "finetune_train_strategy": "fsdp-full-shard", "finetune_warmup_ratio": 0.03, "finetune_weight_decay": 0.1, "image_resize_strategy": "resize-naive", "llm_backbone_id": "phi3_base", "llm_max_length": 2048, "model_id": "dino-siglip-phi3-lora-model", "reduce_in_full_precision": false, "type": "dino-siglip-phi3-lora-model", "vision_backbone_id": "dinosiglip-vit-so-384px" }, "pretrained_checkpoint": null, "run_id": "llava-lrv+dino-siglip-phi3-lora-model+stage-align+x7", "run_root_dir": "/data/projects/12003782/model_weights/runs", "seed": 7, "stage": "align", "trackers": [ "jsonl", "wandb" ], "wandb_entity": null, "wandb_project": "nscc-prismatic-phi3" }