paligemma_VQAv2_ko_v0

This model is a fine-tuned version of google/paligemma-3b-pt-224 on an unknown dataset.

Model description

This model was finetuned using VQAv2 datasets translated into KOREAN.

Intended uses & limitations

Nothing

Training and evaluation data

Nothing

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 1
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 4
  • total_train_batch_size: 4
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 2
  • num_epochs: 1

Training results

Framework versions

  • PEFT 0.8.2
  • Transformers 4.45.0.dev0
  • Pytorch 2.1.2+cu121
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
15
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The HF Inference API does not support visual-question-answering models for peft library.

Model tree for namyi/paligemma_VQAv2_ko_v0

Adapter
(169)
this model

Dataset used to train namyi/paligemma_VQAv2_ko_v0