question

#1
by giannisan - opened

I'm interested in fine-tuning the ViViT model with low rank adaptation, how did you manage this?

Owner

you can use the same lora config as the vit implementation, the target modules should also match the one in your architecture . It works pretty wel
l for fine tuning

Thanks for the feedback!! I'm pretty new to vision models, as I'm coming from LLMs. Do you have documentation for this? Or where can i find how this is done? I saw a post on hf about vit lora, but not sure if that's what you meant. Thanks again for the help

Owner

yeah, i usually work on vision models . This should give you a clear idea
https://huggingface.co/docs/peft/task_guides/image_classification_lora

NiiCole changed discussion status to closed

yeah, i usually work on vision models . This should give you a clear idea
https://huggingface.co/docs/peft/task_guides/image_classification_lora

Thanks!

Sign up or log in to comment