metadata
license: apache-2.0
datasets:
- Ejafa/ye-pop
A ViT-B/32 CLIP model trained for 4 epochs on the ye-pop dataset (491,520 images and their alt-texts). Research artifact of clip-synthetic-captions.
Note: likely not directly useful as it is severely undertrained.