|
--- |
|
license: cc-by-4.0 |
|
datasets: |
|
- imagenet-1k |
|
metrics: |
|
- accuracy |
|
pipeline_tag: image-classification |
|
language: |
|
- en |
|
tags: |
|
- resnet |
|
- convolutional neural network |
|
- simpool |
|
- dino |
|
- computer vision |
|
- deep learning |
|
--- |
|
|
|
# Self-supervised ResNet-50 model |
|
|
|
ResNet-50 official model trained on ImageNet-1k for 100 epochs. Self-supervision with [DINO](https://arxiv.org/abs/2104.14294). Reproduced for ICCV 2023 [SimPool](https://arxiv.org/abs/2309.06891) paper. |
|
|
|
SimPool is a simple attention-based pooling method at the end of network, released in this [repository](https://github.com/billpsomas/simpool/). |
|
Disclaimer: This model card is written by the author of SimPool, i.e. [Bill Psomas](http://users.ntua.gr/psomasbill/). |
|
|
|
## Evaluation with k-NN |
|
|
|
| k | top1 | top5 | |
|
| ------- | ------- | ------- | |
|
| 10 | 61.84 | 80.35 | |
|
| 20 | 62.174 | 82.75 | |
|
| 100 | 60.088 | 84.216 | |
|
| 200 | 58.544 | 83.834 | |
|
|
|
|
|
## BibTeX entry and citation info |
|
|
|
``` |
|
@misc{psomas2023simpool, |
|
title={Keep It SimPool: Who Said Supervised Transformers Suffer from Attention Deficit?}, |
|
author={Bill Psomas and Ioannis Kakogeorgiou and Konstantinos Karantzalos and Yannis Avrithis}, |
|
year={2023}, |
|
eprint={2309.06891}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.CV} |
|
} |
|
``` |
|
|
|
``` |
|
@inproceedings{caron2021emerging, |
|
title={Emerging properties in self-supervised vision transformers}, |
|
author={Caron, Mathilde and Touvron, Hugo and Misra, Ishan and J{\'e}gou, Herv{\'e} and Mairal, Julien and Bojanowski, Piotr and Joulin, Armand}, |
|
booktitle={Proceedings of the IEEE/CVF international conference on computer vision}, |
|
pages={9650--9660}, |
|
year={2021} |
|
} |
|
``` |