reproducing: "Ensemble everything everywhere: Multi-scale aggregation for adversarial robustness" (https://arxiv.org/abs/2408.05446)

source code and usage examples: https://github.com/ETH-DISCO/self-ensembling

architecture based on Torchvision's Resnet152 default implementation

hyperparameters:

  • criterion: torch.nn.CrossEntropyLoss()
  • optimizer: torch.optim.AdamW
  • scaler: GradScaler
  • datasets: ["cifar10", "cirfar100"]
  • lr: 0.0001
  • num_epochs: 16 (higher would be even better, but maybe by <1%)
  • crossmax_k: 2 (difference between crossmax_k=2 and crossmax_k=3 is about 1-2%, so it's not a big deal)
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.