Shortcuts

Model Zoo of Supervised Learning

OpenMixup provides mixup benchmarks on supervised learning on various tasks. Configs, experiments results, training logs will be updated as soon as possible. More mixup variants will be supported!

Now, we have supported 13 popular mixup methods! Notice that * denotes open-source arxiv pre-prints reproduced by us, and :book: denotes original results reproduced by official implementations. We modified the original AttentiveMix by using pre-trained R-18 and sampling $\lambda$ from $\Beta(\alpha,8)$ as AttentiveMix+. Moreover, you can summary experiment results (json files) by tools in openmixup/tools/summary/.

Supported sample mixing policies
Supported label mixing policies

ImageNet Benchmarks

We provide three popular benchmarks on ImageNet-1k based on various backbones. We also provide results on TinyImageNet-200 for fast training. The median of top-1 accuracy in the last 5/10 training epochs for 100/300 epochs is reported for ResNet variants, and the best top-1 accuracy is reported for DeiT training settings.

PyTorch-style Training Settings on ImageNet-1k

Note

  • These benchmarks follow PyTorch-style settings, training 100 and 300 epochs on ImageNet-1k.

  • Please run configs in configs/classification/imagenet/mixups/basic, and modify epochs and mix_mode in auto_train_in_mixups.py to generate proper configs by yourself.

  • Notice that :book: denotes original results reproduced by official implementations.

Backbones ResNet-18 ResNet-34 ResNet-50 ResNet-101 ResNeXt-101
Epochs 100 epochs 100 epochs 100 epochs 100 epochs 100 epochs
Vanilla 70.04 73.85 76.83 78.18 78.71
MixUp 69.98 73.97 77.12 78.97 79.98
CutMix 68.95 73.58 77.17 78.96 80.42
ManifoldMix 69.98 73.98 77.01 79.02 79.93
SaliencyMix 69.16 73.56 77.14 79.32 80.27
AttentiveMix+ 68.57 - 77.28 - -
FMix* 69.96 74.08 77.19 79.09 80.06
PuzzleMix 70.12 74.26 77.54 79.43 80.53
Co-Mixup :book: - - 77.60 - -
SuperMix :book: - - 77.63 - -
ResizeMix* 69.50 73.88 77.42 79.27 80.55
AlignMix :book: - - 78.00 - -
Grafting :book: - - 77.74 - -
AutoMix* 70.50 74.52 77.91 79.87 80.89
SAMix* 70.83 74.95 78.06 80.05 80.98
Backbones ResNet-18 ResNet-34 ResNet-50 ResNet-101
Epochs 300 epochs 300 epochs 300 epochs 300 epochs
Vanilla 71.83 75.29 77.35 78.91
MixUp 71.72 75.73 78.44 80.60
CutMix 71.01 75.16 78.69 80.59
ManifoldMix 71.73 75.44 78.21 80.64
SaliencyMix 70.21 75.01 78.46 80.45
FMix* 70.30 75.12 78.51 80.20
PuzzleMix 71.64 75.84 78.86 80.67
ResizeMix* 71.32 75.64 78.91 80.52
AlignMix :book: - - 79.32 -
AutoMix* 72.05 76.10 79.25 80.98
SAMix* 72.27 76.28 79.39 81.10

Timm RSB A2/A3 Training Settings on ImageNet-1k

Note

  • This benchmark follows timm RSB A2/A3 settings, training 300/100 epochs with the BCE loss on ImageNet-1k. RSB A3 is a fast

  • Please run configs in configs/classification/imagenet/mixups/rsb_a2 and configs/classification/imagenet/mixups/rsb_a3.

Backbones ResNet-50 ResNet-50 Eff-B0 Eff-B0 Mob.V2 1x Mob.V2 1x
Settings A2 A3 A2 A3 A2 A3
RSB 79.80 78.08 77.26 74.02 72.87 69.86
MixUp 77.66 77.19 73.87 72.78 70.17
CutMix 79.38 77.62 77.24 73.46 72.23 69.62
ManifoldMix 79.47 77.78 77.22 73.83 72.34 70.05
SaliencyMix 79.42 77.93 77.67 73.42 72.07 69.69
FMix* 79.05 77.76 77.33 73.71 72.79 70.10
PuzzleMix 79.78 78.02 77.35 74.10 72.85 70.04
ResizeMix* 79.74 77.85 77.27 73.67 72.50 69.94
AutoMix* 78.44 77.58 74.61 73.19 71.16
SAMix 78.64 77.69 75.28 73.42 71.24

DeiT Training Settings on ImageNet-1k

Note

  • Since recently proposed transformer-based archetectures adopt mixups as parts of enssential augmentations, we report the mean of the best performance in trivals as their original paper. Notice that the performances of transformer-based archetectures are more difficult to reproduce than ResNet variants.

  • Please run configs in configs/classification/imagenet/deit/, configs/classification/imagenet/swin/, and configs/classification/imagenet/convnext/.

  • Notice that :book: denotes original results reproduced by official implementations.

Methods DeiT-Small Swin-Tiny ConvNeXt-Tiny
Vanilla 73.57 78.95 79.22
DeiT 79.80 81.20 82.10
MixUp 77.72 80.23 80.88
CutMix 79.54 80.59 81.57
ManifoldMix - - 80.57
AttentiveMix+ 77.63 81.14
SaliencyMix 78.70 81.33
PuzzleMix 79.73 81.48
FMix 77.37 81.04
ResizeMix 76.79 80.73 81.64
TransMix :book: 80.70 81.80 -
AutoMix 80.78 82.28
SAMix 80.94 82.35

TinyImageNet-200

Note

  • This benchmark largely based on CIFAR settings, training 400 epochs on TinyImageNet-200.

  • Please run configs in configs/classification/tiny_imagenet/mixups/.

  • Notice that :book: denotes original results reproduced by official implementations.

Backbones ResNet-18 ResNeXt-50
Vanilla 61.68 65.04
MixUp 63.86 66.36
CutMix 65.53 66.47
ManifoldMix 64.15 67.30
SaliencyMix 64.60 66.55
AttentiveMix+ 64.85 67.42
FMix* 63.47 65.08
GridMix :book: - 69.12
PuzzleMix 65.81 67.83
Co-Mixup :book: 65.92 68.02
ResizeMix* 63.74 65.87
Grafting :book: 64.84 -
AlignMix :book: 66.87 -
AutoMix* 67.33 70.72
SAMix* 68.89 72.18

CIFAR-10/100 Benchmarks

CIFAR benchmarks based on ResNet variants. We report the median of top-1 accuracy in the last 10 training epochs.

CIFAR-10

Note

  • This benchmark follows CutMix settings, training 200/400/800/1200 epochs on CIFAR-10.

  • Please run configs in configs/classification/cifar10/mixups/.

  • Notice that :book: denotes original results reproduced by official implementations.

Backbones ResNet-18 ResNet-18 ResNet-18 ResNet-18
Epochs 200 epochs 400 epochs 800 epochs 1200 epochs
Vanilla 94.87 95.10 95.50 95.59
MixUp 95.70 96.55 96.62 96.84
CutMix 96.11 96.13 96.68 96.56
ManifoldMix 96.04 96.57 96.71 97.02
SaliencyMix 96.05 96.42 96.20 96.18
AttentiveMix+ 96.21 96.45 96.63 96.49
FMix* 96.17 96.53 96.18 96.01
PuzzleMix 96.42 96.87 97.10 97.13
ResizeMix* 96.16 96.91 96.76 97.04
AlignMix :book: 97.05
AutoMix* 96.59 97.08 97.34 97.30
SAMix* 96.67 97.16 97.50 97.41
Backbones ResNeXt-50 ResNeXt-50 ResNeXt-50 ResNeXt-50
Epochs 200 epochs 400 epochs 800 epochs 1200 epochs
Vanilla 95.92 95.81 96.23 96.26
MixUp 96.88 97.19 97.30 97.33
CutMix 96.78 96.54 96.60 96.35
ManifoldMix 96.97 97.39 97.33 97.36
SaliencyMix 96.65 96.89 96.70 96.60
AttentiveMix+ 96.84 96.91 96.87 96.62
FMix* 96.72 96.76 96.76 96.10
PuzzleMix 97.05 97.24 97.37 97.34
ResizeMix* 97.02 97.38 97.21 97.36
AlignMix :book:
AutoMix* 97.19 97.42 97.65 97.51
SAMix* 97.23 97.51 97.93 97.74

CIFAR-100

Note

  • This benchmark follows CutMix settings, training 200/400/800/1200 epochs on CIFAR-100. Please use wd=5e-4 for cutting-based methods (CutMix, AttributeMix+, SaliencyMix, FMix, ResizeMix) based on ResNeXt-50 for better performances.

  • Please run configs in configs/classification/cifar100/mixups/.

  • Notice that :book: denotes original results reproduced by official implementations.

Backbones ResNet-18 ResNet-18 ResNet-18 ResNet-18
Epoch 200 epochs 400 epochs 800 epochs 1200 epochs
Vanilla 76.42 77.73 78.04 78.55
MixUp 78.52 79.34 79.12 79.24
CutMix 79.45 79.58 78.17 78.29
ManifoldMix 79.18 80.18 80.35 80.21
SaliencyMix 79.75 79.64 79.12 77.66
AttentiveMix+ 79.62 80.14 78.91 78.41
FMix* 78.91 79.91 79.69 79.50
PuzzleMix 79.96 80.82 81.13 81.10
Co-Mixup :book: 80.01 80.87 81.17 81.18
ResizeMix* 79.56 79.19 80.01 79.23
AlignMix :book: 81.71
AutoMix* 80.12 81.78 82.04 81.95
SAMix* 81.21 81.97 82.30 82.41
Backbones ResNeXt-50 ResNeXt-50 ResNeXt-50 ResNeXt-50 WRN-28-8
Epoch 200 epochs 400 epochs 800 epochs 1200 epochs 400 epochs
Vanilla 79.37 80.24 81.09 81.32 81.63
MixUp 81.18 82.54 82.10 81.77 82.82
CutMix 81.52 78.52 78.32 77.17 84.45
ManifoldMix 81.59 82.56 82.88 83.28 83.24
SaliencyMix 80.72 78.63 78.77 77.51 84.35
AttentiveMix+ 81.69 81.53 80.54 79.60 84.34
FMix* 79.87 78.99 79.02 78.24 84.21
PuzzleMix 81.69 82.84 82.85 82.93 85.02
Co-Mixup :book: 81.73 82.88 82.91 82.97 85.05
ResizeMix* 79.56 79.78 80.35 79.73 84.87
AlignMix :book:
AutoMix* 82.84 83.32 83.64 83.80 85.18
SAMix* 83.81 84.27 84.42 84.31 85.50

Fine-grained and Scenic Classification Benchmarks

We further provide benchmarks on downstream classification scenarios. We report the median of top-1 accuracy in the last 5/10 training epochs for 100/200 epochs.

Transfer Learning on Small-scale Datasets

Note

  • These benchmarks follow transfer learning settings on fine-grained datasets. use PyTorch pre-trained models as initialization and train 200 epochs on CUB-200 and FGVC-Aircraft.

  • Please run configs in configs/classification/aircrafts/ and configs/classification/cub200/.

Datasets CUB-200 CUB-200 Aircraft Aircraft
Backbones ResNet-18 ResNeXt-50 ResNet-18 ResNeXt-50
Vanilla 77.68 83.01 80.23 85.10
MixUp 78.39 84.58 79.52 85.18
CutMix 78.40 85.68 78.84 84.55
ManifoldMix 79.76 86.38 80.68 86.60
SaliencyMix 77.95 83.29 80.02 84.31
FMix* 77.28 84.06 79.36 86.23
PuzzleMix 78.63 84.51 80.76 86.23
ResizeMix* 78.50 84.77 78.10 84.08
AutoMix* 79.87 86.56 81.37 86.72
SAMix* 81.11 86.83 82.15 86.80

Large-scale Datasets

Note

  • These benchmarks largely based on PyTorch-style ImageNet-1k training settings, training 100 epochs from stretch on iNat2017/2018 and Place205.

  • Please run configs in configs/classification/inaturalist2017/, configs/classification/inaturalist2018/, and configs/classification/place205/.

Datasets iNat2017 iNat2017 iNat2018 iNat2018
Backbones ResNet-50 ResNeXt-101 ResNet-50 ResNeXt-101
Vanilla 60.23 63.70 62.53 66.94
MixUp 61.22 66.27 62.69 67.56
CutMix 62.34 67.59 63.91 69.75
ManifoldMix 61.47 66.08 63.46 69.30
SaliencyMix 62.51 67.20 64.27 70.01
FMix* 61.90 66.64 63.71 69.46
PuzzleMix 62.66 67.72 64.36 70.12
ResizeMix* 62.29 66.82 64.12 69.30
AutoMix* 63.08 68.03 64.73 70.49
SAMix* 63.32 68.26 64.84 70.54
Datasets Place205 Place205
Backbones ResNet-18 ResNet-50
Vanilla 59.63 63.10
MixUp 59.33 63.01
CutMix 59.21 63.75
ManifoldMix 59.46 63.23
SaliencyMix 59.50 63.33
FMix* 59.51 63.63
PuzzleMix 59.62 63.91
ResizeMix* 59.66 63.88
AutoMix* 59.74 64.06
SAMix* 59.86 64.27
Read the Docs v: stable
Versions
latest
stable
Downloads
On Read the Docs
Project Home
Builds

Free document hosting provided by Read the Docs.