๐
๐
Old Age
TokenMixup: Efficient Attention-guided Token-level Data Augmentation for Transformers
October 14, 2022 ยท Entered Twilight ยท ๐ Neural Information Processing Systems
Repo contents: LICENSE, README.md, assets, env.yaml, experiments, tokenmixup
Authors
Hyeong Kyu Choi, Joonmyung Choi, Hyunwoo J. Kim
arXiv ID
2210.07562
Category
cs.CV: Computer Vision
Citations
43
Venue
Neural Information Processing Systems
Repository
https://github.com/mlvlab/TokenMixup
โญ 48
Last Checked
2 months ago
Abstract
Mixup is a commonly adopted data augmentation technique for image classification. Recent advances in mixup methods primarily focus on mixing based on saliency. However, many saliency detectors require intense computation and are especially burdensome for parameter-heavy transformer models. To this end, we propose TokenMixup, an efficient attention-guided token-level data augmentation method that aims to maximize the saliency of a mixed set of tokens. TokenMixup provides x15 faster saliency-aware data augmentation compared to gradient-based methods. Moreover, we introduce a variant of TokenMixup which mixes tokens within a single instance, thereby enabling multi-scale feature augmentation. Experiments show that our methods significantly improve the baseline models' performance on CIFAR and ImageNet-1K, while being more efficient than previous methods. We also reach state-of-the-art performance on CIFAR-100 among from-scratch transformer models. Code is available at https://github.com/mlvlab/TokenMixup.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computer Vision
๐
๐
Old Age
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
R.I.P.
๐ป
Ghosted
You Only Look Once: Unified, Real-Time Object Detection
๐
๐
Old Age
SSD: Single Shot MultiBox Detector
๐
๐
Old Age
Squeeze-and-Excitation Networks
R.I.P.
๐ป
Ghosted