Synthetic Minority Oversampling Technique

The Synthetic Minority Oversampling Technique (SMOTE) addresses the challenge of class imbalance in machine learning datasets, where one class significantly outnumbers others, hindering accurate model training. Current research focuses on improving SMOTE's effectiveness through variations like Minimum Enclosing Ball SMOTE (MEB-SMOTE) and integrating it with other techniques such as deep learning models (e.g., convolutional neural networks, generative adversarial networks) and data augmentation methods (e.g., Mixup). This work is crucial for enhancing the performance of machine learning models across diverse applications, including medical diagnosis, fraud detection, and cybersecurity, where imbalanced datasets are common. Recent studies also explore the theoretical underpinnings of SMOTE and its variants, aiming to better understand its strengths and limitations.

Papers