Structured Dropout
Structured dropout is a regularization technique used in deep learning to improve model generalization and robustness by randomly dropping out neurons or connections during training, forcing the network to learn more robust features. Current research focuses on adapting and extending dropout methods for various architectures, including transformers and convolutional neural networks, exploring optimal dropout rates and strategies (e.g., conditional dropout, layer-wise regularization) to enhance performance in diverse applications such as language modeling, image processing, and reinforcement learning. These advancements contribute to more reliable and efficient deep learning models, impacting fields ranging from autonomous driving (3D object detection) to personalized education (predicting student success). The overall goal is to mitigate overfitting and improve the generalizability of deep learning models across various domains.
Papers
Robust and IP-Protecting Vertical Federated Learning against Unexpected Quitting of Parties
Jingwei Sun, Zhixu Du, Anna Dai, Saleh Baghersalimi, Alireza Amirshahi, David Atienza, Yiran Chen
Do Neural Topic Models Really Need Dropout? Analysis of the Effect of Dropout in Topic Modeling
Suman Adhya, Avishek Lahiri, Debarshi Kumar Sanyal