Slimmable Neural Network
Slimmable neural networks are a class of models designed to dynamically adjust their size and computational complexity based on available resources, aiming to optimize performance while minimizing energy consumption and memory footprint. Current research focuses on developing efficient training algorithms, such as superposition coding and specialized pruning techniques, and applying these networks to resource-constrained environments like mobile and edge devices, autonomous systems, and federated learning settings. This adaptability makes slimmable networks particularly valuable for applications where computational resources are limited or vary significantly, improving efficiency and enabling deployment on a wider range of hardware.
Papers
Communication and Energy Efficient Slimmable Federated Learning via Superposition Coding and Successive Decoding
Hankyul Baek, Won Joon Yun, Soyi Jung, Jihong Park, Mingyue Ji, Joongheon Kim, Mehdi Bennis
Joint Superposition Coding and Training for Federated Learning over Multi-Width Neural Networks
Hankyul Baek, Won Joon Yun, Yunseok Kwak, Soyi Jung, Mingyue Ji, Mehdi Bennis, Jihong Park, Joongheon Kim