Neural Architecture
Neural architecture research focuses on designing and optimizing the structure of artificial neural networks to improve efficiency, accuracy, and interpretability. Current efforts concentrate on developing novel architectures like Kolmogorov-Arnold Networks and transformers, employing efficient search algorithms (e.g., evolutionary algorithms, generative flows) to explore vast design spaces, and analyzing the representational similarity and training efficiency of different models. These advancements are crucial for deploying deep learning in resource-constrained environments and for gaining a deeper understanding of how neural networks learn and generalize, impacting fields ranging from computer vision and natural language processing to scientific computing and edge devices.
Papers
Walking Noise: On Layer-Specific Robustness of Neural Architectures against Noisy Computations and Associated Characteristic Learning Dynamics
Hendrik Borras, Bernhard Klein, Holger Fröning
A deep learning Attention model to solve the Vehicle Routing Problem and the Pick-up and Delivery Problem with Time Windows
Baptiste Rabecq, Rémy Chevrier
AIO-P: Expanding Neural Performance Predictors Beyond Image Classification
Keith G. Mills, Di Niu, Mohammad Salameh, Weichen Qiu, Fred X. Han, Puyuan Liu, Jialin Zhang, Wei Lu, Shangling Jui
GENNAPE: Towards Generalized Neural Architecture Performance Estimators
Keith G. Mills, Fred X. Han, Jialin Zhang, Fabian Chudak, Ali Safari Mamaghani, Mohammad Salameh, Wei Lu, Shangling Jui, Di Niu