Optimal Architecture
Optimal architecture research focuses on automating the design of efficient and high-performing neural networks, addressing the limitations of manual design. Current efforts concentrate on developing efficient search algorithms, including those based on reinforcement learning, evolutionary methods, and differentiable architecture search, often applied to various model architectures like transformers and convolutional neural networks. This research is crucial for advancing deep learning applications by reducing computational costs, improving model performance, and enabling deployment on resource-constrained devices like mobile phones and embedded systems. The resulting optimized architectures are impacting diverse fields, from natural language processing and computer vision to resource-limited applications like autonomous drones.
Papers
$\Lambda$-DARTS: Mitigating Performance Collapse by Harmonizing Operation Selection among Cells
Sajad Movahedi, Melika Adabinejad, Ayyoob Imani, Arezou Keshavarz, Mostafa Dehghani, Azadeh Shakery, Babak N. Araabi
Pareto-aware Neural Architecture Generation for Diverse Computational Budgets
Yong Guo, Yaofo Chen, Yin Zheng, Qi Chen, Peilin Zhao, Jian Chen, Junzhou Huang, Mingkui Tan