Traditional Deep Learning
Traditional deep learning focuses on developing and improving artificial neural networks for various tasks, aiming to enhance accuracy, efficiency, and robustness. Current research emphasizes addressing limitations such as vulnerability to adversarial attacks, high computational costs, and the need for large labeled datasets, exploring solutions like equivariant convolutional networks, physics-informed neural networks, and model compression techniques (e.g., quantization, pruning). These advancements are crucial for deploying deep learning models in resource-constrained environments and improving their reliability and trustworthiness across diverse applications, from image recognition to natural language processing.
Papers
A Survey: Collaborative Hardware and Software Design in the Era of Large Language Models
Cong Guo, Feng Cheng, Zhixu Du, James Kiessling, Jonathan Ku, Shiyu Li, Ziru Li, Mingyuan Ma, Tergel Molom-Ochir, Benjamin Morris, Haoxuan Shan, Jingwei Sun, Yitu Wang, Chiyue Wei, Xueying Wu, Yuhao Wu, Hao Frank Yang, Jingyang Zhang, Junyao Zhang, Qilin Zheng, Guanglei Zhou, Hai (Helen)Li, Yiran Chen
Adaptation Odyssey in LLMs: Why Does Additional Pretraining Sometimes Fail to Improve?
Fırat Öncel, Matthias Bethge, Beyza Ermis, Mirco Ravanelli, Cem Subakan, Çağatay Yıldız