Embracing CompAct
"Embracing CompAct" broadly refers to the ongoing research effort to develop compact and efficient models across various machine learning domains. Current research focuses on creating smaller, faster models while maintaining or improving performance, employing techniques like structured pruning, knowledge distillation, and novel architectures such as lightweight YOLO variants and compact graph neural networks. This pursuit is driven by the need to reduce computational costs, memory requirements, and energy consumption, enabling deployment on resource-constrained devices and improving the scalability of machine learning applications in fields ranging from robotics and medical image analysis to natural language processing and recommendation systems.
Papers
Llama Guard 3-1B-INT4: Compact and Efficient Safeguard for Human-AI Conversations
Igor Fedorov, Kate Plawiak, Lemeng Wu, Tarek Elgamal, Naveen Suda, Eric Smith, Hongyuan Zhan, Jianfeng Chi, Yuriy Hulovatyy, Kimish Patel, Zechun Liu, Changsheng Zhao, Yangyang Shi, Tijmen Blankevoort, Mahesh Pasupuleti, Bilge Soran, Zacharie Delpierre Coudert, Rachad Alao, Raghuraman Krishnamoorthi, Vikas Chandra
LeC$^2$O-NeRF: Learning Continuous and Compact Large-Scale Occupancy for Urban Scenes
Zhenxing Mi, Dan Xu