Embracing CompAct
"Embracing CompAct" broadly refers to the ongoing research effort to develop compact and efficient models across various machine learning domains. Current research focuses on creating smaller, faster models while maintaining or improving performance, employing techniques like structured pruning, knowledge distillation, and novel architectures such as lightweight YOLO variants and compact graph neural networks. This pursuit is driven by the need to reduce computational costs, memory requirements, and energy consumption, enabling deployment on resource-constrained devices and improving the scalability of machine learning applications in fields ranging from robotics and medical image analysis to natural language processing and recommendation systems.
Papers
ViralVectors: Compact and Scalable Alignment-free Virome Feature Generation
Sarwan Ali, Prakash Chourasia, Zahra Tayebi, Babatunde Bello, Murray Patterson
Design and Control of a Ballbot Drivetrain with High Agility, Minimal Footprint, and High Payload
Chenzhang Xiao, Mahshid Mansouri, David Lam, Joao Ramos, Elizabeth T. Hsiao-Wecksler