Weight Sharing Supernet
Weight-sharing supernets are a powerful technique in neural architecture search (NAS), aiming to efficiently train a single large network encompassing many smaller architectures (subnets). Current research focuses on improving the training stability and accuracy of these supernets, often employing techniques like information-based measurements to select optimal subnets and novel loss functions to address weight sharing issues. This approach significantly accelerates the NAS process, enabling the rapid discovery of efficient and accurate models tailored to various resource constraints (e.g., latency, memory, energy), with applications ranging from mobile devices to large-scale datacenter deployments.
Papers
December 27, 2023
December 19, 2023
August 15, 2023
June 21, 2023
June 8, 2023
March 18, 2023