Split Computing
Split computing addresses the challenges of deploying large deep neural networks (DNNs) on resource-constrained devices by partitioning the model between a local device and a remote server. Current research focuses on optimizing model architectures and training strategies for efficient splitting, including techniques like predefined sparsity, multi-task learning, and the use of bottlenecks or slimmable encoders to minimize communication overhead and computational load. This approach is significant for improving the performance and privacy of AI applications in various domains, such as edge computing, IoT devices, and mobile platforms, by enabling the deployment of powerful DNNs while mitigating latency, bandwidth, and security concerns.
Papers
October 31, 2024
September 3, 2024
July 16, 2024
July 8, 2024
April 2, 2024
October 23, 2023
October 12, 2023
September 17, 2023
June 22, 2023
April 19, 2023
September 23, 2022
August 30, 2022
August 24, 2022
May 23, 2022
March 16, 2022
January 7, 2022