Edge Inference
Edge inference focuses on performing machine learning inference directly on resource-constrained devices at the network edge, aiming to reduce latency, bandwidth consumption, and privacy concerns associated with cloud-based processing. Current research emphasizes efficient model architectures (like Vision Transformers and MobileNets), optimization techniques (including quantization, pruning, and model merging), and intelligent task offloading strategies to balance accuracy and resource usage. This field is crucial for enabling real-time AI applications in diverse areas such as video analytics, natural language processing, and robotics, driving advancements in both hardware and software for efficient AI deployment.
Papers
June 21, 2023
June 1, 2023
May 18, 2023
May 1, 2023
March 25, 2023
December 28, 2022
December 21, 2022
December 8, 2022
November 25, 2022
November 15, 2022
September 21, 2022
August 24, 2022
August 22, 2022
August 4, 2022
June 15, 2022
March 3, 2022
January 19, 2022