Edge Inference
Edge inference focuses on performing machine learning inference directly on resource-constrained devices at the network edge, aiming to reduce latency, bandwidth consumption, and privacy concerns associated with cloud-based processing. Current research emphasizes efficient model architectures (like Vision Transformers and MobileNets), optimization techniques (including quantization, pruning, and model merging), and intelligent task offloading strategies to balance accuracy and resource usage. This field is crucial for enabling real-time AI applications in diverse areas such as video analytics, natural language processing, and robotics, driving advancements in both hardware and software for efficient AI deployment.
Papers
December 21, 2022
December 8, 2022
November 25, 2022
November 15, 2022
September 21, 2022
August 24, 2022
August 22, 2022
August 4, 2022
June 15, 2022
March 3, 2022
January 19, 2022