Machine Learning Inference

Machine learning inference focuses on efficiently and reliably using trained models to make predictions or solve tasks with new data. Current research emphasizes improving the trustworthiness and efficiency of inference, exploring diverse model architectures like probabilistic circuits, predictive coding networks, and ordinary differential equations, as well as optimizing inference for resource-constrained environments (e.g., edge devices, microcontrollers). These advancements are crucial for enhancing the reliability and applicability of machine learning across various domains, from medical diagnosis and industrial control to astrophysics and autonomous systems, while addressing concerns about energy efficiency, privacy, and explainability.

Papers