Weak Learner
Weak learners are base classifiers that achieve only slightly better-than-random performance, but are crucial building blocks for powerful ensemble methods. Current research focuses on optimizing the parallel efficiency of boosting algorithms that combine weak learners, improving the integration of diverse weak learners (e.g., using confidence tensors), and exploring their application in various contexts, including dimensionality reduction, malware detection, and explainable AI. This research is significant because it addresses fundamental limitations in machine learning scalability and efficiency, leading to improved model performance and resource utilization across diverse applications.
Papers
December 19, 2022
October 10, 2022
September 24, 2022
September 4, 2022
June 24, 2022
June 3, 2022
May 27, 2022