Trust Penalization
Trust penalization is a technique used to improve the reliability and efficiency of various machine learning models by discouraging undesirable behaviors or outputs. Current research focuses on applying this approach to diverse areas, including large language models (where it mitigates hallucinations by adjusting attention weights during decoding), federated learning (where it enhances the trustworthiness of participating nodes), and traffic assignment (where it optimizes route selection by penalizing congested edges). These advancements demonstrate the broad applicability of trust penalization in improving model accuracy, robustness, and fairness across different domains, leading to more reliable and efficient systems.
Papers
July 21, 2024
October 30, 2023
June 23, 2023
November 29, 2022
November 15, 2022