Hidden Risk
Hidden risks in various applications of artificial intelligence, particularly large language models (LLMs), are a growing concern. Current research focuses on identifying and mitigating these risks, including privacy vulnerabilities stemming from fine-tuning on generated data, the potential for adversarial attacks to subvert safety mechanisms, and biases introduced by training data or evaluation methods. These investigations utilize diverse techniques such as reinforcement learning for adversarial attack generation, deep learning for risk prediction in medical contexts, and knowledge graph construction for analyzing complex systems, highlighting the need for robust and reliable AI systems across numerous domains.
Papers
October 10, 2024
September 12, 2024
August 5, 2024
June 24, 2024
June 11, 2024
April 1, 2024
November 22, 2023
May 15, 2023
January 23, 2023
July 21, 2022
April 30, 2022