Safe Deep

"Safe Deep" research focuses on developing and deploying deep learning models while mitigating risks and ensuring safety, reliability, and trustworthiness. Current efforts concentrate on improving model robustness through techniques like Lyapunov function-based reinforcement learning and adversarial training for imitation learning, as well as optimizing resource usage and latency in edge computing environments using deep reinforcement learning frameworks. This work is crucial for responsible AI development, enabling the safe application of deep learning in high-stakes domains such as autonomous systems, finance, and healthcare, while addressing concerns about bias, instability, and inappropriate content generation.

Papers