Global Impact
Research on global impact examines how various factors influence the performance, fairness, and broader consequences of machine learning models and algorithms across diverse applications. Current investigations focus on understanding the effects of data characteristics (e.g., homophily, outliers, imbalanced classes), model architectures (e.g., CNNs, LLMs, GNNs), and training methodologies (e.g., regularization, transfer learning) on model behavior and outcomes. These studies are crucial for improving model robustness, fairness, and efficiency, ultimately leading to more reliable and beneficial applications in fields ranging from healthcare and autonomous systems to open-source software development and environmental monitoring. The ultimate goal is to develop more responsible and effective AI systems that minimize unintended consequences and maximize societal benefit.
Papers
Symmetric Neural-Collapse Representations with Supervised Contrastive Loss: The Impact of ReLU and Batching
Ganesh Ramachandra Kini, Vala Vakilian, Tina Behnia, Jaidev Gill, Christos Thrampoulidis
For Better or Worse: The Impact of Counterfactual Explanations' Directionality on User Behavior in xAI
Ulrike Kuhl, André Artelt, Barbara Hammer
Impact of Experiencing Misrecognition by Teachable Agents on Learning and Rapport
Yuya Asano, Diane Litman, Mingzhi Yu, Nikki Lobczowski, Timothy Nokes-Malach, Adriana Kovashka, Erin Walker
On Minimizing the Impact of Dataset Shifts on Actionable Explanations
Anna P. Meyer, Dan Ley, Suraj Srinivas, Himabindu Lakkaraju
Beyond Rankings: Exploring the Impact of SERP Features on Organic Click-through Rates
Erik Fubel, Niclas Michael Groll, Patrick Gundlach, Qiwei Han, Maximilian Kaiser
The Impact of Positional Encoding on Length Generalization in Transformers
Amirhossein Kazemnejad, Inkit Padhi, Karthikeyan Natesan Ramamurthy, Payel Das, Siva Reddy