Case Relevance
"Case relevance" broadly refers to research investigating how effectively models, particularly in machine learning and natural language processing, identify and utilize relevant information within data for specific tasks. Current research focuses on improving model explainability, addressing data imbalances, and enhancing the performance of various architectures, including transformers, convolutional neural networks, and ensemble methods, across diverse applications like legal text analysis, weather prediction, and direct mail marketing. This work is significant because it directly impacts the reliability, efficiency, and ethical implications of AI systems across numerous fields, driving improvements in model accuracy, interpretability, and resource utilization.
Papers
Are Quantum Computers Practical Yet? A Case for Feature Selection in Recommender Systems using Tensor Networks
Artyom Nikitin, Andrei Chertkov, Rafael Ballester-Ripoll, Ivan Oseledets, Evgeny Frolov
Counterfactually Augmented Data and Unintended Bias: The Case of Sexism and Hate Speech Detection
Indira Sen, Mattia Samory, Claudia Wagner, Isabelle Augenstein
Contextualize Me -- The Case for Context in Reinforcement Learning
Carolin Benjamins, Theresa Eimer, Frederik Schubert, Aditya Mohan, Sebastian Döhler, André Biedenkapp, Bodo Rosenhahn, Frank Hutter, Marius Lindauer
Generalized Strategic Classification and the Case of Aligned Incentives
Sagi Levanon, Nir Rosenfeld
Information Extraction through AI techniques: The KIDs use case at CONSOB
Domenico Lembo, Alessandra Limosani, Francesca Medda, Alessandra Monaco, Federico Maria Scafoglieri
An Indirect Rate-Distortion Characterization for Semantic Sources: General Model and the Case of Gaussian Observation
Jiakun Liu, Shuo Shao, Wenyi Zhang, H. Vincent Poor