Case Relevance
"Case relevance" broadly refers to research investigating how effectively models, particularly in machine learning and natural language processing, identify and utilize relevant information within data for specific tasks. Current research focuses on improving model explainability, addressing data imbalances, and enhancing the performance of various architectures, including transformers, convolutional neural networks, and ensemble methods, across diverse applications like legal text analysis, weather prediction, and direct mail marketing. This work is significant because it directly impacts the reliability, efficiency, and ethical implications of AI systems across numerous fields, driving improvements in model accuracy, interpretability, and resource utilization.
Papers
Metacognitive AI: Framework and the Case for a Neurosymbolic Approach
Hua Wei, Paulo Shakarian, Christian Lebiere, Bruce Draper, Nikhil Krishnaswamy, Sergei Nirenburg
A Systematic Analysis of Large Language Models as Soft Reasoners: The Case of Syllogistic Inferences
Leonardo Bertolazzi, Albert Gatt, Raffaella Bernardi