Recent Large Language Model
Recent research on large language models (LLMs) centers on improving their capabilities in handling long contexts, multilingual support, and complex reasoning tasks, while also addressing limitations in efficiency, bias, and uncertainty quantification. Current efforts focus on novel architectures like Mamba, enhanced Mixture of Experts models, and improved training methods such as self-contrast learning and fine-grained reward systems. These advancements are crucial for expanding the practical applications of LLMs across diverse fields, from biomedical research and public health interventions to improving the reliability of AI-assisted tools and mitigating the risks associated with misinformation.
Papers
Are Large-Language Models Graph Algorithmic Reasoners?
Alexander K Taylor, Anthony Cuturrufo, Vishal Yathish, Mingyu Derek Ma, Wei Wang
MARCO: Multi-Agent Real-time Chat Orchestration
Anubhav Shrimal, Stanley Kanagaraj, Kriti Biswas, Swarnalatha Raghuraman, Anish Nediyanchath, Yi Zhang, Promod Yenigalla
Uncertainty-Guided Optimization on Large Language Model Search Trees
Julia Grosse, Ruotian Wu, Ahmad Rashid, Philipp Hennig, Pascal Poupart, Agustinus Kristiadi
HYBRINFOX at CheckThat! 2024 -- Task 1: Enhancing Language Models with Structured Information for Check-Worthiness Estimation
Géraud Faye, Morgane Casanova, Benjamin Icard, Julien Chanson, Guillaume Gadek, Guillaume Gravier, Paul Égré