Whispering Llama
Research on "Whispering Llama," a metaphorical term encompassing various advancements in Large Language Models (LLMs) based on the Llama architecture, focuses on improving their capabilities and addressing limitations. Current efforts concentrate on enhancing reasoning abilities through techniques like Monte Carlo Tree Search and pairwise reward models, improving linguistic plausibility via phoneme-based tokenization, and mitigating safety risks by developing robust safeguards against adversarial attacks and biases. These advancements are significant because they contribute to creating more efficient, reliable, and ethically sound LLMs with broader applications in diverse fields, from mathematical problem-solving to improved speech recognition.
Papers
Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
Mengzhou Xia, Tianyu Gao, Zhiyuan Zeng, Danqi Chen
Whispering LLaMA: A Cross-Modal Generative Error Correction Framework for Speech Recognition
Srijith Radhakrishnan, Chao-Han Huck Yang, Sumeer Ahmad Khan, Rohit Kumar, Narsis A. Kiani, David Gomez-Cabrero, Jesper N. Tegner