Mixture Component
Mixture component models are a powerful class of machine learning techniques that combine multiple specialized models (experts) to improve performance and efficiency on complex tasks. Current research focuses on developing novel architectures, such as mixtures of experts (MoE), and applying them to diverse fields including natural language processing, computer vision, and signal processing, often incorporating techniques like low-rank adaptation (LoRA) for parameter efficiency. These advancements are significant because they enable the creation of larger, more capable models while mitigating computational costs and improving generalization across heterogeneous datasets, leading to improved accuracy and efficiency in various applications.
Papers
Leveraging Mixture of Experts for Improved Speech Deepfake Detection
Viola Negroni, Davide Salvi, Alessandro Ilic Mezza, Paolo Bestagini, Stefano Tubaro
Time-MoE: Billion-Scale Time Series Foundation Models with Mixture of Experts
Xiaoming Shi, Shiyu Wang, Yuqi Nie, Dianqi Li, Zhou Ye, Qingsong Wen, Ming Jin
Boosting Code-Switching ASR with Mixture of Experts Enhanced Speech-Conditioned LLM
Fengrun Zhang, Wang Geng, Hukai Huang, Yahui Shan, Cheng Yi, He Qu
Mixture of Prompt Learning for Vision Language Models
Yu Du, Tong Niu, Rong Zhao
Mixture of Experts Fusion for Fake Audio Detection Using Frozen wav2vec 2.0
Zhiyong Wang, Ruibo Fu, Zhengqi Wen, Jianhua Tao, Xiaopeng Wang, Yuankun Xie, Xin Qi, Shuchen Shi, Yi Lu, Yukun Liu, Chenxing Li, Xuefei Liu, Guanjun Li
Mixture of Diverse Size Experts
Manxi Sun, Wei Liu, Jian Luan, Pengzhi Gao, Bin Wang
Duplex: A Device for Large Language Models with Mixture of Experts, Grouped Query Attention, and Continuous Batching
Sungmin Yun, Kwanhee Kyung, Juhwan Cho, Jaewan Choi, Jongmin Kim, Byeongho Kim, Sukhan Lee, Kyomin Sohn, Jung Ho Ahn
Beyond Parameter Count: Implicit Bias in Soft Mixture of Experts
Youngseog Chung, Dhruv Malik, Jeff Schneider, Yuanzhi Li, Aarti Singh
Eagle: Exploring The Design Space for Multimodal LLMs with Mixture of Encoders
Min Shi, Fuxiao Liu, Shihao Wang, Shijia Liao, Subhashree Radhakrishnan, De-An Huang, Hongxu Yin, Karan Sapra, Yaser Yacoob, Humphrey Shi, Bryan Catanzaro, Andrew Tao, Jan Kautz, Zhiding Yu, Guilin Liu
Mamba or Transformer for Time Series Forecasting? Mixture of Universals (MoU) Is All You Need
Sijia Peng, Yun Xiong, Yangyong Zhu, Zhiqiang Shen