Foundation Model
Foundation models are large, pre-trained AI models designed to generalize across diverse tasks and datasets, offering a powerful alternative to task-specific models. Current research emphasizes adapting these models to various domains, including healthcare (e.g., medical image analysis, EEG interpretation), scientific applications (e.g., genomics, weather forecasting), and robotics, often employing architectures like transformers and mixtures of experts with innovative gating functions. This approach promises to improve efficiency and accuracy in numerous fields by leveraging the knowledge embedded within these powerful models, streamlining data analysis and enabling new applications previously hindered by data scarcity or computational limitations.
Papers
SoTaNa: The Open-Source Software Development Assistant
Ensheng Shi, Fengji Zhang, Yanlin Wang, Bei Chen, Lun Du, Hongyu Zhang, Shi Han, Dongmei Zhang, Hongbin Sun
Fine-tuning can cripple your foundation model; preserving features may be the solution
Jishnu Mukhoti, Yarin Gal, Philip H. S. Torr, Puneet K. Dokania
Foundation Model is Efficient Multimodal Multitask Model Selector
Fanqing Meng, Wenqi Shao, Zhanglin Peng, Chonghe Jiang, Kaipeng Zhang, Yu Qiao, Ping Luo
Decentralised Governance-Driven Architecture for Designing Foundation Model based Systems: Exploring the Role of Blockchain in Responsible AI
Yue Liu, Qinghua Lu, Liming Zhu, Hye-Young Paik