Full Model
"Full Model" research encompasses the development and improvement of large-scale machine learning models across diverse applications, aiming to enhance performance, efficiency, and robustness. Current research focuses on addressing model vulnerabilities (e.g., adversarial attacks, hallucinations), improving efficiency for resource-constrained devices, and developing specialized models for specific domains (e.g., finance, astronomy, medical imaging). This work is significant for advancing AI capabilities in various fields and for mitigating potential risks associated with deploying complex models in real-world settings.
Papers
Continual LLaVA: Continual Instruction Tuning in Large Vision-Language Models
Meng Cao, Yuyang Liu, Yingfei Liu, Tiancai Wang, Jiahua Dong, Henghui Ding, Xiangyu Zhang, Ian Reid, Xiaodan Liang
Discrete the solving model of time-variant standard Sylvester-conjugate matrix equations using Euler-forward formula
Jiakuang He, Dongqing Wu
Evaluating Creative Short Story Generation in Humans and Large Language Models
Mete Ismayilzada, Claire Stevenson, Lonneke van der Plas
An Exploration of Higher Education Course Evaluation by Large Language Models
Bo Yuan, Jiazi Hu
Know Where You're Uncertain When Planning with Multimodal Foundation Models: A Formal Framework
Neel P. Bhatt, Yunhao Yang, Rohan Siva, Daniel Milan, Ufuk Topcu, Zhangyang Wang
Adaptive Conformal Inference by Particle Filtering under Hidden Markov Models
Xiaoyi Su, Zhixin Zhou, Rui Luo
Enhancing Diabetic Retinopathy Detection with CNN-Based Models: A Comparative Study of UNET and Stacked UNET Architectures
Ameya Uppina, S Navaneetha Krishnan, Talluri Krishna Sai Teja, Nikhil N Iyer, Joe Dhanith P R
Swan and ArabicMTEB: Dialect-Aware, Arabic-Centric, Cross-Lingual, and Cross-Cultural Embedding Models and Benchmarks
Gagan Bhatia, El Moatez Billah Nagoudi, Abdellah El Mekki, Fakhraddin Alwajih, Muhammad Abdul-Mageed
Hollowed Net for On-Device Personalization of Text-to-Image Diffusion Models
Wonguk Cho, Seokeon Choi, Debasmit Das, Matthias Reisser, Taesup Kim, Sungrack Yun, Fatih Porikli
CmdCaliper: A Semantic-Aware Command-Line Embedding Model and Dataset for Security Research
Sian-Yao Huang, Cheng-Lin Yang, Che-Yu Lin, Chun-Ying Huang
Identify Backdoored Model in Federated Learning via Individual Unlearning
Jiahao Xu, Zikai Zhang, Rui Hu
LLM-KT: A Versatile Framework for Knowledge Transfer from Large Language Models to Collaborative Filtering
Nikita Severin, Aleksei Ziablitsev, Yulia Savelyeva, Valeriy Tashchilin, Ivan Bulychev, Mikhail Yushkov, Artem Kushneruk, Amaliya Zaryvnykh, Dmitrii Kiselev, Andrey Savchenko, Ilya Makarov
Evaluating the Evolution of YOLO (You Only Look Once) Models: A Comprehensive Benchmark Study of YOLO11 and Its Predecessors
Nidhal Jegham, Chan Young Koh, Marwan Abdelatti, Abdeltawab Hendawi
A Geometric Framework for Understanding Memorization in Generative Models
Brendan Leigh Ross, Hamidreza Kamkari, Tongzi Wu, Rasa Hosseinzadeh, Zhaoyan Liu, George Stein, Jesse C. Cresswell, Gabriel Loaiza-Ganem
Hamiltonian Monte Carlo Inference of Marginalized Linear Mixed-Effects Models
Jinlin Lai, Justin Domke, Daniel Sheldon
Approximate attention with MLP: a pruning strategy for attention-based model in multivariate time series forecasting
Suhan Guo, Jiahong Deng, Yi Wei, Hui Dou, Furao Shen, Jian Zhao
Metamorphic Malware Evolution: The Potential and Peril of Large Language Models
Pooria Madani
Transferable Ensemble Black-box Jailbreak Attacks on Large Language Models
Yiqi Yang, Hongye Fu