Full Model
"Full Model" research encompasses the development and improvement of large-scale machine learning models across diverse applications, aiming to enhance performance, efficiency, and robustness. Current research focuses on addressing model vulnerabilities (e.g., adversarial attacks, hallucinations), improving efficiency for resource-constrained devices, and developing specialized models for specific domains (e.g., finance, astronomy, medical imaging). This work is significant for advancing AI capabilities in various fields and for mitigating potential risks associated with deploying complex models in real-world settings.
Papers
HyperSeg: Towards Universal Visual Segmentation with Large Language Model
Cong Wei, Yujie Zhong, Haoxian Tan, Yong Liu, Zheng Zhao, Jie Hu, Yujiu Yang
What's in the Image? A Deep-Dive into the Vision of Vision Language Models
Omri Kaduri, Shai Bagon, Tali Dekel
DiffSLT: Enhancing Diversity in Sign Language Translation via Diffusion Model
JiHwan Moon, Jihoon Park, Jungeun Kim, Jongseong Bae, Hyeongwoo Jeon, Ha Young Kim
An In-depth Investigation of Sparse Rate Reduction in Transformer-like Models
Yunzhe Hu, Difan Zou, Dong Xu
Safe to Serve: Aligning Instruction-Tuned Models for Safety and Helpfulness
Avinash Amballa, Durga Sandeep Saluru, Gayathri Akkinapalli, Abhishek Sureddy, Akshay Kumar Sureddy
On-Road Object Importance Estimation: A New Dataset and A Model with Multi-Fold Top-Down Guidance
Zhixiong Nan, Yilong Chen, Tianfei Zhou, Tao Xiang
ThreatModeling-LLM: Automating Threat Modeling using Large Language Models for Banking System
Shuiqiao Yang, Tingmin Wu, Shigang Liu, David Nguyen, Seung Jang, Alsharif Abuadbba
Human-Calibrated Automated Testing and Validation of Generative Language Models
Agus Sudjianto, Aijun Zhang, Srinivas Neppalli, Tarun Joshi, Michal Malohlava
Towards Efficient Model-Heterogeneity Federated Learning for Large Models
Ruofan Jia, Weiying Xie, Jie Lei, Haonan Qin, Jitao Ma, Leyuan Fang
BlendServe: Optimizing Offline Inference for Auto-regressive Large Models with Resource-aware Batching
Yilong Zhao, Shuo Yang, Kan Zhu, Lianmin Zheng, Baris Kasikci, Yang Zhou, Jiarong Xing, Ion Stoica
A Training-Free Approach for Music Style Transfer with Latent Diffusion Models
Sooyoung Kim, Joonwoo Kwon, Heehwan Wang, Shinjae Yoo, Yuewei Lin, Jiook Cha
Fusion Matters: Learning Fusion in Deep Click-through Rate Prediction Models
Kexin Zhang, Fuyuan Lyu, Xing Tang, Dugang Liu, Chen Ma, Kaize Ding, Xiuqiang He, Xue Liu
Enabling Efficient Serverless Inference Serving for LLM (Large Language Model) in the Cloud
Himel Ghosh
Machine-agnostic Automated Lumbar MRI Segmentation using a Cascaded Model Based on Generative Neurons
Promit Basak, Rusab Sarmun, Saidul Kabir, Israa Al-Hashimi, Enamul Hoque Bhuiyan, Anwarul Hasan, Muhammad Salman Khan, Muhammad E. H. Chowdhury
MolMetaLM: a Physicochemical Knowledge-Guided Molecular Meta Language Model
Yifan Wu, Min Zeng, Yang Li, Yang Zhang, Min Li
Exploiting Watermark-Based Defense Mechanisms in Text-to-Image Diffusion Models for Unauthorized Data Usage
Soumil Datta, Shih-Chieh Dai, Leo Yu, Guanhong Tao
DyCoke: Dynamic Compression of Tokens for Fast Video Large Language Models
Keda Tao, Can Qin, Haoxuan You, Yang Sui, Huan Wang
Large Multi-modal Models Can Interpret Features in Large Multi-modal Models
Kaichen Zhang, Yifei Shen, Bo Li, Ziwei Liu