Hyper Tune
Hyper-tuning, the optimization of model parameters and hyperparameters, is crucial for maximizing the performance of machine learning models across diverse applications. Current research focuses on developing efficient tuning strategies, including Bayesian optimization, importance sampling, and novel algorithms like pairwise sample optimization, particularly for large-scale models and resource-constrained environments. These advancements aim to reduce computational costs and human effort associated with tuning, improving the efficiency and scalability of machine learning workflows in various fields, from recommender systems to image generation and natural language processing. The ultimate goal is to achieve optimal model performance with minimal computational overhead and human intervention.
Papers
MetaMorph: Multimodal Understanding and Generation via Instruction Tuning
Shengbang Tong, David Fan, Jiachen Zhu, Yunyang Xiong, Xinlei Chen, Koustuv Sinha, Michael Rabbat, Yann LeCun, Saining Xie, Zhuang Liu
T$^3$-S2S: Training-free Triplet Tuning for Sketch to Scene Generation
Zhenhong Sun, Yifu Wang, Yonhon Ng, Yunfei Duan, Daoyi Dong, Hongdong Li, Pan Ji