Hyper Parameter
Hyperparameters are the settings of a machine learning model that are not learned from data but are set beforehand, significantly impacting model performance and resource consumption. Current research focuses on optimizing hyperparameter selection across various model architectures, including deep neural networks, large language models, and Gaussian processes, often employing techniques like Bayesian optimization, evolutionary algorithms, and novel mathematical frameworks to improve efficiency and generalization. Effective hyperparameter tuning is crucial for achieving optimal model performance, reducing computational costs (including energy consumption), and enhancing the reliability and reproducibility of machine learning results across diverse applications.
Papers
In defense of parameter sharing for model-compression
Aditya Desai, Anshumali Shrivastava
Thin and Deep Gaussian Processes
Daniel Augusto de Souza, Alexander Nikitin, ST John, Magnus Ross, Mauricio A. Álvarez, Marc Peter Deisenroth, João P. P. Gomes, Diego Mesquita, César Lincoln C. Mattos
ASP: Automatic Selection of Proxy dataset for efficient AutoML
Peng Yao, Chao Liao, Jiyuan Jia, Jianchao Tan, Bin Chen, Chengru Song, Di Zhang