Balancing Weight
Balancing weights in neural networks is a crucial area of research aiming to improve model efficiency, accuracy, and robustness. Current efforts focus on optimizing weight distributions to mitigate issues like class imbalance in training data, the dominance of "massive weights" in large language models, and the impact of quantization on model performance. Researchers are exploring various techniques, including novel training algorithms (e.g., incorporating collaborative filtering into LLMs), weight pruning methods, and innovative weight initialization strategies, across diverse architectures such as transformers and convolutional neural networks. These advancements have significant implications for reducing computational costs, enhancing model generalizability, and improving the privacy and security of deep learning systems.
Papers
Correlative Information Maximization: A Biologically Plausible Approach to Supervised Deep Neural Networks without Weight Symmetry
Bariscan Bozkurt, Cengiz Pehlevan, Alper T Erdogan
Rewarded soups: towards Pareto-optimal alignment by interpolating weights fine-tuned on diverse rewards
Alexandre Ramé, Guillaume Couairon, Mustafa Shukor, Corentin Dancette, Jean-Baptiste Gaya, Laure Soulier, Matthieu Cord