Semantic Unit
Semantic units represent the smallest meaningful components of language, encompassing words, phrases, or even sub-word units depending on the context and application. Current research focuses on identifying and leveraging these units within various language processing tasks, employing techniques like self-supervised learning to create coarse semantic representations for speech and rank-wise clustering to merge and manipulate parameters in large language models. This work aims to improve model efficiency, address limitations like the "reversal curse," and enhance performance in areas such as machine translation and speech recognition, ultimately leading to more robust and nuanced natural language understanding.
Papers
October 28, 2024
October 5, 2024
September 24, 2024
March 1, 2024
October 17, 2023
September 21, 2023
September 4, 2023
February 27, 2023
January 4, 2023
August 16, 2022