Context Bigram
Context bigrams, pairs of consecutive words in text, are crucial for understanding how language models process and generate text. Current research focuses on leveraging bigram statistics to improve model performance, particularly through methods like contrastive learning and the development of "statistical induction heads" within transformer architectures. This work aims to enhance the quality of training data, improve the accuracy of semantic textual relatedness tasks across multiple languages, and ultimately lead to more reliable and efficient large language models.
Papers
October 30, 2024
October 23, 2024
September 9, 2024
April 6, 2024
February 16, 2024
January 19, 2024