Long Span
"Long span" research addresses the limitations of current models in processing and generating lengthy sequences of data, whether text, audio, or video. Current efforts focus on improving large language models (LLMs) and other deep learning architectures like transformers (including Longformer and variations) and LSTMs to handle longer contexts effectively, often employing techniques like coreference resolution, hierarchical attention, and efficient attention mechanisms. This research is crucial for advancing natural language processing, improving video and audio analysis, and enabling more sophisticated applications in diverse fields such as medical diagnosis, legal document processing, and personalized search.
Papers
June 3, 2024
May 31, 2024
February 7, 2024
January 13, 2024
December 8, 2023
November 21, 2023
November 9, 2023
September 16, 2023
September 13, 2023
August 23, 2023
August 7, 2023
August 3, 2023
July 25, 2023
July 20, 2023
July 18, 2023
May 5, 2023
May 4, 2023
May 2, 2023
April 26, 2023
April 20, 2023