Long Span
"Long span" research addresses the limitations of current models in processing and generating lengthy sequences of data, whether text, audio, or video. Current efforts focus on improving large language models (LLMs) and other deep learning architectures like transformers (including Longformer and variations) and LSTMs to handle longer contexts effectively, often employing techniques like coreference resolution, hierarchical attention, and efficient attention mechanisms. This research is crucial for advancing natural language processing, improving video and audio analysis, and enabling more sophisticated applications in diverse fields such as medical diagnosis, legal document processing, and personalized search.
Papers
November 14, 2024
November 8, 2024
November 6, 2024
November 2, 2024
October 30, 2024
October 21, 2024
October 10, 2024
October 2, 2024
September 23, 2024
September 3, 2024
August 13, 2024
July 18, 2024
June 20, 2024
June 3, 2024
May 31, 2024
February 7, 2024
January 13, 2024
December 8, 2023
November 21, 2023