Encoder Decoder Model
Encoder-decoder models are a class of neural networks designed for sequence-to-sequence tasks, aiming to map an input sequence (e.g., an image, audio, or text) to an output sequence (e.g., a caption, translation, or code). Current research emphasizes improving efficiency and robustness, exploring architectures like Transformers and LSTMs, and incorporating techniques such as contrastive learning, adversarial training, and direct preference optimization to enhance performance across diverse applications. These models are proving highly impactful, enabling advancements in various fields including machine translation, speech recognition, image captioning, and even biological sequence analysis.
Papers
January 3, 2025
December 8, 2024
October 26, 2024
October 22, 2024
October 10, 2024
September 23, 2024
September 4, 2024
August 30, 2024
August 7, 2024
July 19, 2024
July 6, 2024
June 15, 2024
June 4, 2024
May 30, 2024
May 2, 2024
May 1, 2024
April 30, 2024
March 19, 2024
February 29, 2024