Encoder Decoder Model
Encoder-decoder models are a class of neural networks designed for sequence-to-sequence tasks, aiming to map an input sequence (e.g., an image, audio, or text) to an output sequence (e.g., a caption, translation, or code). Current research emphasizes improving efficiency and robustness, exploring architectures like Transformers and LSTMs, and incorporating techniques such as contrastive learning, adversarial training, and direct preference optimization to enhance performance across diverse applications. These models are proving highly impactful, enabling advancements in various fields including machine translation, speech recognition, image captioning, and even biological sequence analysis.
Papers
February 3, 2024
January 15, 2024
January 10, 2024
December 12, 2023
November 29, 2023
November 20, 2023
November 16, 2023
November 9, 2023
November 4, 2023
October 16, 2023
October 15, 2023
October 5, 2023
October 2, 2023
September 12, 2023
July 13, 2023
June 24, 2023
June 22, 2023
June 10, 2023