Tensor Train
Tensor Trains (TTs) are a low-rank tensor decomposition method aiming to efficiently represent and manipulate high-dimensional data by exploiting underlying structure. Current research focuses on applying TTs to improve the efficiency and scalability of large language models, neural networks, and other machine learning tasks, often incorporating them into novel architectures like Block Tensor-Train Mixture-of-Experts (BTT-MoE) or using them for parameter-efficient fine-tuning (e.g., TT-LoRA). This approach offers significant potential for reducing computational costs and memory requirements in various applications, from anomaly detection and image processing to solving high-dimensional partial differential equations and accelerating large-scale machine learning training.
Papers
TT-SDF2PC: Registration of Point Cloud and Compressed SDF Directly in the Memory-Efficient Tensor Train Domain
Alexey I. Boyko, Anastasiia Kornilova, Rahim Tariverdizadeh, Mirfarid Musavian, Larisa Markeeva, Ivan Oseledets, Gonzalo Ferrer
Generative Modeling via Hierarchical Tensor Sketching
Yifan Peng, Yian Chen, E. Miles Stoudenmire, Yuehaw Khoo