Time Matter
"Time Matter" encompasses research efforts to effectively incorporate temporal dynamics into various machine learning tasks. Current research focuses on developing novel model architectures, such as recurrent neural networks and transformers adapted for time series analysis, and employing techniques like time-distributed convolutions and Hamiltonian learning to improve temporal modeling. This work is significant because accurately representing and reasoning about time is crucial for improving the performance and reliability of AI systems across diverse applications, from forecasting and risk estimation to medical diagnosis and personalized treatment.
Papers
Winning the Lottery Ahead of Time: Efficient Early Network Pruning
John Rachwan, Daniel Zügner, Bertrand Charpentier, Simon Geisler, Morgane Ayle, Stephan Günnemann
Route to Time and Time to Route: Travel Time Estimation from Sparse Trajectories
Zhiwen Zhang, Hongjun Wang, Zipei Fan, Jiyuan Chen, Xuan Song, Ryosuke Shibasaki