3260 papers • 126 benchmarks • 313 datasets
The goal of Time Series Prediction is to infer the future values of a time series from the past. Source: Orthogonal Echo State Networks and stochastic evaluations of likelihoods
(Image credit: Papersgraph)
These leaderboards are used to track progress in time-series-prediction-22
Use these libraries to find time-series-prediction-22 models and implementations
No subtasks available.
Diffusion Convolutional Recurrent Neural Network (DCRNN), a deep learning framework for traffic forecasting that incorporates both spatial and temporal dependency in the traffic flow and evaluates the framework on two real-world large scale road network traffic datasets and observes consistent improvement.
A dual-stage attention-based recurrent neural network (DA-RNN) to address the long-term temporal dependencies of the Nonlinear autoregressive exogenous model and can outperform state-of-the-art methods for time series prediction.
Novel deep learning models are developed based on Gated Recurrent Unit, a state-of-the-art recurrent neural network that takes two representations of missing patterns, i.e., masking and time interval, and effectively incorporates them into a deep model architecture so that it not only captures the long-term temporal dependencies in time series, but also utilizes the missing patterns to achieve better prediction results.
Gluon Time Series is introduced, a library for deep-learning-based time series modeling that provides all necessary components and tools that scientists need for quickly building new models, for efficiently running and analyzing experiments and for evaluating model accuracy.
A novel end-to-end Bayesian deep model is proposed that provides time series prediction along with uncertainty estimation at Uber and is successfully applied to large-scale time series anomaly detection at Uber.
This paper investigates Long Short-Term Memory neural networks as an approach to build consistently accurate models for a wide range of predictive process monitoring tasks and shows that LSTMs outperform existing techniques to predict the next event of a running case and its timestamp.
A new class of time-continuous recurrent neural network models that exhibit stable and bounded behavior, yield superior expressivity within the family of neural ordinary differential equations, and give rise to improved performance on time-series prediction tasks are introduced.
This paper presents evolutionary state graph, a dynamic graph structure designed to systematically represent the evolving relations (edges) among states (nodes) along time, and proposes a novel graph neural network model, Evolutionary State Graph Network (EvoNet), to encode the evolutionary stategraph for accurate and interpretable time-series event prediction.
The goal of this survey is to provide a selfcontained explication of the state of the art of recurrent neural networks together with a historical perspective and references to primary research.
Adding a benchmark result helps the community track progress.