Darts is a Python library for user-friendly forecasting and anomaly detection
on time series. It contains a variety of models, from classics such as ARIMA to
deep neural networks. The forecasting models can all be used in the same way,
using fit()
and predict()
functions, similar to scikit-learn.
The library also makes it easy to backtest models,
combine the predictions of several models, and take external data into account.
Darts supports both univariate and multivariate time series and models.
The ML-based models can be trained on potentially large datasets containing multiple time
series, and some of the models offer a rich support for probabilistic forecasting.
Darts also offers extensive anomaly detection capabilities. For instance, it is trivial to apply PyOD models on time series to obtain anomaly scores, or to wrap any of Darts forecasting or filtering models to obtain fully fledged anomaly detection models.
- Training Models on Multiple Time Series
- Using Past and Future Covariates
- Temporal Convolutional Networks and Forecasting
- Probabilistic Forecasting
- Transfer Learning for Time Series Forecasting
- Hierarchical Forecast Reconciliation
We recommend to first setup a clean Python environment for your project with Python 3.8+ using your favorite tool (conda, venv, virtualenv with or without virtualenvwrapper).
Once your environment is set up you can install darts using pip:
pip install darts
For more details you can refer to our installation instructions.
Create a TimeSeries
object from a Pandas DataFrame, and split it in train/validation series:
import pandas as pd
from darts import TimeSeries
# Read a pandas DataFrame
df = pd.read_csv("AirPassengers.csv", delimiter=",")
# Create a TimeSeries, specifying the time and value columns
series = TimeSeries.from_dataframe(df, "Month", "#Passengers")
# Set aside the last 36 months as a validation series
train, val = series[:-36], series[-36:]
Fit an exponential smoothing model, and make a (probabilistic) prediction over the validation series' duration:
from darts.models import ExponentialSmoothing
model = ExponentialSmoothing()
model.fit(train)
prediction = model.predict(len(val), num_samples=1000)
Plot the median, 5th and 95th percentiles:
import matplotlib.pyplot as plt
series.plot()
prediction.plot(label="forecast", low_quantile=0.05, high_quantile=0.95)
plt.legend()
Load a multivariate series, trim it, keep 2 components, split train and validation sets:
from darts.datasets import ETTh2Dataset
series = ETTh2Dataset().load()[:10000][["MUFL", "LULL"]]
train, val = series.split_before(0.6)
Build a k-means anomaly scorer, train it on the train set and use it on the validation set to get anomaly scores:
from darts.ad import KMeansScorer
scorer = KMeansScorer(k=2, window=5)
scorer.fit(train)
anom_score = scorer.score(val)
Build a binary anomaly detector and train it over train scores, then use it over validation scores to get binary anomaly classification:
from darts.ad import QuantileDetector
detector = QuantileDetector(high_quantile=0.99)
detector.fit(scorer.score(train))
binary_anom = detector.detect(anom_score)
Plot (shifting and scaling some of the series to make everything appear on the same figure):
import matplotlib.pyplot as plt
series.plot()
(anom_score / 2. - 100).plot(label="computed anomaly score", c="orangered", lw=3)
(binary_anom * 45 - 150).plot(label="detected binary anomaly", lw=4)
-
Forecasting Models: A large collection of forecasting models; from statistical models (such as ARIMA) to deep learning models (such as N-BEATS). See table of models below.
-
Anomaly Detection The
darts.ad
module contains a collection of anomaly scorers, detectors and aggregators, which can all be combined to detect anomalies in time series. It is easy to wrap any of Darts forecasting or filtering models to build a fully fledged anomaly detection model that compares predictions with actuals. ThePyODScorer
makes it trivial to use PyOD detectors on time series. -
Multivariate Support:
TimeSeries
can be multivariate - i.e., contain multiple time-varying dimensions/columns instead of a single scalar value. Many models can consume and produce multivariate series. -
Multiple series training (global models): All machine learning based models (incl. all neural networks) support being trained on multiple (potentially multivariate) series. This can scale to large datasets too.
-
Probabilistic Support:
TimeSeries
objects can (optionally) represent stochastic time series; this can for instance be used to get confidence intervals, and many models support different flavours of probabilistic forecasting (such as estimating parametric distributions or quantiles). Some anomaly detection scorers are also able to exploit these predictive distributions. -
Past and Future Covariates support: Many models in Darts support past-observed and/or future-known covariate (external data) time series as inputs for producing forecasts.
-
Static Covariates support: In addition to time-dependent data,
TimeSeries
can also contain static data for each dimension, which can be exploited by some models. -
Hierarchical Reconciliation: Darts offers transformers to perform reconciliation. These can make the forecasts add up in a way that respects the underlying hierarchy.
-
Regression Models: It is possible to plug-in any scikit-learn compatible model to obtain forecasts as functions of lagged values of the target series and covariates.
-
Training with sample weights: All global models support being trained with sample weights. They can be applied to each observation, forecasted time step and target column.
-
Forecast Start Shifting: All global models support training and prediction on a shifted output window. This is useful for example for Day-Ahead Market forecasts, or when the covariates (or target series) are reported with a delay.
-
Explainability: Darts has the ability to explain some forecasting models using Shap values.
-
Data processing: Tools to easily apply (and revert) common transformations on time series data (scaling, filling missing values, differencing, boxcox, ...)
-
Metrics: A variety of metrics for evaluating time series' goodness of fit; from R2-scores to Mean Absolute Scaled Error.
-
Backtesting: Utilities for simulating historical forecasts, using moving time windows.
-
PyTorch Lightning Support: All deep learning models are implemented using PyTorch Lightning, supporting among other things custom callbacks, GPUs/TPUs training and custom trainers.
-
Filtering Models: Darts offers three filtering models:
KalmanFilter
,GaussianProcessFilter
, andMovingAverageFilter
, which allow to filter time series, and in some cases obtain probabilistic inferences of the underlying states/values. -
Datasets The
darts.datasets
submodule contains some popular time series datasets for rapid and reproducible experimentation.
Here's a breakdown of the forecasting models currently implemented in Darts. We are constantly working on bringing more models and features.
Model | Sources | Target Series Support: Univariate/ Multivariate |
Covariates Support: Past-observed/ Future-known/ Static |
Probabilistic Forecasting: Sampled/ Distribution Parameters |
Training & Forecasting on Multiple Series |
---|---|---|---|---|---|
Baseline Models (LocalForecastingModel) |
|||||
NaiveMean | β β | π΄ π΄ π΄ | π΄ π΄ | π΄ | |
NaiveSeasonal | β β | π΄ π΄ π΄ | π΄ π΄ | π΄ | |
NaiveDrift | β β | π΄ π΄ π΄ | π΄ π΄ | π΄ | |
NaiveMovingAverage | β β | π΄ π΄ π΄ | π΄ π΄ | π΄ | |
Statistical / Classic Models (LocalForecastingModel) |
|||||
ARIMA | β π΄ | π΄ β π΄ | β π΄ | π΄ | |
VARIMA | π΄ β | π΄ β π΄ | β π΄ | π΄ | |
AutoARIMA | β π΄ | π΄ β π΄ | π΄ π΄ | π΄ | |
StatsForecastAutoArima (faster AutoARIMA) | Nixtla's statsforecast | β π΄ | π΄ β π΄ | β π΄ | π΄ |
ExponentialSmoothing | β π΄ | π΄ π΄ π΄ | β π΄ | π΄ | |
StatsforecastAutoETS | Nixtla's statsforecast | β π΄ | π΄ β π΄ | β π΄ | π΄ |
StatsforecastAutoCES | Nixtla's statsforecast | β π΄ | π΄ π΄ π΄ | π΄ π΄ | π΄ |
BATS and TBATS | TBATS paper | β π΄ | π΄ π΄ π΄ | β π΄ | π΄ |
Theta and FourTheta | Theta & 4 Theta | β π΄ | π΄ π΄ π΄ | π΄ π΄ | π΄ |
StatsForecastAutoTheta | Nixtla's statsforecast | β π΄ | π΄ π΄ π΄ | β π΄ | π΄ |
Prophet | Prophet repo | β π΄ | π΄ β π΄ | β π΄ | π΄ |
FFT (Fast Fourier Transform) | β π΄ | π΄ π΄ π΄ | π΄ π΄ | π΄ | |
KalmanForecaster using the Kalman filter and N4SID for system identification | N4SID paper | β β | π΄ β π΄ | β π΄ | π΄ |
Croston method | β π΄ | π΄ π΄ π΄ | π΄ π΄ | π΄ | |
Global Baseline Models (GlobalForecastingModel) |
|||||
GlobalNaiveAggregate | β β | π΄ π΄ π΄ | π΄ π΄ | β | |
GlobalNaiveDrift | β β | π΄ π΄ π΄ | π΄ π΄ | β | |
GlobalNaiveSeasonal | β β | π΄ π΄ π΄ | π΄ π΄ | β | |
Regression Models (GlobalForecastingModel) |
|||||
RegressionModel: generic wrapper around any sklearn regression model | β β | β β β | π΄ π΄ | β | |
LinearRegressionModel | β β | β β β | β β | β | |
RandomForest | β β | β β β | π΄ π΄ | β | |
LightGBMModel | β β | β β β | β β | β | |
XGBModel | β β | β β β | β β | β | |
CatBoostModel | β β | β β β | β β | β | |
PyTorch (Lightning)-based Models (GlobalForecastingModel) |
|||||
RNNModel (incl. LSTM and GRU); equivalent to DeepAR in its probabilistic version | DeepAR paper | β β | π΄ β π΄ | β β | β |
BlockRNNModel (incl. LSTM and GRU) | β β | β π΄ π΄ | β β | β | |
NBEATSModel | N-BEATS paper | β β | β π΄ π΄ | β β | β |
NHiTSModel | N-HiTS paper | β β | β π΄ π΄ | β β | β |
TCNModel | TCN paper, DeepTCN paper, blog post | β β | β π΄ π΄ | β β | β |
TransformerModel | β β | β π΄ π΄ | β β | β | |
TFTModel (Temporal Fusion Transformer) | TFT paper, PyTorch Forecasting | β β | β β β | β β | β |
DLinearModel | DLinear paper | β β | β β β | β β | β |
NLinearModel | NLinear paper | β β | β β β | β β | β |
TiDEModel | TiDE paper | β β | β β β | β β | β |
TSMixerModel | TSMixer paper, PyTorch Implementation | β β | β β β | β β | β |
Ensemble Models (GlobalForecastingModel): Model support is dependent on ensembled forecasting models and the ensemble model itself |
|||||
NaiveEnsembleModel | β β | β β β | β β | β | |
RegressionEnsembleModel | β β | β β β | β β | β |
Anyone is welcome to join our Gitter room to ask questions, make proposals, discuss use-cases, and more. If you spot a bug or have suggestions, GitHub issues are also welcome.
If what you want to tell us is not suitable for Gitter or Github, feel free to send us an email at [email protected] for darts related matters or [email protected] for any other inquiries.
The development is ongoing, and we welcome suggestions, pull requests and issues on GitHub. All contributors will be acknowledged on the change log page.
Before working on a contribution (a new feature or a fix), check our contribution guidelines.
If you are using Darts in your scientific work, we would appreciate citations to the following JMLR paper.
Darts: User-Friendly Modern Machine Learning for Time Series
Bibtex entry:
@article{JMLR:v23:21-1177,
author = {Julien Herzen and Francesco LΓΒ€ssig and Samuele Giuliano Piazzetta and Thomas Neuer and LΓΒ©o Tafti and Guillaume Raille and Tomas Van Pottelbergh and Marek Pasieka and Andrzej Skrodzki and Nicolas Huguenin and Maxime Dumonal and Jan KoΓ
βΊcisz and Dennis Bader and FrΓΒ©dΓΒ©rick Gusset and Mounir Benheddi and Camila Williamson and Michal Kosinski and Matej Petrik and GaΓΒ«l Grosch},
title = {Darts: User-Friendly Modern Machine Learning for Time Series},
journal = {Journal of Machine Learning Research},
year = {2022},
volume = {23},
number = {124},
pages = {1-6},
url = {http://jmlr.org/papers/v23/21-1177.html}
}