binder

Interval based time series classification in sktime#

Interval based approaches look at phase dependent intervals of the full series, calculating summary statistics from selected subseries to be used in classification.

Currently 5 univariate interval based approaches are implemented in sktime. Time Series Forest (TSF) [1], the Random Interval Spectral Ensemble (RISE) [2], Supervised Time Series Forest (STSF) [3], the Canonical Interval Forest (CIF) [4] and the Diverse Representation Canonical Interval Forest (DrCIF). Both CIF and DrCIF have multivariate capabilities.

In this notebook, we will demonstrate how to use these classifiers on the ItalyPowerDemand and BasicMotions datasets.

References:#

[1] Deng, H., Runger, G., Tuv, E., & Vladimir, M. (2013). A time series forest for classification and feature extraction. Information Sciences, 239, 142-153.

[2] Flynn, M., Large, J., & Bagnall, T. (2019). The contract random interval spectral ensemble (c-RISE): the effect of contracting a classifier on accuracy. In International Conference on Hybrid Artificial Intelligence Systems (pp. 381-392). Springer, Cham.

[3] Cabello, N., Naghizade, E., Qi, J., & Kulik, L. (2020). Fast and Accurate Time Series Classification Through Supervised Interval Search. In IEEE International Conference on Data Mining.

[4] Middlehurst, M., Large, J., & Bagnall, A. (2020). The Canonical Interval Forest (CIF) Classifier for Time Series Classification. arXiv preprint arXiv:2008.09172.

[5] Lubba, C. H., Sethi, S. S., Knaute, P., Schultz, S. R., Fulcher, B. D., & Jones, N. S. (2019). catch22: CAnonical Time-series CHaracteristics. Data Mining and Knowledge Discovery, 33(6), 1821-1852.

1. Imports#

[ ]:
import numpy as np
from sklearn import metrics
from sklearn.pipeline import Pipeline

from sktime.classification.interval_based import (
    CanonicalIntervalForest,
    DrCIF,
    RandomIntervalSpectralEnsemble,
    SupervisedTimeSeriesForest,
    TimeSeriesForestClassifier,
)
from sktime.datasets import load_basic_motions, load_italy_power_demand
from sktime.transformations.panel.compose import ColumnConcatenator

2. Load data#

[ ]:
X_train, y_train = load_italy_power_demand(split="train", return_X_y=True)
X_test, y_test = load_italy_power_demand(split="test", return_X_y=True)
X_test = X_test[:50]
y_test = y_test[:50]

print(X_train.shape, y_train.shape, X_test.shape, y_test.shape)

X_train_mv, y_train_mv = load_basic_motions(split="train", return_X_y=True)
X_test_mv, y_test_mv = load_basic_motions(split="test", return_X_y=True)

X_train_mv = X_train_mv[:50]
y_train_mv = y_train_mv[:50]
X_test_mv = X_test_mv[:50]
y_test_mv = y_test_mv[:50]

print(X_train_mv.shape, y_train_mv.shape, X_test_mv.shape, y_test_mv.shape)

3. Time Series Forest (TSF)#

TSF is an ensemble of tree classifiers built on the summary statistics of randomly selected intervals. For each tree sqrt(series_length) intervals are randomly selected. From each of these intervals the mean, standard deviation and slope is extracted from each time series and concatenated into a feature vector. These new features are then used to build a tree, which is added to the ensemble.

[ ]:
tsf = TimeSeriesForestClassifier(n_estimators=50, random_state=47)
tsf.fit(X_train, y_train)

tsf_preds = tsf.predict(X_test)
print("TSF Accuracy: " + str(metrics.accuracy_score(y_test, tsf_preds)))
[ ]:
tsf = Pipeline(
    [
        ("column_concatenar", ColumnConcatenator()),
        ("classify", TimeSeriesForestClassifier(n_estimators=50, random_state=47)),
    ]
)
tsf.fit(X_train_mv, y_train_mv)

tsf_preds = tsf.predict(X_test_mv)
print("TSF Accuracy: " + str(metrics.accuracy_score(y_test_mv, tsf_preds)))
[ ]:
temporal_feature_importance = tsf["classify"].feature_importances_
separators = range(0, tsf["classify"].series_length, len(X_train_mv.iloc[0, 0]))

ax = temporal_feature_importance.plot(figsize=(20, 10))
for index, separator in enumerate(separators):
    ax.vlines(
        separator,
        temporal_feature_importance.min().min(),
        temporal_feature_importance.max().max(),
        color="r",
        alpha=0.3,
    )
    ax.text(
        separator, temporal_feature_importance.max().max(), X_train_mv.columns[index]
    )
[ ]:
X_train_mv_columns = list(X_train_mv.columns)
np.random.shuffle(X_train_mv_columns)

X_train_shuffled = X_train_mv[X_train_mv_columns]
X_train_shuffled.columns = X_train_mv.columns

X_test_shuffled = X_test_mv[X_train_mv_columns]
X_test_shuffled.columns = X_test_mv.columns

tsf = Pipeline(
    [
        ("column_concatenator", ColumnConcatenator()),
        ("classify", TimeSeriesForestClassifier(n_estimators=50, random_state=47)),
    ]
)
tsf.fit(X_train_shuffled, y_train_mv)

tsf_preds = tsf.predict(X_test_shuffled)
print("TSF Accuracy: " + str(metrics.accuracy_score(y_test_mv, tsf_preds)))
[ ]:
temporal_feature_importance = tsf["classify"].feature_importances_
separators = range(0, tsf["classify"].series_length, len(X_train_mv.iloc[0, 0]))

ax = temporal_feature_importance.plot(figsize=(20, 10))
for index, separator in enumerate(separators):
    ax.vlines(
        separator,
        temporal_feature_importance.min().min(),
        temporal_feature_importance.max().max(),
        color="r",
        alpha=0.3,
    )
    ax.text(
        separator, temporal_feature_importance.max().max(), X_train_mv_columns[index]
    )
[ ]:
tsf = Pipeline(
    [
        ("column_concatenator", ColumnConcatenator()),
        (
            "classify",
            TimeSeriesForestClassifier(
                n_estimators=50, random_state=47, inner_series_length=100
            ),
        ),
    ]
)
tsf.fit(X_train_mv, y_train_mv)

tsf_preds = tsf.predict(X_test_mv)
print("TSF Accuracy: " + str(metrics.accuracy_score(y_test_mv, tsf_preds)))
[ ]:
temporal_feature_importance = tsf["classify"].feature_importances_
separators = range(0, tsf["classify"].series_length, len(X_train_mv.iloc[0, 0]))

ax = temporal_feature_importance.plot(figsize=(20, 10))
for index, separator in enumerate(separators):
    ax.vlines(
        separator,
        temporal_feature_importance.min().min(),
        temporal_feature_importance.max().max(),
        color="r",
        alpha=0.3,
    )
    ax.text(
        separator, temporal_feature_importance.max().max(), X_train_mv.columns[index]
    )
[ ]:
X_train_mv_columns = list(X_train_mv.columns)
np.random.shuffle(X_train_mv_columns)

X_train_shuffled = X_train_mv[X_train_mv_columns]
X_train_shuffled.columns = X_train_mv.columns

X_test_shuffled = X_test_mv[X_train_mv_columns]
X_test_shuffled.columns = X_test_mv.columns

tsf = Pipeline(
    [
        ("column_concatenator", ColumnConcatenator()),
        (
            "classify",
            TimeSeriesForestClassifier(
                n_estimators=50, random_state=47, inner_series_length=100
            ),
        ),
    ]
)
tsf.fit(X_train_shuffled, y_train_mv)

tsf_preds = tsf.predict(X_test_shuffled)
print("TSF Accuracy: " + str(metrics.accuracy_score(y_test_mv, tsf_preds)))
[ ]:
temporal_feature_importance = tsf["classify"].feature_importances_
separators = range(0, tsf["classify"].series_length, len(X_train_mv.iloc[0, 0]))

ax = temporal_feature_importance.plot(figsize=(20, 10))
for index, separator in enumerate(separators):
    ax.vlines(
        separator,
        temporal_feature_importance.min().min(),
        temporal_feature_importance.max().max(),
        color="r",
        alpha=0.3,
    )
    ax.text(
        separator, temporal_feature_importance.max().max(), X_train_mv_columns[index]
    )

4. Random Interval Spectral Ensemble (RISE)#

RISE is a tree based interval ensemble aimed at classifying audio data. Unlike TSF, it uses a single interval for each tree, and it uses spectral features rather than summary statistics.

[ ]:
rise = RandomIntervalSpectralEnsemble(n_estimators=50, random_state=47)
rise.fit(X_train, y_train)

rise_preds = rise.predict(X_test)
print("RISE Accuracy: " + str(metrics.accuracy_score(y_test, rise_preds)))

5. Supervised Time Series Forest (STSF)#

STSF makes a number of adjustments from the original TSF algorithm. A supervised method of selecting intervals replaces random selection. Features are extracted from intervals generated from additional representations in periodogram and 1st order differences. Median, min, max and interquartile range are included in the summary statistics extracted.

[ ]:
stsf = SupervisedTimeSeriesForest(n_estimators=50, random_state=47)
stsf.fit(X_train, y_train)

stsf_preds = stsf.predict(X_test)
print("STSF Accuracy: " + str(metrics.accuracy_score(y_test, stsf_preds)))

6. Canonical Interval Forest (CIF)#

CIF extends from the TSF algorithm. In addition to the 3 summary statistics used by TSF, CIF makes use of the features from the Catch22 [5] transform. To increase the diversity of the ensemble, the number of TSF and catch22 attributes is randomly subsampled per tree.

Univariate#

[ ]:
cif = CanonicalIntervalForest(n_estimators=50, att_subsample_size=8, random_state=47)
cif.fit(X_train, y_train)

cif_preds = cif.predict(X_test)
print("CIF Accuracy: " + str(metrics.accuracy_score(y_test, cif_preds)))

Multivariate#

[ ]:
cif_m = CanonicalIntervalForest(n_estimators=50, att_subsample_size=8, random_state=47)
cif_m.fit(X_train_mv, y_train_mv)

cif_m_preds = cif_m.predict(X_test_mv)
print("CIF Accuracy: " + str(metrics.accuracy_score(y_test_mv, cif_m_preds)))

6. Diverse Representation Canonical Interval Forest (DrCIF)#

DrCIF makes use of the periodogram and differences representations used by STSF as well as the addition summary statistics in CIF.

Univariate#

[ ]:
drcif = DrCIF(n_estimators=5, att_subsample_size=10, random_state=47)
drcif.fit(X_train, y_train)

drcif_preds = drcif.predict(X_test)
print("DrCIF Accuracy: " + str(metrics.accuracy_score(y_test, drcif_preds)))

Multivariate#

[ ]:
drcif_m = DrCIF(n_estimators=5, att_subsample_size=10, random_state=47)
drcif_m.fit(X_train_mv, y_train_mv)

drcif_m_preds = drcif_m.predict(X_test_mv)
print("DrCIF Accuracy: " + str(metrics.accuracy_score(y_test_mv, drcif_m_preds)))

Generated using nbsphinx. The Jupyter notebook can be found here.