Unsupervised Feature Based Algorithms for Time Series Extrinsic Regression
Abstract
Time Series Extrinsic Regression (TSER) involves using a set of training time series to form a predictive model of a continuous response variable that is not directly related to the regressor series. The TSER archive for comparing algorithms was released in 2022 with 19 problems. We increase the size of this archive to 63 problems and reproduce the previous comparison of baseline algorithms. We then extend the comparison to include a wider range of standard regressors and the latest versions of TSER models used in the previous study. We show that none of the previously evaluated regressors can outperform a regression adaptation of a standard classifier, rotation forest. We introduce two new TSER algorithms developed from related work in time series classification. FreshPRINCE is a pipeline estimator consisting of a transform into a wide range of summary features followed by a rotation forest regressor. DrCIF is a tree ensemble that creates features from summary statistics over random intervals. Our study demonstrates that both algorithms, along with InceptionTime, exhibit significantly better performance compared to the other 18 regressors tested. More importantly, these two proposals (DrCIF and FreshPRINCE) models are the only ones that significantly outperform the standard rotation forest regressor.
Index Terms:
Time series, Extrinsic regression, Interval ensembles.I Introduction
Time series analysis is a popular topic in machine learning and data mining research. Thousands of research papers in this field have been published in the last decade. Various algorithms have been proposed for disparate tasks across a wide range of applications. The main reason for this development is the increased ability to store data over time and the spread of cheap sensor technology to most fields of science. For example, solar panels depend on sensors to maximise their potential (e.g. to tilt the solar panel so that the sun shines directly on it) and hospitals routinely record and store patient data such as vital signs. This vast wealth of data offers great potential for data mining.
Two of the most researched time series machine learning/analysis tasks are classification [1, 2, 3] and forecasting [4]. Time Series Classification (TSC) involves building a predictive model from (possibly multivariate) time series for a categorical target variable. TSC differs from standard classification in that the discriminatory features are often in the shape of the series or the autocorrelation. Forecasting consists of predicting (usually numeric) values based on past observations. Forecasting is usually approached through a model-based algorithm (e.g., autoregressive or exponential smoothing) or by reducing the forecasting problem to a regression problem through a sliding window then using deep learning or a global model such as XGBoost.
Tan et al. in [5] formally specified a related, but distinct, type of time series regression problem: Time Series Extrinsic Regression (TSER). Rather than being derived from a forecasting problem, TSER involves a predictive model built on time series to predict a real-valued variable distinct from the training input series. For example, Figure 1 shows soil spectrograms which can be used to estimate the potassium concentration. Ground truth is found through expensive lab based experiments that take some time. Spectrograms (ordered data series we treat as time series) are cheap to obtain and the data can be collected in any environment. An accurate regressor from spectrogram to concentration would make land and crop management more efficient.

TSER is related to TSC as traditional regression is to classification: the only difference is that the target variable is real-valued rather than categorical. The distinction extrinsic is required because of the prevalence of the term time series regression in the forecasting literature to mean reduce forecasting to regression through a sliding window.
The first benchmarking work for TSER [5] introduced an archive of 19 TSER problems, including four univariate and 15 multivariate datasets. They performed an experimental comparison of the performance of 13 algorithms on these data. The two algorithms adapted from the TSC literature, the RandOm Convolutional KErnel Transform (ROCKET) [6] and the deep learner InceptionTime [7], were top-ranked. However, there was no significant difference in Root Mean Square Error (RMSE) between the ten best-performing algorithms, possibly because of the relatively small number of datasets and the conservative nature of the adjustment for multiple tests used. The abstract of [5] states that “we show that much research is needed in this field to improve the accuracy of ML models [for TSER]”. Despite the paper’s popularity and the identification of a clear need for novel research, there has been little or no progress in addressing this challenge. We have responded to this call to arms and developed and assessed a range of TSER algorithms. We have proposed new algorithms that are significantly better than the ones evaluated in [5].
Our starting point to TSER is to adapt TSC algorithms for regression. The ROCKET family of classifiers all involve transformations using randomised convolutions and a pooling operation followed by a linear or ridge classifier. The original ROCKET was converted to TSER by switching the classifier for a ridge regressor. We extend this to consider a more recent ROCKET variant, MultiROCKET [8]. Deep learning algorithms are also simple to adapt, and we extend the previous study to include a convolutionary neural network in addition to an ensemble regression version of InceptionTime [7].
An alternative approach to TSC is to use a large number of unsupervised summary features as a transform. A review of a range of alternatives [9] found that the Fresh Pipeline with RotatIoN forest Classifier (FreshPRINCE) was the best transform pipeline for TSC. The FreshPRINCE uses the Time Series Feature Extraction based on Scalable Hypothesis Tests (TSFresh) [10] followed by a Rotation Forest (RotF) classifier [11]. We implement the FreshPRINCE for TSER.
Interval-based classifiers also extract unsupervised features, but they do so by ensembling pipelines with randomly selected intervals and a fast base classifier. The first interval-based approach for TSC was the Time Series Forest (TSF) [12]. TSF generates a set of random intervals and concatenates each interval’s mean, standard deviation and slope to make a unique feature space for every base classifier. A version of TSF for regression is available in the aeon toolkit111https://github.com/aeon-toolkit/aeon.. The Canonical Interval Forest (CIF) [13], and the subsequent Diverse Representation Canonical Interval Forest (DrCIF) [14], adopt a similar model to the TSF but use different summary features and data representations. CIF uses the Canonical Time Series Characteristics (Catch22) [15] feature set. Details of these transformation-based algorithms and how we have adapted them to TSER are provided in Section III. Our main contributions can be summarised as follows:
-
1.
We provide 44 new datasets to the TSER archive, including 24 univariate and 20 multivariate datasets, to take the archive to 63 datasets;
-
2.
We repeat the study from [5] on these new data to examine whether the conclusions translate to the larger collection;
-
3.
We implement recently proposed convolutional-based, feature-based, interval-based, and deep learning-based TSC algorithms to TSER;
-
4.
We conduct an extensive experimental study using 21 regressors and demonstrate that feature-based and interval-based regressors, on average, achieve a significantly better RMSE than any other assessed algorithms;
-
5.
We carry out a comprehensive analysis and an ablative study of the two best proposed approaches: FreshPRINCE and DrCIF.
-
6.
We provide open source implementations of scikit-learn compatible implementations, clear guidance on reproducibility, and detailed results on the associated repository222https://tsml-eval.readthedocs.io/en/latest/publications/2023/tser_archive_expansion/tser_archive_expansion.html.
The rest of this paper is structured as follows. In Section II, the background and related works are exposed. Section III describes our new TSER algorithms in detail. In Section IV, we give an overview of the new archive and describe our experimental setup. In Section V, experimental results for the 21 approaches applied to the total of 63 datasets are presented. Section VI looks at these results in more detail. Finally, Section VII summarises our findings and highlights future work.
II Background and related work
TSER aims to create a mapping function between a time series and a scalar value [5]. A time series is composed of real-valued ordered observations. Formally, a univariate time series of length is defined as . A multivariate time series with channels is specified as , where . We assume we have a list containing time series. Hence, represents the -th observation of the -th case for the -th channel. A dataset is composed of time series samples and an associated response variable, , where are the output continuous values, i. e. the input time series is associated to the output .
A TSER model is a mapping function , where is the space of all time series and a continuous value. A TSER model is trained on a dataset and evaluated on an independent test dataset .
TSER shares some similarities with Scalar-on-Function Regression (SoFR) [16], a functional regression model where basis models are applied to the series prior to regression, i. e. its goal is to fit a regression model with scalar responses and functional data points as predictors [17]. [5] used two SoFR models in their comparison based on [16]. These were Functional Principal Component Regression (FPCR) and FPCR with B-splines (FPCR-Bs).
Time series forecasting is often reduced to a regression window through the application of a sliding window to form time series and a forecast horizon specifying how to select the target . The most common techniques used in Time Series Forecasting Regression (TSFR) include deep learning variants and global models, where channels are concatenated and a standard regressor such as Random Forest or XGBoost is applied.
Our primary source for TSER [5] compared three standard regression algorithms (Support Vector Regressors (SVR) [18], Random Forest (RandF) [19], and eXtreme Gradient Boosting (XGBoost) [20]); two -Nearest Neighbours models using Euclidean and Dynamic Time Warping distances using both one and five neighbours; three deep learning approaches (Fully Convolutional Neural Network (FCN) [21], Residual Network (ResNet) [22] and InceptionTime [7]); two functional analysis approaches (FPCR and FPCR with B-splines [16]); and ROCKET [6]. The standard regressors adopt the approach of global forecasting regressors: time series are flattened into a vector concatenating all the channels. Hence, a multivariate time series of length and channels is converted into a single vector of length . Subsequently, there have been very few algorithmic advances for TSER. Most novel developments are domain specific and not aimed at TSER as a whole. Among them, a Linear Space State Layers (LSSL) model [23] has been tested on three TSER datasets, achieving low error metrics. Another state-space model, Liquid-S4 [24] has also been evaluated on those three datasets, and claims better results. An architecture based on Graph Neural Networks called TISER-GCN [25] has been applied to seismic data as an extrinsic regression task. In a similar context, [26] introduces Septor, a hierarchical neural network model developed to estimate the depth of seismic events from waveform data, i.e. a domain-specific extrinsic regression task. ROCKET-XGBoost [27] has been the only novel algorithm evaluated on the 19 TSER archive datasets, but it offered no significant improvement over the algorithms evaluated in [5].
II-A Time Series Classification (TSC) algorithms
There are a plethora of algorithms for TSC that have been compared in reproducible comparative studies [1, 2, 3]. Broadly, algorithms can be grouped into how they represent time series. We provide a very brief overview with a focus on how classifiers have been or could be adapted to TSER.
Distance-based classifiers use a distance function in conjunction with an algorithm such as Nearest Neighbour (NN) classifier. The two most commonly used distance functions are Euclidean distance and Dynamic Time Warping (DTW). A NN classifier can trivially be adapted for regression by averaging over the target variable of the NN. For multivariate data, using terminology presented in [28], DTW can either be independent (find DTW distance on each channel separately them sum the values) or dependent (use all channels in the point wise distance calculation).
Feature-based algorithms transform series into features using unsupervised descriptive statistics, then complete the pipeline with a classifier trained on the new feature set.
Interval-based classifiers are an extension of feature-based pipeline classifiers where rather than form summary features over the whole series, they concatenate features found over different intervals. They then form an ensemble over different randomised intervals rather than use a single estimator. Together, we group feature-based and interval-based approaches together as unsupervised feature-based classifiers. Adapting these algorithms to TSER is our primary research goal, so we cover this topic in more detail in Section III.
Kernel/convolution-based models find convolutions from the space of all possible subseries and use them to create features through a form of pooling operation. The most popular approach, ROCKET [6] generates random convolutions and is used in conjunction with a ridge classifier in a pipeline. It was adapted for TSER by simply changing the ridge classifier for a ridge regressor [5]. More recently, MultiROCKET [29] was proposed as an improved version of ROCKET. ROCKET uses two pooling operations to generate features: max pooling and the percentage of positive values. MultiROCKET adds three new pooling operations: mean of positive values, mean of indices of positive values, and longest stretch of positive values. It also extracts features from first order differences in addition to the raw data. We adapt MultiROCKET for TSER in exactly the same way as ROCKET.
Deep learning continues to be popular for TSC [30], although to our knowledge InceptionTime [7] is still the best performing deep learner. The study in [5] used Residual Networks (ResNet), Fully Connected Networks (FCN), and InceptionTime. The original InceptionTime paper [7] proposed an ensemble of five InceptionTime classifiers to obtain the final results. However, [5] used a single InceptionTime model for TSER. We evaluate both a single InceptionTime (Inception) and an InceptionTime Ensemble (InceptionTimeE) faithful to the TSC version for TSER. We also evaluate the Convolutional Neural Network (CNN) regressor based on the classifier described in [31].
Shapelet-based approaches [32, 33] base classification on the presence of selected phase-independent subseries found from the training data. For classification, shapelets are assessed with a supervised measure such as information gain. Furthermore, the most accurate shapelet-based approaches [14] evaluate shapelets with a one vs many approach and balance the search procedure between classes to improve diversity. Adapting shapelets for TSER requires significant internal changes and design decisions, being beyond the scope of this paper.
Dictionary based algorithms use a bag of words-like approach to base classification on the number of occurrences of approximated subseries (patterns). The most successful dictionary-based classifiers [34, 35, 36] involve a degree of supervised selection using accuracy for filtering/weighting or feature selection and their adaptation for TSER is also beyond the scope of this paper.
III Unsupervised Feature-Based Regressors
Approaches which extract features from time series in an unsupervised process have been shown to perform well in classification scenarios. ROCKET [6] and CIF [13] perform as well as or better than single representation approaches such as shapelet or dictionary algorithms. These algorithms also have the benefit of lower complexity, essentially consisting of transform to classifier pipelines or an ensemble of pipelines. ROCKET and CIF [13] were also top ranked for Multivariate TSC (MTSC) in a recent survey [3].
We describe the features extracted and our adaptations for two additional algorithms based on unsupervised transformations. The first is the FreshPRINCE [9], a pipeline using the TSFresh [10] feature set. The second is DrCIF [14], an interval-based ensemble.
III-A FreshPRINCE
FreshPRINCE is a pipeline algorithm for regression with two components: the TSFresh feature extraction algorithm that transforms the input time series into a feature vector, and then a Rotation Forest (RotF) [11] estimator that builds a model and makes label predictions. TSFresh [37] is a collection of just under 800 features that can be extracted from time series data. While the features can be used on their own, a feature selection method called Fresh is provided to remove irrelevant features. FreshPRINCE does not make use of this feature extraction, however, keeping the transformation process unsupervised and allowing the RotF decide the worth of features. TSFresh is generally popular within the data science community, and has shown to perform better than other unsupervised transformation pipelines on classification problems as part of FreshPRINCE [9].
RotF is an ensemble of tree classifiers which has been shown to accurately make predictions for problems where the attributes are continuous [38]. The classifier has been used as a benchmark and as a part of other pipeline classifiers in TSC [1, 14], and performed better than a ridge classifier and XGBoost [20] when paired with unsupervised transforms for TSC [9]. Full descriptions of the RotF algorithm are available in [11] and [38]. RotF is easily adaptable for regression: the implementation we developed removes class subsampling [39], replaces the C4.5 decision tree with a Classification and Regression Tree (CART) [40], and averages the label predictions for each tree in the forest. The full TSFresh transformation and altered RotF make up our FreshPRINCE adaptation for TSER.
III-B DrCIF
Interval-based techniques select phase-dependent intervals of fixed offsets from which to extract summary features. These intervals share their position for all time series, with the aim of discovering discriminatory features from particular locations in time. Most interval techniques take the form of a forest of decision trees, using different intervals to achieve diversity in the ensemble. While some interval forests do make use of supervised feature extraction [41], TSF [12] and DrCIF [14] are completely unsupervised in their method for selecting intervals and extracting features from said intervals. All that we change for TSER from the classification implementation is a swapping of the tree algorithm used. TSF can be adapted for the regression task in the same way.
From a series of length , there are possible intervals when considering all interval lengths and positions. Even at small series lengths, it is unfeasible to extract features from or evaluate all possible intervals. To solve the issue of which intervals from this pool to select, DrCIF uses a random forest based approach. An ensemble of CART regressors is formed, built on the output of different random interval transformations. Algorithm 1 describes the full build process for DrCIF. The transformation has three steps. First, the base time series is split into three series representations: the original time series, the first order differences of the series, and the periodogram of the series (characterised in line 3 of Algorithm 1). The differences and periodogram series-to-series transformations have shown to provide useful information in classification approaches [42, 41, 8, 43]. Then, a different transform is created for each base regressor. First, a pool of features is selected from a candidate pool of 29 features (line 6). DrCIF makes use of the CAnonical Time series CHaracteristics (Catch22) [15]. Catch22 is a diverse set of 22 features filtered from the 7000+ available in the Highly Comparative Time Series Analysis (HCTSA) toolbox [44]. The Catch22 features were selected for use on normalised data, but we do not make that assumption. Hence, seven additional summary statistics are also candidates: the mean, standard-deviation, slope, median, interquartile range, min, and max. Then, for each data representation, a set of random intervals are selected (lines 10-13), and the unsupervised features are calculated and concatenated from a randomly selected channel (lines 13-15). Finally, a CART tree is trained on the feature set unique to each ensemble member. Figure 2 visualises the transformation (left) and ensemble (right) process for DrCIF. Predictions for new cases are found by averaging the predictions of the base regressors.
![]() |
![]() |
IV Methodology
We summarise the new problems we have added, the regressors used in experiments, and a description of our experimental method.
The current version of the TSER archive includes 19 different datasets. We have increased the number of datasets in the archive to 63. There are now 28 univariate problems and 35 multivariate, with number of channels ranging from 2 to 24. Dataset size range from 93 to over 16,000. 70% are used for training, 30% for testing. Series length ranges from 14 to 7500. Nine of the problems contain missing values and two have unequal-length series.
The new datasets have been taken from Kaggle competitions and other archives and repositories/websites associated with applied research. Table I summarises the gathered data. More details on the datasets are available on the associated repository. None of the datasets have been normalised. One of the new problems has unequal-length series. For experiments, keeping with the practice in [5], missing values in the series are linearly interpolated, and unequal-length series are truncated to the minimum length series. Full descriptions, and both unequal and missing values series are available on the associated website.
Name | Prediction problem (response variable) |
---|---|
Economic Analysis | |
DailyOilGasPrices | Daily gas price with time series of oil prices3. |
Energy Monitoring | |
Energy building predictors | Estimate the energy consumption of different sorts on buildings4. |
OccupancyDetectionLight | The average hourly occupancy of an office from sensor measurements [45]. |
SolarRadiationAndalusia | The average hourly solar radiation from atmospherical measurements5. |
TetuanEnergyConsumption | The daily average power consumption in three areas of Tetouan from atmospherical measurements[46]. |
WindTurbinePower | The daily power output of a wind turbine based on time series of torque measurements6. |
Environment monitoring . | |
AcousticContaminationMadrid | The 1st percentile of sound pressure levels from LAeq7. |
Africa Soil Chemistry | A set of 12 problems derived from the Africa Soil Information Service (AfSIS) Soil Chemistry 8. |
BeijingIntAirportPM25Quality | The daily average of particulate matter in the Airport of Beijing from atmospherical data. [47]. |
DailyTemperatureLatitude | The latitude of a city based on the annual time series of daily temperature9. |
DhakaHourlyAirQuality | The Air Quality Index in Dhaka using localised particular matter time series10. |
MadridPM10Quality | The weekly average of particulate matter in the city of Madrid, Spain, from measurements of gases11. |
MetroInterstateTrafficVolume | The daily average traffic volume of a road in the USA from atmospherical variables12. |
ParkingBirmingham | The daily occupancy rate from the hourly total number of parked cars [48]. |
PrecipitationAndalusia | The yearly average of rainfall on Andalusia, Spain, from meterological measurements13. |
SierraNevadaMountainsSnow | The amount of snow based on temperature time series [49]. |
Equipment monitoring | |
ElectricMotorTemperature | The temperature of an electric motor based on time series of torque readings14. |
LPGas/Methane MonitoringHomeActivity | The liquefied petroleum and methane concentration from gas sensors [50]. |
GasSensorArray Ethanol/Acetone | The concentrations of two analytes, acetone based on 16 metal-oxide sensors [51]. |
WaveTensionData | The tension of a string based on wave elevation time series15. |
Health Monitoring | |
BarCrawl6min | The transdermal alcohol content by using an accelerometer [52]. |
Covid19Andalusia | The rate of deceased/contagions people from number of contagions in Andalusia, Spain [53]. |
VentilatorPressure | The pressure of the inspiratory solenoid valve from control input and output of the same valve16. |
Sentiment Analysis | |
Crypto Sentiment | The sentiment of four cryptocurrencies based on the same days hourly price17. |
NaturalGasPriceSentiment | Sentiment scores about natural gas prices [54] based on the daily natural gas prices. |
IV-A Regression Algorithms
The full list of the 21 regressors (with associated abbreviation) evaluated in Section V is as follows:
- •
- •
- •
-
•
Distance-based:
-
–
Nearest neighbours with Euclidean distance (NN-ed), with .
-
–
Nearest neighbours with dynamic time warping distance (NN-dtw), with .
-
–
-
•
Feature-based:
-
–
FreshPRINCE [9].
-
–
- •
- •
Parameter settings for all algorithms are in the supplementary material.
IV-B Experimental Design
Each dataset is provided with a default train/test split. We repeat every experiment 30 times to obtain a more significant analysis. The first experiment is with the default data. Subsequent experiments are conducted with data resampled by pooling the train and test and randomly partitioning the data with the same train/test proportions as the original. Performance is measured with the RMSE to conform with [5]. To compare regressors, we first average RMSE over all resamples. We use ranks in all statistical tests. For multiple regressors over multiple datasets we use an adaptation of the critical difference diagram [55], replacing the post-hoc Nemenyi test with a comparison of all classifiers using pairwise Wilcoxon signed-rank tests, and cliques formed using the Holm correction [56, 57].
V Results
![]() |
![]() |
Our experiments are structured as follows. In Section V-A we recreate the results presented in [5] on the original 19 datasets. We then extend the analysis to our larger collection of datasets to test whether the conclusions reached in [5] generalise to the new archive of 63 problems. In Section V-B we include two improvements for regressors used in the previous study and other regressors available in open source toolkits. In Section V-C we compare the best performing regressors from the previous experiments to the new algorithms we are proposing, FreshPRINCE and DrCIF.
V-A Recreating Results on the 19 TSER datasets
We ran the 13 regressors reported in [5] on the current 19 datasets in the archive, using five resamples, keeping in practise with the original work. Figure 3 shows a critical difference diagram of our results alongside the results presented in [5]. Broadly, the ordering of algorithms is the same and the cliques are similar. There are some differences in the ordering, with ROCKET and FCN lower ranked and Inception and ResNet higher in our experiments than the original. We also have more diverse cliques. This is because our adjustment for multiple testing is less conservative than the one used in [5], where a full Bonferonni adjustment is used rather than a Holm correction.

In Figure 4 we compare the 13 regressors used for Figure 3 on the larger archive of 63 datasets. From now on, we also extended the number of resamples from five to 30. We note that RandF is now the best, improving significantly on ROCKET, Inception, and ResNet. Hence, the time series specific methods previously proposed for TSER are not better than using an off the shelf regressor with concatenated features.
V-B Benchmarking the new TSER archive
For the next set of experiments, we take the top five algorithms in Figure 4 and compare these to some alternative adaptations of time series specific algorithms. The good performance of XGBoost and RandF suggests we should not overlook standard classifiers. Rotation Forest (RotF) [11] is a classifier that can be easily adapted to regression by simple averaging [39]. It has been shown to be particularly effective for problems with all real valued attributes, including time series [38, 1]. Hence, we include a regression adaption in this round of experiments. We also add in the standard Ridge regressor for completeness sake. Besides, the open source toolkit aeon181818https://github.com/aeon-toolkit/aeon includes two regression implementations not previously evaluated in the context of TSER. TimeSeriesForestRegressor (TSF) is an adaptation of the Time Series Forest classifier [12] and CNNRegressor (CNN) is Convolutional Neural Network based on the version described in [31]. On further investigation, we found that the results for InceptionTime in [5] were created with a single InceptionTime model (Inception). However, in the original work [7], the results supporting InceptionTime as a classifier are found with an ensemble of five InceptionTime models. We include an InceptionTime ensemble model for regression (InceptionE). Furthermore, an improved version of the ROCKET algorithm has been recently publish, known as MultiROCKET [29]. We adapted it to the TSER paradigm accordingly. We provide implementations of MultiROCKET, RotF, and InceptionE in the associated repository. Figure 5 shows the results of the three best algorithms from experiments presented in Figure 4 and six new regressors. We have had to exclude the AustralianRainfall dataset and hence reduce the number of datasets in our study to 62 because of MultiROCKET. The MultiROCKET approach requires over 600GB memory for this dataset and takes more than 15 days to complete.

The CNN and Ridge regressors are not competitive with the other nine algorithms. InceptionE is significantly better than a single Inception network. As expected, MultiROCKET is significantly better than ROCKET. However, the standout result is that RotF is one of the top performing algorithms along with InceptionE. RotF is the best performing standard algorithm for TSC [1], so this is perhaps not surprising. Nevertheless, the fact that an algorithm for standard regression outperforms all the deep learning and time series specific approaches but InceptionE is indicative of the scope for improvement in the field of TSER.
V-C Evaluation of new TSER algorithms

We now compare the top six algorithms from the previous experiment with our two regressors based on unsupervised feature extraction, DrCIF and FreshPRINCE. Note that the single Inception network is not included as InceptionE always outperforms it. Moreover, we also decided to include ROCKET as it was the top performing approach in the original study of [5]. Figure 6 shows that both of the new regressors are in the top clique, and are significantly better in terms of averaged rank for RMSE than RotF, which is the top performing standard approach of all others we have tried. The InceptionE is the third best algorithm. InceptionE is often very good: it is top ranked on 18 of the 62 problems. However, it also fails spectacularly on many problems. Figure 7 shows the boxplot for the relative deviation of the RMSE over all problems. A figure below 0.5 means the algorithm is better than the median RMSE for that particular problem. We observe that overall, only DrCIF and FreshPRINCE are consistently better than the median performance, and the distribution is tightly coupled. InceptionE has the widest spread.

Figure 8 elaborates on the analysis of the results. In this case, a heatmap with the average RMSE results is shown, done with a novel tool191919https://github.com/MSD-IRIMAS/Multi_Comparison_Matrix. As can be seen, the average results obtained by DrCIF and FreshPRINCE are much better than the rest. Finally, as previously observed, InceptionE shows the worst average RMSE results, which indicates that it is not a very stable method, as its rank is the 3rd, but in terms of mean RMSE it is the 8th.

VI Analysis
We explore our results in more detail to better profile the regressors and gain insights into the drivers behind their performance.
VI-A Run time
Figure 9 shows the average rank RMSE against the run time (on a log scale) for the eight regressors from Figure 7. We see a direct trade off between runtime and performance. All algorithms run on a single thread CPU except for InceptionE, which ran on a GPU. This means the graph is very flattering for InceptionE. Even on a GPU it is slower than RotF on a CPU.

VI-B Performance by Data Characteristics
We break down the performance of regressors by the core characteristics of the data to help gain insights into when different algorithms perform well. We stress this is purely exploratory: the relatively small number of datasets in each category, in brackets, preclude useful significant testing.
300 (15) | 300-999 (23) | 1000-5000 (13) | 5000 (11) | |
---|---|---|---|---|
DrCIF | 2.87 | 3.39 | 2.85 | 3.91 |
FreshPRINCE | 3.60 | 2.87 | 2.92 | 2.27 |
InceptionE | 3.87 | 3.17 | 2.69 | 3.91 |
RotF | 4.67 | 4.57 | 5.46 | 4.73 |
MultiROCKET | 4.27 | 4.30 | 3.85 | 3.73 |
TSF | 5.13 | 5.48 | 5.08 | 5.55 |
RandF | 6.07 | 6.04 | 4.31 | 4.36 |
ROCKET | 5.53 | 6.17 | 5.69 | 5.55 |
Table II shows the average rank RMSE when we group problems by the number of training cases. The pattern is that DrCIF is better with a small number of cases, FreshPRINCE is better with larger train set size.
1 (28) | 2 (9) | 3 or 4 (14) | 5 (11) | |
---|---|---|---|---|
DrCIF | 3.29 | 2.90 | 2.33 | 2.91 |
FreshPRINCE | 2.96 | 3.80 | 4.00 | 2.55 |
InceptionE | 4.21 | 2.90 | 3.58 | 5.09 |
RotF | 4.25 | 5.50 | 4.58 | 2.91 |
MultiROCKET | 4.04 | 4.40 | 5.75 | 5.45 |
TSF | 5.50 | 5.10 | 3.33 | 5.45 |
RandF | 6.21 | 5.70 | 5.92 | 5.18 |
ROCKET | 5.54 | 5.70 | 6.50 | 6.45 |
Table III shows the average rank RMSE when we group problems by the number of channels. DrCIF performs relatively better than FreshPRINCE on multivariate problems with 2, 3 or 4 channels. Whereas FreshPRINCE achieves better results when dealing with univariate and multivariate datasets with more than 5 channels.
50 (14) | 50-150 (13) | 151-365 (16) | 1000 (19) | |
---|---|---|---|---|
DrCIF | 2.69 | 2.00 | 2.64 | 3.95 |
FreshPRINCE | 3.77 | 4.25 | 3.00 | 2.58 |
InceptionE | 4.23 | 5.63 | 4.64 | 2.74 |
RotF | 3.69 | 4.13 | 4.05 | 5.05 |
MultiROCKET | 5.15 | 3.88 | 5.23 | 4.05 |
TSF | 4.85 | 4.63 | 4.41 | 5.79 |
RandF | 5.08 | 5.88 | 6.09 | 6.21 |
ROCKET | 6.54 | 5.63 | 5.95 | 5.63 |
Table IV shows the average rank RMSE when we group problems by the series length. Surprisingly, the interval based DrCIF also performs relatively better than FreshPRINCE with short series but worse with longer series (length 1000).
Finally, we also assessed relative performance for different problem types but did not detect any interesting trends.
VI-C Ablation of FreshPRINCE
FreshPRINCE is a pipeline of a TSFresh transform and a RotF regressor. We address the question of whether the performance of this regressor is due to the transform, the regressor or both. Figure 10 summarises the performance of freshPRINCE, RotF on the raw series and TSFresh transform followed by an alternative regressor. It demonstrates that transforming followed by RandF or XGBoost are no better than simply applying RotF to the raw data. We conclude that it is the combination of transform and regressor that give significantly better performance.



There is very little agreement between InceptionE and the other regressors. We believe one reason InceptionE performance is so variable is it sometimes completely fails to find anything useful in a dataset where other models have at least some predictive power. To demonstrate this, we look at the standardised residuals of DrCIF and InceptionE on the BarCrawl6min dataset. The time series are accelerometer data, and the response variable is the transdermal alcohol concentration of the test subjects. The response variable is bounded below by zero.

In traditional regression, the analyst might look to transform the response with, for example, a Yeo-Johnson transform [58]. We are interested in performance over multiple datasets without hand tailored transforms. The RMSE for the default train-test partition for DrCIF is 0.0017 and for InceptionE it is 0.0045. If we plot the predicted response vs the actual response for DrCIF (Figure 11) we see that DrCIF is making negative predictions for low actual values, and underestimating higher values and there seems to be some heteroscedasticity in the residuals. Nevertheless, it has definitely found some relationship between the regressor series and the response. However, if we see the same plot for InceptionE (Figure 12), we see that InceptionE is nearly always predicting the same value of 0.082. It is likely that careful configuration, tuning and transform of the data may improve InceptionE. However, the same is true for DrCIF and freshPRINCE. We are using InceptionE in the way recommended by its creators [7].
Furthermore, Figure 13 shows the scatter plot of the standardised RMSE used in Figure 7 for DrCIF vs InceptionE. There is no correlation. For context, for our top four regressors, the strongest correlation is between DrCIF and RotF () and DrCIF/freshPRINCE are very weakly correlated (). This diversity suggests there may be some value in ensembling.
VII Conclusions
We have proposed new algorithms for Time Series Extrinsic Regression (TSER) based on classifiers and conducted an extensive experimental study. We have increased the TSER archive size from 19 to 63, introduced improved versions of regressors used in the previous study and shown our new adaptations of classification algorithms are significantly better than the best alternatives. There are several limitations to this study. The data is not randomly sampled (we have taken problems from where ever we can find them) and some domains may be over represented; we have not tuned any of these regressors (the computation required to tune these algorithms over 63 datasets would be prohibitive); we have not looked at more complex diagnostics of performance such as residual analysis. Nevertheless, we believe we have made a significant contribution to advance the new field of TSER. RotF outperforms all previously assessed regressors, and DrCIF and FreshPRINCE are the only TSER algorithms so far proposed to significantly outperform all standard regression algorithms. We have made all our experiments reproducible by releasing structured source code compatible with standard toolkits, guidance on reproducing experiments, and all of our results.
We believe there is scope for further improvement for algorithms for TSER. Adapting supervised Time Series Classification (TSC) approaches may help further leverage of this popular theme for research. InceptionE is the most promising deep learning approach, and perhaps it can be engineered to avoid the catastrophic failure it tends towards with smaller train set sizes. Heterogeneous ensembles are very successful for TSC, and the diversity in performance between the best performing algorithms suggest this may, with careful adaptation, translate to TSER. We will continue to enhance the repository with more problems and would welcome all donations.
Acknowledgments
This work has been partially subsidised by “Agencia Española de Investigación (España)” (grant ref.: PID2020-115454GB-C22 / AEI / 10.13039 / 501100011033). David Guijo-Rubio’s research has been subsidised by the University of Córdoba through grants to Public Universities for the requalification of the Spanish university system of the Ministry of Universities, financed by the European Union - NextGenerationEU (grant reference: UCOR01MS). It is also supported by EPSRC (grant reference EP/W030756/1). Guilherme Arcencio’s research has been subsidised by the São Paulo Research Foundation (FAPESP) (grants references: #2022/12486-4 and #2022/00305-5). Diego Furtado Silva’s research has also been subsidised by FAPESP (grant reference: #2022/03176-1). Some of the experiments were carried out on the High Performance Computing Cluster supported by the Research and Specialist Computing Support service at the University of East Anglia. We would like to thank all those responsible for helping maintain the time series classification archives and those contributing to open source implementations of the algorithms.
Appendix A Data description
The list of all 63 datasets in the archive is shown in Table V. More details on the 44 datasets we have added are given in this section.
Dataset | Train size | Test size | Length | Dims | Missing | Category |
AppliancesEnergy | 95 | 42 | 144 | 24 | No | Energy monitoring |
HouseholdPowerConsumption1 | 745 | 686 | 1440 | 5 | Yes | Energy monitoring |
HouseholdPowerConsumption2 | 745 | 686 | 1440 | 5 | Yes | Energy monitoring |
AustraliaRainfall | 112186 | 48081 | 24 | 3 | No | Environment monitoring |
BeijingPM10Quality | 11918 | 5048 | 24 | 9 | Yes | Environment monitoring |
BeijingPM25Quality | 11918 | 5048 | 24 | 9 | Yes | Environment monitoring |
BenzeneConcentration | 3349 | 5163 | 240 | 8 | Yes | Environment monitoring |
FloodModeling1 | 471 | 202 | 266 | 1 | No | Environment monitoring |
FloodModeling2 | 466 | 201 | 266 | 1 | No | Environment monitoring |
FloodModeling3 | 429 | 184 | 266 | 1 | No | Environment monitoring |
LiveFuelMoistureContent | 3493 | 1510 | 365 | 7 | No | Environment monitoring |
BIDMC32HR | 5550 | 2399 | 4000 | 2 | No | Health monitoring |
BIDMC32RR | 5471 | 2399 | 4000 | 2 | No | Health monitoring |
BIDMC32SpO2 | 5550 | 2399 | 4000 | 2 | No | Health monitoring |
Covid3Month | 140 | 61 | 84 | 1 | No | Health monitoring |
IEEEPPG | 1768 | 1328 | 1000 | 5 | No | Health monitoring |
PPGDalia | 43215 | 21482 | 256-512D | 4 | No | Health monitoring |
NewsHeadlineSentiment | 58213 | 24951 | 144 | 3 | No | Sentiment analysis |
NewsTitleSentiment | 58213 | 24951 | 144 | 3 | No | Sentiment analysis |
DailyOilGasPrices | 133 | 58 | 30 | 2 | No | Economic analysis |
ChilledWaterPredictor | 321 | 138 | 168 | 4 | No | Energy monitoring |
ElectricityPredictor | 567 | 243 | 168 | 4 | No | Energy monitoring |
HotwaterPredictor | 245 | 106 | 168 | 4 | No | Energy monitoring |
OccupancyDetectionLight | 237 | 103 | 60 | 3 | No | Energy monitoring |
SolarRadiationAndalusia | 470 | 202 | 365 | 2 | Yes | Energy monitoring |
SteamPredictor | 210 | 90 | 168 | 4 | No | Energy monitoring |
TetuanEnergyConsumption | 254 | 110 | 144 | 5 | No | Energy monitoring |
WindTurbinePower | 596 | 256 | 144 | 1 | No | Energy monitoring |
AcousticContaminationMadrid | 166 | 72 | 365 | 1 | Yes | Environment monitoring |
AluminiumConcentration | 440 | 189 | 2542 | 1 | No | Environment monitoring |
BeijingIntAirportPM25Quality | 1099 | 472 | 24 | 6 | No | Environment monitoring |
BoronConcentration | 438 | 188 | 2542 | 1 | No | Environment monitoring |
CalciumConcentration | 444 | 191 | 2307 | 1 | No | Environment monitoring |
CopperConcentration | 440 | 189 | 2542 | 1 | No | Environment monitoring |
DailyTemperatureLatitude | 27440 | 11760 | 365 | 1 | No | Environment monitoring |
DhakaHourlyAirQuality | 1447 | 621 | 24 | 1 | No | Environment monitoring |
IronConcentration | 427 | 184 | 1716 | 1 | No | Environment monitoring |
MadridPM10Quality | 4845 | 2077 | 168 | 3 | Yes | Environment monitoring |
MagnesiumConcentration | 1560 | 669 | 3578 | 1 | No | Environment monitoring |
ManganeseConcentration | 427 | 184 | 1716 | 1 | No | Environment monitoring |
MetroInterstateTrafficVolume | 849 | 365 | 24 | 4 | No | Environment monitoring |
ParkingBirmingham | 1391 | 597 | 14-18S | 1 | No | Environment monitoring |
PhosphorusConcentration | 1573 | 675 | 3578 | 1 | No | Environment monitoring |
PotassiumConcentration | 1561 | 669 | 3578 | 1 | No | Environment monitoring |
PrecipitationAndalusia | 470 | 202 | 365 | 4 | Yes | Environment monitoring |
SierraNevadaMountainsSnow | 350 | 150 | 30 | 3 | No | Environment monitoring |
SodiumConcentration | 424 | 183 | 1716 | 1 | No | Environment monitoring |
SulphurConcentration | 444 | 191 | 2307 | 1 | No | Environment monitoring |
ZincConcentration | 445 | 191 | 2307 | 1 | No | Environment monitoring |
ElectricMotorTemperature | 15503 | 6645 | 60 | 6 | No | Equipment monitoring |
GasSensorArrayAcetone | 324 | 140 | 7500 | 1 | No | Equipment monitoring |
GasSensorArrayEthanol | 324 | 140 | 7500 | 1 | No | Equipment monitoring |
LPGasMonitoringHomeActivity | 2017 | 865 | 100 | 1 | No | Equipment monitoring |
MethaneMonitoringHomeActivity | 2017 | 865 | 100 | 1 | No | Equipment monitoring |
WaveDataTension | 1325 | 568 | 57 | 1 | No | Equipment monitoring |
BarCrawl6min | 140 | 61 | 360 | 3 | No | Health monitoring |
Covid19Andalusia | 142 | 62 | 91 | 1 | No | Health monitoring |
VentilatorPressure | 52815 | 22635 | 80 | 2 | No | Health monitoring |
BinanceCoinSentiment | 184 | 79 | 24 | 2 | No | Sentiment analysis |
BitcoinSentiment | 232 | 100 | 24 | 2 | No | Sentiment analysis |
CardanoSentiment | 74 | 33 | 24 | 2 | No | Sentiment analysis |
EthereumSentiment | 249 | 107 | 24 | 2 | No | Sentiment analysis |
NaturalGasPricesSentiment | 65 | 28 | 20 | 1 | No | Sentiment analysis |
For those datasets with unequal-length time series: S indicates that the unequal-length time series are with respect to | ||||||
the samples, whereas D indicates that is with respect to the dimensions. |
A-A Economic analysis
A-A1 Oil and natural gas prices
A dataset published on Kaggle3 consists of historical prices of Brent Oil, Crude Oil WTI, Natural Gas, and Heating Oil from 2000 to 2022. We created the DailyOilGasPrices by using 30 consecutive business days of Crude Oil WTI close prices and traded volumes as predictors and the average natural gas close price during each 30-day time frame as the target variable. The final dataset has 191 2-dimensional time series of length 30, of which 70% were randomly sampled as training data and the remaining 30% as testing data. This type of model could help companies and governments to better analyse and predict economic situations and correlations regarding oil and natural gas.
A-B Energy monitoring
A-B1 ASHRAE - Great Energy Predictor III
This dataset, published on Kaggle4, aims to assess the value of energy efficiency improvements. For that purpose, four types of sources are identified: electricity, chilled water, steam and hot water. The goal is to estimate the energy consumption in kWh.
Dimensions correspond to the air temperature, dew temperature, wind direction and wind speed. These values were taken hourly during a week, and the output is the meter reading of the four aforementioned sources. In this way, we created four datasets: ChilledWaterPredictor, ElectricityPredictor, HotwaterPredictor, and SteamPredictor. Each dataset has a different number of time series as they correspond to different buildings using those sources. In this sense, ChilledWaterPredictor resulted in 459 4-dimensional time series of length 168, ElectricityPredictor resulted in 810 4-dimensional time series of length 168, HotwaterPredictor resulted in 351 4-dimensional time series of length 168, and SteamPredictor resulted in 300 4-dimensional time series of length 168. We randomly sampled 70% of those time series to use as train data and the remaining 30% as test data.
Even though there is a kaggle post indicating that there is one building with meter reading in kBTU they have been transformed into kWh accordingly.
A-B2 Occupancy detection
In a study by Candanedo and Feldheim [45], measurements of temperature, light, , and humidity, collected every minute, were used to detect whether an office room was occupied or not. This data has been made available in the UCI Machine Learning repository20.
We created the OccupancyDetectionLight dataset by reformulating the problem. We used one hour of temperature (in °C), humidity ratio, and concentration (in ppm) as predictors and the average light during that hour (in Lux) as the response variable. This resulted in 340 3-dimensional time series of length 60. We randomly sampled 70% of those time series to use as train data and the remaining 30% as test data. Better models for this data can lead to improvements in energy consumption analysis and prediction.
A-B3 Solar radiation in Andalusia
This dataset has been obtained from the Andalusia Government (Spain)5. Data was retrieved from different stations of the 8 districts of Andalusia: Almeria, Cadiz, Cordoba, Granada, Huelva, Jaen, Malaga, Sevilla, from 2000 until february 2014. The dataset is known as SolarRadiationAndalusia. Dimensions correspond to daily mean of humidity and temperature, whereas the output is the solar radiation for the same day. As time series take daily values during complete years, data from 2014 is not used.
The final dataset includes 672 time series with 2 dimensions with length of 365. The training dataset includes randomly samples 70%, whereas the remaining 30% forms the testing dataset.
A-B4 Energy consumption in Tetuan
This dataset, published on UCI Machine Learning Repository21, aims to estimate the power consumption in three zones in Tetouan [46]. The new dataset is known as TetuanEnergyConsumption. Data has been collected on a ten minute basis. Hence, time series have 144 values (6 values per hour). A total of 5 dimensions have been identified: temperature, humidity, wind speed, general diffuse flows and diffuse flows. The goal is to estimate the daily average power consumption in the three zones of Tetouan.
The aforementioned dataset includes 364 5-dimensional time series of length 144. 70% of those 364 time series have been randomly selected for the training set, whereas the remaining 30% belong to the testing set.
A-B5 Wind turbine power generation
The “Wind Turbine Power (kW) Generation Data” dataset on Kaggle6 consists of large amounts of data collected from a wind turbine, from the temperatures at different parts of the turbines, to the angular position of each blade, and the turbine’s power output. The measurements were made on a 10-minute basis and span from 2019 to 2021.
We used this data to create the WindTurbinePower dataset. Each time series consists of 144 timepoints, i.e. one day of measurements, and its target variable is the turbine’s average power output on that day. From the 76 possible features, we used only the turbine’s torque as a predictor, since it was enough to achieve good results and the other features seemed secondary at best. Some instances were removed due to their respective days having fewer than 144 measurements. The resulting dataset has 852 instances, of which 70% were randomly sampled as training data and the remaining 30% as testing data.
Good regression models for this dataset should improve the logistics of green energy production, by better predicting the energy output of a wind turbine in a given place and/or season.
A-C Environment monitoring
A-C1 Acoustic contamination in Madrid, Spain
This dataset has been made publicly available by the Government of Madrid, Spain7. Data is collected by a number of stations located in the city of Madrid. This dataset is updated daily since 2014. However, we created the AcousticContaminationMadrid dataset with data up to December 2021. The input time series is the LAeq, a fundamental measurement parameter designed to represent a varying sound source over a given time as a single number. Whereas the output time series is the LAS01, the first percentile of sound pressure levels, with A frequency weighting and slow time weighting, recorded during the corresponding period. Examples of such series and outputs are shown in Figure 14.
The final dataset includes 238 univariate time series. Moreover, their length is 365, which corresponds to daily values taken during a year. The training dataset is composed of randomly selected 70% of the samples, whereas the remaining 30% composes the testing dataset.

A-C2 Africa soil chemistry
The Africa Soil Information Service (AfSIS) Soil Chemistry8 dataset contains large amounts of dry and wet chemistry data obtained from soil samples collected from many countries throughout Sub-Saharan Africa, from 2009 to 2013. Dry chemistry analysis, such as infrared spectroscopy and X-ray fluorescence, is comparably less expensive than wet chemistry. Therefore, a model which uses only dry chemistry to predict certain nutrients measured by wet chemical analyses is commercially interesting.
The dry chemical measurements were taken using many different machine models, of which we selected four: “Alpha ZnSe”, “Alpha KBr”, “HTSXT” and “MPA”. The wet chemistry data for each soil sample includes the quantity of 12 nutrients: Aluminium, Boron, Copper, Iron, Manganese, Sodium, Phosphorous, Potassium, Magnesium, Sulphur, Zync and Calcium. Each dry chemical measurement is a time series where the time axis is wavelength and the y-axis is the respective response.
We paired a third of each dry chemistry machine’s experiments to a different nutrient measurement, thus creating 12 datasets. They are AluminiumConcentration (629 cases of length 2542), BoronConcentration (626 cases of length 2542), CopperConcentration (629 cases of length 2542), IronConcentration (611 cases of length 1716), ManganeseConcentration (611 cases of length 1716), SodiumConcentration (607 cases of length 1716), PhosphorusConcentration (2248 cases of length 3578), PotassiumConcentration (2230 cases of length 3578), MagnesiumConcentration (2229 cases of length 3578), SulphurConcentration (635 cases of length 2307), ZincConcentration (636 cases of length 2307), and CalciumConcentration (635 cases of length 2307).
In each dataset, 70% of cases were randomly sampled as training data and the remaining 30% as testing data. Three example soil spectrograms are shown in Figure 1.
A-C3 Beijing Airport PM2.5 contamination
This dataset was obtained from the UCI Machine Learning Repository22. The authors [47], collected hourly data containing the data of US Embassy in Beijing, as well as meteorological data from Beijing Capital International Airport. This dataset, known as BeijingIntAirportPM25Quality, includes 6-dimensional time series of 24 points. The dimensions are the dew point, temperature, pressure, combined wind direction, and accumulated hours of snow and rain measured, as mentioned, in the Beijing Capital International Airport. The output is the data averaged daily.
The aforementioned dataset includes 1571 6-dimensional time series of length 24. 70% of those 1571 time series have been randomly selected for the training set, whereas the remaining 30% belong to the testing set.
A-C4 Daily temperature and latitude
A dataset published on Kaggle9 contains daily temperature data for the 1000 most populous cities in the world, along with their geographic coordinates, from 1980 to 2020.
We used this data to create the DailyTemperatureLatitude dataset. We split each city’s temperature data into 1 year long time series, i.e. 365 timepoints. Leap years were shortened by averaging the temperatures on the 28th and 29th of February. The predictors are the daily temperatures (in °C) recorded during the year and the response variable is the corresponding city’s latitude. The final dataset has 39200 univariate time series of length 365, of which 70% were randomly sampled as training data and the remaining 30% as testing data. Two samples of the constructed dataset are shown in Figure 15.
With exploratory analysis and regression on this dataset, climate change and its effects can be better understood and predicted on a local basis.

A-C5 Air quality in Dhaka
Data sourced from AirNow23 and made available on Kaggle10 comprises 7 years of hourly measurements of fine particulate matter () concentrations at the United States Embassy in Dhaka, Bangladesh, along with each corresponding Air Quality Index (AQI).
We used this data to create the DhakaHourlyAirQuality dataset, in which the predictors are 24 hours of concentrations and the response variable is the average AQI on that respective day. Thus, better models can be created and tested to more cheaply and quickly evaluate air quality in different cities. The resulting dataset consists of 2068 univariate time series of length 24, of which 70% were randomly sampled as training data and the remaining 30% as testing data.
A-C6 Madrid PM10 contamination
This dataset is publicly available on Kaggle11, even though the data was originally published in the public repository of the Madrid Government24. The dataset is known as MadridPM10Quality. Its input data consists in measurements of the level of sulphur dioxide, carbon monoxide and nitric oxide. Time series correspond to hourly values measured during a week. The output value is the weekly averaged .
The final dataset includes 6922 3-dimensional time series with a length of 168. The training dataset is composed of randomly selected 70% of the samples, whereas the remaining 30% forms the testing dataset.
A-C7 Metro Interstate Traffic Volume
This dataset is publicly available on the UCI Machine Learning Repository12. Data consists of hourly traffic volume for MN DoT ATR station 301, roughly midway between Minneapolis and St Paul, collected from 2012 to 2018. The dataset, known as MetroInterstateTrafficVolume, aims to estimate the daily average traffic volume for the aforementioned road. The input dimensions consists of the average temperature, the amount of rain and snow, and the percentage of cloud cover.
The final dataset includes 1214 time series with 4 dimensions. Moreover, their length is 24, which corresponds to hourly measures taken during a day. The training dataset includes a randomly selected 70% of the whole data, whereas the remaining 30% forms the testing dataset.
A-C8 Parking Birmingham
This dataset is publicly available on the UCI Machine Learning Repository25. The authors [48] collected data from car parks in Birmingham (United Kingdom) operated by National Car Parks from Birmingham City Council. The dataset, known as ParkingBirmingham, aims to estimate occupancy rates from 2016/10/04 to 2016/12/19. Input time series is the number of parked cars every hour, whereas the output is the occupancy rate. The total number of hours measured per day varies from 14 to 18.
The aforementioned dataset includes 1988 unequal length time series (with lengths between 14 and 18). 70% of those 1988 time series have been randomly selected for the training set, whereas the remaining 30% belong to the testing set.
A-C9 Precipitation in Andalusia
This dataset has been obtained from the Andalusia Government (Spain)13. Data was retrieved from different stations in the 8 districts of Andalusia: Almeria, Cadiz, Cordoba, Granada, Huelva, Jaen, Malaga, and Sevilla, from 2000 until February 2014. This dataset, known as PrecipitationAndalusia, includes the daily averaged temperature, humidity, and wind speed and direction as inputs, whereas the output is the average precipitation. Two examples are illustrated in Figure 16.
This resulted in 672 4-dimensional time series of length 365. We randomly sampled 70% of those time series to use as train data and the remaining 30% as test data.

A-C10 Sierra Nevada mountains measurements
A dataset published by Osterhuber and Schwartz [49] and made available on Kaggle26 consists of daily measurements of minimum and maximum air temperatures, precipiation, and snowpack characteristics made at a field station in Sierra Nevada, United States, between 1971 and 2019.
We split the measurements into groups of 30 consecutive days and used the minimum and maximum air temperature (in °C) and precipitation (in mm) to create 3-dimensional time series of length 30. Using the amount of new snow (in cm) accumulated during each respective 30-day timeframe as the target variable, we created the SierraNevadaMountainsSnow dataset, with 500 instances in total. Around 20 instances were removed due to missing values. The dataset is split into train and test sets by randomly sampling 30% of the data as test. This data can be used to train models to predict heavy snowfall and prepare cities and roads for harsh weather conditions.
A-D Equipment monitoring
A-D1 Electric motor temperature
Kirchgässner, Wallscheid and Böcker [59] collected large amounts of sensor data, made available on Kaggle14, from a permanent magnet synchronous motor (PMSM) deployed on a test bench. The dataset consists of multiple measurement sessions, each ranging between one and six hour long, whose recordings were sampled at 2Hz. The sensors collected a variety of features, such as current and voltage, ambient, coolant, and rotor temperatures, and motor speed. Rotor temperature, specifically, is not reliably and economically measurable in a commercial vehicle, thus being an interesting candidate for response variable. The data is, therefore, useful for industrial processes and monitoring.
Therefore, we created the ElectricMotorTemperature dataset by first splitting the measurement sessions into groups of 30 consecutive seconds, i.e. 60 timepoints. Then, we used the recorded ambient and coolant temperatures and d and q components of voltage and current as predictors to form 6-dimensional time series of length 60. The target variable is the maximum recorded rotor temperature during each respective 30-second time frame. An example of such a time series is shown in Figure 17. The resulting dataset has 22148 instances, of which 70% were sampled as training data and the remaining 30% as testing data.

A-D2 Home activity monitoring of gases
This dataset has been obtained from UCI Machine Learning Repository27. The authors [50] collected recordings of a gas sensor array composed of 8 MOX gas sensors, as well as a temperature and humidity sensor. Sensors were exposed to three different conditions: presence of wine, banana and background activity. Two datasets have been collated from this repository: LPGasMonitoringHomeActivity and MethaneMonitoringHomeActivity. The first estimates the liquefied petroleum gas concentration from humidity measurements. On the other hand, the second one estimates the methane concentration from temperature measurements. The latter is illustrated in Figure 18.
This resulted in 2882 univariate time series of length 100. We randomly sampled 70% of those time series to use as train data and the remaining 30% as test data.

A-D3 Gas sensor array under flow modulation
Ziyatdinov et al. [51] combined an array of 16 metal-oxide gas sensors and an external mechanical ventilator to simulate sniffling behaviour within the biological respiration cycle. The study extracted high and low frequency features from the signals and proposed a regression problem where the predictors are either of those features and the responses are the concentrations of two analytes used to form test gasses, acetone and ethanol. The data collected by the authors has been made available in the UCI Machine Learning repository28. The development of better regression techniques for this data should lead to improvements in early detection of gases in chemo-sensory systems.
We used the raw signal data with 58 samples, each one consisting of 16 sensors and totalling 928 time series, to create two datasets: GasSensorArrayEthanol, in which the target variable is the ethanol concentration in the tested gas, and GasSensorArrayAcetone, in which the target is similarly the concentration of acetone. Both consist of 464 univariate time series of length 7500 and are split into train and test sets by randomly selecting 30% of instances as test.
A-D4 Wave elevation and line tension
A dataset published on Kaggle15 consists of a simulation of a ship. The dataset’s goal is to predict the tension of a string given temporal wave elevation data. While the data is simulated, well-fitted models should improve monitoring of naval equipment during harsh conditions.
Thus, we created the WaveTensionData dataset by separating the source data into univariate time series of length 57 and using wave height as the predictor and the corresponding average string tension as the target variable. The resulting dataset has 1893 instances, of which 70% were sampled as training data and the remaining 30% as testing data.
A-E Health monitoring
A-E1 Bar Crawl: Detecting Heavy Drinking Episodes
This dataset, made available in the UCI Machine Learning repository29 by their authors [52], consists in predicting heavy drinking episodes via mobile data. Data was collected from smartphones from 13 participants. The goal is to estimate the transdermal alcohol content by using an accelerometer. We have, therefore, created the BarCrawl6min dataset, where each input dimension corresponds to a different axis of the accelerometer. Moreover, even though data was recorded at 30 minutes intervals, in order to accurately estimate the drinking episode, only the last 6 minutes from the recording were kept. Two resulting samples are shown in Figure 19. Note that all data is fully anonymised and that TAC readings were preprocessed/cleaned by the authors.
This resulted in 201 3-dimensional time series of length 360, which correspond to 6 minutes of secondly measurements. We randomly sampled 70% of those time series to use as train data and the remaining 30% as test data.

A-E2 Covid-19 in Andalusia
This dataset consists of estimating the mortality rate during Covid-19 waves and in different districts in eight different areas of Andalusia, its country’s second largest and most populated autonomous region, located in southern Spain. The output is the number of deaths in proportion to the total number of infected people in that district. This dataset, known as Covid19Andalusia, has been made public by the authors of the work [53], who took the data available in the Andalusia Government Website30. This dataset comprises 6 waves from a total of 34 districts. All waves are equal-length (91 points) since we have considered 45 days before and after the peak of the outbreak, as it has been demonstrated to be the most relevant data. Two such waves are illustrated in Figure 20.
The aforementioned dataset includes 204 unidimensional time series of length 91. 70% of those 204 time series have been randomly selected for the training set, whereas the remaining 30% belong to the testing set.

A-E3 Pressure of a ventilator connected to a sedated patient’s lung
This dataset was made available by Kaggle in collaboration with Google Brain16. Data used in this competition was produced by connecting a ventilator to an artificial bellows test lung through a respiratory circuit. The goal is to estimate the pressure, a value ranging from 0 to 100, representing how much the inspiratory solenoid valve is open to let air into the lung. The proposed dataset, named VentilatorPressure, includes time series corresponding to approximately 3-second breaths. Dimensions are the control input and output for the inspiratory solenoid valve.
The final dataset includes 75450 time series with 2 dimensions and length 80. The training dataset comprises randomly sampled 70% of the series, whereas the remaining 30% belong to the testing dataset.
A-F Sentiment analysis
A-F1 Cryptocurrency sentiment
By combining historical sentiment data for 4 cryptocurrencies, extracted from EODHistoricalData31 and made available on Kaggle17, with historical price data for the same cryptocurrencies, extracted from CryptoDataDownload32, we created the BitcoinSentiment, EthereumSentiment, CardanoSentiment, and BinanceCoinSentiment datasets, with 332, 356, 107, and 263 total instances, respectively.
In all four datasets, the predictors are hourly close price (in USD) and traded volume for each respective cryptocurrency during a day, resulting in 2-dimensional time series of length 24. The response variable is the normalized sentiment score on the day spanned by the timepoints. The datasets were split into train and test sets by randomly selecting 30% of each set as test data. Using this data, companies can better prepare for shifts of public perception regarding cryptocurrencies.
A-F2 Sentiment on natural gas prices
Natural gas prices historical data was taken from the U.S. Energy Information Administration33 along with corresponding sentiment scores obtained by analysing relevant tweets on the topic. From this data, we created the NaturalGasPricesSentiment dataset.
We first split the data into groups of 20 consecutive business days. We then used the daily natural gas prices as predictors and the average sentiment score during each 20-day time frame as the response variable. The final dataset has 93 univariate time series of length 20, of which 70% were randomly sampled as training data and the remaining 30% as testing data. Two of those time series are shown in Figure 21. Again, companies and local governments can use the data to analyse and predict shifts in public perception on natural gas.

Appendix B Regressor configurations
Parameter settings for all algorithms is shown in Table VI.
Regressor | Configuration |
---|---|
Ridge | Regularization strength: |
grid-SVR | Kernel |
Regulatisation parameter | |
Kernel Coefficient | |
RandF | Num. Estimators: |
RotF | Num. Estimators: |
XGBoost | Num. Estimators: |
Learning Rate: | |
FPCR | FPCA Num. Components: |
FPCR-Bs | Smooth: B-splines |
FPCA Num. Components: | |
Num. Basis Functions: | |
Order: | |
ROCKET | Num. Kernels: |
MultiROCKET | Num. Kernels: |
Max. Num. Dilations per Kernel: | |
Num. Features per Kernel: | |
NN-ED | Num. Neighbours: — |
NN-ED | Distance measure: Euclidean |
NN-DTW | Num. Neighbours: — |
NN-DTW | Distance measure: DTW |
FreshPRINCE | Num. Estimators: |
TSF | Num. Estimators: |
Num. Intervals per Tree: | |
DrCIF | Num. Estimators: |
Num. Intervals per Representation: | |
Num. Features per Tree: | |
CNN | Num. Epochs: |
Batch size: | |
Kernel size: | |
Num. Convolutional Layers: | |
FCN | Num. Epochs: |
Batch size: | |
Num. Components: (inceptionE: ) | |
Inception | Num. Epochs: |
InceptionE | Batch size: |
Kernel size: | |
ResNet | Num. Epochs: |
Batch size: |
Appendix C Results
RMSE results for the best performing regressors on 63 TSER datasets are available in the accompaning website343434https://tsml-eval.readthedocs.io/en/latest/publications/2023/tser_archive_expansion/tser_archive_expansion.html..
References
- [1] A. Bagnall, J. Lines, A. Bostrom, J. Large, and E. Keogh, “The great time series classification bake off: a review and experimental evaluation of recent algorithmic advances,” Data Mining and Knowledge Discovery, vol. 31, no. 3, pp. 606–660, 2017.
- [2] M. Middlehurst, P. Schäfer, and A. Bagnall, “Bake off redux: a review and experimental evaluation of recent time series classification algorithms,” arXiv preprint arXiv:2304.13029, 2023.
- [3] A. P. Ruiz, M. Flynn, J. Large, M. Middlehurst, and A. Bagnall, “The great multivariate time series classification bake off: a review and experimental evaluation of recent algorithmic advances,” Data Mining and Knowledge Discovery, vol. 35, no. 2, pp. 401–449, 2021.
- [4] S. Makridakis, S. C. Wheelwright, and R. J. Hyndman, Forecasting methods and applications. John wiley & sons, 2008.
- [5] C. W. Tan, C. Bergmeir, F. Petitjean, and G. I. Webb, “Time series extrinsic regression,” Data Mining and Knowledge Discovery, vol. 35, no. 3, pp. 1032–1060, 2021.
- [6] A. Dempster, F. Petitjean, and G. I. Webb, “Rocket: exceptionally fast and accurate time series classification using random convolutional kernels,” Data Mining and Knowledge Discovery, vol. 34, no. 5, pp. 1454–1495, 2020.
- [7] H. Ismail Fawaz, B. Lucas, G. Forestier, C. Pelletier, D. F. Schmidt, J. Weber, G. I. Webb, L. Idoumghar, P.-A. Muller, and F. Petitjean, “Inceptiontime: Finding alexnet for time series classification,” Data Mining and Knowledge Discovery, vol. 34, no. 6, pp. 1936–1962, 2020.
- [8] C. W. Tan, A. Dempster, C. Bergmeir, and G. I. Webb, “Multirocket: multiple pooling operators and transformations for fast and effective time series classification,” Data Mining and Knowledge Discovery, vol. 36, no. 5, pp. 1623–1646, 2022.
- [9] M. Middlehurst and A. Bagnall, “The freshprince: A simple transformation based pipeline time series classifier,” in International Conference on Pattern Recognition and Artificial Intelligence. Springer, 2022, pp. 150–161.
- [10] M. Christ, N. Braun, J. Neuffer, and A. W. Kempa-Liehr, “Time series feature extraction on basis of scalable hypothesis tests (tsfresh–a python package),” Neurocomputing, vol. 307, pp. 72–77, 2018.
- [11] J. J. Rodriguez, L. I. Kuncheva, and C. J. Alonso, “Rotation forest: A new classifier ensemble method,” IEEE transactions on pattern analysis and machine intelligence, vol. 28, no. 10, pp. 1619–1630, 2006.
- [12] H. Deng, G. Runger, E. Tuv, and M. Vladimir, “A time series forest for classification and feature extraction,” Information Sciences, vol. 239, pp. 142–153, 2013.
- [13] M. Middlehurst, J. Large, and A. Bagnall, “The canonical interval forest (cif) classifier for time series classification,” in 2020 IEEE international conference on big data (big data). IEEE, 2020, pp. 188–195.
- [14] M. Middlehurst, J. Large, M. Flynn, J. Lines, A. Bostrom, and A. Bagnall, “Hive-cote 2.0: a new meta ensemble for time series classification,” Machine Learning, vol. 110, no. 11, pp. 3211–3243, 2021.
- [15] C. H. Lubba, S. S. Sethi, P. Knaute, S. R. Schultz, B. D. Fulcher, and N. S. Jones, “catch22: Canonical time-series characteristics,” Data Mining and Knowledge Discovery, vol. 33, no. 6, pp. 1821–1852, 2019.
- [16] J. Goldsmith and F. Scheipl, “Estimator selection and combination in scalar-on-function regression,” Computational Statistics & Data Analysis, vol. 70, pp. 362–372, 2014.
- [17] P. T. Reiss, J. Goldsmith, H. L. Shang, and R. T. Ogden, “Methods for scalar-on-function regression,” International Statistical Review, vol. 85, no. 2, pp. 228–249, 2017.
- [18] H. Drucker, C. J. Burges, L. Kaufman, A. Smola, and V. Vapnik, “Support vector regression machines,” Advances in neural information processing systems, vol. 9, 1996. [Online]. Available: https://proceedings.neurips.cc/paper/1996/file/d38901788c533e8286cb6400b40b386d-Paper.pdf
- [19] L. Breiman, “Random forests,” Machine learning, vol. 45, no. 1, pp. 5–32, 2001.
- [20] T. Chen and C. Guestrin, “Xgboost: A scalable tree boosting system,” in Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, 2016, pp. 785–794.
- [21] Z. Wang, W. Yan, and T. Oates, “Time series classification from scratch with deep neural networks: A strong baseline,” in Proceedings of the IEEE international joint conference on neural networks, 2017, pp. 1578–1585.
- [22] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
- [23] A. Gu, I. Johnson, K. Goel, K. Saab, T. Dao, A. Rudra, and C. Ré, “Combining recurrent, convolutional, and continuous-time models with linear state space layers,” in Advances in Neural Information Processing Systems, M. Ranzato, A. Beygelzimer, Y. Dauphin, P. Liang, and J. W. Vaughan, Eds., vol. 34. Curran Associates, Inc., 2021, pp. 572–585. [Online]. Available: https://proceedings.neurips.cc/paper/2021/file/05546b0e38ab9175cd905eebcc6ebb76-Paper.pdf
- [24] R. Hasani, M. Lechner, T.-H. Wang, M. Chahine, A. Amini, and D. Rus, “Liquid structural state-space models,” 2022.
- [25] S. Bloemheuvel, J. van den Hoogen, D. Jozinović, A. Michelini, and M. Atzmueller, “Graph neural networks for multivariate time series regression with application to seismic data,” International Journal of Data Science and Analytics, 2022.
- [26] M. A. Siddiquee, V. M. A. Souza, G. E. Baker, and A. Mueen, “Septor: Seismic depth estimation using hierarchical neural networks,” in Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2022, p. 3889–3897.
- [27] M. Bayani, “Essays on machine learning methods in economics,” Ph.D. dissertation, City University of New York, 2022. [Online]. Available: https://academicworks.cuny.edu/cgi/viewcontent.cgi?article=6069&context=gc_etds
- [28] M. Shokoohi-Yekta, B. Hu, H. Jin, J. Wang, and E. Keogh, “Generalizing DTW to the multi-dimensional case requires an adaptive approach,” Data Mining and Knowledge Discovery, vol. 31, no. 1, pp. 1–31, 2017.
- [29] C. W. Tan, A. Dempster, C. Bergmeir, and G. Webb, “MultiRocket: multiple pooling operators and transformations for fast and effective time series classification,” Data Mining and Knowledge Discovery, vol. 36, p. 1623–1646, 2022.
- [30] H. Ismail Fawaz, G. Forestier, J. Weber, L. Idoumghar, and P.-A. Muller, “Deep learning for time series classification: a review,” Data Mining and Knowledge Discovery, vol. 33, no. 4, pp. 917–963, 2019.
- [31] B. Zhao, H. Lu, S. Chen, J. Liu, and D. Wu, “Convolutional neural networks for time series classification,” Journal of Systems Engineering and Electronics, vol. 28, no. 1, pp. 162–169, 2017.
- [32] A. Bostrom and A. Bagnall, “Binary shapelet transform for multiclass time series classification,” Transactions on Large-Scale Data and Knowledge Centered Systems, vol. 32, pp. 24–46, 2017.
- [33] L. Ye and E. Keogh, “Time series shapelets: a novel technique that allows accurate, interpretable and fast classification,” Data Mining and Knowledge Discovery, vol. 22, no. 1-2, pp. 149–182, 2011.
- [34] P. Schäfer, “The BOSS is concerned with time series classification in the presence of noise,” Data Mining and Knowledge Discovery, vol. 29, no. 6, pp. 1505–1530, 2015.
- [35] P. Schäfer and U. Leser, “Weasel 2.0–a random dilated dictionary transform for fast, accurate and memory constrained time series classification,” arXiv preprint arXiv:2301.10194, 2023.
- [36] M. Middlehurst, J. Large, G. Cawley, and A. Bagnall, “The temporal dictionary ensemble (TDE) classifier for time series classification,” in Proceedings of European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ser. Lecture Notes in Computer Science, vol. 12457, 2020, pp. 660–676.
- [37] M. Christ, N. Braun, J. Neuffer, and A. W. Kempa-Liehr, “Time series feature extraction on basis of scalable hypothesis tests (tsfresh–a python package),” Neurocomputing, vol. 307, pp. 72–77, 2018.
- [38] A. Bagnall, M. Flynn, J. Large, J. Line, A. Bostrom, and G. Cawley, “Is rotation forest the best classifier for problems with continuous features?” arXiv preprint arXiv:1809.06705, 2018.
- [39] C. Pardo, J. F. Diez-Pastor, C. García-Osorio, and J. J. Rodríguez, “Rotation forests for regression,” Applied Mathematics and Computation, vol. 219, no. 19, pp. 9914–9924, 2013.
- [40] L. Breiman, J. Friedman, R. Olshen, and C. Stone, “Classification and regression trees,” 1984.
- [41] N. Cabello, E. Naghizade, J. Qi, and L. Kulik, “Fast and accurate time series classification through supervised interval search,” in 2020 IEEE International Conference on Data Mining (ICDM). IEEE, 2020, pp. 948–953.
- [42] M. Flynn, J. Large, and A. Bagnall, “The contract random interval spectral ensemble (c-rise): the effect of contracting a classifier on accuracy,” in International Conference on Hybrid Artificial Intelligence Systems. Springer, 2019, pp. 381–392.
- [43] E. J. Keogh and M. J. Pazzani, “Derivative dynamic time warping,” in Proceedings of the 2001 SIAM international conference on data mining. SIAM, 2001, pp. 1–11.
- [44] B. D. Fulcher and N. S. Jones, “hctsa: A computational framework for automated time-series phenotyping using massive feature extraction,” Cell Systems, vol. 5, no. 5, pp. 527–531, 2017.
- [45] L. M. Candanedo and V. Feldheim, “Accurate occupancy detection of an office room from light, temperature, humidity and measurements using statistical learning models,” Energy and Buildings, vol. 112, pp. 28–39, 2016.
- [46] A. Salam and A. El Hibaoui, “Comparison of machine learning algorithms for the power consumption prediction:-case study of tetouan city–,” in 2018 6th International Renewable and Sustainable Energy Conference (IRSEC). IEEE, 2018, pp. 1–5.
- [47] X. Liang, T. Zou, B. Guo, S. Li, H. Zhang, S. Zhang, H. Huang, and S. X. Chen, “Assessing beijing’s pm2. 5 pollution: severity, weather impact, apec and winter heating,” Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, vol. 471, no. 2182, p. 20150257, 2015.
- [48] D. H. Stolfi, E. Alba, and X. Yao, “Predicting car park occupancy rates in smart cities,” in Smart Cities: Second International Conference, Smart-CT 2017. Springer, 2017, pp. 107–117.
- [49] R. Osterhuber and A. Schwartz, “Snowpack, precipitation, and temperature measurements at the Central Sierra Snow Laboratory for water years 1971 to 2019,” 2021, dryad.
- [50] R. Huerta, T. Mosqueiro, J. Fonollosa, N. F. Rulkov, and I. Rodriguez-Lujan, “Online decorrelation of humidity and temperature in chemical sensors for continuous monitoring,” Chemometrics and Intelligent Laboratory Systems, vol. 157, pp. 169–176, 2016.
- [51] A. Ziyatdinov, J. Fonollosa, L. Fernández, A. Gutierrez-Gálvez, S. Marco, and A. Perera, “Bioinspired early detection through gas flow modulation in chemo-sensory systems,” Sensors and Actuators B: Chemical, vol. 206, pp. 538–547, 2015.
- [52] J. A. Killian, K. M. Passino, A. Nandi, D. R. Madden, J. D. Clapp, N. Wiratunga, F. Coenen, and S. Sani, “Learning to detect heavy drinking episodes using smartphone accelerometer data.” in KHD@ IJCAI, 2019, pp. 35–42. [Online]. Available: https://ceur-ws.org/Vol-2429/paper6.pdf
- [53] M. Díaz-Lozano, D. Guijo-Rubio, P. A. Gutiérrez, A. M. Gómez-Orellana, I. Túñez, L. Ortigosa-Moreno, A. Romanos-Rodríguez, J. Padillo-Ruiz, and C. Hervás-Martínez, “Covid-19 contagion forecasting framework based on curve decomposition and evolutionary artificial neural networks: A case study in andalusia, spain,” Expert Systems with Applications, vol. 207, p. 117977, 2022.
- [54] R. Ghosh, “Natural gas prices with Twitter sentiment scores,” 2022, kaggle.
- [55] J. Demšar, “Statistical comparisons of classifiers over multiple data sets,” Journal of Machine Learning Research, vol. 7, pp. 1–30, 2006.
- [56] S. García and F. Herrera, “An extension on “statistical comparisons of classifiers over multiple data sets” for all pairwise comparisons,” Journal of Machine Learning Research, vol. 9, pp. 2677–2694, 2008. [Online]. Available: https://www.jmlr.org/papers/v9/garcia08a.html
- [57] A. Benavoli, G. Corani, and F. Mangili, “Should we really use post-hoc tests based on mean-ranks?” Journal of Machine Learning Research, vol. 17, pp. 1–10, 2016. [Online]. Available: https://jmlr.org/papers/v17/benavoli16a.html
- [58] I. Yeo and R. A. Johnson, “A new family of power transformations to improve normality or symmetry,” Biometrika, vol. 87, no. 4, pp. 954–959, 2000.
- [59] W. Kirchgässner, O. Wallscheid, and J. Böcker, “Estimating electric motor temperatures with deep residual machine learning,” IEEE Transactions on Power Electronics, vol. 36, no. 7, pp. 7480–7488, 2021.
![]() |
David Guijo-Rubio received the BSc degree in Computer Science in 2016 and the MSc degree in Artificial Intelligence in 2017. In 2021, he received the PhD degree in Computer Science from the University of Córdoba (Spain). He is working as an Postdoctoral Researcher at both the University of Córdoba and the University of East Anglia (UK). His current interests include ordinal classification, pattern recognition and different tasks applied to time series (classification -nominal and ordinal, clustering and extrinsic regression). |
![]() |
Matthew Middlehurst received the BSc degree in Computer Science in 2018 and the PhD degree in Computer Science in 2023 from the University of East Anglia (UEA), Norwich, UK. He is working as a Senior Research Associate at UEA in the time series machine learning group. His research interests are primarily time series classification related, with recent expansion to time series clustering and time series extrinsic regression. He is interested in the provision and maintenance of open source software for researchers in time series based fields. |
![]() |
Guilherme Arcencio is pursuing the BSc in Computer Engineering from the Federal University of São Carlos (UFSCar), São Carlos, Brazil. He worked as a Research Intern at the University of East Anglia (UEA) in the time series machine learning group. His current research interests include time series machine learning with a focus on extrinsic regression, as well as the development and maintenance of open source software. |
![]() |
Diego Furtado Silva received the BSc degree in Computer Science at the Institute of Mathematics and Computer Sciences (ICMC), University of São Paulo (USP), São Carlos, as well as the MSc and PhD degree in Computer Science and Computational Mathematics, in time series and digital signal analysis. Currently, he is an Assistant Professor at ICMC-USP and collaborate with the Computational Intelligence Laboratory (LABIC) and the Data Mining and Application Group at the Federal University of São Carlos (MIDAS-UFSCar). His main research interests are machine learning foundations and applications. |
![]() |
Anthony Bagnall received the PhD degree in Computer Science from the University of East Anglia (UEA), Norwich, UK in 2001. Since 2018 he has been a Full Professor of Computer Science at UEA, where he leads the time series machine learning group. His primary research interest is in time series machine learning, with a historic focus on classification, but more recently looking at clustering and regression. He has a side interest in ensemble design. |