Course: L4-TS Introduction to Time Series Analysis

This course introduces the main concepts behind Time Series Analysis, with an emphasis on forecasting applications: data cleaning, missing value imputation, time-based aggregation techniques, creation of a vector/tensor of past values, descriptive analysis, model training (from simple basic models to more complex statistics and machine learning based models), hyperparameter optimization, and model evaluation.

Learn how to implement all these steps using real-world time series datasets. Put what you’ve learnt into practice with the hands-on exercises.

This course consists of four, 90 minutes online sessions run by Professor Daniele Tonini and two of our KNIME data scientists. Each session has an exercise for you to complete at home. The course concludes with a 15 to 30 minute wrap up session.

  • Session 1: Introduction to Time Series Analysis and KNIME Components
  • Session 2: Understanding Stationarity, Trend and Seasonality
  • Session 3: Naive Method, ARIMA models, Residual Analysis
  • Session 4: Machine Learning, Model Optimization, Deployment
  • Session 5: Recap and final Q&A

Instructors:
Prof. Daniele Tonini (University of Bocconi)
Corey Weisinger
Maarit Widmann

2 Likes

Hi,

I enjoyed the first session quite alot and I’m looking forward to the deep dives!

Out of curiosity:
Is there a plan to develop the time series nodes as generic KNIME nodes, as well? What is the advantage of using components rather than programming them as generic nodes?

Personally, I am used to python, but it is always a bit cumbersome when introducing colleagues to a low coding tool KNIME and then explain that they would also need to install scripting libraries for certain features :smiley:

Hi, thank you for your questions and ideas.

In the long term we hope to provide more and more time series functionalities as KNIME native nodes. In the short term, we’re developing and enhancing our time series components.

Components are KNIME nodes that encapsulate functionalities built by other KNIME nodes. For a data scientist, building a workflow or writing a script is often easy, but some other types of users prefer a component with clearly defined task and configuration settings in the configuration dialog. On the other side, if you’d like to, you can also customize the functionalities of the components. For example, you as a Python programmer could add functionalities to our time series components that are based on Python code.

You’re right, it’s not ideal at the moment that the Python libraries need to be installed separately. But at the moment this is required to excute Python code from within KNIME without writing any code.

Some of you were asking for additional materials about clustering time series. See two good references below:


1 Like

Hi, in Exercise 1 the “Missing Value” node produces an error message:

The current settings use missing value handling methods that cannot be represented in PMML 4.2

Could that be prevented? And what is causing this?

Thanks
Thomas

The one-pager for setting up the python environment is very useful. It is probably beyond the scope of this great course, but would TSA with R require a similar setup? If so, is there such a one-pager or cheat sheet available for using R?

Thanks

Thomas

1 Like

Hi Thomas, this is because not all techniques for missing value handling available in the missing value node can be exported as PMML (the model output of the node). If you configured the node to use such a technique, you will get this notification, but it’s only a warning, not an error.

1 Like

Hi Thomas, many Windows users say integrating R statistics in KNIME is more straightforward than integrating Python, and indeed it requires fewer steps, with the local installation of R and Rserve package installed. You can check the required steps for installing R Statistics Integration in KNIME on this documentation page: https://docs.knime.com/2019-12/r_installation_guide/index.html

2 Likes

In yesterday’s session you asked about anomaly detection techniques. Below a few links to example workflows on the KNIME Hub:

Anomaly Detection. Control Chart
Simple Anomaly Detection Using a Convolutional Network
Anomaly Detection with Time Series Analysis
Example workflows for fraud detection with Random Forest, Autoencoder, and Isolation Forest

1 Like

Hi guys,

Few more questions from my side.

ARIMA Modelling:
About decomposing a time series, what would be the best strategy here? Decompose it myself or just let ARIMA do the job? My first impression is that an automated ARIMA (learner) does a much better job here? Manual decomposition would be rather a technique to analyze the time series and describe it?

Models for heteroskedasticity:
ARIMA does not assume shocks in other statistical moments, e.g. Variance, am I right? The ARCH model “zoology” would come into play here. Are there also components covering ARCH processes? But similar to the ARIMA universe, one could just use ML and Deep Learning models for these cases as well?

Components in general (off topic):
If a component gets uploaded by the community to the hub, is there a kind of quality gate, that tests the component for functionality and also for the theory behind it?

Machine Learning & Deep Learning:
Also my (beginner’s) first impression about ML is, that ML beats classical models by haveing a better fit to the data and also much less restrictions / assumptions when implementing them. What speaks for classical models, such as ARIMA then? I guess one point would be the analytical traceability? Are there cases where classical and parsimonious models beat ML models?

Deep Learning (maybe a more philosophical question):
The power of DL is quite impressive. However, analytical traceability is limited here. I am a little bit worried about users/colleagues (including myself) using it negligently and forming false conclusions about the reality by trusting DL algorithms too much. What would be here the best strategy, when to apply Deep Learning techniques? For example, if the DL model performs only slightly better than a parsimonious ML model, wouldn’t it be more reasonable to go with the parsimonious model, where analytics and math behind the model can be better understood?

Thanks a lot for the great sessions. They were really insightful!

1 Like

Hi, here answers to your questions. We’re happy to elaborate them in today’s Q&A session if needed.

ARIMA Modeling
ARIMA models assume stationary series, and therefore we first manually decompose the series into a trend, first and second seasonalities, and residual, which is supposed to be stationary. ARIMA model can only make a time series stationary by first order differencing (I order 1), maximum two times (I order 2), but it cannot perform the seasonal differencing like we perform manually at lags 24 and 168. Besides that, the manual decomposition is also used to inspect and describe the time series.

Models for heteroscedasticity
Unfortunately, we don’t have components for ARCH models

Components in general
Components uploaded by the KNIME community are not quality checked by KNIME. Components uploaded by KNIME (those also available on the EXAMPLES Server) are tested regularly.

Machine Learning & Deep Learning
Classical models often work the best with small amounts of data available, and with frequent changes in the data

Deep Learning
Deep learning techniques have the disadvantage of low interpretability and slow computation. If these are not an issue in your use case, the deep learning model might be the right choice. However, like you said, the highest accuracy is not always the most important criterion when selecting a model. Deep learning models also require a lot of training data to perform well. If you have a time series with frequently changing dynamics, enough training data are often not available.

Hi guys,

@Maarit: Thanks alot for the answers! Things became much clearer now.

I had the chance now to put the components into action and few more questions popped up.

  1. Is there a possibility to check if a timeseries itself is stationary? A common statistical tool would be the (augmented) dickey fuller test. Is there an ADF node already available?

  2. I have a time series, which I log differenced and it looks quite stationary to me:

I applied the decompose node to further reduce seasonality and trend and ACF and PACF decay initially into the 95% CI Bounds. What worries me a little is the PACF spike at leg 11, 12 and 16. How can I interpret it and do I have to remove it before applying ARIMA? And how could I remove it? See screen below (left ACF, middle PACF and right residual plot -> residual after decomposing)

  1. Let’s assume the time series is optimal for ARIMA modelling and the auto Arima Model suggests the best fit Model. In my case this is an ARMA (0,4) and the ARIMA residuals are stationary (see screen below).

I now want to deploy the suggested model and want to forecast let’s say 9 months.

My approach for deployment would be to take the Hyper Parameters of the ARIMA Learner (ARMA 0,4) and enter them into ARIMA Learner within the loop deployment metanode. My question: Since I have initially (log) differenced the time series to make it stationary, do I have to enter also the integration Hyperparameter I=1 into the ARIMA learner for deployment, i.e. ARIMA (0,1,4), or is this already considered and I just have to enter ARMA (0,4), which the auto ARIMA learner suggests?

Sometimes it seems that the ARIMA learner wants me to enter a Hyperparameter I, even though I have differenced the timeseries before and I get the error message listed here:

  1. And last question so far: For the ARIMA Learner, it sometimes happens that I get the error message:

numpy.linalg.LinAlgError: SVD did not converge

I found this post, however, I cannot imagine where “NaN” or “inf” values would appear. Do you know more about this error message?

Thanks again for your support! I hope the questions are not to much of a hassle. If there are any unclear points, feel free to reach out :slight_smile:

Best regards,

Stiefel