A Survey of Time Series Analysis Methods and Approaches

Introduction

Time series refers to data collected with respect to time and Time series analysis pertains to the methods and techniques we employ to understand and analyse time series data. The need to understand the universe, physical systems and processes over time has made time series analysis a mature branch of mathematics and statistics and has also catapulted its development as a tool that is widely used in various academic fields, sciences and industries. The continuous development of time series analysis in the era of computational advances has been driven by the significant value it brings across industries such as business, finance, the energy sector, and the water industry. In the water industry, measuring consumption over time has been vital for understanding the usage patterns of consumers which aids in forecasting demand and managing resources, as well as conservation strategies such as detecting leakage.

In this blog series, we will present a survey of approaches and methods used in traditional and modern time series analysis. We will then propose a novel hybrid approach applied to the time series data of Continuously Logged Users (CLU) in a District Metered Area (DMA) as it pertains to Paradigmโ€™s purpose of forecasting consumer demand.

Understanding CLU Data

CLU data consists flow rate in l/s and recorded at a 15 – minute interval. A time series data of this frequency can provide powerful insights into water usage patterns. By using exploratory analysis of the CLU data across different water companies, we have observed types of patterns that are typical of time series data that we can generalize for many of the CLUs across different water companies.

  • CLU data exhibits seasonal patterns or regular variations that repeat over a period.
  • CLU data exhibits cyclical trends or longerโ€“term fluctuations influenced by some factors.
  • CLU data shows clear trends or persistent movement of consumption over time.
  • CLU data shows noisy and irregular variations.
  • CLU data integrity suffers when there are missing data.

Figure 1: A few representative time series and their weekly and monthly distribution showcasing the types of time series pattern of CLU data.

Time Series Analysis of CLU Data: Prediction, Forecasting and Imputation

The value of time series analysis lies in its ability to uncover underlying patterns and trends within temporal data, enabling us to predict and forecast future values accurately. Furthermore, time series analysis plays a critical role in imputing missing data, thereby preserving the integrity and usability of the dataset.

In the context of time series analysis, prediction or forecasting involves estimating future values ๐’€๐’•+๐’Œ of the series based on its past values ๐’€๐’• , ๐’€๐’•โˆ’๐Ÿ, โ€ฆ , ๐’€๐’•โˆ’๐’. Here, ๐’• represent the current time, and ๐’Œ is the number of future time steps we aim to predict. On the other hand, imputation refers to filling in missing values, ๐’€๐’•๐’Š within the existing dataset, where ๐’•๐’Š represents the time points with missing data. By reconstructing a complete dataset, imputation ensures continuity of the series.

Consider the animation below as a demonstration of prediction, imputation, and forecasting in action. This visual illustrates how time series analysis can be applied to effectively predict future values and handle missing data.

The quest to develop accurate prediction and forecasting models is driven by the immense benefits they offer. We are getting access to more data than ever before, and the availability of advanced computing power and tools have revolutionized time series prediction. However, real-world time series data is also complex and messy. Our aim is to develop an approach or a combination of approaches that can take care of prediction, forecasting and imputation relevant to the various types of patterns, complexity, and state of the CLU data.

Traditional Methods in Time Series Analysis

Traditional time series analysis methods have been foundational in understanding and forecasting time series data. They are effective for making short-term forecasts and handling missing data. However, these methods may not fully capture complex patterns in data. We list out some of these methods below.

Moving Averages (MA)

MA is known to be simple, yet powerful technique used to forecast time series data. The Simple Moving Average (SMA)

where ๐‘Œ_๐‘ก is the value at time ๐‘ก, ๐‘› is the number of data points in the moving window, calculates the average of the last ๐‘› data points. This helps to smooth the data by averaging out short-term fluctuations, making it easier to see long-term trends.

Given a simulated โ€œTrueโ€ data, we present the animation below to demonstrate SMA using three different windows and how each fare in predicting the True data.

Exponential Smoothing

Exponential Smoothing assigns decreasing weights (importance) to older observations, giving more importance to recent data. It includes Single Exponential Smoothing (SES), Double Exponential Smoothing (DES) and Triple Exponential Smoothing (TES). It is known to be suitable for time series data with a level, trend, and seasonal components.

Single Exponential Smoothing (SES)

Where ๐‘†๐‘ก is the smoothed value at time ๐‘ก, ๐›ผ is the smoothing factor, and 0 < ๐›ผ < 1. SES updates the smoothed value ๐‘†๐‘ก by combining the actual value ๐‘Œ๐‘ก and the previous smoothed value ๐‘†๐‘กโˆ’1 . The smoothing factor determines how much importance is given to the recent observation compared to the past smoothed values.

Double Exponential Smoothing (DES)

Where ๐‘‡๐‘ก is the trend estimate at time t, ๐›พ is the trend smoothing factor. DES is used for time series with trends. It calculates both the smoothed value and trend separately. The smoothed value considers the current observation and the previous smoothed value plus the trend. The trend is updated based on the change in smoothed value.

Triple Exponential Smoothing (TES)

Where ๐ถ๐‘ก is the seasonal component, ๐ฟ is the length of the seasonal cycle, and ๐›ฝ and ๐›พ are smoothing factors for trend and seasonality. TES, otherwise known as Holt-Winters methods, account for seasonality in addition to trends. It uses three equations to update the smoothed value, trend, and seasonal components, capturing more complex patterns in data.

Spline Interpolation

Spline interpolation fits a curve through the known data points, making it useful for interpolating missing values in time series data. Spline interpolations make use of polynomial functions. For instance, a linear spline uses a linear function to connect each data point with a straight line, a cubic spline uses a smoother piecewise cubic function that passes through the known data points.

Kalman Filter

The Kalman Filter is a probabilistic approach used to estimate the state of a dynamic system from a series of noisy measurements. It is an iterative process using a series of inputs to produce estimates of unknown variables. It is effective in smoothing time series data and filling missing values by considering statistical noise and inaccuracies. The Kalman Filter operates in two steps:

  1. Predict: Project the current state and error covariance forward to obtain the โ€˜a prioriโ€™ estimate for the next time step.
  2. Update: Incorporate new measurements to obtain the โ€˜a posterioriโ€™ estimate.

Modern Methods in Time Series Analysis

Machine learning and deep learning techniques offer advanced capabilities for time series predictions and forecasting. These techniques can capture patterns and relationships in the data, making them suitable for long-term forecasts and more intricate time series. Machine learning and deep learning in the context of time series involve using algorithms that can learn from and make predictions concerning sequential data. Unlike traditional statistical methods, which often rely on the predefined models and assumptions (equations and its parameters), machine learning techniques adaptively learn patterns from the data itself. This flexibility allows machine learning models to capture nonlinear relationships and interactions within the data, making them powerful tools for time series forecasting.

We present some of the most prominent machine learning and deep learning techniques being used across different domains.

Support Vector Machine (SVM)

SVM is a type of supervised learning method that is known to be effective in capturing non-linear relationships. For time series, SVM Regression aims to fit the data and predict future values. For a set of training data points (๐‘ฅ๐‘– , ๐‘ฆ๐‘– ) where ๐‘ฅ๐‘– represents the input features and ๐‘ฆ๐‘– represents the target values.

Random Forests

Random Forests use multiple decision trees to make predictions, averaging the results to improve accuracy. They are very effective for time series forecasting due to their ability to capture various patterns and handle large datasets.

Gradient Boosting Method

GBM builds models sequentially, each new model aiming to reduce errors from previous ones. This approach makes GBMs highly accurate and suitable for time series forecasting where capturing patterns is essential.

Long Short - Term Memory (LSTM) Network

Deep Learning, a subset of machine learning, involves neural networks with multiple layers that can automatically learn feature representations from data. Deep learning models are particularly effective at capturing complex temporal dependencies and patterns that may be difficult to identify with traditional methods.

LSTM networks are a type of recurrent neural network (RNN) that are wellsuited for time series forecasting because they can learn long-term dependencies and remember information for extended periods. The key components of LSTM include:

  • The โ€œCell Stateโ€ which is essentially the memory of a network. It is designed to carry information across sequences, helping network remember context over many times steps.
  • The โ€œForget Gateโ€ that determines what information should be discarded in the cell state.
  • The โ€œInput Gateโ€ that decides which new information should be stored in the cell state.
  • The โ€œUpdate Stageโ€ where the cell state is computed by combining the old cell state and new candidate values.
  • The โ€œOutput Gateโ€ that determine the output based on the cell state.

These components allow LSTMs to learn complex patterns in time series data.

Conclusion

In this blog, we presented various methods for time series analysis, compared their effectiveness, discussed some of the underlying mathematics, and provided animations demonstrating the implementation of each of the methods. We have reviewed Moving Averages, Exponential Smoothing, Spline Interpolation, and Kalman Filter, which are foundational methods employed for prediction, imputation and forecasting of time series with varying types and behaviour. We also took a look at Support Vector Machines, Random Forests, Gradient Boosting Machines, and LSTM. These methods provide advanced capabilities for capturing complex patterns and improving the accuracy of predictions.

The limitations of both traditional and modern methods highlights the need for a hybrid approach. A hybrid method combines the strengths of multiple approaches for better imputation and improves the accuracy and robustness of predictions. The varying types of patterns and behaviour of CLU data warrants a combination of approaches that will leverage their collective strengths and adapt to different data characteristics.

In my next blog, we will dive deeper into the hybrid approach, building the hybrid models and exploring comprehensive comparisons between traditional and modern methods with examples and case studies.

References:

  1. BV Vishwas& Ashish Patel, Hands-on Time Series Analysis with Python, Springer Link, Appres, 2020
  2. Evangelos Spilliotis, Time Series Forecasting with Statistical, Machine Learning, Deep Learning Methods: Past, Present, and Future, Springer Link, September 2023

Related articles