Overview of the Different Algorithms

The models were trained on three years of weather and generation data from 2015 - 2017. The algorithms were then tested on the results from 2018 and the predictions were plotted against the actual values for this period to give an indication of performance accuracy. During testing four models were run: a stacking ensemble, two transformer variants and an LSTM network. After training the transformer and LSTM models, an XGBoost algorithm was trained on the residuals to capture underlying patterns which may not have been captured by the original model. This significantly improved overall performance. This approach was excluded from the ensemble method as it already included XGBoost within the ensemble.

Stacking Ensemble

Best performer

The stacking method involved a two-phase process. Four prediction models were trained independently: XGBoost, LightGBM, Ridge Regression and Random Forest. A second-stage model was then used to identify how to combine the different predictions, specifically what weight to give each model. The Stacking Ensemble produced the lowest error across all five targets.

DeepAR

A deep learning algorithm developed by Amazon which uses Long Short-Term Memory (LSTM) networks to produce forecasts for related time series, learning shared patterns. It outputs a full probability distribution at each time step, providing confidence intervals around predictions.

Temporal Fusion Transformer

An architecture which combines LSTM-based recurrent layers with self-attention mechanisms within transformers. It automatically learns which input variables and time steps matter most, making it one of the more interpretable deep learning approaches.

PatchTST

A transformer model which splits the context window into 24-hour patches and processes them as tokens, rather than treating each hour individually. This patching approach makes the model more efficient and better at capturing daily patterns.