Ensemble models for time series
In this short post, we summarize Yue’s talk from May 19th and share some of the resources.
Time series data comes as a sequence of values . The goal of forecasting is to be able to predict the outcome at time if all the previous outcomes are known. Currently, there are many techniques for doing this; you can find an excellent summary in this paper (about 50 pages long). Coming back to our goal, we want to find a predictor (function) such that .
Recursive Neural Networks
RNN are useful to predict sequential data, they are powerful but training them can be hard. This problem is somehow fixed with the introduction of Lont Short Term Memory. They are powerful, like really powerful. We don’t dwell into the explanations here, but refer the reader to (the excellent) Colah’s post for an explanation or the more technical Andrej Karpathy Blog. Luckily for us, some people have written wrappers that allow us to create an LSTM network with five lines of code using Keras, to see how check Siraj’s video on using LSTM to predict the stock market.
Ensemble models
The idea is simple; we have some classifiers that by themselves aren’t doing well, maybe together they can do better. This is usually understood as building strong classifiers out of weak classifiers. We can do this in different ways…
Random Forest
The decision of the majority…
Boosting
Penalizing bad decisions…
Why linear regression shouldn’t work and what to do about it.
The data is correlated…