Ensemble Learning
How combining several models via ensemble learning improves accuracy—illustrated with a simple temperature-forecast example.
Understanding Ensemble Learning
Let’s say you want to predict the weather forecast for tomorrow. If you rely on only one forecaster or one statistical model, you might run into two common problems:
Bias: where the forecaster is systematically off because their method is too simple or mis-calibrated.
Variance: where their prediction swings wildly because it overreacts to fluctuations in the latest training data.
Ensemble learning fixes this by consulting several forecasters, each trained a bit differently or looking at different signals (satellites, historical averages, local sensors).
Their individual mistakes tend to point in different directions, so when you combine—by averaging or voting—those random errors partially cancel out, sharply lowering variance.
Consider three separate forecasts for tomorrow’s temperature:
One model might overshoot tomorrow’s temperature by +2 °C
Another might undershoot by –3 °C
A third could be off by +1 °C
These errors differ in both size and sign: some positive, some negative.
When you merge the forecasts by averaging (or voting in a classification example), the positive and negative errors offset each other, so the combined prediction ends up much nearer to the true value than most of the individual forecasts.
Meaning, the combined prediction is closer to the truth and less sensitive to any single model’s overshoot or undershoot. This canceling effect is what lowers variance: the ensemble’s prediction varies far less across different datasets than the output of any individual model.
That same principle underlies many of today’s best-performing data-science tools where multiple decision trees or other models are blended to curb variance and capture complex patterns:
Bagging -> Training many versions of the same model on different random resamples of the data, then averaging (or voting) on their predictions to smooth out individual errors.
Random forests -> A bagging method that uses lots of decision trees, each grown on a bootstrapped dataset and a random subset of features, then averages their votes for a stronger, less-noisy result.
Gradient boosting -> Building multiple models in sequence, where each new model focuses on correcting the mistakes of the combined models so far; the final prediction is a weighted sum of all these “error-fixing” models.
Stacking -> Train several diverse models in parallel, then feed their predictions into a separate “meta-model” that learns the best way to blend them for the final answer.


Ensemble Learning Example
Connect with Me!
LinkedIn: caroline-rennier
Email: caroline@rennier.com