Uncertainty Quantification and Why You Should Care

How to improve your ML model with three lines of code

Jonte Dancker
Towards Data Science

--

Turning a point prediction into a prediction set for classification or a prediction interval for regression models to quantify the uncerftainty of the underlying ML model.
Turning a point prediction into a prediction region to quantify the model’s uncertainty to give us more information (Image by the author).

Prediction models are trained to predict well and give us point forecasts.

Let’s assume we want to buy a house. Before we do so, we want to verify that the advertised price of 400,000 € is reasonable. For this, we use a model that, based on the number of rooms, the size and the location of the house, predicts that the house is worth 500,232.12 €.

Should we buy this house? It seems like a good deal, doesn’t it? But would our decision be different if the model instead had predicted a price of 340,021.34 €? Probably, right?

Hence, should we trust the model when making the decision to buy this house? What is the probability that the house is worth exactly 500,232.12 € or 340,021.34 €?

As the probability is close to zero, we should rather see the prediction as a best guess of the true outcome. Moreover, there is a risk for us when making a decision based on the model.

In the housing example, we could pay too much, resulting in a financial loss. If a self-driving car does not classify a person crossing the street as an obstacle, the result might be deadly.

--

--

Expert in time series forecasting and analysis | Writing about my data science side projects and sharing my learnings