Model Evaluation, Monitoring, and Error Analysis in Machine Learning. Is it all the same?

A gentle introduction to what model evaluation means in the ML development lifecycle, the requirements to implement it, and how to avoid common pitfalls.

Facundo Santiago
12 min readMar 20, 2023
Source: Stable diffusion 1.5 for the prompt “a clipart showing a machine learning model being evaluated for performance”.

One of the hottest topics in machine learning (besides the fact ChatGPT can explain jokes) is probably model monitoring and evaluation (read how much funding startups are getting to solve it). However, what exactly is Model Evaluation? How is it related to monitoring, error analysis, concept drift, and all those buzzwords?

In this post, we will go over the topic and focus on the role of Model Monitoring in the ML development lifecycle. We will also cover why some answers are not that straightforward (isn’t evaluation just calculated the ROC curve?) by combining them with some statistics (because statistics is what? … fun-da-mental!)

I will go over the concepts briefly in this post so it doesn’t become endless. However, I will create a series with implementation examples so you can see them in action. Stay tuned!

--

--

Facundo Santiago
Facundo Santiago

Written by Facundo Santiago

Product Manager @ Microsoft AI. Graduate adjunct professor at University of Buenos Aires. Frustrated sociologist.

No responses yet