Monitoring is just a tool: necessary but not sufficient. It is useless without people committed to AI system maintenance as well as processes & tools in case things break down. Mature organizations set up rolling on-call schedules for their engineers and invest in debugging tools to solve issues faster.
It is also about the need for monitoring. đ
Four years ago, I worked on my first end-to-end AI pipeline. It involved a year of development to:
- speak to business experts to craft good features
- select & tune the best algorithm
- meet with IT engineers to understand how to integrate with the real-time systems
- create an online feedback loop with automated model retraining
At the end of the project, I added monitoring to this production system. It was no easy task! đ¤
I built everything from scratch: input quality checks, model performance thresholds, and drift scores. All metrics were available in a dashboard and sent automatically to stakeholders by email.
Lastly, I onboarded other data scientists to maintain the system and moved on to other projects.
Four years later, I learned that the system I had built had somewhat failed. Some input changes in the IT system went undetected. Online performance metrics didnât match offline metrics. Few people used the monitoring dashboard. đ˘
Why?
We data scientists rely too much on statistical KPIs such as concept drift. They are hard to interpret, even by data scientists and especially by business users. If I were to do the project again, I would add more simple input checks and business-oriented KPIs.
After a few months since deployment, the infamous problem of âdashboard fatigueâ arises. People pay less attention to monitoring. To overcome this, you need to set up alerting mechanisms. From a user perspective, alerts are an entry point to monitoring. The hard part is to make sure you donât trigger too many alerts.
Monitoring is just a tool: necessary but not sufficient. It is useless without people committed to AI system maintenance as well as processes & tools in case things break down. Mature organizations set up rolling on-call schedules for their engineers and invest in debugging tools to solve issues faster.
With Giskard AI, we want to make it easy to monitor AI models and get actionable alerts.
Whatâs your experience? Do you invest a lot of time in monitoring? What works and what doesnât?
Why not start a discussion in the comments or contact us at hello@giskard.ai?
Share the love đ
Alex
Original post published on LinkedIn on October 28, 2021
â