AI is characterized by many feedback loops and interactions between components. The output of one model can be ingested into the training base of another. As a result, AI errors can be difficult to identify, measure, and correct.In this new series, each week we are going to present the most famous AI testing methods, showing illustrative examples and practical methods to implement them
While regulators are asking for Quality management systems for AI (article 17 from the European AI Act), the capacity to create tests is becoming crucial in the AI industry. But ML testing is highly complex and itās still an active research area. Here are some reasons:
1ļøā£ AI follows a data-driven programming paradigm
According to the paper from Paleyes (2021), unlike in traditional software products where changes only happen in the code, AI systems change along 3 axes: the code, the model, and the data. The modelās behavior evolves in response to the frequent provision of new data.
2ļøā£ AI is not easily breakable in small unit components
Some AI properties (e.g., accuracy) only emerge as a combination of different components such as the training data, the learning program, and the learning library. It is hard to break the AI system into smaller components that can be tested in isolation.
3ļøā£ AI errors are systemic and self-amplifying
AI is characterized by many feedback loops and interactions between components. The output of one model can be ingested into the training base of another. As a result, AI errors can be difficult to identify, measure, and correct.
In this new series, each week we are going to present the most famous AI testing methods, showing illustrative examples and practical methods to implement them. Weāll cover concepts such as:
Stay tuned!
Jean-Marie John-Mathews
ā