Testing & Monitoring framework for ML models
Documentation


Detect hidden vulnerabilities in your ML model
.png)
Performance bias
Identify discrepancies in accuracy, precision, recall, or other evaluation metrics on specific data slices.
.png)
Unrobustness
Detect when your model is sensitive to small perturbations in the input data.
.png)
Overconfidence
Avoid incorrect predictions when your model is overly confident.
.png)
Data leakage
Detect inflated performance metrics and inaccuracy due to unintentional external data used in your model.
.png)
Unethical behavior
Identify perturbations in your model behavior when switching input data (gender, ethnicity...).
.png)
Stochasticity
Detect inherent randomness in your model and avoid variations in your results.


Automatically scan your model to find hidden vulnerabilities
In a few lines of code, identify vulnerabilities that may affect the performance of your model, such as data leakage, non-robustness, ethical biases, and overconfidence. Directly in your notebook.

‍
model, df = giskard.demo.titanic()
dataset = giskard.Dataset(df, target="Survived")
model = giskard.Model(model, model_type="classification")
giskard.scan(model, dataset)
test_suite = scan.generate_test_suite()
test_suite.run()
Design & run custom ML tests
If the scan found some issues with your model, you can automatically generate a test suite, add your own custom tests, and execute them with our native Python API.
Centralize tests, compare models & share results with your team
Upload the generated test suite to the Giskard server, create reusable test suites, and get ready-made dashboards you can share with the rest of your team. Compare different model versions over time.

‍
giskard server start
Collaborative and shareable testing
Leverage the power of our open-source community by easily uploading test fixtures, including AI detectors for identifying issues like hate speech, toxicity, and more.
Access data transformation tools for tasks like rewriting and introducing typos, allowing you to simulate a wide range of real-world scenarios for comprehensive testing.
Benefit from the collective knowledge and resources of our community while enhancing your AI model testing and evaluation processes. (product page)

Collaborative and shareable testing
Leverage the power of our open-source community by easily uploading test fixtures, including AI detectors for identifying issues like hate speech, toxicity, and more.
Access data transformation tools for tasks like rewriting and introducing typos, allowing you to simulate a wide range of real-world scenarios for comprehensive testing.
Benefit from the collective knowledge and resources of our community while enhancing your AI model testing and evaluation processes. (product page)

Save time with our catalog of
reusable test components

Stop wasting time creating new testing components for every new ML use case. Use our ready-made tests, create new ones, and easily add them to your test suite.
Apply data slicing functions for identifying issues like hate speech, toxicity, and more. Access data transformation tools for tasks like rewriting and introducing typos, allowing you to simulate a wide range of real-world scenarios for comprehensive testing.
.webp)

All resources
Thought leadership articles about ML Quality: Risk Management, Robustness, Efficiency, Reliability & Ethics
Ready. Set. Test!
‍Get started today
Get started