If you like what we're working on, please  star us on GitHub. This enables us to continue to give back to the community.
validating your machine learning models

Deepchecks
Open Source

Deepchecks Open Source is a Python package for comprehensively validating your machine-learning models and data with minimal effort.
Open Source

Key Capabilities of Deepchecks Open Source

Data Integrity

Data Integrity

When you have a fresh dataset and want to validate your data’s correctness and uncover inconsistencies such as conflicting labels or data duplicates.
ML Model Evaluation

Model Evaluation

When you have a trained model and want to examine performance metrics, compare it to various benchmarks, and create a clear and granular picture for validating the model’s behavior (e.g., are there segments where it under-performs).
ML Train-Test Validation

Train-Test Validation

When you have separate datasets (such as train and test or training data collected at different times) and want to validate that they are representative of each other and don’t have issues such as drift or leakage.

Deepchecks Open Source: For ML Practitioners During the Research Phase

Deepchecks Open Source is a Python package for comprehensively validating your machine-learning models and data with minimal effort. It includes checks related to various issues, such as model performance, data integrity, distribution mismatches, and more. Model and data validation is one of the most important processes that data scientists and ML engineers deal with while scaling up from the “laboratory phase” to ML Systems that provide continuous value. Whether your main interest is testing, CI/CD, model auditing or plain old testing, we recommend “kicking the tires” with Deepchecks Open Source as a first step.
Deepchecks Deepchecks

How Does It Work?

Suites are composed of checks. Each check contains outputs displayed in a notebook and/or conditions with a pass/fail output.

Conditions can be added or removed from a check;

Checks can be edited or added/removed to a suite;

Suites can be created from scratch or forked from an existing suite.


The “Report” beings with a list of the Conditions (passed or not) followed by the Display Data (from the various checks).
How Does It Work?
The “Report” beings with a list of the Conditions (passed or not) followed by the Display Data (from the various checks).

Key Features & Checks


Data Integrity

Data Integrity
from deepchecks.tabular.checks import StringMismatch StringMismatch().run(ds)
Data Integrity

Methodology Issues

Methodology Issues
from deepchecks.tabular.checks import BoostingOverfit BoostingOverfit().run(train_ds, validation_ds, clf)
Methodology Issues

Distribution Checks

Distribution Checks
from deepchecks.tabular.checks import FeatureDrift FeatureDrift().run(train_dataset, test_dataset, model)
Distribution Checks

Performance Checks

Performance Checks
from deepchecks.tabular.checks import WeakSegmentPerformance WeakSegmentPerformance().run(validation_ds, model)
Performance Checks

Model Explainability Checks

Model Explainability Checks
Model Explainability Checks

Suites of Checks

General Suite
from deepchecks.tabular.suites import train_test_validation train_test_validation().run(ds_train, ds_test)
General Suite

Open Source & Community

Deepchecks is committed to keeping the ML validation package open-source and community-focused.

Subscribe to Our Newsletter

Do you want to stay informed? Keep up-to-date with industry news, the latest trends in MLOps, and the observability of ML systems.