Note
Go to the end to download the full example code
Full Suite Quickstart#
In order to run your first Deepchecks Suite all you need to have is the data and model that you wish to validate. More specifically, you need:
Your train and test data (in Pandas DataFrames or Numpy Arrays)
(optional) A Working with Models and Predictions (including XGBoost, scikit-learn models, and many more). Required for running checks that need the model’s predictions for running.
To run your first suite on your data and model, you need only a few lines of code, that start here: Define a Dataset Object.
# If you don’t have deepchecks installed yet:
# If you don't have deepchecks installed yet:
import sys
!{sys.executable} -m pip install deepchecks -U --quiet #--user
Load Data, Split Train-Val, and Train a Simple Model#
For the purpose of this guide we’ll use the simple iris dataset and train a simple random forest model for multiclass classification:
import numpy as np
# General imports
import pandas as pd
from sklearn.ensemble import RandomForestClassifier
from sklearn.model_selection import train_test_split
from deepchecks.tabular.datasets.classification import iris
# Load Data
iris_df = iris.load_data(data_format='Dataframe', as_train_test=False)
label_col = 'target'
df_train, df_test = train_test_split(iris_df, stratify=iris_df[label_col], random_state=0)
# Train Model
rf_clf = RandomForestClassifier(random_state=0)
rf_clf.fit(df_train.drop(label_col, axis=1), df_train[label_col]);
RandomForestClassifier(random_state=0)
Define a Dataset Object#
Initialize the Dataset object, stating the relevant metadata about the dataset (e.g. the name for the label column)
Check out the Dataset’s attributes to see which additional special columns can be declared and used (e.g. date column, index column).
from deepchecks.tabular import Dataset
# We explicitly state that this dataset has no categorical features, otherwise they will be automatically inferred
# If the dataset has categorical features, the best practice is to pass a list with their names
ds_train = Dataset(df_train, label=label_col, cat_features=[])
ds_test = Dataset(df_test, label=label_col, cat_features=[])
Run a Deepchecks Suite#
Run the full suite#
Use the full_suite
that is a collection of (most of) the prebuilt checks.
Check out the when you should use deepchecks guide for some more info about the existing suites and when to use them.
from deepchecks.tabular.suites import full_suite
suite = full_suite()
Full Suite:
| | 0/35 [Time: 00:00]
Full Suite:
|█ | 1/35 [Time: 00:00, Check=Train Test Performance]
Full Suite:
|█████ | 5/35 [Time: 00:00, Check=Simple Model Comparison]
Full Suite:
|███████ | 7/35 [Time: 00:01, Check=Calibration Score]
Full Suite:
|████████████████████ | 20/35 [Time: 00:01, Check=Feature Label Correlation Change]
Full Suite:
|████████████████████████ | 24/35 [Time: 00:01, Check=Is Single Value]
Full Suite:
|█████████████████████████████████ | 33/35 [Time: 00:01, Check=Feature Label Correlation]
Run the integrity suite#
If you still haven’t started modeling and just have a single dataset, you
can use the data_integrity
:
from deepchecks.tabular.suites import data_integrity
integ_suite = data_integrity()
integ_suite.run(ds_train)
Data Integrity Suite:
| | 0/12 [Time: 00:00]
Data Integrity Suite:
|███████████ | 11/12 [Time: 00:00, Check=Feature Feature Correlation]
Run a Deepchecks Check#
If you want to run a specific check, you can just import it and run it directly.
Check out the Check Gallery or the API Reference for more info about the existing checks and their parameters.
from deepchecks.tabular.checks import LabelDrift
check = LabelDrift()
result = check.run(ds_train, ds_test)
result
and also inspect the result value which has a check-dependant structure:
result.value
{'Drift score': 0.0, 'Method': "Cramer's V"}
Edit an Existing Suite#
Inspect suite and remove condition#
We can see that the Feature Label Correlation check failed, both for test and for train. Since this is a very simple dataset with few features and this behavior is not necessarily problematic, we will remove the existing conditions for the PPS
# Lets first print the suite to find the conditions that we want to change:
suite
Full Suite: [
0: TrainTestPerformance
Conditions:
0: Train-Test scores relative degradation is less than 0.1
1: RocReport
Conditions:
0: AUC score for all the classes is greater than 0.7
2: ConfusionMatrixReport
3: PredictionDrift
Conditions:
0: Prediction drift score < 0.15
4: SimpleModelComparison
Conditions:
0: Model performance gain over simple model is greater than 10%
5: WeakSegmentsPerformance(n_to_show=5)
Conditions:
0: The relative performance of weakest segment is greater than 80% of average model performance.
6: CalibrationScore
7: RegressionErrorDistribution
Conditions:
0: Kurtosis value higher than -0.1
1: Systematic error ratio lower than 0.01
8: UnusedFeatures
Conditions:
0: Number of high variance unused features is less or equal to 5
9: BoostingOverfit
Conditions:
0: Test score over iterations is less than 5% from the best score
10: ModelInferenceTime
Conditions:
0: Average model inference time for one sample is less than 0.001
11: DatasetsSizeComparison
Conditions:
0: Test-Train size ratio is greater than 0.01
12: NewLabelTrainTest
Conditions:
0: Number of new label values is less or equal to 0
13: NewCategoryTrainTest
Conditions:
0: Ratio of samples with a new category is less or equal to 0%
14: StringMismatchComparison
Conditions:
0: No new variants allowed in test data
15: DateTrainTestLeakageDuplicates
Conditions:
0: Date leakage ratio is less or equal to 0%
16: DateTrainTestLeakageOverlap
Conditions:
0: Date leakage ratio is less or equal to 0%
17: IndexTrainTestLeakage
Conditions:
0: Ratio of leaking indices is less or equal to 0%
18: TrainTestSamplesMix(n_to_show=5)
Conditions:
0: Percentage of test data samples that appear in train data is less or equal to 5%
19: FeatureLabelCorrelationChange(ppscore_params={}, random_state=42)
Conditions:
0: Train-Test features' Predictive Power Score difference is less than 0.2
1: Train features' Predictive Power Score is less than 0.7
20: FeatureDrift
Conditions:
0: categorical drift score < 0.2 and numerical drift score < 0.2
21: LabelDrift
Conditions:
0: Label drift score < 0.15
22: MultivariateDrift
Conditions:
0: Drift value is less than 0.25
23: IsSingleValue
Conditions:
0: Does not contain only a single value
24: SpecialCharacters
Conditions:
0: Ratio of samples containing solely special character is less or equal to 0.1%
25: MixedNulls
Conditions:
0: Number of different null types is less or equal to 1
26: MixedDataTypes
Conditions:
0: Rare data types in column are either more than 10% or less than 1% of the data
27: StringMismatch
Conditions:
0: No string variants
28: DataDuplicates
Conditions:
0: Duplicate data ratio is less or equal to 5%
29: StringLengthOutOfBounds
Conditions:
0: Ratio of string length outliers is less or equal to 0%
30: ConflictingLabels
Conditions:
0: Ambiguous sample ratio is less or equal to 0%
31: OutlierSampleDetection
32: FeatureLabelCorrelation(ppscore_params={}, random_state=42)
Conditions:
0: Features' Predictive Power Score is less than 0.8
33: FeatureFeatureCorrelation
Conditions:
0: Not more than 0 pairs are correlated above 0.9
34: IdentifierLabelCorrelation(ppscore_params={})
Conditions:
0: Identifier columns PPS is less or equal to 0
]
WeakSegmentsPerformance(n_to_show=5)
Conditions:
0: The relative performance of weakest segment is greater than 80% of average model performance.
# print and see that the condition was removed
suite[5]
WeakSegmentsPerformance(n_to_show=5)
If we now re-run the suite, all of the existing conditions will pass.
Note: the check we manipulated will still run as part of the Suite, however it won’t appear in the Conditions Summary since it no longer has any conditions defined on it. You can still see its display results in the Additional Outputs section
For more info about working with conditions, see the detailed configuring conditions guide.
Total running time of the script: (0 minutes 3.619 seconds)