PropertySegmentsPerformance#
- class PropertySegmentsPerformance[source]#
Search for segments with low performance scores.
The check is designed to help you easily identify weak spots of your model and provide a deepdive analysis into its performance on different segments of your data. Specifically, it is designed to help you identify the model weakest segments in the data distribution for further improvement and visibility purposes.
The segments are based on the text properties - which are features extracted from the text, such as “language” and “number of words”. For more on properties, see the NLP Properties Guide.
In order to achieve this, the check trains several simple tree based models which try to predict the error of the user provided model on the dataset. The relevant segments are detected by analyzing the different leafs of the trained trees.
- Parameters
- propertiesUnion[Hashable, List[Hashable]] , default: None
Properties to check, if none are given checks all properties except ignored ones.
- ignore_propertiesUnion[Hashable, List[Hashable]] , default: None
Properties to ignore, if none given checks based on properties variable
- n_top_propertiesOptional[int] , default: 10
Number of properties to use for segment search. Selected at random.
- segment_minimum_size_ratio: float , default: 0.05
Minimum size ratio for segments. Will only search for segments of size >= segment_minimum_size_ratio * data_size.
- alternative_scorerDict[str, Union[str, Callable]] , default: None
Scorer to use as performance measure, either function or sklearn scorer name. If None, a default scorer (per the model type) will be used.
- score_per_sample: Optional[np.array, pd.Series, None], default: None
Score per sample are required to detect relevant weak segments. Should follow the convention that a sample with a higher score mean better model performance on that sample. If provided, the check will also use provided score per sample as a scoring function for segments. if None the check calculates score per sample by via neg cross entropy for classification.
- n_samplesint , default: 5_000
Maximum number of samples to use for this check.
- n_to_showint , default: 3
number of segments with the weakest performance to show.
- categorical_aggregation_thresholdfloat , default: 0.05
In each categorical column, categories with frequency below threshold will be merged into “Other” category.
- multiple_segments_per_propertybool , default: False
If True, will allow the same property to be a segmenting feature in multiple segments, otherwise each property can appear in one segment at most.
- __init__(properties: Optional[Union[Hashable, List[Hashable]]] = None, ignore_properties: Optional[Union[Hashable, List[Hashable]]] = None, n_top_properties: Optional[int] = 10, segment_minimum_size_ratio: float = 0.05, alternative_scorer: Optional[Dict[str, Union[Callable, str]]] = None, score_per_sample: Optional[Union[ndarray, Series]] = None, n_samples: int = 5000, categorical_aggregation_threshold: float = 0.05, n_to_show: int = 3, multiple_segments_per_property: bool = False, **kwargs)[source]#
- __new__(*args, **kwargs)#
Attributes
|
|
Methods
Add new condition function to the check. |
|
|
Add condition - check that the score of the weakest segment is greater than supplied relative threshold. |
Remove all conditions from this check instance. |
|
Run conditions on given result. |
|
Return check configuration (conditions' configuration not yet supported). |
|
Return check object from a CheckConfig object. |
|
Deserialize check instance from JSON string. |
|
Return check metadata. |
|
Name of class in split camel case. |
|
Return parameters to show when printing the check. |
|
Remove given condition by index. |
|
|
Run check. |
Run check. |
|
Serialize check instance to JSON string. |