TabularPredictor.evaluate_predictions¶
- TabularPredictor.evaluate_predictions(y_true, y_pred, sample_weight=None, decision_threshold=None, display: bool = False, auxiliary_metrics=True, detailed_report=False, **kwargs) dict[source]¶
- Evaluate the provided prediction probabilities against ground truth labels. Evaluation is based on the eval_metric previously specified in init, or default metrics if none was specified. - Parameters:
- y_true ( - np.arrayor- pd.Series) – The ordered collection of ground-truth labels.
- y_pred ( - pd.Seriesor- pd.DataFrame) – The ordered collection of prediction probabilities or predictions. Obtainable via the output of predictor.predict_proba. Caution: For certain types of eval_metric (such as ‘roc_auc’), y_pred must be predicted-probabilities rather than predicted labels.
- sample_weight ( - pd.Series, default = None) – Sample weight for each row of data. If None, uniform sample weights are used.
- decision_threshold (float, default = None) – The decision threshold to use when converting prediction probabilities to predictions. This will impact the scores of metrics such as f1 and accuracy. If None, defaults to predictor.decision_threshold. Ignored unless problem_type=’binary’. Refer to the predictor.decision_threshold docstring for more information. 
- display (bool, default = False) – If True, performance results are printed. 
- auxiliary_metrics (bool, default = True) – Should we compute other (problem_type specific) metrics in addition to the default metric? 
- detailed_report (bool, default = False) – Should we computed more detailed versions of the auxiliary_metrics? (requires auxiliary_metrics = True) 
 
- Returns:
- Returns dict where keys = metrics, values = performance along each metric. 
- NOTE (Metrics scores always show in higher is better form.) 
- This means that metrics such as log_loss and root_mean_squared_error will have their signs FLIPPED, and values will be negative.