Metrics for classification evaluation results.
Metrics for each confidenceThreshold in 0.00,0.05,0.10,...,0.95,0.96,0.97,0.98,0.99 and positionThreshold = INT32_MAX_VALUE.
ROC and precision-recall curves, and other aggregated metrics are derived from them. The confidence metrics entries may also be supplied for additional values of positionThreshold, but from these no aggregated metrics are computed.
Confusion matrix of the evaluation.
auPrcnumber
The Area Under Precision-Recall Curve metric. Micro-averaged for the overall evaluation.
auRocnumber
The Area Under Receiver Operating Characteristic curve metric. Micro-averaged for the overall evaluation.
logLossnumber
The log Loss metric.
| JSON representation |
|---|
{ "confidenceMetrics": [ { object ( |
ConfidenceMetrics
Confusion matrix of the evaluation for this confidenceThreshold.
confidenceThresholdnumber
Metrics are computed with an assumption that the Model never returns predictions with score lower than this value.
maxPredictionsinteger
Metrics are computed with an assumption that the Model always returns at most this many predictions (ordered by their score, descendingly), but they all still need to meet the confidenceThreshold.
recallnumber
Recall (True Positive Rate) for the given confidence threshold.
precisionnumber
Precision for the given confidence threshold.
falsePositiveRatenumber
False Positive Rate for the given confidence threshold.
f1Scorenumber
The harmonic mean of recall and precision. For summary metrics, it computes the micro-averaged F1 score.
f1ScoreMicronumber
Micro-averaged F1 Score.
f1ScoreMacronumber
Macro-averaged F1 Score.
recallAt1number
The Recall (True Positive Rate) when only considering the label that has the highest prediction score and not below the confidence threshold for each DataItem.
precisionAt1number
The precision when only considering the label that has the highest prediction score and not below the confidence threshold for each DataItem.
falsePositiveRateAt1number
The False Positive Rate when only considering the label that has the highest prediction score and not below the confidence threshold for each DataItem.
f1ScoreAt1number
The harmonic mean of recallAt1 and precisionAt1.
The number of Model created labels that match a ground truth label.
The number of Model created labels that do not match a ground truth label.
The number of ground truth labels that are not matched by a Model created label.
The number of labels that were not created by the Model, but if they would, they would not match a ground truth label.
| JSON representation |
|---|
{
"confusionMatrix": {
object ( |