Area under PR evaluation metric

The area under PR metric measures how well your model balances correctly identifying positive classes with finding all positive classes.

Metric details

Area under precision recall (PR) is quality evaluation metric that measures the quality of the performance of binary classification machine learning models in watsonx.governance.

Scope

The area under PR metric evaluates machine learning models only.

  • Types of AI assets: Machine learning models

  • Machine learning problem type: Binary classification

Scores and values

The area under PR metric metric score indicates how well the model balances precision and recall. Higher scores indicate better model performance with identifying and finding positive classes.

  • Range of values: 0.0-1.0
  • Best possible score: 1.0
  • Chart values: Last value in the timeframe

A score of 0.5 suggests random guessing, while a score of 1.0 represents perfect classification.

Settings

Default threshold: Lower limit = 80%

Evaluation process

Area under PR is calculated by plotting precision against recall for different threshold values. For each threshold, a confusion matrix is generated that specifies classes of true positives, false positives, and false negatives.

Precision and recall are calculated with these classes and plotted to create the PR curve. The area under this curve is calculated to generate the metric score.

Do the math

Area under PR calculates the total for precision and recall with the following formula:

area under PR formula is displayed

Precision (P) is calculated as the number of true positives (Tp) over the number of true positives plus the number of false positives (Fp) with the following formula:

precision formula is displayed

Recall (R) is calculated as the number of true positives (Tp) over the number of true positives plus the number of false negatives (Fn) with the following formula:

recall formula is displayed

Parent topic: Evaluation metrics