0 / 0

Micro f1 score evaluation metric

Last updated: May 08, 2025
Micro f1 score evaluation metric

The micro f1 score calculates the harmonic mean of precision and recall.

Metric details

Micro f1 score is a multi-label/class metric for generative AI quality evaluations that measures how well generative AI assets perform entity extraction tasks for multi-label/multi-class predictions.

Scope

The micro F1 score metric evaluates generative AI assets only.

  • Types of AI assets: Prompt templates
  • Generative AI tasks: Entity extraction
  • Supported languages: English

Scores and values

The micro F1 score metric indicates the harmonic mean of precision and recall. Higher scores indicate that predictions are more accurate.

  • Range of values: 0.0-1.0
  • Best possible score: 1.0

Settings

  • Thresholds:
    • Lower limit: 0.8
    • Upper limit: 1

Parent topic: Evaluation metrics