Evaluation Metrics

Short Definition

Evaluation metrics measure model performance.

Definition

Evaluation metrics quantify how well a model performs on unseen data. Unlike loss functions, metrics are often task-specific and designed for human interpretation.

Choosing the right metric is crucial for meaningful evaluation.

Why It Matters

A model can optimize loss while performing poorly on the real objective.

How It Works (Conceptually)

  • Select task-appropriate metrics
  • Evaluate on validation or test data
  • Compare models fairly

Minimal Python Example

metric = correct_predictions / total

Common Pitfalls

  • Using the wrong metric
  • Evaluating on training data
  • Ignoring class imbalance

Related Concepts