Underfitting

Understanding underfitting in machine learning - Neural Network Lexicon
Understanding underfitting in machine learning – Neural Network Lexicon

Short Definition

Underfitting occurs when a model is too simple to capture meaningful patterns in the data.

Definition

Underfitting is a failure mode in which a neural network fails to learn the underlying structure of the data. The model performs poorly not only on unseen data but also on the training data itself. This typically happens when the model has insufficient capacity, is trained for too few steps, or is constrained by overly strong regularization.

An underfitted model has high bias and low variance.

Why It Matters

Underfitting means the model has not learned enough to be useful. While it is often less dangerous than overfitting, it results in wasted data, wasted training time, and misleading conclusions about model or data quality.

Understanding underfitting is essential for diagnosing whether poor performance is due to:

  • model simplicity
  • training procedure
  • data representation issues

How It Works (Conceptually)

  • The model’s hypothesis space is too limited
  • Learned patterns are overly simplistic
  • Training loss remains high
  • Validation loss is also high and close to training loss

Underfitting indicates that the model cannot represent the target function adequately.

Minimal Python Example

train_accuracy = 0.55
val_accuracy = 0.54 # both low → underfitting

Common Pitfalls

  • Assuming more data will fix underfitting
  • Using overly simple models for complex tasks
  • Stopping training too early
  • Applying excessive regularization
  • Confusing underfitting with noisy data

Related Concepts

  • Generalization
  • Overfitting
  • Bias–Variance Tradeoff
  • Model Capacity
  • Regularization
  • Training Dynamics