Image Quality Filtering for Fine-Grained Classification

March 11, 2026 · 12 min read

Robustness FGVC Image Quality
Figure illustrating how image quality affects deep neural network classification
Representative visual from the quality-effect analysis in my thesis and VISAPP 2024 paper.

Why This Problem Matters

Fine-grained visual classification (FGVC) is often presented as a model architecture problem: better backbones, better attention modules, better losses. In real deployment, however, one of the dominant failure drivers is much simpler: image quality mismatch between training and inference.

In industrial and mobile settings, images are captured under motion blur, inconsistent focus, variable illumination, compression artefacts, and changing devices. If training data contains too much low-information content, the model can learn brittle shortcuts or simply fail to learn stable class cues.

What I Investigated in Chapter 3

In Chapter 3 of my thesis (How Quality Affects Deep Neural Networks in Fine-Grained Image Classification), I evaluated whether no-reference image quality assessment (NRIQA) can be used as a practical data selection mechanism before training.

The core idea was to avoid a binary split of "perfect" versus "imperfect" images. Instead, I used quality scores plus confidence-quality analysis to identify and remove the most harmful quality tail while preserving enough variation for robust learning.

Method Summary

The procedure was based on three linked components:

This makes the setup operationally useful: the method can be applied in data pipelines where no pristine reference image exists.

Key Results

The main result was consistent: quality-filtered training subsets improved classification performance relative to all-quality training. One representative outcome was ResNet-34 improving from 81.2% to 85.4% when trained on filtered high-quality data.

This is a 4.2-point absolute gain in a practical FGVC setting. In larger decision pipelines, even modest absolute gains can materially reduce false alarms or re-inspection workload.

Practical takeaway: a targeted quality gate can provide stronger reliability gains than a costly architecture swap, especially when the deployment domain is mobile or uncontrolled.

What This Means for Deployment

There are several deployment-level lessons that generalize beyond this specific thesis domain:

Recommended Minimal Workflow

If you want a lightweight replication path in another CV project, this sequence usually works:

  1. Compute NRIQA scores on candidate training images.
  2. Train baseline on full data.
  3. Train one filtered model after removing only the lowest quality tail.
  4. Compare not only accuracy but error profile and confidence calibration.
  5. Choose the threshold that improves stability, not just peak score.

Related Papers and Thesis Source