Diffusion models, known for their generative capabilities, have recently shown unexpected potential in image classification tasks by using Bayes’ theorem. However, most diffusion classifiers require evaluating all class labels for a single classification, leading to significant computational costs that can hinder their application in large-scale scenarios. To address this, we present a Hierarchical Diffusion Classifier (HDC) that exploits the inherent hierarchical label structure of a dataset. By progressively pruning irrelevant high-level categories and refining predictions only within relevant subcategories, i.e., leaf nodes, HDC reduces the total number of class evaluations. As a result, HDC can accelerate inference by up to 60% while maintaining and, in some cases, improving classification accuracy. Our work enables a new control mechanism of the trade-off between speed and precision, making diffusion-based classification more viable for real-world applications, particularly in large-scale image classification tasks.
If you use this information, method or the associated code, please cite our paper:
@misc{shanbhag2024justleafit,
title = {Just Leaf It: Accelerating Diffusion Classifiers with Hierarchical Class Pruning},
author = {Arundhati S. Shanbhag and Brian B. Moser and Tobias C. Nauen and Stanislav Frolov and Federico Raue and Andreas Dengel},
year = {2024},
eprint = {2411.12073},
archiveprefix = {arXiv},
primaryclass = {cs.CV}
}