Work in progress: we have compiled and summarized relevant papers in this field by year and are continuing to improve the categorization and organization of the collection to help newcomers quickly understand the area. Feel free to suggest improvements or add new papers via a pull request.
The Concept Bottleneck Model (CBM) is an emerging self-explainable architecture that first maps inputs to a set of human-interpretable concepts before making predictions using an interpretable classifier, typically a single-layer linear model. Beyond inherent interpretability, CBMs provide an intervention interface through the concept bottleneck layer, allowing users to directly modify concept activations to refine model predictions, and this serves as the most significant difference between CBMs and other explainable models, such as the CapsulesNet and ProtoPNet.
(images from IntCEMs, highlighting the interpretability and intervention ability of CBM architectures)
Improving Concept Representations
The original Concept Bottleneck Model maps each concept to a single (probabilistic) value to construct the concept bottleneck layer, followed by a linear layer that predicts image-level class labels based on these concept values. However, the semantics of individual concepts, the relationships and hierarchies among different concepts, and the dependencies between concepts and class labels are inherently complex. Therefore, to address the need for richer, more expressive concept representations and to model the intricate concept–concept and concept–class relationships, many studies have proposed improvements to the representation methods used in the concept bottleneck layer.
Method | Publication | Concept Representation | Highlight | Code/Project |
---|---|---|---|---|
Concept Embedding Models (CEMs) | NeurIPS 2022 | high-dimensional embeddings | representing each concept as a supervised high-dimensional embeddings to preserve high performance and interpretability under incomplete concept annotations | Code |
Probabilistic Concept Bottleneck Models (PCBMs) | ICML 2023 | probabilistic embeddings | leveraging probabilistic concept embeddings to model uncertainty in concept predictions and provide more reliable explanations with uncertainty | Code |
Energy-based Concept Bottleneck Models (ECBMs) | ICLR 2024 | high-dimensional embeddings + energy networks | using a set of networks to define the joint energy of the (input, concept, class) triplet, therefore providing a unified way for prediction, concept intervention, and probabilistic explanation via minimizing energy. | Code |
Logic-enhanced CBMs | ICML W 2024 | augmented with propositional logic rules | using the propositional logic derived from the concepts to model the relationships between concepts | - |
EQ-CBM | ACCV 2024 | quantized probabilistic embeddings | enhances CBMs through probabilistic concept encoding using energy-based models with quantized concept activation vectors to capture uncertainties | - |
Improving Intervention Ability / Interactivity
Method | Publication | Highlight | Code/Project |
---|---|---|---|
Method | Publication | Concept Source | Code/Project |
---|---|---|---|
LF-CBMs | ICLR 2023 | LLM | Code |
Post-hoc CBMs | ICLR 2023 | LLM / TCAV | Code |
LaBo | CVPR 2023 | LLM | Code |
BotCL | CVPR 2023 | Concept Prototypes (images + heatmap) | Code |
LM4CV | ICCV 2023 | LLM | Code |
CDMs | ICCV 2023 Worshop | LLM + VLMs | Code |
Res-CBM | CVPR 2024 | LLM + Visual genome | [Code](https://github.com/HelloSCM/ Res-CBM) |
DN-CBMs | ECCV 2024 | Sparse Auto Encoder + Words | Code |
CF-CBMs | NeurIPS 2024 | LLM + VLMs | Code |
VLG-CBM | NeurIPS 2024 | LLM + Object Detectors | Code |
BC-LLM | NeurIPS 2024 Workshop | LLM + Bayesian search framework | Code |
CCBM | Arxiv 2024 | Heatmaps | - |
CCPM | IEEE TMM | LLM, learnable | - |
XBMs | AAAI 2025 | MLLM (LLaVA) | Code |
V2C-CBM | AAAI 2025 | VLM (CLIP) + Common words | Code |
UBMs | TMLR 2025 | Concept discovery (image patch) | Code |
Text
Table
Scientific Data
Concept Annotated Datasets
Name | Task | N. of concepts | N. of classes |
---|---|---|---|
CUB | birds classification | 312 | 200 |
AwA2 | animals classification | 85 | 50 |
CelebA | identities classification | 6 | 1,000 |
OAI | x-ray grading | 10 | 4 |
WBCAtt | white blood cells classification | 31 | 5 |
Fitzpatrick 17k (subset) | skin diseases classification | 48 | 2 |
Diverse Dermatology Images (DDI) | skin diseases classification | 48 | 2 |
Skincon (Fitz sub + DDI annotated) | skin diseases classification | 48 | 2 |
DermaCon-IN | skin diseases classification | 47 | 8 |
Substitutions on CUB (SUB) | synthetic bird classification | 312 | 200 |
Publication | Paper Title | Code/Project |
---|---|---|
ICML | Concept bottleneck models | Code |
NMI | Concept whitening for interpretable image recognition | Code |
Acknowledgement
This project was originally inspired by https://github.com/kkzhang95/Awesome_Concept_Bottleneck_Models. We thank the authors for their contributions. Our main motivation is to provide an additional architecture organized by research focus, supplement it with more recent papers, and sort them by conference name for easier navigation.