This is a list of some awesome works on mitigating hallucination in large multimodal models.
- Hallucination of Multimodal Large Language Models: A Survey (Apr. 30, 2024)
- A Survey on Hallucination in Large Vision-Language Models (Feb. 1, 2024)
- Object Hallucination in Image Captioning(Sep. 6, 2018, EMNLP 2018)
- Evaluating Object Hallucination in Large Vision-Language Models (May. 17, 2023, EMNLP 2023)
- Eyes Wide Shut? Exploring the Visual Shortcomings of Multimodal LLMs (Jan. 11, 2024, CVPR 2024)
- MME: A Comprehensive Evaluation Benchmark for Multimodal Large Language Models (Jun. 23, 2023)
- MASH-VLM: Mitigating Action-Scene Hallucination in Video-LLMs through Disentangled Spatial-Temporal Representations (Mar. 20, 2025, CVPR 2025)
- AMBER: An LLM-free Multi-dimensional Benchmark for MLLMs Hallucination Evaluation (Nov. 13, 2023)
- Aligning Large Multimodal Models with Factually Augmented RLHF (Sep. 25, 2023, ACL 2024)
- MASH-VLM: Mitigating Action-Scene Hallucination in Video-LLMs through Disentangled Spatial-Temporal Representations (Mar. 20, 2025, CVPR 2025)
- ClearSight: Visual Signal Enhancement for Object Hallucination Mitigation in Multimodal Large language Models (Mar. 17, 2025)
- Through the Magnifying Glass: Adaptive Perception Magnification for Hallucination-Free VLM Decoding (Mar. 13, 2025)
- EAZY: Eliminating Hallucinations in LVLMs by Zeroing out Hallucinatory Image Tokens (Mar. 10, 2025)
- Mask-DPO: Generalizable Fine-grained Factuality Alignment of LLMs (Mar. 4, 2025, ICLR 2025)
- PerturboLLaVA: Reducing Multimodal Hallucinations with Perturbative Visual Training (Mar. 9, 2025, ICLR 2025)
- Octopus: Alleviating Hallucination via Dynamic Contrastive Decoding (Mar. 1, 2025, CVPR 2025)
- Exploring Causes and Mitigation of Hallucinations in Large Vision Language Models (Feb. 24, 2025)
- Refine Knowledge of Large Language Models via Adaptive Contrastive Learning (Feb. 11, 2025, ICLR 2025)
- Mitigating Hallucinations in Large Vision-Language Models via DPO: On-Policy Data Hold the Key (Jan. 16, 2025, CVPR 2025)
- VASparse: Towards Efficient Visual Hallucination Mitigation for Large Vision-Language Model via Visual-Aware Sparsification (Jan. 11, 2025, CVPR 2025)
- Nullu: Mitigating Object Hallucinations in Large Vision-Language Models via HalluSpace Projection (Dec. 18, 2024, CVPR 2025)
- Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations (Oct. 3, 2024, ICLR 2025)
- Devils in Middle Layers of Large Vision-Language Models: Interpreting, Detecting and Mitigating Object Hallucinations via Attention Lens (Nov. 23, 2024, CVPR 2025)
- ICT: Image-Object Cross-Level Trusted Intervention for Mitigating Object Hallucination in Large Vision-Language Models (Nov. 22, 2024, CVPR 2025)
- Eagle: Exploring The Design Space for Multimodal LLMs with Mixture of Encoders (Aug. 28, 2024, ICLR 2025)
- Self-Introspective Decoding: Alleviating Hallucinations for Large Vision-Language Models (Aug. 4, 2024, ICLR 2025)
- Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMs (Jul. 31, 2024, ECCV 2024)
- Mitigating Object Hallucinations in Large Vision-Language Models with Assembly of Global and Local Attention (Jun. 18, 2024, CVPR 2025)
- Reducing Hallucinations in Vision-Language Models via Latent Space Steering (Oct. 21, 2024, ICLR 2025)
- Woodpecker: Hallucination Correction for Multimodal Large Language Models (Oct. 10, 2024)
- Alleviating Hallucinations in Large Vision-Language Models through Hallucination-Induced Optimization (May. 24, 2024, NeurIPS 2024)
- Mitigating Hallucinations in Large Vision-Language Models with Instruction Contrastive Decoding (Mar. 27, 2024, ACL 2024)
- Mitigating Object Hallucination via Concentric Causal Attention (Oct. 21, 2024, NeurIPS 2024)
- DAMRO: Dive into the Attention Mechanism of LVLM to Reduce Object Hallucination (Oct. 6, 2024 EMNLP 2024)
- Skip \n: A Simple Method to Reduce Hallucination in Large Vision-Language Models (Feb. 2, 2024, ICLR 2024)
- Eyes Wide Shut? Exploring the Visual Shortcomings of Multimodal LLMs (Jan. 11, 2024, CVPR 2024)
- Hallucination Augmented Contrastive Learning for Multimodal Large Language Model (Dec. 12, 2023, CVPR 2024)
- OPERA: Alleviating Hallucination in Multi-Modal Large Language Models via Over-Trust Penalty and Retrospection-Allocation (Nov. 29, 2023, CVPR 2024)
- Mitigating Object Hallucinations in Large Vision-Language Models through Visual Contrastive Decoding (Nov. 28, 2023, CVPR 2024)
- Analyzing and Mitigating Object Hallucination in Large Vision-Language Models (Oct. 1, 2023, ICLR 2024)
- DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models (Sep. 7, 2023, ICLR 2024)
- Inference-Time Intervention: Eliciting Truthful Answers from a Language Model (Jun. 26, 2023, NeurIPS 2023)
This project is inspired by Awesome-MLLM-Hallucination and Awesome-Multimodal-Large-Language-Models. Thanks for their contribution to the research community.