Repository navigation
hallucination
- Website
- Wikipedia
Loki: Open-source solution designed to automate the process of verifying factuality
Awesome-LLM-Robustness: a curated list of Uncertainty, Reliability and Robustness in Large Language Models
✨✨Woodpecker: Hallucination Correction for Multimodal Large Language Models
RefChecker provides automatic checking pipeline and benchmark dataset for detecting fine-grained hallucinations generated by Large Language Models.
[CVPR'24] HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(ision), LLaVA-1.5, and Other Multi-modality Models
[ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning
Explore concepts like Self-Correct, Self-Refine, Self-Improve, Self-Contradict, Self-Play, and Self-Knowledge, alongside o1-like reasoning elevation🍓 and hallucination alleviation🍄.
[ACL 2024] User-friendly evaluation framework: Eval Suite & Benchmarks: UHGEval, HaluEval, HalluQA, etc.
😎 curated list of awesome LMM hallucinations papers, methods & resources.
Code for ACL 2024 paper "TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space"
[NeurIPS 2024] Knowledge Circuits in Pretrained Transformers
[ICLR 2025] LLaVA-MoD: Making LLaVA Tiny via MoE-Knowledge Distillation
up-to-date curated list of state-of-the-art Large vision language models hallucinations research work, papers & resources
[IJCAI 2024] FactCHD: Benchmarking Fact-Conflicting Hallucination Detection
This is the official repo for Debiasing Large Visual Language Models, including a Post-Hoc debias method and Visual Debias Decoding strategy.
Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"
Code scanner to check for issues in prompts and LLM calls
[ICLR 2025] MLLM can see? Dynamic Correction Decoding for Hallucination Mitigation
Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute, relative and much more. It contains a list of all the available tool, methods, repo, code etc to detect hallucination, LLM evaluation, grading and much more.