LLM Hallucination Mitigation LLM Tools

Tools and techniques for detecting, measuring, and correcting hallucinations in large language models across text and multimodal outputs. Does NOT include general LLM evaluation, factuality benchmarks, or non-hallucination-specific safety measures.

There are 46 llm hallucination mitigation tools tracked. 1 score above 50 (established tier). The highest-rated is vectara/hallucination-leaderboard at 62/100 with 3,122 stars. 1 of the top 10 are actively maintained.

Get all 46 projects as JSON

curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=llm-tools&subcategory=llm-hallucination-mitigation&limit=20"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.

# Tool Score Tier
1 vectara/hallucination-leaderboard

Leaderboard Comparing LLM Performance at Producing Hallucinations when...

62
Established
2 amir-hameed-mir/Sirraya_LSD_Code

Layer-wise Semantic Dynamics (LSD) is a model-agnostic framework for...

36
Emerging
3 Amirhosein-gh98/Gnosis

Can LLMs Predict Their Own Failures? Self-Awareness via Internal Circuits

35
Emerging
4 PKU-YuanGroup/Hallucination-Attack

Attack to induce LLMs within hallucinations

34
Emerging
5 MemTensor/HaluMem

HaluMem is the first operation level hallucination evaluation benchmark...

34
Emerging
6 NishilBalar/Awesome-LVLM-Hallucination

up-to-date curated list of state-of-the-art Large vision language models...

32
Emerging
7 intuit/sac3

Official repo for SAC3: Reliable Hallucination Detection in Black-Box...

31
Emerging
8 HillZhang1999/llm-hallucination-survey

Reading list of hallucination in LLMs. Check out our new survey paper:...

28
Experimental
9 OpenMOSS/HalluQA

Dataset and evaluation script for "Evaluating Hallucinations in Chinese...

27
Experimental
10 Mattbusel/LLM-Hallucination-Detection-Script

A comprehensive toolkit for detecting potential hallucinations in LLM...

25
Experimental
11 hongcheki/sweet-watermark

Official repository of the paper: Who Wrote this Code? Watermarking for Code...

25
Experimental
12 plll4zzx/Awesome-LLM-Watermark

A collection list for Large Language Model (LLM) Watermark

24
Experimental
13 VITA-MLLM/Woodpecker

✨✨Woodpecker: Hallucination Correction for Multimodal Large Language Models

23
Experimental
14 hzy312/Awesome-LLM-Watermark

UP-TO-DATE LLM Watermark paper. 🔥🔥🔥

23
Experimental
15 10nc0/Nyan-Protocol

Hallucination guard for AI — one invariant, any model, no training required.

23
Experimental
16 18907305772/KCA

EMNLP'2024: Knowledge Verification to Nip Hallucination in the Bud

23
Experimental
17 hallucinatemd/hallucinate.md

The open standard for telling AI not to hallucinate.

23
Experimental
18 amazon-science/THRONE

Code release for THRONE, a CVPR 2024 paper on measuring object...

23
Experimental
19 oumi-ai/halloumi-demo

Try out HallOumi, a state-of-the-art claim verification model in a simple UI!

22
Experimental
20 destination-earth/DestinE_ESA_DUA

DestinE Usage Assessment (DUA) framework services provides a system capable...

22
Experimental
21 hongbinye/Cognitive-Mirage-Hallucinations-in-LLMs

Repository for the paper "Cognitive Mirage: A Review of Hallucinations in...

22
Experimental
22 zjunlp/FactCHD

[IJCAI 2024] FactCHD: Benchmarking Fact-Conflicting Hallucination Detection

22
Experimental
23 utkukose/llm_persona_hallucination_study

Code for the study on persona vectors in controling / understanding...

22
Experimental
24 WindJammer6/37.-A-Hallucination-Mitigation-Scheme-in-Security-Policy-Generation-with-Large-Language-Models

Source code for the paper: A Hallucination Mitigation Scheme in Security...

21
Experimental
25 llaraspata/HallucinationDetection

Analyzing the correlation between Hallucinations and Knowledge Conflicts in...

21
Experimental
26 Sayar-212/TheLucidityBound

A Unified Framework for Hallucination Benchmarking and Mitigation in Large...

20
Experimental
27 MoNejjar/echo-hallucination-detect

Mitigating Hallucination Potential in User Prompts Through AI-Guided...

20
Experimental
28 Intelligent-Computing-Research-Group/HaVen

[DATE 2025] haven: hallucination-mitigated llm for verilog code generation...

20
Experimental
29 schipp/hallucination_guarantees

A pipeline that gives probabilistic guarantees for reducing contextual...

19
Experimental
30 stbiadmin/diu-halu

Framework for evaluating LLM hallucination detection in Department of...

19
Experimental
31 KhoiBui16/UIT_CS221_Basic_Natural_Language_Processing

The project focuses on classifying hallucinations in Vietnamese LLM outputs...

19
Experimental
32 lilakk/PostMark

Official repository for "PostMark: A Robust Blackbox Watermark for Large...

18
Experimental
33 rkhokhla/kakeya

When AI makes $10M decisions, hallucinations aren't bugs—they're business...

18
Experimental
34 IAAR-Shanghai/ICSFSurvey

Explore concepts like Self-Correct, Self-Refine, Self-Improve,...

17
Experimental
35 141forever/UncerSema4HalluDetec

This is the repository for the paper 'Enhancing Uncertainty Modeling with...

15
Experimental
36 tranhoangtu-it/halluciguard-api

HalluciGuard API — AI Hallucination Firewall as a Service. Detect and filter...

14
Experimental
37 strayfear/HalluWorld

🌍 Explore the HalluWorld project, a benchmark for understanding and defining...

14
Experimental
38 DegenAI-Labs/HalluWorld

Repository for the paper "A Unified Definition of Hallucination: It’s The...

14
Experimental
39 ruisizhang123/REMARK-LLM

[USENIX Security'24] REMARK-LLM: A robust and efficient watermarking...

12
Experimental
40 kjgpta/WhoDunIt-Evaluation_benchmark_for_culprit_detection_in_mystery_stories

WHODUNIT is a benchmark repository for evaluating large language models'...

12
Experimental
41 lasithadilshan/Hallucination-Detector-App

A Hallucination Detection Tool powered by UQML, designed to identify whether...

12
Experimental
42 akborsusom/watermark-ai-analysis

Reproduction and attack analysis of LLM text watermarking (Kirchenbauer et...

12
Experimental
43 Xieyangxinyu/Watermark-in-the-Classroom

This repository reproduces the results reported in Watermark in the...

11
Experimental
44 AInnovateLab/watermark-collision

[NAACL'25 Findings] Lost in Overlap: Exploring Logit-based Watermark...

11
Experimental
45 serhanylmz/pas2

PAS2: A Python-based hallucination detection system that evaluates AI...

10
Experimental
46 pranav-kural/llm-hallucination-detection-service

Build your own open-source REST API endpoint to detect hallucination in LLM...

10
Experimental