open-compass/ANAH
[ACL 2024] ANAH & [NeurIPS 2024] ANAH-v2 & [ICLR 2025] Mask-DPO
This project helps anyone evaluating or building large language models (LLMs) to identify and reduce 'hallucinations' — instances where LLMs generate factually incorrect information. It takes LLM-generated text as input and provides detailed annotations about hallucinated sentences, along with a factuality score. The output can be used to improve the reliability of LLMs or to compare the performance of different models.
No commits in the last 6 months.
Use this if you need to precisely measure the factual accuracy of LLM responses, fine-tune an LLM to reduce its tendency to hallucinate, or benchmark different LLMs for factual correctness.
Not ideal if you are looking for a simple, out-of-the-box solution for general text generation without a focus on factuality or advanced LLM development.
Stars
63
Forks
3
Language
Python
License
Apache-2.0
Category
Last pushed
Apr 30, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/open-compass/ANAH"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
THU-BPM/MarkLLM
MarkLLM: An Open-Source Toolkit for LLM Watermarking.(EMNLP 2024 System Demonstration)
git-disl/Vaccine
This is the official code for the paper "Vaccine: Perturbation-aware Alignment for Large...
zjunlp/Deco
[ICLR 2025] MLLM can see? Dynamic Correction Decoding for Hallucination Mitigation
HillZhang1999/ICD
Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced...
voidism/DoLa
Official implementation for the paper "DoLa: Decoding by Contrasting Layers Improves Factuality...