zjunlp/BiasEdit
[TrustNLP@NAACL 2025] BiasEdit: Debiasing Stereotyped Language Models via Model Editing
Language models sometimes generate biased or stereotypical text. This project helps researchers and developers remove harmful stereotypes, such as gender or race bias, from large language models without compromising their overall language abilities. You input a pre-trained language model and a dataset designed to identify bias, and it outputs a refined, less-biased language model ready for use in applications.
No commits in the last 6 months.
Use this if you need to reduce or eliminate specific biases from your language models to ensure fair and ethical AI outputs.
Not ideal if you are looking for a general-purpose language model fine-tuning tool rather than a specialized bias mitigation solution.
Stars
18
Forks
3
Language
Python
License
—
Category
Last pushed
Sep 30, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/zjunlp/BiasEdit"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
yueqingliang1/UNBench
Data and code for paper "𝗕𝗲𝗻𝗰𝗵𝗺𝗮𝗿𝗸𝗶𝗻𝗴 𝗟𝗟𝗠𝘀 𝗳𝗼𝗿 𝗣𝗼𝗹𝗶𝘁𝗶𝗰𝗮𝗹 𝗦𝗰𝗶𝗲𝗻𝗰𝗲: 𝗔 𝗨𝗻𝗶𝘁𝗲𝗱 𝗡𝗮𝘁𝗶𝗼𝗻𝘀 𝗣𝗲𝗿𝘀𝗽𝗲𝗰𝘁𝗶𝘃𝗲".
neha13rana/Stereotypical-Bias-Analyzer
In this project, we analyzed biases in ten domains using four datasets and created a useful...
MiuLab/FactAlign
Source code of our EMNLP 2024 paper "FactAlign: Long-form Factuality Alignment of Large Language Models"
josmarios/textbias-edu
Code for the article "From hype to evidence: exploring large language models for inter-group...