OpenMOSS/Say-I-Dont-Know
[ICML'2024] Can AI Assistants Know What They Don't Know?
This project helps AI developers and researchers train large language models (LLMs) to recognize and communicate when they don't know the answer to a question. It provides specialized datasets and training methods to improve an AI assistant's ability to truthfully say "I don't know" rather than hallucinating responses. The end user is an AI developer or researcher aiming to build more reliable and honest AI assistants.
No commits in the last 6 months.
Use this if you are a developer working with large language models and want to enhance their capability to identify and articulate the boundaries of their knowledge.
Not ideal if you are an end-user simply looking to interact with an AI assistant, as this project focuses on the underlying training of such models.
Stars
85
Forks
10
Language
Python
License
—
Category
Last pushed
Feb 05, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/OpenMOSS/Say-I-Dont-Know"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
filipnaudot/llmSHAP
llmSHAP: a multi-threaded explainability framework using Shapley values for LLM-based outputs.
microsoft/automated-brain-explanations
Generating and validating natural-language explanations for the brain.
CAS-SIAT-XinHai/CPsyCoun
[ACL 2024] CPsyCoun: A Report-based Multi-turn Dialogue Reconstruction and Evaluation Framework...
wesg52/universal-neurons
Universal Neurons in GPT2 Language Models
ICTMCG/LLM-for-misinformation-research
Paper list of misinformation research using (multi-modal) large language models, i.e., (M)LLMs.