Yushi-Hu/tifa
TIFA: Accurate and Interpretable Text-to-Image Faithfulness Evaluation with Question Answering
This tool helps researchers and developers working with text-to-image models understand how accurately generated images reflect their input text. You provide text prompts and the images your model creates, and it returns a detailed score and explanation of how well the image aligns with the text. This is for AI researchers or machine learning engineers who need to benchmark and improve their image generation models.
182 stars. No commits in the last 6 months.
Use this if you need an objective, detailed, and interpretable way to evaluate whether your text-to-image model is truly creating images that match the given text descriptions.
Not ideal if you are a casual user generating images and simply want to know if an image looks good, rather than conducting a fine-grained evaluation of model faithfulness.
Stars
182
Forks
12
Language
Python
License
Apache-2.0
Category
Last pushed
Apr 29, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/Yushi-Hu/tifa"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
zai-org/CogVideo
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
zhaorw02/DeepMesh
[ICCV 2025] Official code of DeepMesh: Auto-Regressive Artist-mesh Creation with Reinforcement Learning
YangLing0818/RPG-DiffusionMaster
[ICML 2024] Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with...
thu-nics/FrameFusion
[ICCV'25] The official code of paper "Combining Similarity and Importance for Video Token...
OpenMeshLab/MeshXL
[NeurIPS 2024] MeshXL: Neural Coordinate Field for Generative 3D Foundation Models, a 3D...