lamalab-org/chembench
How good are LLMs at chemistry?
ChemBench helps chemists and materials scientists evaluate how well large language models (LLMs) and multimodal models perform on chemistry-related tasks. You provide a language model (or a vision-language model) and it outputs detailed reports on the model's accuracy across various chemistry topics. This is for researchers and developers working with AI in chemistry who need to assess model capabilities.
134 stars.
Use this if you need to systematically test and compare the performance of different AI models on chemical problems and datasets.
Not ideal if you are looking for a tool to perform chemical simulations or analyze experimental data directly without involving AI model evaluation.
Stars
134
Forks
16
Language
Python
License
MIT
Category
Last pushed
Jan 26, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/lamalab-org/chembench"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
theochem/ModelHamiltonian
Generate 1- and 2-electron integrals so that molecular quantum chemistry software can be used...
pnnl/cactus
LLM Agent that leverages cheminformatics tools to provide informed responses.
jan-janssen/LangSim
Application of Large Language Models (LLM) for computational materials science - visit...
MasterAI-EAM/Darwin
An open-source project dedicated to build foundational large language model for natural science,...
andresilvapimentel/AI4Chem
AI4Chem is a code to test the ability of large language models (ChatGPT) to comprehend Chemistry.