lamalab-org/chembench

How good are LLMs at chemistry?

50
/ 100
Established

ChemBench helps chemists and materials scientists evaluate how well large language models (LLMs) and multimodal models perform on chemistry-related tasks. You provide a language model (or a vision-language model) and it outputs detailed reports on the model's accuracy across various chemistry topics. This is for researchers and developers working with AI in chemistry who need to assess model capabilities.

134 stars.

Use this if you need to systematically test and compare the performance of different AI models on chemical problems and datasets.

Not ideal if you are looking for a tool to perform chemical simulations or analyze experimental data directly without involving AI model evaluation.

computational chemistry materials science AI model evaluation chemical informatics drug discovery
No Package No Dependents
Maintenance 10 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 14 / 25

How are scores calculated?

Stars

134

Forks

16

Language

Python

License

MIT

Last pushed

Jan 26, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/lamalab-org/chembench"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.