lamalab-org/macbench

Probing the limitations of multimodal language models for chemistry and materials research

36
/ 100
Emerging

This tool helps chemistry and materials science researchers evaluate how well advanced AI models (multimodal language models) understand and respond to questions using both text and images in your field. You input a multimodal language model and a set of chemistry/materials research tasks, and it provides a report on the model's performance across various stages of scientific work. This is designed for scientists, engineers, and researchers who want to assess or compare AI models for scientific discovery workflows.

Use this if you need to objectively benchmark and understand the strengths and weaknesses of different multimodal AI models for tasks in chemistry and materials research.

Not ideal if you are looking for an AI model to directly perform scientific experiments or data analysis, rather than evaluating the models themselves.

chemistry-research materials-science AI-model-evaluation scientific-discovery computational-chemistry
No Package No Dependents
Maintenance 10 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 4 / 25

How are scores calculated?

Stars

23

Forks

1

Language

Python

License

MIT

Last pushed

Feb 01, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/lamalab-org/macbench"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.