braingpt-lovelab/BrainBench
Source code for
This project helps neuroscience researchers replicate and extend findings from a study that used Large Language Models (LLMs) to predict neuroscience experimental results. It takes raw experimental data and generates analyses and plots, allowing researchers to reproduce the original paper's figures and findings. This is for neuroscientists and cognitive scientists interested in using or evaluating AI for scientific discovery.
No commits in the last 6 months.
Use this if you are a neuroscience researcher looking to reproduce or build upon the results of the "Large language models surpass human experts in predicting neuroscience results" paper.
Not ideal if you are a general AI developer looking for a framework unrelated to neuroscience research replication.
Stars
85
Forks
12
Language
—
License
Apache-2.0
Category
Last pushed
Nov 29, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/braingpt-lovelab/BrainBench"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Featured in
Higher-rated alternatives
sierra-research/tau2-bench
τ²-Bench: Evaluating Conversational Agents in a Dual-Control Environment
xlang-ai/OSWorld
[NeurIPS 2024] OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments
bigcode-project/bigcodebench
[ICLR'25] BigCodeBench: Benchmarking Code Generation Towards AGI
THUDM/AgentBench
A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)
scicode-bench/SciCode
A benchmark that challenges language models to code solutions for scientific problems