lechmazur/goods
LLM public goods game
This project helps researchers and evaluators understand how different large language models (LLMs) behave in a classic economic game. It simulates a "public goods game" where LLMs decide to contribute tokens to a shared pool or keep them. The output is a leaderboard showing which LLMs are more cooperative (contribute more) and which tend to "free-ride" across various game conditions. Anyone evaluating LLM behavior for research or practical applications would find this useful.
No commits in the last 6 months.
Use this if you need to compare how different large language models cooperate or free-ride in a group economic scenario.
Not ideal if you are looking to benchmark LLMs on tasks like code generation, creative writing, or factual recall.
Stars
8
Forks
—
Language
—
License
—
Category
Last pushed
Feb 22, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/lechmazur/goods"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ai4co/reevo
[NeurIPS 2024] ReEvo: Large Language Models as Hyper-Heuristics with Reflective Evolution
SALT-NLP/collaborative-gym
Framework and toolkits for building and evaluating collaborative agents that can work together...
Gen-Verse/LatentMAS
Latent Collaboration in Multi-Agent Systems
lean-dojo/LeanCopilot
LLMs as Copilots for Theorem Proving in Lean
WooooDyy/AgentGym-RL
Code and implementations for the paper "AgentGym-RL: Training LLM Agents for Long-Horizon...