ltzheng/agent-studio

[ICLR 2025] A trinity of environments, tools, and benchmarks for general virtual agents

45
/ 100
Emerging

This project provides a comprehensive toolkit for AI researchers and developers working on virtual agents. It helps in creating, evaluating, and benchmarking AI agents that can interact with various computer software using visual observations (like video) and actions (like GUI clicks or API calls). The user provides agent code, and the toolkit outputs performance metrics and detailed insights into agent capabilities.

229 stars. No commits in the last 6 months.

Use this if you are developing or studying general-purpose virtual agents and need a standardized environment, tools, and benchmarks to test their ability to interact with diverse software, from terminal commands to graphical user interfaces.

Not ideal if you are looking for a pre-built agent to solve a specific problem or if you are not involved in the research and development of AI agents.

AI-agent-development virtual-agent-benchmarking human-computer-interaction-automation intelligent-system-evaluation
Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 17 / 25

How are scores calculated?

Stars

229

Forks

30

Language

Python

License

AGPL-3.0

Last pushed

Jun 16, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/agents/ltzheng/agent-studio"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.