bigai-nlco/langsuite

Official Repo of LangSuitE

31
/ 100
Emerging

This project helps evaluate how well large language models (LLMs) can control and interact within text-based environments for tasks like navigation and object manipulation. It takes in configurations for agents and environments, and a chosen LLM, to simulate and assess its performance in these 'embodied' scenarios. Researchers and AI developers working on improving LLM interaction with virtual worlds would find this useful.

No commits in the last 6 months.

Use this if you need a systematic way to test and benchmark large language models on their ability to perform actions and communicate in simulated text-based environments without a visual simulator.

Not ideal if you are looking for a visual simulation environment or a tool to directly deploy LLMs into real-world physical robots.

LLM evaluation embodied AI agent simulation AI research natural language understanding
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 9 / 25
Maturity 16 / 25
Community 6 / 25

How are scores calculated?

Stars

84

Forks

3

Language

Python

License

MIT

Last pushed

Aug 15, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/bigai-nlco/langsuite"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.