Skip to content

A Comprehensive Benchmark to Evaluate LLMs as Agents

Notifications You must be signed in to change notification settings

XueyangFeng/AgentBench

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

1 Commit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

AgentBench

🌐 Website | 🐦 Twitter | ✉️ Google Group | 📃 Paper | 🌏中文文档

👋 Join our Slack for Q & A or collaboration on next version of AgentBench!

📌Introducing AgentBench v0.2🎉

You are now browsing AgentBench v0.2. If you wish to use the older version, you can revert to v0.1.

Based on v0.1, we:

  • Updated the framework architecture for easier use and extension
  • Adjusted some task settings
  • Added test results for more models
  • Released the full data for the Dev and Test sets

AgentBench: Evaluating LLMs as Agents

agentbench-cover.mp4

AgentBench is the first benchmark designed to evaluate LLM-as-Agent across a diverse spectrum of different environments. It encompasses 8 distinct environments to provide a more comprehensive evaluation of the LLMs' ability to operate as autonomous agents in various scenarios. These environments include 5 freshly created domains, namely

  • Operating System (OS)
  • Database (DB)
  • Knowledge Graph (KG)
  • Digital Card Game (DCG)
  • Lateral Thinking Puzzles (LTP)

as well as 3 recompiled from published datasets:

Table of Contents

Dataset Summary

We offer two splits for each dataset: Dev and Test. The multi-turn interaction requires an LLMs to generate around 4k and 13k times respectively.

Leaderboard

Here is the scores on test set (standard) results of AgentBench.

While LLMs begin to manifest their proficiency in LLM-as-Agent, gaps between models and the distance towards practical usability are significant.

Prerequisites

Install the dependencies.

pip install -r requirements.txt

Also, ensure that Docker is properly installed. And locally, you have images for mysql and ubuntu.

Quick Start

This section will guide you on how to quickly use gpt-3.5-turbo-0613 as an agent to launch the dbbench-std, os-std, and kg-std tasks. For the specific framework structure, please refer to Framework Introduction. For more detailed configuration and launch methods, please check Configuration Guide and Program Entrance Guide.

Configure the Agent

Fill in your OpenAI API Key at the correct location in configs/agents/openai-chat.yaml.

You can try using python -m src.client.agent_test to check if your Agent is configured correctly.

Start the task server

Starting the task worker involves specific tasks. Manual starting might be cumbersome; hence, we provide an automated script.

The assumption for this step is that ports from 5000 to 5015 are available.

python -m src.start_task -a

This will launch five task_workers each for dbbench-std, os-std, and kg-std tasks and automatically connect them to the controller on port 5000.

Start the assigner

This step is to actually start the tasks.

If everything is correctly configured so far, you can now initiate the task tests.

python -m src.assigner

Next Steps

If you wish to launch more tasks or use other models, you can refer to the content in Configuration Guide and Program Entrance Guide.

For the environment of the remaining five tasks, you will need to download the Docker images we provide.

longinyu/agentbench-ltp
longinyu/agentbench-webshop
longinyu/agentbench-mind2web
longinyu/agentbench-card_game
longinyu/agentbench-alfworld

The resource consumption of a single task_worker for the eight tasks is roughly as follows; consider this when launching:

Task Name Start-up Speed Memory Consumption
webshop ~3min ~15G
mind2web ~5min ~1G
db ~20s < 500M
alfworld ~10s < 500M
card_game ~5s < 500M
ltp ~5s < 500M
os ~5s < 500M
kd ~5s < 500M

Citation

@article{liu2023agentbench,
  title   = {AgentBench: Evaluating LLMs as Agents},
  author  = {Xiao Liu and Hao Yu and Hanchen Zhang and Yifan Xu and Xuanyu Lei and Hanyu Lai and Yu Gu and Hangliang Ding and Kaiwen Men and Kejuan Yang and Shudan Zhang and Xiang Deng and Aohan Zeng and Zhengxiao Du and Chenhui Zhang and Sheng Shen and Tianjun Zhang and Yu Su and Huan Sun and Minlie Huang and Yuxiao Dong and Jie Tang},
  year    = {2023},
  journal = {arXiv preprint arXiv: 2308.03688}
}

About

A Comprehensive Benchmark to Evaluate LLMs as Agents

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • C++ 46.4%
  • Python 45.7%
  • Makefile 3.3%
  • HTML 1.5%
  • CMake 1.5%
  • Shell 1.4%
  • Other 0.2%