---
language:
- en
license: apache-2.0
task_categories:
- question-answering
tags:
- agent
- benchmark
- tool-use
- korean
---
# **🇰🇷 Ko-AgentBench v1**
**"Korean Agent Benchmark Project"**
**English | [한국어](README.md)**
As AI agents become more sophisticated, it has become crucial to precisely measure their performance under conditions similar to real-world environments. However, most benchmarks are designed based on English-speaking environments, which limits their ability to reflect Korea's unique usage contexts.
To address this issue, we have developed a high-quality agent benchmark specialized for the Korean real-world usage environment.
# Ko-AgentBench Key Features ✨
**1. Step-by-step Task Design**
We have comprehensively analyzed agent capabilities across 7 levels, from simple tool calls to long-term contextual abilities and robustness handling capabilities.
**2. 18 Korean-specific APIs and High-quality Scenarios Tailored to Real-life Environments**
Based on APIs from Korean real-world usage environments such as Naver, Maps, Kakao, and websites, we have implemented realistic problem-solving scenarios closely related to domestic users' daily lives, such as 'appointment booking' and 'blog review search'.
**3. Cache-based Iterative Evaluation and Robustness Testing**
We solve chronic problems of existing benchmarks, such as 'information attribute inconsistency changes'.
By improving failed API responses, we ensure benchmark consistency and reliability.
By evaluating error recognition/response capabilities (strategies) in intentional error situations, we select models that operate stably even in real-world environments.
**4. Step-specific Precision Metrics**
We evaluate the necessity/requirements of problem-solving step by step, including tool selection, parameter configuration, and data flow. Through this, we quantitatively identify the strengths and weaknesses of models.
## **Data Loading**
```bash
from datasets import load_dataset
# Load specific level
dataset = load_dataset("Hugging-Face-KREW/Ko-AgentBench", data_files="L1.json")
# Or load all levels
dataset = load_dataset("Hugging-Face-KREW/Ko-AgentBench", data_files="*.json")
```
# Dataset Overview
- Define task classification system for agent benchmark design
- Design to evaluate agent's tool calling capabilities in a step-by-step manner
## Dataset Scope
- Evaluation Target: Open-weight sLLM (supports tool calling), Commercial APIs
- Evaluation Scope: Agent tool calling performance in single-turn and multi-turn conversation situations
- Applied APIs: 18 Korean-specific open APIs
# Task Levels
## Single-Turn
**L1. Single Tool Call**
- Goal: Verify the most basic API calling capability
- Description: Check if the given tool can be executed with correct parameters
- Feature: Evaluate "accuracy only" by performing requests with specified API names or natural language requests as-is
- Example: "Search for 'Rapid Current' using Naver Book API and tell me the price."
- Example: "Tell me the price of the 'Rapid Current' book"
**L2. Tool Selection**
- Goal: Verify the ability to select the optimal API among multiple candidate tools
- Description: Users make requests in natural language, and the model must select the most suitable tool from the given tool list
- Feature: Evaluate accurate tool mapping with input natural language
- Example: "Check the price of the 'All Back English Middle 2-1 Cheonjae (Kim)' book."
- Candidate tools: `hotel_booking_api`, `aladin_books_api`
- Candidate tools must have no mutual correlation.
**L3. Sequential Tool Reasoning**
- Goal: Verify planning and execution capabilities through multi-step reasoning
- Description: Check if a correct pipeline can be constructed by connecting the results of one tool as input to another tool
- Feature: Evaluate "planned chain-of-tools" rather than simple calls
- Example: "Tell me when the Instax11 I bought from 11st Amazon will be delivered"
- Candidate tools: `11st_order_api`, `customs_api`, `cj_delivery_api`
- Tools must be callable sequentially (11st delivery number inquiry → customs clearance → courier company)
**L4. Parallel Tool Reasoning**
- Goal: Collect information in parallel and derive conclusions by synthesizing it
- Description: Simultaneously call multiple independent tools, compare and analyze results, then produce final answers
- Feature: Evaluate multi-source aggregation (information synthesis and comparison ability)
- Example: "Check the stock of the 'Hanroro Grapefruit Apricot Club' book."
- Candidate tools: `kyobo_books_api`, `aladin_books_api`
- Expected answer: There are 12 books at Kyobo Book Centre and 18 books at Aladin, totaling 30 books.
- At this time, candidate tools must handle the same function in parallel.
**L5. Error Handling and Robustness**
- Goal: Verify coping ability in error situations
- Description: Evaluate how various failure modes are handled, not just "failed"
- **Sub-items:**
- A. Request for additional questions
- Guide users to make clearer requests when information is insufficient
- B. Hallucination prevention
- Prohibit calling non-existent APIs
- Prohibit "pretending to succeed" answers when failed
- C. Fallback maneuvers
- Whether alternative APIs with the same function can be utilized when specific API errors occur
- Example: "When Naver Movie API call fails → Report 'API call failed' or call Kakao Movie API as alternative"
## Multi-Turn
**L6. Efficient Tool Utilization**
- Goal: Verify the ability to efficiently reuse previous tool results
- Description: While recalling APIs in all situations is accurate, it's inefficient in terms of cost and delay. Conversely, unconditionally reusing old information also causes accuracy problems.
- Feature: Evaluate whether reasonable choices can be made between "recall vs reuse"
- Example:
- User: "Compare Coupang and Naver prices." → Result: Coupang 80, Naver 85
- User: "What was the Naver price?"
- Correct answer: 85 (utilize past information, avoid unnecessary recalls)
- Wrong answer: Call API again or "I don't know"
**L7. Long-Context Reasoning**
- Goal: Verify the ability to maintain long-term context in multi-turn conversations
- Description: Remember information from several turns ago and correctly perform tool calling by connecting it with new questions
- Example:
- User's first question: "I'm going to travel to Jeju Island."
- Later: "How's the weather?" → Call weather API using Jeju Island context
- (Additional turn) "If it rains, find places where I can buy an umbrella." → Utilize all previous Jeju Island + weather context
## Links
You can check more detailed information about Ko-AgentBench.
- 🏆 [Live Leaderboard](https://huggingface.co/spaces/huggingface-KREW/Ko-AgentBench)
- 📊 [Dataset](https://huggingface.co/datasets/huggingface-KREW/Ko-AgentBench)
- 📝 [Github](https://github.com/Hugging-Face-KREW/Ko-AgentBench)
## Contact
If you have any questions about the dataset and benchmark, please contact us!
Hugging Face KREW is a Korean non-profit research organization that strives to deeply understand artificial intelligence through Hugging Face and contribute to open source.
- ✍🏻 Blog: [KREW-blog](https://hugging-face-krew.github.io/)
- 🐦 HuggingFace Community: [@huggingface-KREW](https://huggingface.co/huggingface-KREW)
- 💼 LinkedIn: [Hugging Face KREW](https://www.linkedin.com/company/hugging-face-krew/)