Original PDF: TOOLLLM: FACILITATING LARGE LANGUAGE MODELS TO MASTER 16000+ REAL-WORLD APIS
Author: Yujia Qin , Shihao Liang, Yining Ye, Kunlun Zhu, Lan Yan, Yaxi Lu and more
Summary & Commentary
Introduction
The paper titled "ToolLLM: Eliciting Tool-Use Capabilities within Large Language Models" introduces ToolLLM, a general tool-use framework for data construction, model training, and evaluation. The authors have created an instruction-tuning dataset called ToolBench, which was automatically generated using ChatGPT. ToolBench comprises 16,464 real-world RESTful APIs across 49 categories from RapidAPI Hub. The authors aim to enhance the tool-use capabilities of open-source Large Language Models (LLMs), such as LLaMA and Vicuna, which currently lack the sophistication to understand human instructions and interact appropriately with APIs. The authors argue that the current instruction tuning largely focuses on closed-source models like ChatGPT, which have demonstrated excellent tool-use capabilities (Page 1).