Abstract

Large language models (LLMs) can interact with external tools to retrieve real-time information, make reservations, etc. While complete tool specifications for the external tools can be included in the prompt of an LLM, doing so increases context size and latency. This disclosure describes techniques that use small tool-selector models alongside large language models to select tools that are appropriate to the task specified in the prompt. The techniques preserve prompt space by ensuring that the LLM accesses the full specifications of a relevant tool only when necessary. The techniques enable an LLM to handle complex requests, such as multi-step tasks or interactions with external services, while maintaining high responsiveness and reducing the computational overhead associated with processing extensive tool documentation.

Creative Commons License

Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.

Share

COinS