on Mar 29, 2025 i took...

LLM — Tool calling

Tool calling allows a chat model to respond to a given prompt by "calling a tool". That enables LLMs to interact with enternal functions, like search web, fetch data, or executing taks. That enhance their utility beyond the text genenration.

How it works?

LLMs can't directly execute functions. However, there are two approaches that help LLMs to identify when and how to call a tool:

  1. LLMs with specifically trained or fine-tuned to recognize when the tool should be used. This involves teaching the machine to detect patterns in input that signal a need for tool interference
  2. Tools are provided to the LLMs via system prompts, usually is a list of structure functions with description. LLM can generate a structured format, indicating which tool will be call with what parameters. A framework, such as Ollama, must know how to parse structured response and handle the execution
Example how simulate tool calling works in LLM

System prompts using in Ollama/LMStudio example:

For each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:
<tool_call>
{"name": <function-name>, "arguments": <args-json-object>}
</tool_call>
{{- end }}<|im_end|>
{{ end }}
 
...
 
<|im_start|>assistant
{{ if .Content }}{{ .Content }}
{{- else if .ToolCalls }}<tool_call>
{{ range .ToolCalls }}{"name": "{{ .Function.Name }}", "arguments": {{ .Function.Arguments }}}
{{ end }}</tool_call>
{{- end }}{{ if not $last }}<|im_end|>
{{ end }}

References

Thanks huytdhuytd for sharing this

Edit