Co-authored-by: Pierre F <redacted>
- Use `--chat-template-file` to override the template when appropriate (see examples below)
- Generic support may consume more tokens and be less efficient than a model's native format.
+- Multiple/parallel tool calling is supported on some models but disabled by default, enable it by passing `"parallel_tool_calls": true` in the completion endpoint payload.
+
<details>
<summary>Show some common templates and which format handler they use</summary>
`parse_tool_calls`: Whether to parse the generated tool call.
+`parallel_tool_calls` : Whether to enable parallel/multiple tool calls (only supported on some models, verification is based on jinja template).
+
*Examples:*
You can use either Python `openai` library with appropriate checkpoints: