Skip to content

Commit 4299404

Browse files
authored
update function-calling.md w/ template override for functionary-small-v3.2 (ggml-org#12214)
1 parent e9b2f84 commit 4299404

File tree

1 file changed

+12
-8
lines changed

1 file changed

+12
-8
lines changed

docs/function-calling.md

+12-8
Original file line numberDiff line numberDiff line change
@@ -287,37 +287,41 @@ Here are some models known to work (w/ chat template override when needed):
287287

288288
llama-server --jinja -fa -hf bartowski/Qwen2.5-7B-Instruct-GGUF:Q4_K_M
289289
llama-server --jinja -fa -hf bartowski/Mistral-Nemo-Instruct-2407-GGUF:Q6_K_L
290-
llama-server --jinja -fa -hf bartowski/functionary-small-v3.2-GGUF:Q4_K_M
291290
llama-server --jinja -fa -hf bartowski/Llama-3.3-70B-Instruct-GGUF:Q4_K_M
292291

293-
# Native support for DeepSeek R1 works best w/ our own template (official template buggy)
292+
# Native support for DeepSeek R1 works best w/ our template override (official template is buggy, although we do work around it)
294293

295294
llama-server --jinja -fa -hf bartowski/DeepSeek-R1-Distill-Qwen-7B-GGUF:Q6_K_L \
296-
--chat-template-file models/templates/llama-cpp-deepseek-r1.jinja
295+
--chat-template-file models/templates/llama-cpp-deepseek-r1.jinja
297296

298297
llama-server --jinja -fa -hf bartowski/DeepSeek-R1-Distill-Qwen-32B-GGUF:Q4_K_M \
299-
--chat-template-file models/templates/llama-cpp-deepseek-r1.jinja
298+
--chat-template-file models/templates/llama-cpp-deepseek-r1.jinja
300299

301300
# Native support requires the right template for these GGUFs:
302301

302+
llama-server --jinja -fa -hf bartowski/functionary-small-v3.2-GGUF:Q4_K_M
303+
--chat-template-file models/templates/meetkai-functionary-medium-v3.2.jinja
304+
303305
llama-server --jinja -fa -hf bartowski/Hermes-2-Pro-Llama-3-8B-GGUF:Q4_K_M \
304-
--chat-template-file <( python scripts/get_chat_template.py NousResearch/Hermes-2-Pro-Llama-3-8B tool_use )
306+
--chat-template-file models/templates/NousResearch-Hermes-2-Pro-Llama-3-8B-tool_use.jinja
305307

306308
llama-server --jinja -fa -hf bartowski/Hermes-3-Llama-3.1-8B-GGUF:Q4_K_M \
307-
--chat-template-file <( python scripts/get_chat_template.py NousResearch/Hermes-3-Llama-3.1-8B tool_use )
309+
--chat-template-file models/templates/NousResearch-Hermes-3-Llama-3.1-8B-tool_use.jinja
308310

309311
llama-server --jinja -fa -hf bartowski/firefunction-v2-GGUF -hff firefunction-v2-IQ1_M.gguf \
310-
--chat-template-file <( python scripts/get_chat_template.py fireworks-ai/llama-3-firefunction-v2 tool_use )
312+
--chat-template-file models/templates/fireworks-ai-llama-3-firefunction-v2.jinja
311313

312314
llama-server --jinja -fa -hf bartowski/c4ai-command-r7b-12-2024-GGUF:Q6_K_L \
313-
--chat-template-file <( python scripts/get_chat_template.py CohereForAI/c4ai-command-r7b-12-2024 tool_use )
315+
--chat-template-file models/templates/CohereForAI-c4ai-command-r7b-12-2024-tool_use.jinja
314316

315317
# Generic format support
316318
llama-server --jinja -fa -hf bartowski/phi-4-GGUF:Q4_0
317319
llama-server --jinja -fa -hf bartowski/gemma-2-2b-it-GGUF:Q8_0
318320
llama-server --jinja -fa -hf bartowski/c4ai-command-r-v01-GGUF:Q2_K
319321
```
320322

323+
To get the official template from original HuggingFace repos, you can use [scripts/get_chat_template.py](../scripts/get_chat_template.py) (see examples invocations in [models/templates/README.md](../models/templates/README.md))
324+
321325
> [!TIP]
322326
> If there is no official `tool_use` Jinja template, you may want to set `--chat-template chatml` to use a default that works with many models (YMMV!), or write your own (e.g. we provide a custom [llama-cpp-deepseek-r1.jinja](../models/templates/llama-cpp-deepseek-r1.jinja) for DeepSeek R1 distills)
323327

0 commit comments

Comments
 (0)