diff --git a/examples/agent/README.md b/examples/agent/README.md index aee17fa2f..f2fcc6667 100644 --- a/examples/agent/README.md +++ b/examples/agent/README.md @@ -27,19 +27,19 @@ Here's how to run an agent w/ local tool call: ./llama-server --jinja -fa --verbose \ -hfr NousResearch/Hermes-3-Llama-3.1-8B-GGUF -hff Hermes-3-Llama-3.1-8B.Q4_K_M.gguf \ - --chat-template-file <( python scripts/get_hf_chat_template.py NousResearch/Hermes-3-Llama-3.1-8B tool_use ) + --chat-template-file tests/chat/templates/NousResearch-Hermes-3-Llama-3.1-8B-tool_use.jinja ./llama-server --jinja -fa --verbose \ -hfr meetkai/functionary-small-v3.2-GGUF -hff functionary-small-v3.2.Q8_0.gguf \ - --chat-template-file <( python scripts/get_hf_chat_template.py meetkai/functionary-medium-v3.2 ) + --chat-template-file tests/chat/templates/meetkai-functionary-medium-v3.2.jinja ./llama-server --jinja -fa --verbose \ -hfr lmstudio-community/Llama-3.2-3B-Instruct-GGUF -hff Llama-3.2-3B-Instruct-Q6_K.gguf \ - --chat-template-file <( python scripts/get_hf_chat_template.py meta-llama/Llama-3.2-3B-Instruct ) + --chat-template-file tests/chat/templates/meta-llama-Llama-3.2-3B-Instruct.jinja ./llama-server --jinja -fa --verbose \ -hfr bartowski/Mistral-Nemo-Instruct-2407-GGUF -hff Mistral-Nemo-Instruct-2407-Q8_0.gguf \ - --chat-template-file <( python scripts/get_hf_chat_template.py mistralai/Mistral-Nemo-Instruct-2407 ) + --chat-template-file tests/chat/templates/mistralai-Mistral-Nemo-Instruct-2407.jinja # Generic support, e.g. Phi 3.5, Gemma 2b, but really anything goes