From 2ceabee0f884856b1b264794722c1e9f9b5b96f5 Mon Sep 17 00:00:00 2001 From: ochafik Date: Sat, 18 Jan 2025 01:36:46 +0000 Subject: [PATCH] Fix fetch_server_test_models.py (avoid conv trap) --- scripts/fetch_server_test_models.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/scripts/fetch_server_test_models.py b/scripts/fetch_server_test_models.py index 80c532bdd..a0783ce3c 100755 --- a/scripts/fetch_server_test_models.py +++ b/scripts/fetch_server_test_models.py @@ -86,7 +86,7 @@ if __name__ == '__main__': logging.warning(f'Skipping model at {m.hf_repo} / {m.hf_file} because it is a split file') continue logging.info(f'Using llama-cli to ensure model {m.hf_repo}/{m.hf_file} was fetched') - cmd = [cli_path, '-hfr', m.hf_repo, '-hff', m.hf_file, '-n', '1', '-p', 'Hey', '--no-warmup', '--log-disable'] + cmd = [cli_path, '-hfr', m.hf_repo, '-hff', m.hf_file, '-n', '1', '-p', 'Hey', '--no-warmup', '--log-disable', '-no-cnv'] if m.hf_file != 'tinyllamas/stories260K.gguf' and not m.hf_file.startswith('Mistral-Nemo'): cmd.append('-fa') try: