From 029d911cd1ea12dd957078d6b0b66773fb52c1e1 Mon Sep 17 00:00:00 2001 From: niansa/tuxifan Date: Fri, 28 Jul 2023 02:01:29 +0200 Subject: [PATCH] Added links to LLaMA 2 70B models --- README.md | 2 ++ 1 file changed, 2 insertions(+) diff --git a/README.md b/README.md index 6c49526a3..e20f57131 100644 --- a/README.md +++ b/README.md @@ -640,8 +640,10 @@ python3 convert.py pygmalion-7b/ --outtype q4_1 - Alternatively, if you want to save time and space, you can download already converted and quantized models from [TheBloke](https://huggingface.co/TheBloke), including: - [LLaMA 2 7B base](https://huggingface.co/TheBloke/Llama-2-7B-GGML) - [LLaMA 2 13B base](https://huggingface.co/TheBloke/Llama-2-13B-GGML) + - [LLaMA 2 70B base](https://huggingface.co/TheBloke/Llama-2-70B-GGML) - [LLaMA 2 7B chat](https://huggingface.co/TheBloke/Llama-2-7B-chat-GGML) - [LLaMA 2 13B chat](https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML) + - [LLaMA 2 70B chat](https://huggingface.co/TheBloke/Llama-2-70B-chat-GGML) - Specify `-eps 1e-5` for best generation quality - Specify `-gqa 8` for 70B models to work