Find a file
2024-06-28 15:39:19 +01:00
.devops devops : remove clblast + LLAMA_CUDA -> GGML_CUDA (#8139) 2024-06-26 19:32:07 +03:00
.github CI: fix release build (Ubuntu+Mac) (#8170) 2024-06-27 21:01:23 +02:00
ci llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
cmake llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
common Add SPM infill support (#8016) 2024-06-28 12:53:43 +02:00
docs llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
examples Add SPM infill support (#8016) 2024-06-28 12:53:43 +02:00
ggml CUDA: fix MMQ stream-k for --split-mode row (#8167) 2024-06-27 16:26:05 +02:00
gguf-py fixing dowble key error 2024-06-28 15:39:19 +01:00
grammars json: restore default additionalProperties to false, fix some pattern escapes (#8180) 2024-06-28 09:26:45 +01:00
include Added support for Viking pre-tokenizer (#8135) 2024-06-27 10:58:54 +02:00
media README: add graphic for matrix multiplication (#6881) 2024-04-24 21:29:13 +02:00
models Per token attributes (#7685) 2024-06-04 09:17:17 +02:00
pocs build: rename main → llama-cli, server → llama-server, llava-cli → llama-llava-cli, etc... (#7809) 2024-06-13 00:41:52 +01:00
prompts llama : add Qwen support (#4281) 2023-12-01 20:16:31 +02:00
requirements requirements : Bump torch and numpy for python3.12 (#8041) 2024-06-20 22:01:15 +02:00
scripts scripts : fix filename sync 2024-06-26 23:25:22 +03:00
spm-headers llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
src Add MiniCPM, Deepseek V2 chat template + clean up llama_chat_apply_template_internal (#8172) 2024-06-28 15:11:44 +02:00
tests Add MiniCPM, Deepseek V2 chat template + clean up llama_chat_apply_template_internal (#8172) 2024-06-28 15:11:44 +02:00
.clang-tidy cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
.dockerignore build: rename main → llama-cli, server → llama-server, llava-cli → llama-llava-cli, etc... (#7809) 2024-06-13 00:41:52 +01:00
.ecrc Nomic Vulkan backend (#4456) 2024-01-29 15:50:50 -05:00
.editorconfig cvector: fix CI + correct help message (#8064) 2024-06-22 18:11:30 +02:00
.flake8 py : logging and flake8 suppression refactoring (#7081) 2024-05-05 08:07:48 +03:00
.gitignore llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
.gitmodules llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
.pre-commit-config.yaml convert.py : add python logging instead of print() (#6511) 2024-05-03 22:36:41 +03:00
AUTHORS authors : regen 2024-06-26 19:36:44 +03:00
CMakeLists.txt cmake : allow user to override default options (#8178) 2024-06-28 12:37:45 +02:00
CMakePresets.json llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
CODEOWNERS fixing dowble key error 2024-06-28 15:39:19 +01:00
CONTRIBUTING.md docs: Added initial PR template with directions for doc only changes and squash merges [no ci] (#7700) 2024-06-10 01:24:29 +10:00
convert-hf-to-gguf-update.py Added support for Viking pre-tokenizer (#8135) 2024-06-27 10:58:54 +02:00
convert-hf-to-gguf.py llama: Add support for Gemma2ForCausalLM (#8156) 2024-06-27 21:00:43 -07:00
convert-llama-ggml-to-gguf.py convert.py : add python logging instead of print() (#6511) 2024-05-03 22:36:41 +03:00
flake.lock flake.lock: Update (#8071) 2024-06-27 08:37:29 -07:00
flake.nix build: rename main → llama-cli, server → llama-server, llava-cli → llama-llava-cli, etc... (#7809) 2024-06-13 00:41:52 +01:00
LICENSE license : update copyright notice + add AUTHORS (#6405) 2024-04-09 09:23:19 +03:00
Makefile Add missing items in makefile (#8177) 2024-06-28 02:19:11 +02:00
mypy.ini convert : partially revert PR #4818 (#5041) 2024-01-20 18:14:18 -05:00
Package.swift llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
pyrightconfig.json convert-hf : save memory with lazy evaluation (#7075) 2024-05-08 18:16:38 -04:00
README-sycl.md llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
README.md fixing dowble key error 2024-06-28 15:39:19 +01:00
requirements.txt Move convert.py to examples/convert-legacy-llama.py (#7430) 2024-05-30 21:40:00 +10:00
SECURITY.md chore: Fix markdown warnings (#6625) 2024-04-12 10:52:36 +02:00

this repo if full copy of https://github.com/ggerganov/llama.cpp

this repo was forked because of issue if doube key and still not merged pr: https://github.com/ggerganov/llama.cpp/pull/7992