.devops
|
devops : remove clblast + LLAMA_CUDA -> GGML_CUDA (#8139)
|
2024-06-26 19:32:07 +03:00 |
.github
|
CI: fix release build (Ubuntu+Mac) (#8170)
|
2024-06-27 21:01:23 +02:00 |
ci
|
llama : reorganize source code + improve CMake (#8006)
|
2024-06-26 18:33:02 +03:00 |
cmake
|
llama : reorganize source code + improve CMake (#8006)
|
2024-06-26 18:33:02 +03:00 |
common
|
Add SPM infill support (#8016)
|
2024-06-28 12:53:43 +02:00 |
docs
|
llama : reorganize source code + improve CMake (#8006)
|
2024-06-26 18:33:02 +03:00 |
examples
|
Add SPM infill support (#8016)
|
2024-06-28 12:53:43 +02:00 |
ggml
|
CUDA: fix MMQ stream-k for --split-mode row (#8167)
|
2024-06-27 16:26:05 +02:00 |
gguf-py
|
fixing dowble key error
|
2024-06-28 15:39:19 +01:00 |
grammars
|
json : restore default additionalProperties to false, fix some pattern escapes (#8180)
|
2024-06-28 09:26:45 +01:00 |
include
|
Added support for Viking pre-tokenizer (#8135)
|
2024-06-27 10:58:54 +02:00 |
media
|
README: add graphic for matrix multiplication (#6881)
|
2024-04-24 21:29:13 +02:00 |
models
|
Per token attributes (#7685)
|
2024-06-04 09:17:17 +02:00 |
pocs
|
build : rename main → llama-cli, server → llama-server, llava-cli → llama-llava-cli, etc... (#7809)
|
2024-06-13 00:41:52 +01:00 |
prompts
|
llama : add Qwen support (#4281)
|
2023-12-01 20:16:31 +02:00 |
requirements
|
requirements : Bump torch and numpy for python3.12 (#8041)
|
2024-06-20 22:01:15 +02:00 |
scripts
|
scripts : fix filename sync
|
2024-06-26 23:25:22 +03:00 |
spm-headers
|
llama : reorganize source code + improve CMake (#8006)
|
2024-06-26 18:33:02 +03:00 |
src
|
Add MiniCPM, Deepseek V2 chat template + clean up llama_chat_apply_template_internal (#8172)
|
2024-06-28 15:11:44 +02:00 |
tests
|
Add MiniCPM, Deepseek V2 chat template + clean up llama_chat_apply_template_internal (#8172)
|
2024-06-28 15:11:44 +02:00 |
.clang-tidy
|
cuda : refactor into multiple files (#6269)
|
2024-03-25 13:50:23 +01:00 |
.dockerignore
|
build : rename main → llama-cli, server → llama-server, llava-cli → llama-llava-cli, etc... (#7809)
|
2024-06-13 00:41:52 +01:00 |
.ecrc
|
Nomic Vulkan backend (#4456)
|
2024-01-29 15:50:50 -05:00 |
.editorconfig
|
cvector: fix CI + correct help message (#8064)
|
2024-06-22 18:11:30 +02:00 |
.flake8
|
py : logging and flake8 suppression refactoring (#7081)
|
2024-05-05 08:07:48 +03:00 |
.gitignore
|
llama : reorganize source code + improve CMake (#8006)
|
2024-06-26 18:33:02 +03:00 |
.gitmodules
|
llama : reorganize source code + improve CMake (#8006)
|
2024-06-26 18:33:02 +03:00 |
.pre-commit-config.yaml
|
convert.py : add python logging instead of print() (#6511)
|
2024-05-03 22:36:41 +03:00 |
AUTHORS
|
authors : regen
|
2024-06-26 19:36:44 +03:00 |
CMakeLists.txt
|
cmake : allow user to override default options (#8178)
|
2024-06-28 12:37:45 +02:00 |
CMakePresets.json
|
llama : reorganize source code + improve CMake (#8006)
|
2024-06-26 18:33:02 +03:00 |
CODEOWNERS
|
fixing dowble key error
|
2024-06-28 15:39:19 +01:00 |
CONTRIBUTING.md
|
docs: Added initial PR template with directions for doc only changes and squash merges [no ci] (#7700)
|
2024-06-10 01:24:29 +10:00 |
convert-hf-to-gguf-update.py
|
Added support for Viking pre-tokenizer (#8135)
|
2024-06-27 10:58:54 +02:00 |
convert-hf-to-gguf.py
|
llama: Add support for Gemma2ForCausalLM (#8156)
|
2024-06-27 21:00:43 -07:00 |
convert-llama-ggml-to-gguf.py
|
convert.py : add python logging instead of print() (#6511)
|
2024-05-03 22:36:41 +03:00 |
flake.lock
|
flake.lock: Update (#8071)
|
2024-06-27 08:37:29 -07:00 |
flake.nix
|
build : rename main → llama-cli, server → llama-server, llava-cli → llama-llava-cli, etc... (#7809)
|
2024-06-13 00:41:52 +01:00 |
LICENSE
|
license : update copyright notice + add AUTHORS (#6405)
|
2024-04-09 09:23:19 +03:00 |
Makefile
|
Add missing items in makefile (#8177)
|
2024-06-28 02:19:11 +02:00 |
mypy.ini
|
convert : partially revert PR #4818 (#5041)
|
2024-01-20 18:14:18 -05:00 |
Package.swift
|
llama : reorganize source code + improve CMake (#8006)
|
2024-06-26 18:33:02 +03:00 |
pyrightconfig.json
|
convert-hf : save memory with lazy evaluation (#7075)
|
2024-05-08 18:16:38 -04:00 |
README-sycl.md
|
llama : reorganize source code + improve CMake (#8006)
|
2024-06-26 18:33:02 +03:00 |
README.md
|
fixing dowble key error
|
2024-06-28 15:39:19 +01:00 |
requirements.txt
|
Move convert.py to examples/convert-legacy-llama.py (#7430)
|
2024-05-30 21:40:00 +10:00 |
SECURITY.md
|
chore: Fix markdown warnings (#6625)
|
2024-04-12 10:52:36 +02:00 |