Commit graph

3103 commits

Author SHA1 Message Date
Yazan Agha-Schrader
3e26cd0055
Merge branch 'ggerganov:master' into server-ui-pr 2024-05-31 14:45:35 +02:00
Yazan Agha-Schrader
21138dd2da add new ui files to makefile 2024-05-31 14:39:43 +02:00
Yazan Agha-Schrader
c0b154a7a8 use correct indent 2024-05-31 14:24:56 +02:00
Georgi Gerganov
16926dff92
readme : link homebrew discussion 2024-05-31 15:04:58 +03:00
Georgi Gerganov
0c27e6f62e
ggml : fix loongson compile warnings (#7537)
* ggml : fix loongson compile warnings

ggml-ci

* Fix loongarch quantize test fail.

Fix unexpected error introduced during rebase code.

* tests : disable json test due to lack of python on the CI node

ggml-ci

---------

Co-authored-by: junchao-loongson <zhaojunchao@loongson.cn>
2024-05-31 14:17:10 +03:00
Galunid
2e32f874e6
Somehow '**' got lost (#7663) 2024-05-31 18:24:41 +10:00
Galunid
1af511fc22
Add convert.py removal to hot topics (#7662) 2024-05-31 10:09:20 +02:00
Yazan Agha-Schrader
bc69a1e977 fix typos "prompt-format" -> "prompt-formats" 2024-05-31 06:31:55 +02:00
Yazan Agha-Schrader
80888e93cc renaming to ensure consistency 2024-05-31 06:17:40 +02:00
Yazan Agha-Schrader
d9742fbf4e fix wrong link to old ui 2024-05-31 05:37:55 +02:00
Yazan Agha-Schrader
bb9542b54f include new ui in cpp 2024-05-31 05:37:55 +02:00
Sertaç Özercan
0541f06296
[no ci] docs: add aikit to readme (#7650)
Signed-off-by: Sertac Ozercan <sozercan@gmail.com>
2024-05-31 09:57:16 +10:00
JohnnyB
9022c33646
Fixed painfully slow single process builds. (#7326)
* Fixed painfully slow single process builds.

* Added nproc for systems that don't default to nproc
2024-05-30 22:32:38 +02:00
Georgi Gerganov
5921b8f089
llama : cache llama_token_to_piece (#7587)
* llama : cache llama_token_to_piece

ggml-ci

* llama : use vectors and avoid has_cache

ggml-ci

* llama : throw on unknown tokenizer types

ggml-ci

* llama : print a log of the total cache size
2024-05-31 02:01:41 +10:00
Martin Delille
5dcdf94676
Fix conan badge display [no ci] (#7645) 2024-05-31 01:07:39 +10:00
Manuel
2e2340de17
Add brew installation instruction to README [no ci] (#7616) 2024-05-31 00:58:15 +10:00
Martin Delille
7846540bd2
readme : add Conan badge (#7638) 2024-05-30 15:52:50 +03:00
Brian
e6157f94c8
github: add contact links to issues and convert question into research [no ci] (#7612) 2024-05-30 21:55:36 +10:00
Galunid
9c4c9cc83f
Move convert.py to examples/convert-legacy-llama.py (#7430)
* Move convert.py to examples/convert-no-torch.py

* Fix CI, scripts, readme files

* convert-no-torch -> convert-legacy-llama

* Move vocab thing to vocab.py

* Fix convert-no-torch -> convert-legacy-llama

* Fix lost convert.py in ci/run.sh

* Fix imports

* Fix gguf not imported correctly

* Fix flake8 complaints

* Fix check-requirements.sh

* Get rid of ADDED_TOKENS_FILE, FAST_TOKENIZER_FILE

* Review fixes
2024-05-30 21:40:00 +10:00
Chris Elrod
59b0d07766
faster avx512 exp implementation (#7551)
* faster avx512 exp implementation

* x->r

* improve accuracy, handle special cases

* remove `e`
2024-05-30 21:32:55 +10:00
junchao-loongson
d5c05821f3
ggml : fix loongarch build (O2 issue) (#7636) 2024-05-30 12:30:10 +03:00
Johannes Gäßler
972b555ab9
README: explain parallel build [no ci] (#7618) 2024-05-30 09:52:39 +02:00
Yazan Agha-Schrader
0d75e07bd9
Merge branch 'ggerganov:master' into server-ui-pr 2024-05-30 08:28:26 +02:00
Meng, Hengyu
3854c9d07f
[SYCL] fix intel docker (#7630)
* Update main-intel.Dockerfile

* workaround for https://github.com/intel/oneapi-containers/issues/70

* reset intel docker in CI

* add missed in server
2024-05-30 16:19:08 +10:00
Yazan Agha-Schrader
505d0a3346 move new ui to "/public" due to otherwise problematic CORS behaviour 2024-05-30 04:00:56 +02:00
Yazan Agha-Schrader
8b937a1a71 add a button to the new ui 2024-05-30 03:59:28 +02:00
Galunid
eb57fee51f
gguf-py : Add tokenizer.ggml.pre to gguf-new-metadata.py (#7627) 2024-05-30 02:10:40 +02:00
Yazan Agha-Schrader
734be4dcc9 Merge branch 'master' into server-ui-pr 2024-05-30 01:47:22 +02:00
Yazan Agha-Schrader
d55081767c fix css path 2024-05-30 01:17:47 +02:00
Yazan Agha-Schrader
89b1b38144 move files, clean code 2024-05-30 01:13:10 +02:00
Yazan Agha-Schrader
63de7201fa set default prompt to empty 2024-05-29 22:34:15 +02:00
Yazan Agha-Schrader
dcdc11a5c4 add cmd-r prompt et reduce redundancy 2024-05-29 22:24:24 +02:00
Yazan Agha-Schrader
87bcbbb6c2 fix toggle state localstorage 2024-05-29 22:23:40 +02:00
Georgi Gerganov
55d62262a9
metal : remove invalid asserts (#7617) 2024-05-29 22:21:20 +03:00
Yazan Agha-Schrader
c2badb4697 add hacky llama2 prompt solution, reduce redundancy in promptFormats.js 2024-05-29 20:03:20 +02:00
Georgi Gerganov
975ec63ff2
metal : add missing asserts (#7617) 2024-05-29 20:45:25 +03:00
Georgi Gerganov
fb76ec31a9
ggml : fix YARN + add tests + add asserts (#7617)
* tests : add rope tests

ggml-ci

* ggml : fixes (hopefully)

ggml-ci

* tests : add non-cont tests

ggml-ci

* cuda : add asserts for rope/norm + fix DS2

ggml-ci

* ggml : assert contiguousness

* tests : reduce RoPE tests

ggml-ci
2024-05-29 20:17:31 +03:00
Georgi Gerganov
cce3dcffc5
cuda : non-cont concat support (#7610)
* tests : add non-cont concat tests

* cuda : non-cont concat support

ggml-ci
2024-05-29 15:38:26 +03:00
Yazan Agha-Schrader
1c24ab6e20 move prompt style 2024-05-29 14:09:19 +02:00
Radoslav Gerganov
210d99173d
llama-bench : add support for the RPC backend (#7435) 2024-05-29 14:45:44 +03:00
slaren
87bdf2a199
ggml : use atomic_flag for critical section (#7598)
* ggml : use atomic_flag for critical section

* add windows shims
2024-05-29 13:36:39 +02:00
Yazan Agha-Schrader
f2ef89415c do not separate with new line or comma 2024-05-29 13:36:07 +02:00
Yazan Agha-Schrader
39a163f76e add missing char 2024-05-29 13:32:33 +02:00
Georgi Gerganov
00281b7be3
scripts : remove mpi remnants 2024-05-29 14:31:18 +03:00
Georgi Gerganov
2ab977282b
sync : ggml 2024-05-29 14:29:52 +03:00
Georgi Gerganov
72de268bec
ggml : restore ggml_rope_xpos_inplace (ggml/0)
ggml-ci
2024-05-29 14:29:33 +03:00
Yazan Agha-Schrader
513406ab60 add more comon stop tokens 2024-05-29 13:29:00 +02:00
Yazan Agha-Schrader
80b6143f78 more prompt format fixes 2024-05-29 13:19:22 +02:00
Yazan Agha-Schrader
ca565f4ed6 fix llama3 prompt template 2024-05-29 12:08:39 +02:00
Yazan Agha-Schrader
9fa0aa53f5 fix chatml & add llama3 format 2024-05-29 11:26:34 +02:00