Justine Tunney
|
95fab334e4
|
Use yield on aarch in spin locks
|
2023-05-11 19:57:09 -07:00 |
|
Justine Tunney
|
1f6f9e6701
|
Remove division from matrix multiplication
This change reduces llama.com CPU cycles systemically by 2.5% according
to the Linux Kernel `perf stat -Bddd` utility.
|
2023-05-10 21:19:54 -07:00 |
|
Justine Tunney
|
5f57fc1f59
|
Upgrade llama.cpp to e6a46b0ed1884c77267dc70693183e3b7164e0e0
|
2023-05-10 04:20:48 -07:00 |
|
Justine Tunney
|
a0237a017c
|
Get llama.com working on aarch64
|
2023-05-10 04:20:47 -07:00 |
|
Justine Tunney
|
4c093155a3
|
Get llama.com building as an aarch64 native binary
|
2023-05-10 04:20:47 -07:00 |
|
Justine Tunney
|
d9e27203d4
|
Incorporate some fixes and updates for GGML
|
2023-04-28 20:24:55 -07:00 |
|
Justine Tunney
|
420f889ac3
|
Further optimize the math library
The sincosf() function is now twice as fast, thanks to ARM Limited. The
same might also be true of logf() and expm1f() which have been updated.
|
2023-04-28 01:20:47 -07:00 |
|
Justine Tunney
|
e8b43903b2
|
Import llama.cpp
https://github.com/ggerganov/llama.cpp
0b2da20538d01926b77ea237dd1c930c4d20b686
See third_party/ggml/README.cosmo for changes
|
2023-04-27 14:37:14 -07:00 |
|