mirror of
https://github.com/jart/cosmopolitan.git
synced 2025-07-12 14:09:12 +00:00
Get POSIX threads working on Apple Silicon
It's now possible to run a working ape-m1 o/aarch64/third_party/ggml/llama.com on Apple M1 hardware running XNU!
This commit is contained in:
parent
8fdb31681a
commit
b5eab2b0b7
3 changed files with 78 additions and 16 deletions
2
third_party/ggml/llama.cc
vendored
2
third_party/ggml/llama.cc
vendored
|
@ -1128,8 +1128,10 @@ static void llama_model_load_internal(
|
|||
const size_t mem_required_state =
|
||||
scale*MEM_REQ_KV_SELF().at(model.type);
|
||||
|
||||
if (verbose > 0) {
|
||||
fprintf(stderr, "%s: mem required = %7.2f MB (+ %7.2f MB per state)\n", __func__,
|
||||
mem_required / 1024.0 / 1024.0, mem_required_state / 1024.0 / 1024.0);
|
||||
}
|
||||
|
||||
#ifdef GGML_USE_CUBLAS
|
||||
const int n_gpu = std::min(n_gpu_layers, int(hparams.n_layer));
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue