llama : remove experimental stuff
This commit is contained in:
parent
e343b8b4d8
commit
e7e7b11455
2 changed files with 1 additions and 47 deletions
|
@ -925,9 +925,7 @@ void ggml_metal_graph_compute(
|
|||
nth1 = 1;
|
||||
if (ne11 * ne12 < 4) {
|
||||
[encoder setComputePipelineState:ctx->pipeline_mul_mat_f16_f32_1row];
|
||||
//} else if (ne00 >= 128 && ne01 >= 8 && ne00%4 == 0) {
|
||||
} else if (false) {
|
||||
// TODO: with ggml_mul_mat_pad this kernel no longer seems to be needed
|
||||
} else if (ne00 >= 128 && ne01 >= 8 && ne00%4 == 0) {
|
||||
[encoder setComputePipelineState:ctx->pipeline_mul_mat_f16_f32_l4];
|
||||
nrows = ne11;
|
||||
} else {
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue