kompute : remove Q6_K from list of supported quant types
This commit is contained in:
parent
24a4a5956a
commit
3d850db767
1 changed files with 1 additions and 2 deletions
|
@ -6498,8 +6498,7 @@ struct llama_context * llama_new_context_with_model(
|
|||
&& (model->ftype == LLAMA_FTYPE_ALL_F32
|
||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_F16
|
||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_0
|
||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_1
|
||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q6_K)) {
|
||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_1)) {
|
||||
// this allocates all Vulkan resources and memory buffers
|
||||
ctx->ctx_kompute = ggml_vk_init();
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue