kompute : remove Q6_K from list of supported quant types
This commit is contained in:
parent
24a4a5956a
commit
3d850db767
1 changed files with 1 additions and 2 deletions
|
@ -6498,8 +6498,7 @@ struct llama_context * llama_new_context_with_model(
|
||||||
&& (model->ftype == LLAMA_FTYPE_ALL_F32
|
&& (model->ftype == LLAMA_FTYPE_ALL_F32
|
||||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_F16
|
|| model->ftype == LLAMA_FTYPE_MOSTLY_F16
|
||||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_0
|
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_0
|
||||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_1
|
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q4_1)) {
|
||||||
|| model->ftype == LLAMA_FTYPE_MOSTLY_Q6_K)) {
|
|
||||||
// this allocates all Vulkan resources and memory buffers
|
// this allocates all Vulkan resources and memory buffers
|
||||||
ctx->ctx_kompute = ggml_vk_init();
|
ctx->ctx_kompute = ggml_vk_init();
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue