gguf : quantization is working

This commit is contained in:
M. Yusuf Sarıgöz 2023-08-12 16:39:05 +03:00
parent 1fc3d30b71
commit 202eab04d3
3 changed files with 7 additions and 6 deletions

View file

@ -421,7 +421,7 @@ int main(int argc, char ** argv) {
GGML_ASSERT(gguf_ex_write(fname) && "failed to write gguf file");
} else if (mode == "r") {
GGML_ASSERT(gguf_ex_read_0(fname) && "failed to read gguf file");
//GGML_ASSERT(gguf_ex_read_1(fname) && "failed to read gguf file");
GGML_ASSERT(gguf_ex_read_1(fname) && "failed to read gguf file");
GGML_ASSERT(gguf_ex_read_2(fname) && "failed to read gguf file");
} else if (mode == "q") {
llama_model_quantize_params params = llama_model_quantize_default_params();