diff --git a/NOTES.md b/NOTES.md new file mode 100644 index 000000000..9deb5bcd7 --- /dev/null +++ b/NOTES.md @@ -0,0 +1,7 @@ +# TODOs + +1. How to debug mat_mul (run tests in cpp?) +2. How to wrap the suggestion from lauren on matmul (need to see how to find the llora info to pick up). Something about lora being loaded in the context? How to pick a specifi LoRA +3. check the PR "It was removed in [#7204](https://github.com/ggerganov/llama.cpp/pull/7204). `convert-lora-to-ggml.py` seems to write loras to gguf witouth the model? Should check the train script and see how they match lora with base layers +4. https://github.com/ggerganov/llama.cpp/discussions/3489 +5. check lora example in examples \ No newline at end of file diff --git a/convert-hf-to-gguf.py b/convert-hf-to-gguf.py index 025405a2c..f42967f08 100755 --- a/convert-hf-to-gguf.py +++ b/convert-hf-to-gguf.py @@ -2782,7 +2782,7 @@ def parse_args() -> argparse.Namespace: help="model is executed on big endian machine", ) parser.add_argument( - "model", type=Path, + "--model",dest="model", type=Path, default="models/phi-3-mini/", help="directory containing model file", ) parser.add_argument(