docs : s/LLaMa/LLaMA/
This commit is contained in:
parent
17d22efa40
commit
767aef90be
3 changed files with 5 additions and 5 deletions
|
@ -1376,7 +1376,7 @@ def main(args_in: list[str] | None = None) -> None:
|
|||
if np.uint32(1) == np.uint32(1).newbyteorder("<"):
|
||||
# We currently only support Q8_0 output on little endian systems.
|
||||
output_choices.append("q8_0")
|
||||
parser = argparse.ArgumentParser(description="Convert a LLaMa model to a GGML compatible file")
|
||||
parser = argparse.ArgumentParser(description="Convert a LLaMA model to a GGML compatible file")
|
||||
parser.add_argument("--awq-path", type=Path, help="Path to scale awq cache file", default=None)
|
||||
parser.add_argument("--dump", action="store_true", help="don't convert, just show what's in the model")
|
||||
parser.add_argument("--dump-single", action="store_true", help="don't convert, just show what's in a single model file")
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue