* make : deprecate ggml-ci * ci : disable Makefile builds ggml-ci * docs : remove make references [no ci] * ci : disable swift build ggml-ci * docs : remove obsolete make references, scripts, examples ggml-ci * basic fix for compare-commits.sh * update build.md * more build.md updates * more build.md updates * more build.md updates * Update Makefile Co-authored-by: Diego Devesa <slarengh@gmail.com> --------- Co-authored-by: slaren <slarengh@gmail.com>
		
			
				
	
	
		
			25 lines
		
	
	
	
		
			1.4 KiB
		
	
	
	
		
			Markdown
		
	
	
	
	
	
			
		
		
	
	
			25 lines
		
	
	
	
		
			1.4 KiB
		
	
	
	
		
			Markdown
		
	
	
	
	
	
| ## Convert llama2.c model to ggml
 | |
| 
 | |
| This example reads weights from project [llama2.c](https://github.com/karpathy/llama2.c) and saves them in ggml compatible format. The vocab that is available in `models/ggml-vocab.bin` is used by default.
 | |
| 
 | |
| To convert the model first download the models from the [llama2.c](https://github.com/karpathy/llama2.c) repository.
 | |
| 
 | |
| ```
 | |
| usage: ./llama-convert-llama2c-to-ggml [options]
 | |
| 
 | |
| options:
 | |
|   -h, --help                       show this help message and exit
 | |
|   --copy-vocab-from-model FNAME    path of gguf llama model or llama2.c vocabulary from which to copy vocab (default 'models/7B/ggml-model-f16.gguf')
 | |
|   --llama2c-model FNAME            [REQUIRED] model path from which to load Karpathy's llama2.c model
 | |
|   --llama2c-output-model FNAME     model path to save the converted llama2.c model (default ak_llama_model.bin')
 | |
| ```
 | |
| 
 | |
| An example command using a model from [karpathy/tinyllamas](https://huggingface.co/karpathy/tinyllamas) is as follows:
 | |
| 
 | |
| `$ ./llama-convert-llama2c-to-ggml --copy-vocab-from-model llama-2-7b-chat.gguf.q2_K.bin --llama2c-model stories42M.bin --llama2c-output-model stories42M.gguf.bin`
 | |
| 
 | |
| Note: The vocabulary for `stories260K.bin` should be its own tokenizer `tok512.bin` found in [karpathy/tinyllamas/stories260K](https://huggingface.co/karpathy/tinyllamas/tree/main/stories260K).
 | |
| 
 | |
| Now you can use the model with a command like:
 | |
| 
 | |
| `$ ./llama-cli -m stories42M.gguf.bin -p "One day, Lily met a Shoggoth" -n 500 -c 256`
 |