* Initial conversion to customtkinter.
* Initial conversion to customtkinter.
* Additions to UI, still non-functional
* UI now functional, untested
* UI now functional, untested
* Added saving configs
* Saving and loading now functional
* Fixed sliders not loading
* Cleaned up duplicate arrays
* Cleaned up duplicate arrays
* Fixed loading bugs
* wip fixing all the broken parameters. PLEASE test before you commit
* further cleaning
* bugfix completed for gui. now evaluating save and load
* cleanup prepare to merge
---------
Co-authored-by: Concedo <39025047+LostRuins@users.noreply.github.com>
* server: add option to output probabilities for completion
* server: fix issue when handling probability output for incomplete tokens for multibyte character generation
* server: fix llama_sample_top_k order
* examples/common.h: put all bool variables in gpt_params together
It's currently not possible to cross-compile llama.cpp for aarch64
because CMakeLists.txt forces -mcpu=native for that target.
-mcpu=native doesn't make sense if your build host is not the
target architecture, and clang rejects it for that reason, aborting the
build. This can be easily reproduced using the current Android NDK to build
for aarch64 on an x86_64 host.
If there is not a specific CPU-tuning target for aarch64 then -mcpu
should be omitted completely. I think that makes sense, there is not
enough variance in the aarch64 instruction set to warrant a fixed -mcpu
optimization at this point. And if someone is building natively and wishes
to enable any possible optimizations for the host device, then there is
already the LLAMA_NATIVE option available.
Fixes#495.
* convert checks in llama_load_session_file to throw and handle them
* make llama_load_session_file_internal static
* address feedbacks to avoid using exceptions