Nomic vulkan backend licensed under the Software for Open Models License (SOM), version 1.0.
This commit is contained in:
parent
acfc5478ff
commit
4cdaa3c9cb
97 changed files with 13550 additions and 26 deletions
|
@ -36,6 +36,10 @@
|
|||
#pragma warning(disable: 4244 4267) // possible loss of data
|
||||
#endif
|
||||
|
||||
#if defined(GGML_USE_KOMPUTE)
|
||||
#include "ggml-vulkan.h"
|
||||
#endif
|
||||
|
||||
static llama_context ** g_ctx;
|
||||
static bool is_interacting = false;
|
||||
|
||||
|
@ -118,6 +122,10 @@ int main(int argc, char ** argv) {
|
|||
llama_context * ctx_guidance = NULL;
|
||||
g_ctx = &ctx;
|
||||
|
||||
#if defined(GGML_USE_KOMPUTE)
|
||||
ggml_vk_init_device(0, "gpu");
|
||||
#endif
|
||||
|
||||
// load the model and apply lora adapter, if any
|
||||
std::tie(model, ctx) = llama_init_from_gpt_params(params);
|
||||
if (params.cfg_scale > 1.f) {
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue