llama : try to fix build
This commit is contained in:
parent
25cfbf6776
commit
739b85c985
1 changed files with 1 additions and 1 deletions
|
@ -5544,7 +5544,7 @@ static struct ggml_cgraph * llama_build_graph(
|
||||||
// should we offload the final norm? yes if we are not computing embeddings
|
// should we offload the final norm? yes if we are not computing embeddings
|
||||||
const bool offload_emb = lctx.embedding.empty();
|
const bool offload_emb = lctx.embedding.empty();
|
||||||
|
|
||||||
static const std::unordered_map<offload_func_e, std::string> k_offload_func_name = {
|
static const std::unordered_map<offload_func_e, std::string, std::hash<int>> k_offload_func_name = {
|
||||||
{ OFFLOAD_FUNC_NOP, "CPU" },
|
{ OFFLOAD_FUNC_NOP, "CPU" },
|
||||||
#ifdef GGML_USE_CUBLAS
|
#ifdef GGML_USE_CUBLAS
|
||||||
{ OFFLOAD_FUNC, "GPU (CUDA)" },
|
{ OFFLOAD_FUNC, "GPU (CUDA)" },
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue