llama : try to fix build
This commit is contained in:
parent
25cfbf6776
commit
739b85c985
1 changed files with 1 additions and 1 deletions
|
@ -5544,7 +5544,7 @@ static struct ggml_cgraph * llama_build_graph(
|
|||
// should we offload the final norm? yes if we are not computing embeddings
|
||||
const bool offload_emb = lctx.embedding.empty();
|
||||
|
||||
static const std::unordered_map<offload_func_e, std::string> k_offload_func_name = {
|
||||
static const std::unordered_map<offload_func_e, std::string, std::hash<int>> k_offload_func_name = {
|
||||
{ OFFLOAD_FUNC_NOP, "CPU" },
|
||||
#ifdef GGML_USE_CUBLAS
|
||||
{ OFFLOAD_FUNC, "GPU (CUDA)" },
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue