mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-30 13:53:03 +01:00
llama : try to fix build
This commit is contained in:
parent
25cfbf6776
commit
739b85c985
@ -5544,7 +5544,7 @@ static struct ggml_cgraph * llama_build_graph(
|
|||||||
// should we offload the final norm? yes if we are not computing embeddings
|
// should we offload the final norm? yes if we are not computing embeddings
|
||||||
const bool offload_emb = lctx.embedding.empty();
|
const bool offload_emb = lctx.embedding.empty();
|
||||||
|
|
||||||
static const std::unordered_map<offload_func_e, std::string> k_offload_func_name = {
|
static const std::unordered_map<offload_func_e, std::string, std::hash<int>> k_offload_func_name = {
|
||||||
{ OFFLOAD_FUNC_NOP, "CPU" },
|
{ OFFLOAD_FUNC_NOP, "CPU" },
|
||||||
#ifdef GGML_USE_CUBLAS
|
#ifdef GGML_USE_CUBLAS
|
||||||
{ OFFLOAD_FUNC, "GPU (CUDA)" },
|
{ OFFLOAD_FUNC, "GPU (CUDA)" },
|
||||||
|
Loading…
Reference in New Issue
Block a user