From fc5a26aadea54e2bcf6dd384e1ca0c846575bc0c Mon Sep 17 00:00:00 2001 From: Georgi Gerganov Date: Tue, 31 Oct 2023 08:57:10 +0200 Subject: [PATCH] llama : enable warning about not offloaded tensors --- llama.cpp | 2 -- 1 file changed, 2 deletions(-) diff --git a/llama.cpp b/llama.cpp index f69af36ec..68cb835e1 100644 --- a/llama.cpp +++ b/llama.cpp @@ -5235,7 +5235,6 @@ static struct ggml_cgraph * llama_build_graph( LLAMA_LOG_INFO("%s: non-view tensors processed: %d/%d\n", __func__, n_non_view, n_non_view_total); -#ifdef LLAMA_OFFLOAD_DEBUG if (n_non_view != n_non_view_total) { LLAMA_LOG_WARN("%s: ****************************************************************\n", __func__); LLAMA_LOG_WARN("%s: not all non-view tensors have been processed with a callback\n", __func__); @@ -5244,7 +5243,6 @@ static struct ggml_cgraph * llama_build_graph( LLAMA_LOG_WARN("%s: ref: https://github.com/ggerganov/llama.cpp/pull/3837\n", __func__); LLAMA_LOG_WARN("%s: ****************************************************************\n", __func__); } -#endif } return result;