From 8f70fc3d1b1d3c17b61842330dd106d391cc1227 Mon Sep 17 00:00:00 2001 From: Daniel Bevenius Date: Mon, 13 Jan 2025 13:38:20 +0100 Subject: [PATCH] llama : remove 'd' from bad special token log (#11212) This commit removes the 'd' from the log message in llama-vocab.cpp when logging a bad special token. The motivation for this is that currently the output can look something like the following: ```console load: bad special token: 'tokenizer.ggml.image_token_id' = 128256d, using default id -1 ``` --- src/llama-vocab.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llama-vocab.cpp b/src/llama-vocab.cpp index d0fb85cea..96b74e93a 100644 --- a/src/llama-vocab.cpp +++ b/src/llama-vocab.cpp @@ -1729,7 +1729,7 @@ void llama_vocab::impl::load(llama_model_loader & ml, const LLM_KV & kv) { continue; } if (new_id >= id_to_token.size()) { - LLAMA_LOG_WARN("%s: bad special token: '%s' = %ud, using default id %d\n", + LLAMA_LOG_WARN("%s: bad special token: '%s' = %u, using default id %d\n", __func__, key.c_str(), new_id, id); } else { id = new_id;