mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-12 21:37:19 +01:00
llama : prevent system info string accumulation across calls (#11101)
This commit is contained in:
parent
6369f867a4
commit
96a1dc27c3
@ -12458,6 +12458,8 @@ int llama_split_prefix(char * dest, size_t maxlen, const char * split_path, int
|
|||||||
|
|
||||||
const char * llama_print_system_info(void) {
|
const char * llama_print_system_info(void) {
|
||||||
static std::string s;
|
static std::string s;
|
||||||
|
s.clear(); // Clear the string, since it's static, otherwise it will accumulate data from previous calls.
|
||||||
|
|
||||||
|
|
||||||
for (size_t i = 0; i < ggml_backend_reg_count(); i++) {
|
for (size_t i = 0; i < ggml_backend_reg_count(); i++) {
|
||||||
auto * reg = ggml_backend_reg_get(i);
|
auto * reg = ggml_backend_reg_get(i);
|
||||||
|
Loading…
x
Reference in New Issue
Block a user