mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-26 14:20:31 +01:00
llama : fix compile warnings when reading the vocab
This commit is contained in:
parent
cea1c85948
commit
0ba76c1e73
@ -1444,7 +1444,7 @@ static bool llama_model_quantize_internal(const std::string & fname_inp, const s
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string word;
|
std::vector<char> word(32);
|
||||||
vocab.id_to_token.resize(n_vocab);
|
vocab.id_to_token.resize(n_vocab);
|
||||||
for (int i = 0; i < n_vocab; i++) {
|
for (int i = 0; i < n_vocab; i++) {
|
||||||
uint32_t len;
|
uint32_t len;
|
||||||
@ -1459,10 +1459,10 @@ static bool llama_model_quantize_internal(const std::string & fname_inp, const s
|
|||||||
finp.read ((char *) &score, sizeof(score));
|
finp.read ((char *) &score, sizeof(score));
|
||||||
fout.write((char *) &score, sizeof(score));
|
fout.write((char *) &score, sizeof(score));
|
||||||
|
|
||||||
vocab.token_to_id[word] = i;
|
vocab.token_to_id[word.data()] = i;
|
||||||
|
|
||||||
auto &tok_score = vocab.id_to_token[i];
|
auto &tok_score = vocab.id_to_token[i];
|
||||||
tok_score.tok = word;
|
tok_score.tok = word.data();
|
||||||
tok_score.score = score;
|
tok_score.score = score;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user