mirror of
https://gitlab.com/niansa/libjustlm.git
synced 2025-03-06 20:49:17 +01:00
Check for errors during llama evaluation properly
This commit is contained in:
parent
087fe1396b
commit
59a6a8b1d1
1 changed files with 3 additions and 3 deletions
|
@ -82,7 +82,7 @@ class LLaMaInference final : public Inference {
|
|||
if (it + params.n_batch >= ssize_t(state->tokens.size())) break;
|
||||
|
||||
// Evaluate
|
||||
if (!llama_eval(state->ctx, state->tokens.data()+it, params.n_batch, it, params.n_threads)) {
|
||||
if (llama_eval(state->ctx, state->tokens.data()+it, params.n_batch, it, params.n_threads)) {
|
||||
LM_COTHROW("Failed to evaluate tokens in batches", LM_BOOL_ERROR);
|
||||
}
|
||||
|
||||
|
@ -99,7 +99,7 @@ class LLaMaInference final : public Inference {
|
|||
// Evaluate remaining tokens
|
||||
if (it < state->tokens.size()) {
|
||||
for (; it != state->tokens.size(); it++) {
|
||||
if (!llama_eval(state->ctx, state->tokens.data()+it, 1, it, params.n_threads)) {
|
||||
if (llama_eval(state->ctx, state->tokens.data()+it, 1, it, params.n_threads)) {
|
||||
LM_COTHROW("Failed to evaluate individual tokens", LM_BOOL_ERROR);
|
||||
}
|
||||
}
|
||||
|
@ -186,7 +186,7 @@ public:
|
|||
|
||||
// Evaluate token
|
||||
// TODO: Respect batch size
|
||||
if (!llama_eval(state->ctx, state->tokens.data()+state->tokens.size()-1, 1, state->tokens.size()-1, params.n_threads)) {
|
||||
if (llama_eval(state->ctx, state->tokens.data()+state->tokens.size()-1, 1, state->tokens.size()-1, params.n_threads)) {
|
||||
LM_COTHROW("Failed to evaluate new tokens", "");
|
||||
}
|
||||
|
||||
|
|
Loading…
Add table
Reference in a new issue