|
|
@@ -153,7 +153,7 @@ while n_cur <= n_len {
|
|
|
// const llama_token new_token_id = llama_sample_token_greedy(ctx, &candidates_p);
|
|
|
|
|
|
// is it an end of stream? -> mark the stream as finished
|
|
|
- if new_token_id == llama_token_eos(context) || n_cur == n_len {
|
|
|
+ if new_token_id == llama_token_eos(model) || n_cur == n_len {
|
|
|
i_batch[i] = -1
|
|
|
// print("")
|
|
|
if n_parallel > 1 {
|