|
@@ -579,6 +579,8 @@ int main(int argc, char ** argv) {
|
|
|
|
|
|
|
|
params.model = params.vocoder.model;
|
|
params.model = params.vocoder.model;
|
|
|
params.embedding = true;
|
|
params.embedding = true;
|
|
|
|
|
+ params.ctx_shift = false; // silence warning
|
|
|
|
|
+ params.n_ubatch = params.n_batch;
|
|
|
|
|
|
|
|
common_init_result llama_init_cts = common_init_from_params(params);
|
|
common_init_result llama_init_cts = common_init_from_params(params);
|
|
|
|
|
|
|
@@ -1020,8 +1022,8 @@ lovely<|t_0.56|><|code_start|><|634|><|596|><|1766|><|1556|><|1306|><|1285|><|14
|
|
|
}
|
|
}
|
|
|
GGML_ASSERT(batch.n_tokens == n_codes);
|
|
GGML_ASSERT(batch.n_tokens == n_codes);
|
|
|
|
|
|
|
|
- if (llama_decode(ctx_cts, batch) != 0) {
|
|
|
|
|
- LOG_ERR("%s: llama_decode() failed\n", __func__);
|
|
|
|
|
|
|
+ if (llama_encode(ctx_cts, batch) != 0) {
|
|
|
|
|
+ LOG_ERR("%s: llama_encode() failed\n", __func__);
|
|
|
return 1;
|
|
return 1;
|
|
|
}
|
|
}
|
|
|
|
|
|