@@ -94,6 +94,10 @@ export async function* llama(prompt, params = {}, config = {}) {
break;
}
+ if (result.error) {
+ result.error = JSON.parse(result.error);
+ console.error(`llama.cpp error: ${result.error.content}`);
+ }
@@ -2368,6 +2368,17 @@ int main(int argc, char **argv)
} else {
+ const std::string str =
+ "error: " +
+ result.result_json.dump(-1, ' ', false, json::error_handler_t::replace) +
+ "\n\n";
+ LOG_VERBOSE("data stream", {
+ { "to_send", str }
+ });
+ if (!sink.write(str.c_str(), str.size()))
+ {
+ return false;