LOG_INF("%s: llama threadpool init, n_threads = %d\n", __func__, (int) params.cpuparams.n_threads); @@ -207,7 +201,7 @@ int main(int argc, char ** argv) { } // auto enable conversation mode if chat template is available const bool has_chat_template = !common_get_built...