From 8a9d7ce6245ea7059d7226bd823a1748d1a05b8d Mon Sep 17 00:00:00 2001 From: JohannesGaessler Date: Thu, 11 May 2023 07:05:52 +0200 Subject: [PATCH] fixup! Store layers in VRAM --- examples/common.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/common.cpp b/examples/common.cpp index 7107f1ffb..e8107399d 100644 --- a/examples/common.cpp +++ b/examples/common.cpp @@ -412,7 +412,7 @@ void gpt_print_usage(int /*argc*/, char ** argv, const gpt_params & params) { if (llama_mmap_supported()) { fprintf(stderr, " --no-mmap do not memory-map model (slower load but may reduce pageouts if not using mlock)\n"); } - fprintf(stderr, " --gpu_layers number of layers to store in VRAM"); + fprintf(stderr, " --gpu_layers number of layers to store in VRAM\n"); fprintf(stderr, " --mtest compute maximum memory usage\n"); fprintf(stderr, " --verbose-prompt print prompt before generation\n"); fprintf(stderr, " --lora FNAME apply LoRA adapter (implies --no-mmap)\n");