name: Bug (model use) description: Something goes wrong when using a model (in general, not specific to a single llama.cpp module). title: "Eval bug: " labels: ["bug-unconfirmed", "model evaluation"] body: - type: markdown attributes: value: > Thanks for taking the time to fill out this bug report! This issue template is intended for bug reports where the model evaluation results (i.e. the generated text) are incorrect or llama.cpp crashes during model evaluation. If you encountered the issue while using an external UI (e.g. ollama), please reproduce your issue using one of the examples/binaries in this repository. The `llama-cli` binary can be used for simple and reproducible model inference. - type: textarea id: version attributes: label: Name and Version description: Which version of our software are you running? (use `--version` to get a version string) placeholder: | $./llama-cli --version version: 2999 (42b4109e) built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu validations: required: true - type: dropdown id: operating-system attributes: label: Which operating systems do you know to be affected? multiple: true options: - Linux - Mac - Windows - BSD - Other? (Please let us know in description) validations: required: true - type: dropdown id: backends attributes: label: GGML backends description: Which GGML backends do you know to be affected? options: [AMX, BLAS, CPU, CUDA, HIP, Kompute, Metal, Musa, RPC, SYCL, Vulkan] multiple: true - type: textarea id: hardware attributes: label: Hardware description: Which CPUs/GPUs are you using? placeholder: > e.g. Ryzen 5950X + 2x RTX 4090 validations: required: true - type: textarea id: model attributes: label: Model description: > Which model at which quantization were you using when encountering the bug? If you downloaded a GGUF file off of Huggingface, please provide a link. placeholder: > e.g. Meta LLaMA 3.1 Instruct 8b q4_K_M validations: required: false - type: textarea id: steps_to_reproduce attributes: label: Steps to Reproduce description: > Please tell us how to reproduce the bug and any additional information that you think could be useful for fixing it. If you can narrow down the bug to specific hardware, compile flags, or command line arguments, that information would be very much appreciated by us. placeholder: > e.g. when I run llama-cli with -ngl 99 I get garbled outputs. When I use -ngl 0 it works correctly. Here are the exact commands that I used: ... validations: required: true - type: textarea id: first_bad_commit attributes: label: First Bad Commit description: > If the bug was not present on an earlier version: when did it start appearing? If possible, please do a git bisect and identify the exact commit that introduced the bug. validations: required: false - type: textarea id: logs attributes: label: Relevant log output description: > Please copy and paste any relevant log output, including the command that you entered and any generated text. This will be automatically formatted into code, so no need for backticks. render: shell validations: required: true