| name: Bug (model use) |
| description: Something goes wrong when using a model (in general, not specific to a single llama.cpp module). |
| title: "Eval bug: " |
| labels: ["bug-unconfirmed", "model evaluation"] |
| body: |
| - type: markdown |
| attributes: |
| value: > |
| Thanks for taking the time to fill out this bug report! |
| This issue template is intended for bug reports where the model evaluation results |
| (i.e. the generated text) are incorrect or llama.cpp crashes during model evaluation. |
| If you encountered the issue while using an external UI (e.g. ollama), |
| please reproduce your issue using one of the examples/binaries in this repository. |
| The `llama-completion` binary can be used for simple and reproducible model inference. |
| - type: textarea |
| id: version |
| attributes: |
| label: Name and Version |
| description: Which version of our software are you running? (use `--version` to get a version string) |
| placeholder: | |
| $./llama-cli --version |
| version: 2999 (42b4109e) |
| built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu |
| validations: |
| required: true |
| - type: dropdown |
| id: operating-system |
| attributes: |
| label: Operating systems |
| description: Which operating systems do you know to be affected? |
| multiple: true |
| options: |
| - Linux |
| - Mac |
| - Windows |
| - BSD |
| - Other? (Please let us know in description) |
| validations: |
| required: true |
| - type: dropdown |
| id: backends |
| attributes: |
| label: GGML backends |
| description: Which GGML backends do you know to be affected? |
| options: [AMX, BLAS, CANN, CPU, CUDA, Hexagon, HIP, Metal, Musa, OpenCL, RPC, SYCL, VirtGPU, Vulkan, WebGPU, zDNN, ZenDNN] |
| multiple: true |
| validations: |
| required: true |
| - type: textarea |
| id: hardware |
| attributes: |
| label: Hardware |
| description: Which CPUs/GPUs are you using? |
| placeholder: > |
| e.g. Ryzen 5950X + 2x RTX 4090 |
| validations: |
| required: true |
| - type: textarea |
| id: model |
| attributes: |
| label: Models |
| description: > |
| Which model(s) at which quantization were you using when encountering the bug? |
| If you downloaded a GGUF file off of Huggingface, please provide a link. |
| placeholder: > |
| e.g. Meta LLaMA 3.1 Instruct 8b q4_K_M |
| validations: |
| required: false |
| - type: textarea |
| id: info |
| attributes: |
| label: Problem description & steps to reproduce |
| description: > |
| Please give us a summary of the problem and tell us how to reproduce it. |
| If you can narrow down the bug to specific hardware, compile flags, or command line arguments, |
| that information would be very much appreciated by us. |
| |
| If possible, please try to reproduce the issue using `llama-completion` with `-fit off`. |
| If you can only reproduce the issue with `-fit on`, please provide logs both with and without `--verbose`. |
| placeholder: > |
| e.g. when I run llama-completion with `-fa on` I get garbled outputs for very long prompts. |
| With short prompts or `-fa off` it works correctly. |
| Here are the exact commands that I used: ... |
| validations: |
| required: true |
| - type: textarea |
| id: first_bad_commit |
| attributes: |
| label: First Bad Commit |
| description: > |
| If the bug was not present on an earlier version: when did it start appearing? |
| If possible, please do a git bisect and identify the exact commit that introduced the bug. |
| validations: |
| required: false |
| - type: textarea |
| id: logs |
| attributes: |
| label: Relevant log output |
| description: > |
| Please copy and paste any relevant log output, including the command that you entered and any generated text. |
| For very long logs (thousands of lines), preferably upload them as files instead. |
| On Linux you can redirect console output into a file by appending ` > llama.log 2>&1` to your command. |
| value: | |
| <details> |
| <summary>Logs</summary> |
| <!-- Copy-pasted short logs go into the "console" area here --> |
| |
| ```console |
|
|
| ``` |
| </details> |
|
|
| <!-- Long logs that you upload as files go here, outside the "console" area --> |
| validations: |
| required: true |
|
|