| name: Bug (misc.) |
| description: Something is not working the way it should (and it's not covered by any of the above cases). |
| title: "Misc. bug: " |
| labels: ["bug-unconfirmed"] |
| body: |
| - type: markdown |
| attributes: |
| value: > |
| Thanks for taking the time to fill out this bug report! |
| This issue template is intended for miscellaneous bugs that don't fit into any other category. |
| If you encountered the issue while using an external UI (e.g. ollama), |
| please reproduce your issue using one of the examples/binaries in this repository. |
| - type: textarea |
| id: version |
| attributes: |
| label: Name and Version |
| description: Which version of our software is affected? (You can use `--version` to get a version string.) |
| placeholder: | |
| $./llama-cli --version |
| version: 2999 (42b4109e) |
| built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu |
| validations: |
| required: true |
| - type: dropdown |
| id: operating-system |
| attributes: |
| label: Operating systems |
| description: Which operating systems do you know to be affected? |
| multiple: true |
| options: |
| - Linux |
| - Mac |
| - Windows |
| - BSD |
| - Other? (Please let us know in description) |
| validations: |
| required: false |
| - type: dropdown |
| id: module |
| attributes: |
| label: Which llama.cpp modules do you know to be affected? |
| multiple: true |
| options: |
| - Documentation/Github |
| - libllama (core library) |
| - llama-cli |
| - llama-server |
| - llama-bench |
| - llama-quantize |
| - Python/Bash scripts |
| - Test code |
| - Other (Please specify in the next section) |
| validations: |
| required: false |
| - type: textarea |
| id: command |
| attributes: |
| label: Command line |
| description: > |
| Please provide the exact commands you entered, if applicable. For example: `llama-server -m ... -c ...`, `llama-cli -m ...`, etc. |
| This will be automatically formatted into code, so no need for backticks. |
| render: shell |
| validations: |
| required: false |
| - type: textarea |
| id: info |
| attributes: |
| label: Problem description & steps to reproduce |
| description: > |
| Please give us a summary of the problem and tell us how to reproduce it (if applicable). |
| validations: |
| required: true |
| - type: textarea |
| id: first_bad_commit |
| attributes: |
| label: First Bad Commit |
| description: > |
| If the bug was not present on an earlier version and it's not trivial to track down: when did it start appearing? |
| If possible, please do a git bisect and identify the exact commit that introduced the bug. |
| validations: |
| required: false |
| - type: textarea |
| id: logs |
| attributes: |
| label: Relevant log output |
| description: > |
| If applicable, please copy and paste any relevant log output, including any generated text. |
| If you are encountering problems specifically with the `llama_params_fit` module, always upload `--verbose` logs as well. |
| For very long logs (thousands of lines), please upload them as files instead. |
| On Linux you can redirect console output into a file by appending ` > llama.log 2>&1` to your command. |
| value: | |
| <details> |
| <summary>Logs</summary> |
| <!-- Copy-pasted short logs go into the "console" area here --> |
| |
| ```console |
|
|
| ``` |
| </details> |
|
|
| <!-- Long logs that you upload as files go here, outside the "console" area --> |
| validations: |
| required: false |
|
|