| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104 |
- name: Bug (model use)
- description: Something goes wrong when using a model (in general, not specific to a single llama.cpp module).
- title: "Eval bug: "
- labels: ["bug-unconfirmed", "model evaluation"]
- body:
- - type: markdown
- attributes:
- value: >
- Thanks for taking the time to fill out this bug report!
- This issue template is intended for bug reports where the model evaluation results
- (i.e. the generated text) are incorrect or llama.cpp crashes during model evaluation.
- If you encountered the issue while using an external UI (e.g. ollama),
- please reproduce your issue using one of the examples/binaries in this repository.
- The `llama-completion` binary can be used for simple and reproducible model inference.
- - type: textarea
- id: version
- attributes:
- label: Name and Version
- description: Which version of our software are you running? (use `--version` to get a version string)
- placeholder: |
- $./llama-cli --version
- version: 2999 (42b4109e)
- built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu
- validations:
- required: true
- - type: dropdown
- id: operating-system
- attributes:
- label: Operating systems
- description: Which operating systems do you know to be affected?
- multiple: true
- options:
- - Linux
- - Mac
- - Windows
- - BSD
- - Other? (Please let us know in description)
- validations:
- required: true
- - type: dropdown
- id: backends
- attributes:
- label: GGML backends
- description: Which GGML backends do you know to be affected?
- options: [AMX, BLAS, CPU, CUDA, HIP, Metal, Musa, RPC, SYCL, Vulkan, OpenCL, zDNN]
- multiple: true
- validations:
- required: true
- - type: textarea
- id: hardware
- attributes:
- label: Hardware
- description: Which CPUs/GPUs are you using?
- placeholder: >
- e.g. Ryzen 5950X + 2x RTX 4090
- validations:
- required: true
- - type: textarea
- id: model
- attributes:
- label: Models
- description: >
- Which model(s) at which quantization were you using when encountering the bug?
- If you downloaded a GGUF file off of Huggingface, please provide a link.
- placeholder: >
- e.g. Meta LLaMA 3.1 Instruct 8b q4_K_M
- validations:
- required: false
- - type: textarea
- id: info
- attributes:
- label: Problem description & steps to reproduce
- description: >
- Please give us a summary of the problem and tell us how to reproduce it.
- If you can narrow down the bug to specific hardware, compile flags, or command line arguments,
- that information would be very much appreciated by us.
- If possible, please try to reproduce the issue using `llama-completion` with `-fit off`.
- If you can only reproduce the issue with `-fit on`, please provide logs both with and without `--verbose`.
- placeholder: >
- e.g. when I run llama-completion with `-fa on` I get garbled outputs for very long prompts.
- With short prompts or `-fa off` it works correctly.
- Here are the exact commands that I used: ...
- validations:
- required: true
- - type: textarea
- id: first_bad_commit
- attributes:
- label: First Bad Commit
- description: >
- If the bug was not present on an earlier version: when did it start appearing?
- If possible, please do a git bisect and identify the exact commit that introduced the bug.
- validations:
- required: false
- - type: textarea
- id: logs
- attributes:
- label: Relevant log output
- description: >
- Please copy and paste any relevant log output, including the command that you entered and any generated text.
- This will be automatically formatted into code, so no need for backticks.
- render: shell
- validations:
- required: true
|