Przeglądaj źródła

Github: update issue templates [no ci] (#10489)

Johannes Gäßler 1 rok temu
rodzic
commit
1f922254f0

+ 8 - 4
.github/ISSUE_TEMPLATE/010-bug-compilation.yml

@@ -24,7 +24,8 @@ body:
   - type: dropdown
   - type: dropdown
     id: operating-system
     id: operating-system
     attributes:
     attributes:
-      label: Which operating systems do you know to be affected?
+      label: Operating systems
+      description: Which operating systems do you know to be affected?
       multiple: true
       multiple: true
       options:
       options:
         - Linux
         - Linux
@@ -41,14 +42,17 @@ body:
         description: Which GGML backends do you know to be affected?
         description: Which GGML backends do you know to be affected?
         options: [AMX, BLAS, CPU, CUDA, HIP, Kompute, Metal, Musa, RPC, SYCL, Vulkan]
         options: [AMX, BLAS, CPU, CUDA, HIP, Kompute, Metal, Musa, RPC, SYCL, Vulkan]
         multiple: true
         multiple: true
+    validations:
+      required: true
   - type: textarea
   - type: textarea
-    id: steps_to_reproduce
+    id: info
     attributes:
     attributes:
-      label: Steps to Reproduce
+      label: Problem description & steps to reproduce
       description: >
       description: >
-        Please tell us how to reproduce the bug and any additional information that you think could be useful for fixing it.
+        Please give us a summary of the problem and tell us how to reproduce it.
         If you can narrow down the bug to specific compile flags, that information would be very much appreciated by us.
         If you can narrow down the bug to specific compile flags, that information would be very much appreciated by us.
       placeholder: >
       placeholder: >
+        I'm trying to compile llama.cpp with CUDA support on a fresh install of Ubuntu and get error XY.
         Here are the exact commands that I used: ...
         Here are the exact commands that I used: ...
     validations:
     validations:
       required: true
       required: true

+ 9 - 6
.github/ISSUE_TEMPLATE/011-bug-results.yml

@@ -26,7 +26,8 @@ body:
   - type: dropdown
   - type: dropdown
     id: operating-system
     id: operating-system
     attributes:
     attributes:
-      label: Which operating systems do you know to be affected?
+      label: Operating systems
+      description: Which operating systems do you know to be affected?
       multiple: true
       multiple: true
       options:
       options:
         - Linux
         - Linux
@@ -43,6 +44,8 @@ body:
         description: Which GGML backends do you know to be affected?
         description: Which GGML backends do you know to be affected?
         options: [AMX, BLAS, CPU, CUDA, HIP, Kompute, Metal, Musa, RPC, SYCL, Vulkan]
         options: [AMX, BLAS, CPU, CUDA, HIP, Kompute, Metal, Musa, RPC, SYCL, Vulkan]
         multiple: true
         multiple: true
+    validations:
+      required: true
   - type: textarea
   - type: textarea
     id: hardware
     id: hardware
     attributes:
     attributes:
@@ -55,20 +58,20 @@ body:
   - type: textarea
   - type: textarea
     id: model
     id: model
     attributes:
     attributes:
-      label: Model
+      label: Models
       description: >
       description: >
-        Which model at which quantization were you using when encountering the bug?
+        Which model(s) at which quantization were you using when encountering the bug?
         If you downloaded a GGUF file off of Huggingface, please provide a link.
         If you downloaded a GGUF file off of Huggingface, please provide a link.
       placeholder: >
       placeholder: >
         e.g. Meta LLaMA 3.1 Instruct 8b q4_K_M
         e.g. Meta LLaMA 3.1 Instruct 8b q4_K_M
     validations:
     validations:
       required: false
       required: false
   - type: textarea
   - type: textarea
-    id: steps_to_reproduce
+    id: info
     attributes:
     attributes:
-      label: Steps to Reproduce
+      label: Problem description & steps to reproduce
       description: >
       description: >
-        Please tell us how to reproduce the bug and any additional information that you think could be useful for fixing it.
+        Please give us a summary of the problem and tell us how to reproduce it.
         If you can narrow down the bug to specific hardware, compile flags, or command line arguments,
         If you can narrow down the bug to specific hardware, compile flags, or command line arguments,
         that information would be very much appreciated by us.
         that information would be very much appreciated by us.
       placeholder: >
       placeholder: >

+ 13 - 10
.github/ISSUE_TEMPLATE/019-bug-misc.yml

@@ -14,7 +14,7 @@ body:
     id: version
     id: version
     attributes:
     attributes:
       label: Name and Version
       label: Name and Version
-      description: Which version of our software are you running? (use `--version` to get a version string)
+      description: Which version of our software is affected? (You can use `--version` to get a version string.)
       placeholder: |
       placeholder: |
         $./llama-cli --version
         $./llama-cli --version
         version: 2999 (42b4109e)
         version: 2999 (42b4109e)
@@ -24,7 +24,8 @@ body:
   - type: dropdown
   - type: dropdown
     id: operating-system
     id: operating-system
     attributes:
     attributes:
-      label: Which operating systems do you know to be affected?
+      label: Operating systems
+      description: Which operating systems do you know to be affected?
       multiple: true
       multiple: true
       options:
       options:
         - Linux
         - Linux
@@ -33,28 +34,30 @@ body:
         - BSD
         - BSD
         - Other? (Please let us know in description)
         - Other? (Please let us know in description)
     validations:
     validations:
-      required: true
+      required: false
   - type: dropdown
   - type: dropdown
     id: module
     id: module
     attributes:
     attributes:
       label: Which llama.cpp modules do you know to be affected?
       label: Which llama.cpp modules do you know to be affected?
       multiple: true
       multiple: true
       options:
       options:
+        - Documentation/Github
         - libllama (core library)
         - libllama (core library)
         - llama-cli
         - llama-cli
         - llama-server
         - llama-server
         - llama-bench
         - llama-bench
         - llama-quantize
         - llama-quantize
         - Python/Bash scripts
         - Python/Bash scripts
+        - Test code
         - Other (Please specify in the next section)
         - Other (Please specify in the next section)
     validations:
     validations:
-      required: true
+      required: false
   - type: textarea
   - type: textarea
-    id: steps_to_reproduce
+    id: info
     attributes:
     attributes:
-      label: Steps to Reproduce
+      label: Problem description & steps to reproduce
       description: >
       description: >
-        Please tell us how to reproduce the bug and any additional information that you think could be useful for fixing it.
+        Please give us a summary of the problem and tell us how to reproduce it (if applicable).
     validations:
     validations:
       required: true
       required: true
   - type: textarea
   - type: textarea
@@ -62,7 +65,7 @@ body:
     attributes:
     attributes:
       label: First Bad Commit
       label: First Bad Commit
       description: >
       description: >
-        If the bug was not present on an earlier version: when did it start appearing?
+        If the bug was not present on an earlier version and it's not trivial to track down: when did it start appearing?
         If possible, please do a git bisect and identify the exact commit that introduced the bug.
         If possible, please do a git bisect and identify the exact commit that introduced the bug.
     validations:
     validations:
       required: false
       required: false
@@ -71,8 +74,8 @@ body:
     attributes:
     attributes:
       label: Relevant log output
       label: Relevant log output
       description: >
       description: >
-          Please copy and paste any relevant log output, including the command that you entered and any generated text.
+          If applicable, please copy and paste any relevant log output, including the command that you entered and any generated text.
           This will be automatically formatted into code, so no need for backticks.
           This will be automatically formatted into code, so no need for backticks.
       render: shell
       render: shell
     validations:
     validations:
-      required: true
+      required: false