Parcourir la source

context : fix reorder logic (#13267)

ggml-ci
Georgi Gerganov il y a 8 mois
Parent
commit
a75cb30dc9
1 fichiers modifiés avec 3 ajouts et 3 suppressions
  1. 3 3
      src/llama-context.cpp

+ 3 - 3
src/llama-context.cpp

@@ -1050,6 +1050,9 @@ int llama_context::decode(llama_batch & inp_batch) {
     // finalize the batch processing
     kv_guard.commit();
 
+    // set to total number of outputs in the batch, for use in llama_get_logits_ith
+    n_outputs = n_outputs_all;
+
     // set output mappings
     {
         bool sorted_output = true;
@@ -1103,9 +1106,6 @@ int llama_context::decode(llama_batch & inp_batch) {
         }
     }
 
-    // set to total number of outputs in the batch, for use in llama_get_logits_ith
-    n_outputs = n_outputs_all;
-
     // wait for the computation to finish (automatically done when obtaining the model output)
     //synchronize();