Sfoglia il codice sorgente

readme : add note that LLaMA 3 is not supported with convert.py (#7065)

Lyle Dean 1 anno fa
parent
commit
ca36326020
1 ha cambiato i file con 2 aggiunte e 0 eliminazioni
  1. 2 0
      README.md

+ 2 - 0
README.md

@@ -712,6 +712,8 @@ Building the program with BLAS support may lead to some performance improvements
 
 To obtain the official LLaMA 2 weights please see the <a href="#obtaining-and-using-the-facebook-llama-2-model">Obtaining and using the Facebook LLaMA 2 model</a> section. There is also a large selection of pre-quantized `gguf` models available on Hugging Face.
 
+Note: `convert.py` does not support LLaMA 3, you can use `convert-hf-to-gguf.py` with LLaMA 3 downloaded from Hugging Face. 
+
 ```bash
 # obtain the official LLaMA model weights and place them in ./models
 ls ./models