|
|
@@ -29,19 +29,25 @@ git clone https://huggingface.co/liuhaotian/llava-v1.5-7b
|
|
|
git clone https://huggingface.co/openai/clip-vit-large-patch14-336
|
|
|
```
|
|
|
|
|
|
-2. Use `llava-surgery.py` to split the LLaVA model to LLaMA and multimodel projector constituents:
|
|
|
+2. Install the required Python packages:
|
|
|
+
|
|
|
+```sh
|
|
|
+pip install -r examples/llava/requirements.txt
|
|
|
+```
|
|
|
+
|
|
|
+3. Use `llava-surgery.py` to split the LLaVA model to LLaMA and multimodel projector constituents:
|
|
|
|
|
|
```sh
|
|
|
python ./examples/llava/llava-surgery.py -m ../llava-v1.5-7b
|
|
|
```
|
|
|
|
|
|
-3. Use `convert-image-encoder-to-gguf.py` to convert the LLaVA image encoder to GGUF:
|
|
|
+4. Use `convert-image-encoder-to-gguf.py` to convert the LLaVA image encoder to GGUF:
|
|
|
|
|
|
```sh
|
|
|
python ./examples/llava/convert-image-encoder-to-gguf.py -m ../clip-vit-large-patch14-336 --llava-projector ../llava-v1.5-7b/llava.projector --output-dir ../llava-v1.5-7b
|
|
|
```
|
|
|
|
|
|
-4. Use `convert.py` to convert the LLaMA part of LLaVA to GGUF:
|
|
|
+5. Use `convert.py` to convert the LLaMA part of LLaVA to GGUF:
|
|
|
|
|
|
```sh
|
|
|
python ./convert.py ../llava-v1.5-7b
|