minkdank commited on
Commit
fac60a8
·
verified ·
1 Parent(s): e04b50d

Add README

Browse files
Files changed (1) hide show
  1. README.md +15 -4
README.md CHANGED
@@ -3,7 +3,7 @@ tags:
3
  - gguf
4
  - llama.cpp
5
  - unsloth
6
-
7
  ---
8
 
9
  # LLAMA-JSON-data-extration - GGUF
@@ -15,7 +15,18 @@ This model was finetuned and converted to GGUF format using [Unsloth](https://gi
15
  - For multimodal models: **llama-mtmd-cli** **-m** model_name.gguf **--mmproj** mmproj_file.gguf
16
 
17
  ## Available Model files:
18
- - `llama-3.2-3b-instruct.Q8_0.gguf`
 
 
 
 
 
 
 
 
 
 
 
19
 
20
- ## Ollama
21
- An Ollama Modelfile is included for easy deployment.
 
3
  - gguf
4
  - llama.cpp
5
  - unsloth
6
+ - vision-language-model
7
  ---
8
 
9
  # LLAMA-JSON-data-extration - GGUF
 
15
  - For multimodal models: **llama-mtmd-cli** **-m** model_name.gguf **--mmproj** mmproj_file.gguf
16
 
17
  ## Available Model files:
18
+ - `gemma-3-4b-it.Q8_0.gguf`
19
+ - `gemma-3-4b-it.BF16-mmproj.gguf`
20
+
21
+ ## ⚠️ Ollama Note for Vision Models
22
+ **Important:** Ollama currently does not support separate mmproj files for vision models.
23
+
24
+ To create an Ollama model from this vision model:
25
+ 1. Place the `Modelfile` in the same directory as the finetuned bf16 merged model
26
+ 3. Run: `ollama create model_name -f ./Modelfile`
27
+ (Replace `model_name` with your desired name)
28
+
29
+ This will create a unified bf16 model that Ollama can use.
30
 
31
+ ## Note
32
+ The model's BOS token behavior was adjusted for GGUF compatibility.