Parth673 commited on
Commit
eeb8d5f
·
verified ·
1 Parent(s): 2b3ac1e

Add README

Browse files
Files changed (1) hide show
  1. README.md +25 -13
README.md CHANGED
@@ -1,22 +1,34 @@
1
  ---
2
- base_model: unsloth/gemma-3-4b-it-unsloth-bnb-4bit
3
  tags:
4
- - text-generation-inference
5
- - transformers
6
  - unsloth
7
- - gemma3
8
- - trl
9
- license: apache-2.0
10
- language:
11
- - en
12
  ---
13
 
14
- # Uploaded model
15
 
16
- - **Developed by:** Parth673
17
- - **License:** apache-2.0
18
- - **Finetuned from model :** unsloth/gemma-3-4b-it-unsloth-bnb-4bit
19
 
20
- This gemma3 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
 
 
21
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
  [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
1
  ---
 
2
  tags:
3
+ - gguf
4
+ - llama.cpp
5
  - unsloth
6
+ - vision-language-model
 
 
 
 
7
  ---
8
 
9
+ # gemma3-4b-interview-eval-quantized : GGUF
10
 
11
+ This model was finetuned and converted to GGUF format using [Unsloth](https://github.com/unslothai/unsloth).
 
 
12
 
13
+ **Example usage**:
14
+ - For text only LLMs: `./llama.cpp/llama-cli -hf Parth673/gemma3-4b-interview-eval-quantized --jinja`
15
+ - For multimodal models: `./llama.cpp/llama-mtmd-cli -hf Parth673/gemma3-4b-interview-eval-quantized --jinja`
16
 
17
+ ## Available Model files:
18
+ - `gemma-3-4b-it.Q4_K_M.gguf`
19
+ - `gemma-3-4b-it.F16-mmproj.gguf`
20
+
21
+ ## ⚠️ Ollama Note for Vision Models
22
+ **Important:** Ollama currently does not support separate mmproj files for vision models.
23
+
24
+ To create an Ollama model from this vision model:
25
+ 1. Place the `Modelfile` in the same directory as the finetuned bf16 merged model
26
+ 3. Run: `ollama create model_name -f ./Modelfile`
27
+ (Replace `model_name` with your desired name)
28
+
29
+ This will create a unified bf16 model that Ollama can use.
30
+
31
+ ## Note
32
+ The model's BOS token behavior was adjusted for GGUF compatibility.
33
+ This was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth)
34
  [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)