OmAlve commited on
Commit
e2de43d
·
verified ·
1 Parent(s): e35abd8

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +90 -0
README.md ADDED
@@ -0,0 +1,90 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - mr
4
+ license: llama3.2
5
+ base_model: meta-llama/Llama-3.2-3B
6
+ tags:
7
+ - llama-cpp
8
+ - gguf
9
+ - marathi
10
+ - alpaca
11
+ - android
12
+ ---
13
+
14
+ # vaarta GGUF — Marathi Instruction Model
15
+
16
+ Quantized GGUF files for [OmAlve/vaarta-llama-instruct](https://huggingface.co/OmAlve/vaarta-llama-instruct), a Llama-3.2-3B model fine-tuned on Marathi instruction data. Optimized for on-device inference with [llama.cpp](https://github.com/ggerganov/llama.cpp).
17
+
18
+ ## Files
19
+
20
+ | File | Quant | Size | Recommended for |
21
+ |---|---|---|---|
22
+ | `vaarta-Q4_K_M.gguf` | Q4_K_M | 1.9 GB | ✅ Most Android phones (4 GB+ RAM) |
23
+ | `vaarta-Q5_K_M.gguf` | Q5_K_M | 2.3 GB | Better quality (6 GB+ RAM) |
24
+
25
+ ## Chat Template
26
+
27
+ This model uses a **custom Marathi Alpaca template** — NOT the standard Llama-3 chat format.
28
+
29
+ ```
30
+ खाली एक सूचना दिली आहे. विनंती पूर्ण करणारे योग्य उत्तर लिहा.
31
+
32
+ ### सूचना:
33
+ {instruction}
34
+
35
+ ### इनपुट:
36
+ {input}
37
+
38
+ ### उत्तर:
39
+ {output}
40
+ ```
41
+
42
+ The template is embedded in the GGUF metadata.
43
+
44
+ ## Running on Android (Termux)
45
+
46
+ ### 1. Install Termux (from F-Droid)
47
+
48
+ ### 2. Build llama.cpp
49
+ ```bash
50
+ pkg update && pkg upgrade -y
51
+ pkg install -y clang cmake git wget
52
+ git clone https://github.com/ggerganov/llama.cpp
53
+ cd llama.cpp
54
+ cmake -B build -DGGML_NATIVE=OFF -DCMAKE_BUILD_TYPE=Release
55
+ cmake --build build --target llama-cli -j4
56
+ ```
57
+
58
+ ### 3. Download the model
59
+ ```bash
60
+ cd ~
61
+ wget https://huggingface.co/OmAlve/vaarta-gguf/resolve/main/vaarta-Q4_K_M.gguf
62
+ ```
63
+
64
+ ### 4. Run
65
+ ```bash
66
+ ~/llama.cpp/build/bin/llama-cli \
67
+ -m vaarta-Q4_K_M.gguf \
68
+ --ctx-size 2048 --threads 4 \
69
+ -i \
70
+ -r "### सूचना:" \
71
+ --in-prefix "खाली एक सूचना दिली आहे. विनंती पूर्ण करणारे योग्य उत्तर लिहा.\n\n### सूचना:\n" \
72
+ --in-suffix "\n\n### इनपुट:\n\n\n### उत्तर:\n"
73
+ ```
74
+
75
+ Type your Marathi question and press Enter.
76
+
77
+ ## Running on Desktop
78
+
79
+ ```bash
80
+ ./llama-cli -m vaarta-Q4_K_M.gguf -p "खाली एक सूचना दिली आहे. विनंती पूर्ण करणारे योग्य उत्तर लिहा.
81
+
82
+ ### सूचना:
83
+ पृथ्वीबद्दल थोडक्यात माहिती द्या.
84
+
85
+ ### इनपुट:
86
+
87
+
88
+ ### उत्तर:
89
+ " -n 256
90
+ ```