haidar038 commited on
Commit
4486620
·
verified ·
1 Parent(s): a19357a

Add model card

Browse files
Files changed (1) hide show
  1. README.md +103 -14
README.md CHANGED
@@ -1,21 +1,110 @@
1
  ---
2
- base_model: unsloth/meta-llama-3.1-8b-instruct-bnb-4bit
3
- tags:
4
- - text-generation-inference
5
- - transformers
6
- - unsloth
7
- - llama
8
- license: apache-2.0
9
  language:
10
- - en
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
- # Uploaded finetuned model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
- - **Developed by:** haidar038
16
- - **License:** apache-2.0
17
- - **Finetuned from model :** unsloth/meta-llama-3.1-8b-instruct-bnb-4bit
18
 
19
- This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
20
 
21
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
1
  ---
 
 
 
 
 
 
 
2
  language:
3
+ - id
4
+ - ms
5
+ base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
6
+ tags:
7
+ - llama-3
8
+ - maluku
9
+ - melayu-maluku-utara
10
+ - fine-tuned
11
+ - unsloth
12
+ - lora
13
+ - gguf
14
+ - indonesian
15
+ license: llama3
16
+ pipeline_tag: text-generation
17
  ---
18
 
19
+ # 🌴 Utu — Asisten Melayu Maluku Utara (Llama-3.1-8B)
20
+
21
+ Model fine-tuned dari `meta-llama/Meta-Llama-3.1-8B-Instruct` untuk memahami dan merespons
22
+ dalam **Bahasa Melayu Maluku Utara** (Bahasa Pasar/Bahasa Ternate).
23
+
24
+ Fine-tuned menggunakan **Unsloth** untuk efisiensi maksimal di GPU T4.
25
+
26
+ ## Kosakata Lokal yang Dipahami
27
+
28
+ | Kata | Arti | Contoh |
29
+ |------|------|--------|
30
+ | ngana / nga | kamu | "Ngana mo pi mana?" |
31
+ | kita | saya | "Kita tra tau" |
32
+ | torang / tong | kami/kita semua | "Torang mo pigi pasar" |
33
+ | dorang / dong | mereka | "Dorang su pulang" |
34
+ | pigi | pergi | "Kita mo pigi" |
35
+ | tra / tara | tidak | "Kita tra mau" |
36
+ | su / so | sudah | "Kita su makan" |
37
+ | mo | mau/akan | "Ngana mo makan?" |
38
+ | deng | dengan | "Kita deng ngana" |
39
+ | bolom | belum | "Kita bolom makan" |
40
+ | kobong | kebun | "Pigi di kobong" |
41
+ | foya | bohong | "Jang foya" |
42
+ | pe / p | kepemilikan | "Ini kita p buku" |
43
+ | bkiapa | kenapa | "Bkiapa ngana sedih?" |
44
+
45
+ ## Cara Penggunaan
46
+
47
+ ### Python (Transformers)
48
+ ```python
49
+ from transformers import AutoModelForCausalLM, AutoTokenizer
50
+ import torch
51
+
52
+ model_id = "haidar038/utu-malut"
53
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
54
+ model = AutoModelForCausalLM.from_pretrained(
55
+ model_id, torch_dtype=torch.float16, device_map="auto"
56
+ )
57
+
58
+ messages = [
59
+ {"role": "system", "content": "Ngana adalah Utu, asisten AI dari Ternate."},
60
+ {"role": "user", "content": "Ngana mau pigi mana sore ini?"}
61
+ ]
62
+ inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True).to(model.device)
63
+ outputs = model.generate(inputs, max_new_tokens=200, temperature=0.7, do_sample=True)
64
+ print(tokenizer.decode(outputs[0][inputs.shape[-1]:], skip_special_tokens=True))
65
+ ```
66
+
67
+ ### GGUF (CPU, via llama-cpp-python)
68
+ ```python
69
+ from llama_cpp import Llama
70
+
71
+ llm = Llama.from_pretrained(
72
+ repo_id = "haidar038/utu-malut-GGUF",
73
+ filename = "*q4_k_m*.gguf",
74
+ n_ctx = 512,
75
+ )
76
+ output = llm.create_chat_completion(messages=[
77
+ {"role": "system", "content": "Ngana adalah Utu, asisten AI dari Ternate."},
78
+ {"role": "user", "content": "Ngana mau pigi mana?"},
79
+ ])
80
+ print(output["choices"][0]["message"]["content"])
81
+ ```
82
+
83
+ ## Detail Training
84
+
85
+ | Parameter | Nilai |
86
+ |-----------|-------|
87
+ | Base Model | meta-llama/Meta-Llama-3.1-8B-Instruct |
88
+ | Fine-tuning | QLoRA 4-bit (Unsloth) |
89
+ | LoRA r / alpha | 16 / 32 |
90
+ | Dataset | ~450 train samples |
91
+ | Epochs | 3 |
92
+ | Learning rate | 0.0002 |
93
+ | Max seq length | 512 |
94
+ | Eval loss | 1.1841 |
95
+ | Perplexity | 3.27 |
96
+ | Platform | Kaggle (T4 GPU) |
97
+
98
+ ## Deployment
99
+
100
+ Tersedia sebagai HF Space: [haidar038/utu-malut-chat](https://huggingface.co/spaces/haidar038/utu-malut-chat)
101
+
102
+ ## Limitasi
103
 
104
+ - Dataset ~500 baris; belum mencakup semua variasi dialek
105
+ - Untuk riset dan pengembangan NLP bahasa daerah
106
+ - Verifikasi output sebelum penggunaan produksi
107
 
108
+ ## Kredit
109
 
110
+ Fine-tuned dengan [Unsloth](https://github.com/unslothai/unsloth) + [TRL](https://github.com/huggingface/trl) di Kaggle.