--- library_name: transformers tags: - creole - haitian license: mit language: - ht base_model: - jsbeaudry/makandal-pre-trained pipeline_tag: text-generation --- # Makandal Continue Pre-trained from qwen3-0.6b ## Model Details This model has been continued pre-trained from qwen3-0.6b by Palmis Labs AI. . ### Model Description - **Developed by:** Palmis Labs AI - **Funded by:** Jean Sauvenel Beaudry - **Model type:** GPT (Generative Pre-trained Transformer) - **Language(s) (NLP):** Haitian Creole - **License:** MIT - **Model size:** 0.6B parameters - **Architecture:** qwen3 ### Direct Use ```python from transformers import AutoTokenizer, AutoModelForCausalLM import torch def generate(model, tokenizer, prompt, device): inputs = tokenizer(prompt, return_tensors="pt", padding=True).to(device) output = model.generate( **inputs, max_new_tokens=100, do_sample=True, repetition_penalty=1.2, no_repeat_ngram_size=3, temperature=0.9, top_k=40, top_p=0.85, pad_token_id=tokenizer.pad_token_id, eos_token_id=tokenizer.eos_token_id ) return tokenizer.decode(output[0], skip_special_tokens=True) # Load model and tokenizer tokenizer = AutoTokenizer.from_pretrained("jsbeaudry/makandal-v2") model = AutoModelForCausalLM.from_pretrained("jsbeaudry/makandal-v2") # Set device device = "cuda" if torch.cuda.is_available() else "cpu" model.to(device) # Generate text prompt = "matematik" response = generate(model, tokenizer, prompt, device) print(response) # Answer: # Matematik se yon disiplin matematik ki konsantre sou kalkil, estatistik, ak analiz matematik. # Li pèmèt nou konprann enfòmasyon ak fòmèlman analize done pou jwenn pwopriyete oswa fòmèlman verifye yon konpreyansyon. ``` ### Out-of-Scope Use This model should **NOT** be used for: - Critical decision-making systems - Any application requiring reliable or factual outputs - Commercial deployment without significant additional training ## Bias, Risks, and Limitations - **Insufficient training data**: Only 4.7 MB of training data used - **Limited training time**: Only 4.5 hours of training - **High hallucination rate**: Model frequently generates inaccurate or nonsensical content - **Language coverage**: Limited Haitian Creole language understanding due to minimal dataset - **Bias**: May reflect biases present in the small training dataset ### Recommendations - Do not rely on outputs for factual information - Supervise usage in educational settings ### Training Infrastructure - **GPU:** Tesla T4 (15GB) - **Framework:** Transformers/PyTorch ## Citation ```bibtex @misc{makandal2025, title={Makandal-pretrain: An Educational Haitian Creole Language Model}, author={Jean Sauvenel Beaudry}, year={2025}, howpublished={\url{https://huggingface.co/jsbeaudry/makandal-pre-trained}}, note={Educational demonstration model} } ``` ## Glossary **Makandal**: Named after François Makandal, an 18th-century Haitian revolutionary leader, symbolizing the model's connection to Haitian culture and education.