--- base_model: unsloth/gemma-2b-bnb-4bit library_name: peft pipeline_tag: text-generation tags: - base_model:adapter:unsloth/gemma-2b-bnb-4bit - lora - sft - transformers - trl - unsloth --- # Model Card for ScoLaM ## Model Details ### Model Description ScoLaM is a fine-tuned language model based on the **unsloth/gemma-2b-bnb-4bit** base model. It uses Parameter-Efficient Fine-Tuning (PEFT) techniques, specifically LoRA (Low-Rank Adaptation), to enable efficient adaptation with reduced compute and storage requirements. ScoLaM is designed primarily for text-generation tasks and can be applied in domains requiring lightweight, performant language modeling. - **Developed by:** Team Scorton - **Funded by:** SchoolyAI - **Shared by:** https://github.com/scorton - **Model type:** Transformer-based causal language model with LoRA fine-tuning - **Language(s):** English (primarily), French (second), Spanish (Second) - **License:** [Specify license, e.g., Apache 2.0, MIT, etc.] - **Finetuned from model:** unsloth/gemma-2b-bnb-4bit (4-bit quantized base model) ### Model Sources - **Repository:** hugging.co/schooly - **Paper:** [Link to relevant publication if any] - **Demo:** [URL to demo application if any] ## Uses ### Direct Use ScoLaM is intended for general-purpose text generation tasks such as drafting, creative writing, summarization, or chatbot dialogue generation. It can be used directly via text-generation pipelines in Hugging Face Transformers using PEFT adapters. ### Downstream Use ScoLaM can serve as a base for further fine-tuning on domain-specific datasets or for integration into larger NLP systems, chatbots, or AI assistants that benefit from efficient fine-tuning and inference. ### Out-of-Scope Use - Use in highly safety-critical or sensitive applications without further validation. - Generation of misleading, harmful, or biased content. - Applications requiring strong factual accuracy without additional grounding. ## Bias, Risks, and Limitations ScoLaM inherits biases present in the base model and training data. It may produce biased, harmful, or nonsensical outputs if used improperly. Its quantized 4-bit format may also affect precision in some use cases. ### Recommendations Users should evaluate outputs carefully, especially in high-stakes scenarios. Fine-tuning or prompt engineering may be needed to mitigate undesired behavior. ## How to Get Started with the Model ```python from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline from peft import PeftModel base_model = "unsloth/gemma-2b-bnb-4bit" adapter_model = "path_or_id_to_scolam_adapter" tokenizer = AutoTokenizer.from_pretrained(base_model) model = AutoModelForCausalLM.from_pretrained(base_model) model = PeftModel.from_pretrained(model, adapter_model) text_gen = pipeline("text-generation", model=model, tokenizer=tokenizer) output = text_gen("Your prompt here", max_length=50) print(output)