Model Card for ScoLaM

Model Details

Model Description

ScoLaM is a fine-tuned language model based on the unsloth/gemma-2b-bnb-4bit base model. It uses Parameter-Efficient Fine-Tuning (PEFT) techniques, specifically LoRA (Low-Rank Adaptation), to enable efficient adaptation with reduced compute and storage requirements. ScoLaM is designed primarily for text-generation tasks and can be applied in domains requiring lightweight, performant language modeling.

  • Developed by: Team Scorton
  • Funded by: SchoolyAI
  • Shared by: https://github.com/scorton
  • Model type: Transformer-based causal language model with LoRA fine-tuning
  • Language(s): English (primarily), French (second), Spanish (Second)
  • License: [Specify license, e.g., Apache 2.0, MIT, etc.]
  • Finetuned from model: unsloth/gemma-2b-bnb-4bit (4-bit quantized base model)

Model Sources

  • Repository: hugging.co/schooly
  • Paper: [Link to relevant publication if any]
  • Demo: [URL to demo application if any]

Uses

Direct Use

ScoLaM is intended for general-purpose text generation tasks such as drafting, creative writing, summarization, or chatbot dialogue generation. It can be used directly via text-generation pipelines in Hugging Face Transformers using PEFT adapters.

Downstream Use

ScoLaM can serve as a base for further fine-tuning on domain-specific datasets or for integration into larger NLP systems, chatbots, or AI assistants that benefit from efficient fine-tuning and inference.

Out-of-Scope Use

  • Use in highly safety-critical or sensitive applications without further validation.
  • Generation of misleading, harmful, or biased content.
  • Applications requiring strong factual accuracy without additional grounding.

Bias, Risks, and Limitations

ScoLaM inherits biases present in the base model and training data. It may produce biased, harmful, or nonsensical outputs if used improperly. Its quantized 4-bit format may also affect precision in some use cases.

Recommendations

Users should evaluate outputs carefully, especially in high-stakes scenarios. Fine-tuning or prompt engineering may be needed to mitigate undesired behavior.

How to Get Started with the Model

from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
from peft import PeftModel

base_model = "unsloth/gemma-2b-bnb-4bit"
adapter_model = "path_or_id_to_scolam_adapter"

tokenizer = AutoTokenizer.from_pretrained(base_model)
model = AutoModelForCausalLM.from_pretrained(base_model)
model = PeftModel.from_pretrained(model, adapter_model)

text_gen = pipeline("text-generation", model=model, tokenizer=tokenizer)
output = text_gen("Your prompt here", max_length=50)
print(output)
Downloads last month
5
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for schooly/scolam-instruct

Adapter
(19)
this model