|
|
--- |
|
|
base_model: unsloth/gemma-2b-bnb-4bit |
|
|
library_name: peft |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- base_model:adapter:unsloth/gemma-2b-bnb-4bit |
|
|
- lora |
|
|
- sft |
|
|
- transformers |
|
|
- trl |
|
|
- unsloth |
|
|
--- |
|
|
|
|
|
# Model Card for ScoLaM |
|
|
|
|
|
## Model Details |
|
|
|
|
|
### Model Description |
|
|
|
|
|
ScoLaM is a fine-tuned language model based on the **unsloth/gemma-2b-bnb-4bit** base model. It uses Parameter-Efficient Fine-Tuning (PEFT) techniques, specifically LoRA (Low-Rank Adaptation), to enable efficient adaptation with reduced compute and storage requirements. ScoLaM is designed primarily for text-generation tasks and can be applied in domains requiring lightweight, performant language modeling. |
|
|
|
|
|
- **Developed by:** Team Scorton |
|
|
- **Funded by:** SchoolyAI |
|
|
- **Shared by:** https://github.com/scorton |
|
|
- **Model type:** Transformer-based causal language model with LoRA fine-tuning |
|
|
- **Language(s):** English (primarily), French (second), Spanish (Second) |
|
|
- **License:** [Specify license, e.g., Apache 2.0, MIT, etc.] |
|
|
- **Finetuned from model:** unsloth/gemma-2b-bnb-4bit (4-bit quantized base model) |
|
|
|
|
|
### Model Sources |
|
|
|
|
|
- **Repository:** hugging.co/schooly |
|
|
- **Paper:** [Link to relevant publication if any] |
|
|
- **Demo:** [URL to demo application if any] |
|
|
|
|
|
## Uses |
|
|
|
|
|
### Direct Use |
|
|
|
|
|
ScoLaM is intended for general-purpose text generation tasks such as drafting, creative writing, summarization, or chatbot dialogue generation. It can be used directly via text-generation pipelines in Hugging Face Transformers using PEFT adapters. |
|
|
|
|
|
### Downstream Use |
|
|
|
|
|
ScoLaM can serve as a base for further fine-tuning on domain-specific datasets or for integration into larger NLP systems, chatbots, or AI assistants that benefit from efficient fine-tuning and inference. |
|
|
|
|
|
### Out-of-Scope Use |
|
|
|
|
|
- Use in highly safety-critical or sensitive applications without further validation. |
|
|
- Generation of misleading, harmful, or biased content. |
|
|
- Applications requiring strong factual accuracy without additional grounding. |
|
|
|
|
|
## Bias, Risks, and Limitations |
|
|
|
|
|
ScoLaM inherits biases present in the base model and training data. It may produce biased, harmful, or nonsensical outputs if used improperly. Its quantized 4-bit format may also affect precision in some use cases. |
|
|
|
|
|
### Recommendations |
|
|
|
|
|
Users should evaluate outputs carefully, especially in high-stakes scenarios. Fine-tuning or prompt engineering may be needed to mitigate undesired behavior. |
|
|
|
|
|
## How to Get Started with the Model |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline |
|
|
from peft import PeftModel |
|
|
|
|
|
base_model = "unsloth/gemma-2b-bnb-4bit" |
|
|
adapter_model = "path_or_id_to_scolam_adapter" |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained(base_model) |
|
|
model = AutoModelForCausalLM.from_pretrained(base_model) |
|
|
model = PeftModel.from_pretrained(model, adapter_model) |
|
|
|
|
|
text_gen = pipeline("text-generation", model=model, tokenizer=tokenizer) |
|
|
output = text_gen("Your prompt here", max_length=50) |
|
|
print(output) |
|
|
|