π₯ Aras-Ember v2
Aras-Ember v2 is a lightweight conversational AI model developed by Sparrow AI Solutions.
It is built by fine-tuning Gemma-2-2B, a language model created by Google, using the Ember dataset.
Aras-Ember v2 is designed for research, experimentation, and lightweight conversational AI applications.
This project is independent and not affiliated with Google.
β οΈ Important Notice
This model is a derivative work of the Gemma model family released by Google.
Use of this model is subject to:
Gemma Terms of Use https://ai.google.dev/gemma/terms
and
Gemma Prohibited Use Policy https://ai.google.dev/gemma/prohibited_use_policy
By downloading, using, or distributing this model you agree to comply with those terms.
[]
(https://colab.research.google.com/drive/1wjdkj7niIiKeBpVVjnyBrSYu88lCtZxn?usp=sharing)
π§ Model Details
Model name: Aras-Ember v2 Developer: Sparrow AI Solutions Base model: google/gemma-2-2b Model architecture: Gemma decoder-only transformer Parameter count: ~2 Billion
Training Approach
- LoRA instruction tuning
- Conversational fine-tuning
- LoRA weights merged into the base model
- Exported as a standalone full model
Frameworks Used
- Transformers
- PEFT
- PyTorch
- Hugging Face Datasets
π Dataset
Training dataset:
https://huggingface.co/datasets/sparrowaisolutions/ember-dataset
Dataset Structure
{
"instruction": "...",
"response": "..."
}
The dataset consists of instructionβresponse conversational pairs designed for training instructionβfollowing language models.
Dataset Intended Uses
- Conversational AI
- Creative generation
- Instruction following
βοΈ Training Details
Training Configuration
- Base model: Gemmaβ2β2B
- Training method: LoRA fineβtuning
- Dataset size: ~30,000 examples
- Epochs: 2
Optimization
- Mixed precision (FP16)
- Gradient accumulation
- LoRA adapters merged after training
Training objective:
Instructionβfollowing conversational generation.
π Usage
Install Dependencies
pip install transformers torch
Example Inference
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model_name = "sparrowaisolutions/aras-ember-v2"
tokenizer = AutoTokenizer.from_pretrained(model_name)
tokenizer.pad_token = tokenizer.eos_token
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.float16,
device_map="auto"
)
prompt = """
You are Aras-Ember, a creative AI assistant.
Write a short poem about the sea and the moon.
"""
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(
**inputs,
max_new_tokens=150,
temperature=1.1,
top_p=0.95,
do_sample=True,
pad_token_id=tokenizer.eos_token_id
)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
π¬ Prompt Format
Recommended prompt format:
You are Aras-Ember, a creative AI assistant.
Prompt: <instruction>
Response:
Example:
Prompt: Explain black holes simply.
Response:
π§ͺ Intended Use
ArasβEmber v2 is intended for:
- Conversational AI research
- Educational projects
- Experimentation with LLM fineβtuning
- Creative writing generation
- Lightweight chatbots
- AI development testing
β Out-of-Scope Uses
This model should not be used for:
- Medical advice
- Legal advice
- Safetyβcritical systems
- Automated decisionβmaking
- Misinformation generation
- Illegal or harmful activities
Users are responsible for how the model is used.
β οΈ Limitations
Because this is a relatively small language model:
- May generate incorrect or fabricated information
- Limited reasoning ability
- Limited longβcontext understanding
- Performance depends on prompt quality
- Not suitable for highβstakes applications
π Architecture
Base architecture:
Gemmaβ2β2B by Google
Reference model:
https://huggingface.co/google/gemma-2-2b
This project modifies the base model through instruction fineβtuning only.
No architectural changes were made.
π¨βπ» Authors
Developed by:
Sparrow AI Solutions
Hugging Face profile:
https://huggingface.co/sparrowaisolutions
β€οΈ Acknowledgements
Special thanks to:
- Google Gemma team
- Hugging Face
- Openβsource AI community
π License
This project is released under the Apache License 2.0.
However, because the model is derived from Gemma, use of the model is also subject to:
Gemma Terms of Use https://ai.google.dev/gemma/terms
Gemma Prohibited Use Policy https://ai.google.dev/gemma/prohibited_use_policy
Users must comply with both licenses when using or distributing this model.
βοΈ Disclaimer
The model is provided "AS IS", without warranty of any kind.
The developers are not responsible for any damages, misuse, or consequences resulting from the use of this model.
Users assume full responsibility for ensuring compliance with applicable laws and policies.
Research Paper
EMBER Dataset and ARAS-EMBER Models: Open Lightweight AI Systems for Creative and Conversational Language Generation
DOI: https://doi.org/10.6084/m9.figshare.31617994
π Version
Current version: ArasβEmber v2 Release date: 2026
- Downloads last month
- 91
Model tree for sparrowaisolutions/aras-ember-v2
Base model
google/gemma-2-2b