πŸ”₯ Aras-Ember v2

Aras-Ember v2 is a lightweight conversational AI model developed by Sparrow AI Solutions.

It is built by fine-tuning Gemma-2-2B, a language model created by Google, using the Ember dataset.

Aras-Ember v2 is designed for research, experimentation, and lightweight conversational AI applications.

This project is independent and not affiliated with Google.


⚠️ Important Notice

This model is a derivative work of the Gemma model family released by Google.

Use of this model is subject to:

Gemma Terms of Use https://ai.google.dev/gemma/terms

and

Gemma Prohibited Use Policy https://ai.google.dev/gemma/prohibited_use_policy

By downloading, using, or distributing this model you agree to comply with those terms.


[Open In Colab] (https://colab.research.google.com/drive/1wjdkj7niIiKeBpVVjnyBrSYu88lCtZxn?usp=sharing)

🧠 Model Details

Model name: Aras-Ember v2 Developer: Sparrow AI Solutions Base model: google/gemma-2-2b Model architecture: Gemma decoder-only transformer Parameter count: ~2 Billion

Training Approach

  • LoRA instruction tuning
  • Conversational fine-tuning
  • LoRA weights merged into the base model
  • Exported as a standalone full model

Frameworks Used

  • Transformers
  • PEFT
  • PyTorch
  • Hugging Face Datasets

πŸ“š Dataset

Training dataset:

https://huggingface.co/datasets/sparrowaisolutions/ember-dataset

Dataset Structure

{
  "instruction": "...",
  "response": "..."
}

The dataset consists of instruction–response conversational pairs designed for training instruction‑following language models.

Dataset Intended Uses

  • Conversational AI
  • Creative generation
  • Instruction following

βš™οΈ Training Details

Training Configuration

  • Base model: Gemma‑2‑2B
  • Training method: LoRA fine‑tuning
  • Dataset size: ~30,000 examples
  • Epochs: 2

Optimization

  • Mixed precision (FP16)
  • Gradient accumulation
  • LoRA adapters merged after training

Training objective:

Instruction‑following conversational generation.


πŸš€ Usage

Install Dependencies

pip install transformers torch

Example Inference

import torch
from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "sparrowaisolutions/aras-ember-v2"

tokenizer = AutoTokenizer.from_pretrained(model_name)
tokenizer.pad_token = tokenizer.eos_token

model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype=torch.float16,
    device_map="auto"
)

prompt = """
You are Aras-Ember, a creative AI assistant.

Write a short poem about the sea and the moon.
"""

inputs = tokenizer(prompt, return_tensors="pt").to(model.device)

outputs = model.generate(
    **inputs,
    max_new_tokens=150,
    temperature=1.1,
    top_p=0.95,
    do_sample=True,
    pad_token_id=tokenizer.eos_token_id
)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))

πŸ’¬ Prompt Format

Recommended prompt format:

You are Aras-Ember, a creative AI assistant.

Prompt: <instruction>
Response:

Example:

Prompt: Explain black holes simply.
Response:

πŸ§ͺ Intended Use

Aras‑Ember v2 is intended for:

  • Conversational AI research
  • Educational projects
  • Experimentation with LLM fine‑tuning
  • Creative writing generation
  • Lightweight chatbots
  • AI development testing

❌ Out-of-Scope Uses

This model should not be used for:

  • Medical advice
  • Legal advice
  • Safety‑critical systems
  • Automated decision‑making
  • Misinformation generation
  • Illegal or harmful activities

Users are responsible for how the model is used.


⚠️ Limitations

Because this is a relatively small language model:

  • May generate incorrect or fabricated information
  • Limited reasoning ability
  • Limited long‑context understanding
  • Performance depends on prompt quality
  • Not suitable for high‑stakes applications

πŸ— Architecture

Base architecture:

Gemma‑2‑2B by Google

Reference model:

https://huggingface.co/google/gemma-2-2b

This project modifies the base model through instruction fine‑tuning only.

No architectural changes were made.


πŸ‘¨β€πŸ’» Authors

Developed by:

Sparrow AI Solutions

Hugging Face profile:

https://huggingface.co/sparrowaisolutions


❀️ Acknowledgements

Special thanks to:

  • Google Gemma team
  • Hugging Face
  • Open‑source AI community

πŸ“œ License

This project is released under the Apache License 2.0.

However, because the model is derived from Gemma, use of the model is also subject to:

Gemma Terms of Use https://ai.google.dev/gemma/terms

Gemma Prohibited Use Policy https://ai.google.dev/gemma/prohibited_use_policy

Users must comply with both licenses when using or distributing this model.


βš–οΈ Disclaimer

The model is provided "AS IS", without warranty of any kind.

The developers are not responsible for any damages, misuse, or consequences resulting from the use of this model.

Users assume full responsibility for ensuring compliance with applicable laws and policies.


Research Paper

EMBER Dataset and ARAS-EMBER Models: Open Lightweight AI Systems for Creative and Conversational Language Generation

DOI: https://doi.org/10.6084/m9.figshare.31617994

πŸ“… Version

Current version: Aras‑Ember v2 Release date: 2026

Downloads last month
91
Safetensors
Model size
3B params
Tensor type
F16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for sparrowaisolutions/aras-ember-v2

Finetuned
(428)
this model

Dataset used to train sparrowaisolutions/aras-ember-v2

Space using sparrowaisolutions/aras-ember-v2 1