Zyora-Byte-32B
An AI Teaching Assistant for Anna University Engineering Students, fine-tuned from Qwen2.5-Coder-32B-Instruct using QLoRA.
Model Details
Model Description
Zyora-Byte-32B (ByteBuddy) is a 32B parameter language model fine-tuned to help engineering students at Anna University with syllabus information, course details, problem-solving, and coding assistance. It understands both formal queries and casual student language.
- Developed by: Zyora Labs
- Model type: Causal Language Model (LoRA Adapter)
- Language(s): English
- License: Apache 2.0
- Finetuned from: Qwen/Qwen2.5-Coder-32B-Instruct
Model Sources
- Repository: zyoralabs/zyora-Byte-32B
Uses
Direct Use
- Answer syllabus questions for Anna University R2021 curriculum
- Provide course information (credits, semesters, prerequisites)
- Solve math problems (calculus, Laplace transforms, differential equations)
- Generate code with explanations (C, Python)
- Academic guidance for engineering students
Supported Branches
- Computer Science and Engineering (CSE)
- Electronics and Communication Engineering (ECE)
- Electrical and Electronics Engineering (EEE)
- Mechanical Engineering
- Civil Engineering
- Information Technology (IT)
- Computer Science - Data Science
- Computer Science - Cyber Security
- Computer Science - Business Systems (CSBS)
Out-of-Scope Use
- Non-Anna University curriculum queries
- Medical, legal, or financial advice
- Content generation for harmful purposes
Bias, Risks, and Limitations
- Focused specifically on Anna University R2021 curriculum
- Best performance for Semesters 1-4 (based on training data)
- May hallucinate for courses not in the training dataset
- Requires the base model (Qwen2.5-Coder-32B-Instruct) for inference
Recommendations
Users should verify critical academic information with official university sources. This model is meant to assist, not replace, official curriculum documents.
How to Get Started with the Model
from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig
from peft import PeftModel
import torch
# Load base model with 4-bit quantization
bnb_config = BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_quant_type="nf4",
bnb_4bit_compute_dtype=torch.bfloat16,
bnb_4bit_use_double_quant=True,
)
base_model = AutoModelForCausalLM.from_pretrained(
"Qwen/Qwen2.5-Coder-32B-Instruct",
quantization_config=bnb_config,
device_map="auto",
trust_remote_code=True,
)
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-Coder-32B-Instruct")
# Load LoRA adapters
model = PeftModel.from_pretrained(base_model, "zyoralabs/zyora-Byte-32B")
# Generate response
prompt = """<|im_start|>system
You are ByteBuddy, an AI teaching assistant helping engineering students understand their coursework.<|im_end|>
<|im_start|>user
What are the subjects in CSE Semester 1?<|im_end|>
<|im_start|>assistant
"""
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512, temperature=0.7, top_p=0.9)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Training Details
Training Data
Source: Custom dataset extracted from Anna University R2021 syllabus PDFs
Samples: 16,109 training examples
Content: Course information, credits, semesters, study tips, project ideas, placement prep, problem-solving examples
Training Procedure
Training Hyperparameters
Training regime: bf16 mixed precision
Batch size: 2 per device
Gradient accumulation: 4 steps
Effective batch size: 8
Learning rate: 2e-4
Scheduler: Cosine with 3% warmup
Optimizer: paged_adamw_8bit
Epochs: 3
Final loss: 0.00844
LoRA Configuration
LoraConfig(
r=32,
lora_alpha=64,
target_modules=["q_proj", "k_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj"],
lora_dropout=0.05,
bias="none",
task_type="CAUSAL_LM",
)
Speeds, Sizes, Times
Training time: ~9 hours total
Adapter size: 1.0 GB
Training cost: ~$35 (Modal)
Evaluation
Example Queries & Results
Query Response
"List all courses in Semester 1 for CSE" Full course list with codes and credits
"How many credits is MA25C01?" 4 credits
"bro what are the lab subjects in mech sem 2?" Lab subjects list (understands casual language)
"Find the derivative of x³ + 2x² - 5x + 7" 3x² + 4x - 5
"Find the Laplace transform of e^(-2t)" 1/(s+2)
"Write a C program for factorial" Complete working code with explanation
Environmental Impact
Hardware Type: NVIDIA A100-80GB
Hours used: ~9 hours
Cloud Provider: Modal
Compute Region: US
Technical Specifications
Model Architecture and Objective
Architecture: Transformer (Qwen2.5 architecture)
Parameters: 32B base + ~1GB LoRA adapters
Objective: Causal Language Modeling
Compute Infrastructure
Hardware
1x NVIDIA A100-80GB GPU
64GB System RAM
8 CPU cores
Software
PyTorch 2.5.0
Transformers 4.46.3
PEFT 0.18.1
bitsandbytes 0.44.0
Hardware Requirements for Inference
Setup VRAM Required
4-bit Quantized ~18GB
8-bit Quantized ~32GB
Full Precision ~64GB
Citation
@misc{zyora-byte-32b-2025,
author = {Zyora Labs},
title = {Zyora-Byte-32B: AI Teaching Assistant for Anna University},
year = {2025},
publisher = {Hugging Face},
url = {https://huggingface.co/zyoralabs/zyora-Byte-32B}
}
Model Card Authors
Zyora Labs
Model Card Contact
Zyora Labs on Hugging Face
Framework versions
PEFT 0.18.1
Transformers 4.46.3
PyTorch 2.5.0
- Downloads last month
- 19
Model tree for zyoralabs/zyora-Byte-32B
Base model
Qwen/Qwen2.5-32B
Finetuned
Qwen/Qwen2.5-Coder-32B
Finetuned
Qwen/Qwen2.5-Coder-32B-Instruct