--- base_model: Qwen/Qwen2.5-3B-Instruct library_name: transformers license: apache-2.0 tags: - qwen - conversational - reasoning - math - code-generation - css - javascript - html - physics --- # 💎 Geode Onyx 2 (3B) Onyx 2 is a 3-billion parameter conversational AI model, fine-tuned as part of the second generation of the Geode model family. ## Model Details - **Base Model:** Qwen 2.5 3B Instruct - **Parameters:** 3 Billion - **Fine-Tuning:** LoRA (r=32, alpha=64) - **Training Loss:** 0.40 - **Precision:** FP16 - **License:** Apache 2.0 ## The Geode Family (Second Generation) The Geode family is Genue AI's lineup of locally-runnable conversational models. In the second generation, Beryl has been retired and replaced by Pyrite, a specialized coding model: | Model | Parameters | Role | |-------|------------|------| | Pyrite | 7B | Coding specialist | | Onyx | 3B | Balanced logic & personality | | Thaumite | 8B | Flagship, highest capability | **Note:** Beryl (0.5B) was the original lightweight experimental model in the first generation and has been replaced by Pyrite, which focuses specifically on code generation tasks. ## Usage Onyx 2 uses the Qwen Instruct prompt format: ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch model = AutoModelForCausalLM.from_pretrained( "GenueAI/Geode-Onyx-2", torch_dtype=torch.float16, device_map="auto", trust_remote_code=True ) tokenizer = AutoTokenizer.from_pretrained("GenueAI/Geode-Onyx-2") prompt = "<|im_start|>user\nWhat is your name?<|im_end|>\n<|im_start|>assistant\n" inputs = tokenizer(prompt, return_tensors="pt").to("cuda") outputs = model.generate(**inputs, max_new_tokens=256, temperature=0.7) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ## Training Data Fine-tuned on a curated dataset of 1,013 examples covering: - **Identity & self-awareness** - AI assistant identity and capabilities - **Mathematical reasoning** - Arithmetic, algebra, word problems - **General knowledge** - Broad factual knowledge - **HTML/CSS/JavaScript code generation** - Web development tasks - **Physics problems** - Falling objects, thermodynamics - **Genue AI ecosystem knowledge** - Company information, model family details - **Conversational generalization** - Natural dialogue patterns - **Anti-hallucination training** - Proper handling of unknown information (time, location, preferences) ## Model Architecture - Base: Qwen 2.5 3B Instruct - Adapter: LoRA with r=32, alpha=64 - Target modules: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj - Trainable parameters: 59.9M (1.9% of total) ## Training Details - **Training regime:** FP16 mixed precision - **Epochs:** 2 - **Batch size:** 8 - **Learning rate:** 2e-4 - **Training time:** ~8 minutes on RTX 3090 ## Developed By Genue AI — Founded by Brybod123 (Bradar) ## Model Card Contact For questions or issues, contact Genue AI through the HuggingFace repository.