--- base_model: google/gemma-3-270m-it library_name: peft pipeline_tag: text-generation tags: - base_model:adapter:google/gemma-3-270m-it - lora - transformers --- 🧠 MODEL CARD — DogeAI-v1.0-instruct Model Details Model Description DogeAI-v1.0-instruct is an early-stage instruction-following language model fine-tuned for conversational use and experimentation. This version is intended as a proof of concept (v1) and focuses on language generation rather than reliable logical reasoning. Developed by: Arthur(loboGOAT) Funded by: Independent / Community-driven Shared by: Arthur(loboGOAT) Model type: Small Instruction-Tuned Language Model Language(s): Portuguese (primary), multilingual tendencies inherited from base model License: Apache 2.0 (or the same license as the base model, if different) Finetuned from model: Gemma-3-270M-it Model Sources Repository: loboGOAT/DogeAI-v1.0-instruct Paper: Not available Demo: Not available Uses Direct Use Conversational experiments Text generation and rewriting Prompt testing and evaluation Educational use to study limitations of small LLMs Downstream Use (Optional) Further fine-tuning Research on alignment, reasoning, and instruction-following Benchmarking small models Out-of-Scope Use Tasks requiring reliable logical reasoning Mathematical proof or formal logic Decision-making systems Safety-critical or automated validation tasks Recommendations This model should not be relied upon for reasoning-intensive tasks. Users are encouraged to treat DogeAI-v1.0-instruct as an experimental model and expect occasional logical inconsistencies, multilingual drift, or overgeneration. Future versions aim to address these limitations through: cleaner datasets improved stopping criteria alternative base models How to Get Started with the Model from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("loboGOAT/DogeAI-v1.0-instruct") model = AutoModelForCausalLM.from_pretrained("loboGOAT/DogeAI-v1.0-instruct") inputs = tokenizer("Olá! Vamos conversar?", return_tensors="pt") outputs = model.generate( **inputs, max_new_tokens=128, temperature=0.65, top_p=0.95 ) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) Training Details Training Data The model was fine-tuned on a custom instruction-style dataset, primarily in Portuguese, designed to encourage conversational responses. The dataset does not focus on formal logic or structured reasoning. Training Procedure Preprocessing Instruction–response formatting Text normalization No explicit chain-of-thought supervision Training Hyperparameters Training regime: Supervised fine-tuning (SFT) PEFT: Yes (LoRA-based fine-tuning) Evaluation Testing Data Manual testing and prompt-based evaluation. Factors Logical consistency Instruction-following Language fluency Metrics No automated benchmarks were used for this version. Results Strong conversational fluency for model size Inconsistent logical reasoning Occasional overgeneration beyond intended response Summary Model Examination DogeAI-v1.0-instruct demonstrates the strengths and limitations of small instruction-tuned language models. While capable of natural conversation, it lacks robust reasoning abilities, which will be a focus of future iterations. Environmental Impact Hardware Type: Consumer GPU / Local Machine Hours used: Low Cloud Provider: None Compute Region: Local Carbon Emitted: Negligible Technical Specifications Model Architecture and Objective Decoder-only Transformer Next-token prediction Instruction-following objective Compute Infrastructure Local training environment. Hardware Consumer-grade GPU / CPU Software Transformers PEFT 0.18.0 PyTorch Citation BibTeX: @misc{dogeai_v1_2025, title={DogeAI-v1.0-instruct}, author={Arthur}, year={2025}, note={Early experimental instruction-tuned language model} } APA: Arthur (2025). DogeAI-v1.0-instruct: An experimental instruction-tuned language model. Model Card Authors Arthur Model Card Contact (your Hugging Face profile or GitHub)