MiniGPT-v1.0-10M-BrazilianPortuguese

Hello! My name is Arthur Samuel, I am 13 years old, and I am the creator of this project.

This repository contains MiniGPT, a small transformer-based language model with 10 million parameters. The model is designed to learn reasoning and conversational skills using a custom dataset in Brazilian Portuguese(loboGOAT/Reasoning-MiniGPT-brazilian-portuguese).

About the Model

  • Architecture: Decoder-only Transformer (GPT-style)
  • Parameters: 10M
  • Purpose: To generate text, reason logically, and have casual conversations in Brazilian Portuguese.
  • Special Tokens Used: <|PROMPT|>, <|THINK|>, <|MINIGPT|>
  • Training Dataset: Brazilian Portuguese dataset with dialogues, reasoning steps, daily-life conversations, jokes, school math (8th grade+), and tips for studying, all formatted using special tokens.

Example Usage

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("loboGOAT/MiniGPT-v1.0-10M-BrazilianPortuguese")
model = AutoModelForCausalLM.from_pretrained("loboGOAT/MiniGPT-v1.0-10M-BrazilianPortuguese")

prompt = "<|PROMPT|> Olá, MiniGPT! Como você está hoje?\n<|THINK|>"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=50)
print(tokenizer.decode(outputs[0]))
License
This project is released under the Apache-2.0 License.

⚠️ REMEMBER: This model is still in training. This is a commercial beta test.

Future plans: An English version will be released, trained on English datasets and optimized for reasoning in English as well.

Created by Arthur Samuel, age 13. MiniGPT is a small but capable transformer trained specifically for reasoning and conversation in Brazilian Portuguese. 🚀
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train AxionLab-official/MiniGPT-v1.0-10M-BrazilianPortuguese