YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

๐Ÿ•Š๏ธ TagoreX โ€“ A Bengali Text Generator Inspired by Tagore

Model name: SwastikGuhaRoy/TagoreX Base model: GPT-2 with LoRA adapters (based on AddaGPT2.0) Language: Bengali Author: Swastik Guha Roy (@SwastikGuhaRoy) License: MIT Model size: ~124M parameters Trained on: Curated (but imperfect) corpus of Rabindranath Tagoreโ€™s writings Intended use: Poetic and philosophical Bengali text generation Demo app: TagoreX + Gemini Streamlit App


๐Ÿ“˜ Model Description

TagoreX is a fine-tuned version of AddaGPT2.0 โ€” a small GPT-2 model adapted for Bengali using LoRA (Low-Rank Adaptation). This model was trained on literary works of Rabindranath Tagore as a tribute.

The model continues a given Bengali prompt in a Tagore-like poetic tone. It generates ~256 tokens, which are then optionally refined by Gemini AI in a downstream application.


๐Ÿ”ง Technical Details

  • Architecture: GPT-2 (117M parameters)
  • Training strategy: Full fine-tuning
  • Epochs: 22 (symbolically referencing โ€œเงจเงจเฆถเง‡ เฆถเงเฆฐเฆพเฆฌเฆฃโ€)
  • Max sequence length: 256 tokens
  • Tokenizer: AutoTokenizer from the base model
  • Framework: PyTorch + Transformers

๐Ÿ“‚ Training Data

The dataset includes poems, prose and other works from Rabindranath Tagore which is publicly available. The dataset can be accessed in a consolidated .txt format from here

โš ๏ธ Note: The data may and DOES contain:

  • Typos, formatting errors
  • OCR issues
  • Incomplete or duplicated lines

This model is not a scholarly curation, but an experimental artistic rendering.


๐ŸŽฏ Intended Use

You can use this model to:

  • Experiment with Bengali poetic text generation
  • Create creative writing prompts in Bengali
  • Explore Indic LLM capabilities in low-resource settings

This model is not suitable for:

  • Any commercial or sensitive deployment
  • Factual or linguistic accuracy tasks
  • Scholarly representation of Tagoreโ€™s works

๐Ÿ’ฌ How to Prompt

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("SwastikGuhaRoy/TagoreX")
model = AutoModelForCausalLM.from_pretrained("SwastikGuhaRoy/TagoreX")

prompt = "เฆคเงเฆฎเฆฟ เฆฐเฆฌเง‡ เฆจเง€เฆฐเฆฌเง‡"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=256, do_sample=True, temperature=0.7)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

๐Ÿšซ Limitations & Disclaimer

  • Not aligned, filtered, or safety-trained.
  • Most outputs may be incoherent, repetitive, or nonsensical.
  • This is not meant to reproduce or replace Tagore's literary work.
  • The generation reflects training data and randomness โ€” not any human author.

๐ŸŒ Why It Matters

TagoreX demonstrates how even small-scale, open models can express poetic and cultural essence in Indic languages โ€” using limited compute and a lot of curiosity.

It aims to inspire communities to build Indic LLMs, especially in low-resource and rural settings.

"AI doesnโ€™t have to be massive. It can be local, soulful, and deeply human."



๐Ÿ“ซ Contact

๐Ÿ“ง Email: swastikguharoy@googlemail.com ๐Ÿ’ฌ Feedback, bugs, or nice generations? I'd love to hear from you!


Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for SwastikGuhaRoy/TagoreX

Quantizations
1 model