Venomoussaversai

#3
by Ananthusajeev190 - opened

0.1B INT8 AI Model (Experimental)

This repository contains an experimental 0.1 billion (100M) parameter AI model, optimized for low-memory and edge-device inference.

The model is quantized to INT8 and stored using SafeTensors, making it suitable for deployment on CPU-only systems and mobile-class hardware.


πŸ“¦ Repository Structure

2b/
β”œβ”€β”€ model-00001-of-000xx.safetensors # INT8 weights (≀100MB shards)
β”œβ”€β”€ config.json # Model architecture config
β”œβ”€β”€ generation_config.json # Inference settings
β”œβ”€β”€ tokenizer.json # Custom tokenizer
β”œβ”€β”€ vocab.json # Vocabulary
β”œβ”€β”€ merges.txt # BPE merges
β”œβ”€β”€ tokenizer_config.json
β”œβ”€β”€ special_tokens_map.json
└── README.md


πŸ”§ Model Specifications

  • Parameter Count: ~100,000,000 (0.1B)
  • Model Type: Transformer (GPT-style)
  • Precision: INT8 (quantized)
  • Serialized Size: ~249 MB (including metadata)
  • Weight Format: ".safetensors"
  • Tokenizer: Custom BPE (domain-trained)
  • Framework: PyTorch + Hugging Face

🎯 Intended Use Cases

  • Pattern recognition
  • Trend and probability forecasting
  • Lightweight reasoning tasks
  • Offline and edge-device inference

Not intended for:

  • Real-time future prediction
  • Medical, legal, or financial decision-making
  • Autonomous control systems

πŸš€ Loading the Model

from transformers import AutoModelForCausalLM, AutoTokenizer

model_path = "./2b"

tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
model_path,
load_in_8bit=True,
device_map="auto"
)


⚠️ Limitations

  • Outputs are probabilistic, not deterministic
  • Performance depends on tokenizer quality and training data
  • Some layers may remain in higher precision for stability
  • No real-time data access

πŸ“œ License & Disclaimer

This model is provided for research and experimentation purposes only.
The author assumes no responsibility for misuse or misinterpretation of the model’s outputs.


✨ Author

Ananthu Sajeev
AI Research & Edge Model Development


🧠 Notes

Smaller models prioritize efficiency and control over scale.
At this size, data quality > parameter count.

Sign up or log in to comment