Instructions to use FrontiersMind/Nandi-Mini-150M-Instruct with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use FrontiersMind/Nandi-Mini-150M-Instruct with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="FrontiersMind/Nandi-Mini-150M-Instruct", trust_remote_code=True) messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)# Load model directly from transformers import AutoModelForCausalLM model = AutoModelForCausalLM.from_pretrained("FrontiersMind/Nandi-Mini-150M-Instruct", trust_remote_code=True, dtype="auto") - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use FrontiersMind/Nandi-Mini-150M-Instruct with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "FrontiersMind/Nandi-Mini-150M-Instruct" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "FrontiersMind/Nandi-Mini-150M-Instruct", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'Use Docker
docker model run hf.co/FrontiersMind/Nandi-Mini-150M-Instruct
- SGLang
How to use FrontiersMind/Nandi-Mini-150M-Instruct with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "FrontiersMind/Nandi-Mini-150M-Instruct" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "FrontiersMind/Nandi-Mini-150M-Instruct", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "FrontiersMind/Nandi-Mini-150M-Instruct" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "FrontiersMind/Nandi-Mini-150M-Instruct", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }' - Docker Model Runner
How to use FrontiersMind/Nandi-Mini-150M-Instruct with Docker Model Runner:
docker model run hf.co/FrontiersMind/Nandi-Mini-150M-Instruct
| license: apache-2.0 | |
| language: | |
| - en | |
| - hi | |
| - mr | |
| - ta | |
| - te | |
| - kn | |
| - ml | |
| - bn | |
| - pa | |
| - gu | |
| - or | |
| pipeline_tag: text-generation | |
| library_name: transformers | |
| base_model: | |
| - FrontiersMind/Nandi-Mini-150M | |
| # Nandi-Mini-150M-Instruct | |
| ## Introduction | |
| Nandi-Mini-150M-Instruct is a compact, efficient multilingual language model designed for strong performance in resource-constrained environments. It is pre-trained from scratch on 525 billion tokens and further enhanced through instruction tuning and Direct Preference Optimization (DPO). The model supports English and 10 Indic languages. | |
| Nandi-Mini-150M-Instruct focuses on maximizing performance per parameter through architectural efficiency rather than scale. It is optimized for edge devices, on-prem deployments, and low-latency applications, making it ideal for resource-constrained environments. | |
| Nandi-Mini-150M-Instruct brings the following key features: | |
| - Strong **multilingual capability** across English and Indic languages | |
| - Efficient design enabling **high performance at small scale (150M parameters)** | |
| - Reduced memory footprint using **factorized embeddings** | |
| - Better parameter efficiency through **layer sharing** | |
| ## π Upcoming Releases & Roadmap | |
| Weβre just getting started with the Nandi series π | |
| - **Nandi-Mini-150M-Tool-Calling (Specialized-Model)** β Coming Soon this week | |
| - **Nandi-Mini-500M (Base + Instruct)** β Pre-Training Going On | |
| - **Nandi-Mini-1B (Base + Instruct)** β Pre-Training Going On | |
| π’ **Blogs & technical deep-dives coming soon**, where weβll share: | |
| - Architecture decisions and design trade-offs | |
| - Training insights and dataset composition | |
| - Benchmarks and real-world applications | |
| Stay tuned! | |
| ## π Supported Languages | |
| The model is trained on English and a diverse set of Indic languages, including: | |
| - Hindi, Bengali, Tamil, Telugu, Marathi, Gujarati, Kannada, Malayalam, Punjabi, Odia | |
| ## π Usage | |
| ```python | |
| !pip install transformers=='5.4.0' | |
| from transformers import AutoTokenizer, AutoModelForCausalLM | |
| import torch | |
| model_name = "FrontiersMind/Nandi-Mini-150M-Instruct" | |
| device = "cuda" if torch.cuda.is_available() else "cpu" | |
| tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True) | |
| model = AutoModelForCausalLM.from_pretrained( | |
| model_name, | |
| trust_remote_code=True, | |
| dtype=torch.bfloat16 | |
| ).to(device).eval() | |
| prompt = "Explain newton's second law of motion" | |
| messages = [ | |
| {"role": "user", "content": prompt} | |
| ] | |
| prompt = tokenizer.apply_chat_template(messages, tokenize=False) | |
| inputs = tokenizer(prompt, return_tensors="pt").to(model.device) | |
| generated_ids = model.generate( | |
| **inputs, | |
| max_new_tokens=500, | |
| do_sample=True, | |
| temperature=0.3, | |
| top_p=0.90, | |
| top_k=20, | |
| repetition_penalty=1.1, | |
| ) | |
| generated_ids = [ | |
| output_ids[len(input_ids):] for input_ids, output_ids in zip(inputs.input_ids, generated_ids) | |
| ] | |
| response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0] | |
| print(response) | |
| ``` | |
| ## π¬ Feedback & Suggestions | |
| Weβd love to hear your thoughts, feedback, and ideas! | |
| - **Discord**: https://discord.gg/ZGdjCdRt | |
| - **Email:** support@frontiersmind.ai | |
| - **Official Website** https://www.frontiersmind.ai/ | |
| - **LinkedIn:** https://www.linkedin.com/company/frontiersmind/ | |
| - **X (Twitter):** https://x.com/FrontiersMind |