Instructions to use QuantFactory/Neeto-1.0-8b-GGUF with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use QuantFactory/Neeto-1.0-8b-GGUF with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="QuantFactory/Neeto-1.0-8b-GGUF")# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("QuantFactory/Neeto-1.0-8b-GGUF", dtype="auto") - llama-cpp-python
How to use QuantFactory/Neeto-1.0-8b-GGUF with llama-cpp-python:
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="QuantFactory/Neeto-1.0-8b-GGUF", filename="Neeto-1.0-8b.Q2_K.gguf", )
output = llm( "Once upon a time,", max_tokens=512, echo=True ) print(output)
- Notebooks
- Google Colab
- Kaggle
- Local Apps
- llama.cpp
How to use QuantFactory/Neeto-1.0-8b-GGUF with llama.cpp:
Install from brew
brew install llama.cpp # Start a local OpenAI-compatible server with a web UI: llama-server -hf QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M # Run inference directly in the terminal: llama-cli -hf QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M
Install from WinGet (Windows)
winget install llama.cpp # Start a local OpenAI-compatible server with a web UI: llama-server -hf QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M # Run inference directly in the terminal: llama-cli -hf QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M
Use pre-built binary
# Download pre-built binary from: # https://github.com/ggerganov/llama.cpp/releases # Start a local OpenAI-compatible server with a web UI: ./llama-server -hf QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M # Run inference directly in the terminal: ./llama-cli -hf QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M
Build from source code
git clone https://github.com/ggerganov/llama.cpp.git cd llama.cpp cmake -B build cmake --build build -j --target llama-server llama-cli # Start a local OpenAI-compatible server with a web UI: ./build/bin/llama-server -hf QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M # Run inference directly in the terminal: ./build/bin/llama-cli -hf QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M
Use Docker
docker model run hf.co/QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M
- LM Studio
- Jan
- vLLM
How to use QuantFactory/Neeto-1.0-8b-GGUF with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "QuantFactory/Neeto-1.0-8b-GGUF" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "QuantFactory/Neeto-1.0-8b-GGUF", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker
docker model run hf.co/QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M
- SGLang
How to use QuantFactory/Neeto-1.0-8b-GGUF with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "QuantFactory/Neeto-1.0-8b-GGUF" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "QuantFactory/Neeto-1.0-8b-GGUF", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "QuantFactory/Neeto-1.0-8b-GGUF" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "QuantFactory/Neeto-1.0-8b-GGUF", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }' - Ollama
How to use QuantFactory/Neeto-1.0-8b-GGUF with Ollama:
ollama run hf.co/QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M
- Unsloth Studio new
How to use QuantFactory/Neeto-1.0-8b-GGUF with Unsloth Studio:
Install Unsloth Studio (macOS, Linux, WSL)
curl -fsSL https://unsloth.ai/install.sh | sh # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for QuantFactory/Neeto-1.0-8b-GGUF to start chatting
Install Unsloth Studio (Windows)
irm https://unsloth.ai/install.ps1 | iex # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for QuantFactory/Neeto-1.0-8b-GGUF to start chatting
Using HuggingFace Spaces for Unsloth
# No setup required # Open https://huggingface.co/spaces/unsloth/studio in your browser # Search for QuantFactory/Neeto-1.0-8b-GGUF to start chatting
- Docker Model Runner
How to use QuantFactory/Neeto-1.0-8b-GGUF with Docker Model Runner:
docker model run hf.co/QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M
- Lemonade
How to use QuantFactory/Neeto-1.0-8b-GGUF with Lemonade:
Pull the model
# Download Lemonade from https://lemonade-server.ai/ lemonade pull QuantFactory/Neeto-1.0-8b-GGUF:Q4_K_M
Run and chat with the model
lemonade run user.Neeto-1.0-8b-GGUF-Q4_K_M
List all available models
lemonade list
Install from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf QuantFactory/Neeto-1.0-8b-GGUF:# Run inference directly in the terminal:
llama-cli -hf QuantFactory/Neeto-1.0-8b-GGUF:Use pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf QuantFactory/Neeto-1.0-8b-GGUF:# Run inference directly in the terminal:
./llama-cli -hf QuantFactory/Neeto-1.0-8b-GGUF:Build from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf QuantFactory/Neeto-1.0-8b-GGUF:# Run inference directly in the terminal:
./build/bin/llama-cli -hf QuantFactory/Neeto-1.0-8b-GGUF:Use Docker
docker model run hf.co/QuantFactory/Neeto-1.0-8b-GGUF:- QuantFactory/Neeto-1.0-8b-GGUF
- Original Model Card
- Neeto-1.0-8b - A Specialized Medical LLM for NEET-PG/UKMLE/USMLE preparation
QuantFactory/Neeto-1.0-8b-GGUF
This is quantized version of S4nfs/Neeto-1.0-8b created using llama.cpp
Original Model Card
Neeto-1.0-8b - A Specialized Medical LLM for NEET-PG/UKMLE/USMLE preparation
Neeto-1.0-8b is an openly released biomedical large language model (LLM) created by BYOL Academy to assist learners and practitioners with medical exam study, literature understanding, and structured clinical reasoning.
The model was adapted on a curated mixture (≈410K items) blending synthetic generations and hand-audited instructional / multiple‑choice / rationale samples. The objective was balanced: retain broad linguistic competence while strengthening factual recall, differential diagnostics framing, and question dissection for exams such as NEET‑PG, UKMLE, and USMLE.
Across widely used evaluation suites (MedQA, MedMCQA, PubMedQA, MMLU medical subsets), Neeto‑1.0‑8b attains strong 7B‑class results. Public benchmark numbers (table below) show it standing ahead of several prior open biomedical baselines of similar scale. The model will be used on our platform Medicoplasma as for exam preparation and powering medical applications.
How to Use
Important Note:
This model has been strictly trained on medical datasets only. It is not designed for general chit-chat or off-topic questions.
For example, it may not respond meaningfully to prompts like:
- "Hello"
- "Tell me a joke"
- "What’s the weather today?"
👉 Please use the model only for medical-related tasks, as that is its intended purpose.
The model follows the default Llama‑3 chat message formatting (no explicit system prompt required). Provide a single user turn containing the question or case vignette; the model returns an answer (option selection, rationale, or free-form explanation depending on the prompt style).
Below are illustrative input patterns for multi‑choice items (MedQA / MedMCQA), PubMedQA‑style reasoning, and open clinical queries. For reproducibility of benchmark-style MCQ evaluation, keep choices clearly enumerated (A./B./C./D.) and avoid extra prose.
Example (MedQA / MedMCQA style)
A 55-year-old male presents with sudden onset of severe unilateral flank pain radiating to the groin, accompanied by hematuria. Imaging reveals a calculus in the proximal ureter. Given the high prevalence of anatomical variations in the renal arteries and their proximity to the ureters, what is the primary clinical concern regarding surgical or interventional management of this patient's ureteral calculus, and which specific anatomical variation would most significantly complicate access or increase the risk of iatrogenic injury?
A. Aberrant accessory renal artery crossing the ureter, causing obstruction and risk of vascular injury during intervention.
B. Early bifurcation of the main renal artery within the hilum, increasing the risk of ureteral devascularization.
C. Dual renal veins draining into the inferior vena cava, raising concern for venous congestion during stone removal.
D. Persistent fetal renal lobulations that distort the renal pelvis and complicate stent placement.
🛠️ Run with vLLM (Method 1)
The easiest way to run this model locally or on a server is using vLLM.
1. Install vLLM
pip install vllm
2. Start the vLLM Server
vllm serve S4nfs/Neeto-1.0-8b
3. Interact with the Model
With the server running, you can send requests to it using curl or a Python script.
Using curl Open a new terminal and send a request to the server:
curl http://localhost:8000/v1/completions -H "Content-Type: application/json" -d '{
"model": "S4nfs/Neeto-1.0-8b",
"prompt": "Why is the sky blue?",
"max_tokens": 4096,
"temperature": 0.7
}'
🛠️ Using Transformers (Method 2)
from transformers import AutoTokenizer
from vllm import LLM, SamplingParams
llm = LLM(model="S4nfs/Neeto-1.0-8b", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("S4nfs/Neeto-1.0-8b")
sampling_params = SamplingParams(temperature=0.7, top_p=0.9, max_tokens=1024, stop=["<|eot_id|>"])
messages = [
{"role": "user", "content": """The question format used in the above input examples。"""},
]
prompts = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
print(prompts[0])
"""
<|begin_of_text|><|start_header_id|>user<|end_header_id|>
{question}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
"""
outputs = llm.generate(prompts=prompts, sampling_params=sampling_params)
print(outputs[0].outputs[0].text)
Note: Current release is optimized for single‑turn exchanges. Multi‑turn conversational coherence will be improved in an upcoming iteration.
Benchmark with Major Models
Benchmark Snapshot Among Similar Size Models
Neeto‑1.0‑8b delivers the following published scores:
| Released Date | Model | Average | MedQA | MedMCQA | PubMedQA | MMLU.ck | MMLU.mg | MMLU.an | MMLU.pm | MMLU.cb | MMLU.cm |
|---|---|---|---|---|---|---|---|---|---|---|---|
| 2025.08 | Neeto-1.0-8b | 80.69 | 85.80 | 66.20 | 79.00 | 79.40 | 87.10 | 79.10 | 89.60 | 81.40 | 78.60 |
| 2024.04 | OpenBioLM-8B | 72.47 | 59.07 | 56.83 | 74.18 | 76.02 | 86.15 | 69.72 | 78.11 | 84.24 | 67.94 |
| 2024.04 | Llama-3-8B-Instruct | 71.23 | 62.33 | 56.58 | 75.74 | 72.59 | 83.91 | 71.18 | 70.54 | 80.52 | 67.68 |
| 2024.04 | Internist-7B | 67.77 | 60.42 | 55.88 | 79.31 | 70.53 | 71.10 | 65.81 | 76.18 | - | 62.92 |
| 2024.02 | Gemma-7B | 64.16 | 47.27 | 48.91 | 76.12 | 69.72 | 70.10 | 59.22 | 66.12 | 80.00 | 60.02 |
| 2024.03 | Meerkat-7B | 63.92 | 74.22 | 60.62 | - | 61.99 | 70.31 | 61.58 | 69.42 | 55.49 | 57.72 |
| 2023.03 | MedAlpaca | 58.02 | 41.63 | 37.59 | 72.72 | 57.48 | 68.91 | 57.08 | 67.21 | 65.22 | 54.38 |
| 2024.02 | BioMistral-7B | 57.25 | 46.52 | 45.79 | 68.02 | 63.19 | 63.21 | 49.98 | 57.31 | 63.48 | 57.72 |
Interpretation & Methodology:
- MedQA uses the US 4‑option subset; MedMCQA uses the Dev split; PubMedQA reflects the “reasoning required” subset.
- MMLU medical grouping here incorporates: Clinical Knowledge (CK), Medical Genetics (MG), Anatomy (An), Professional Medicine (PM), College Biology (CB), College Medicine (CM).
- Greedy decoding was the baseline; ensemble self‑consistency scores (not shown) were generated via 10 samples (temperature 0.7, top_p 0.9) with majority voting.
- Comparative baselines partially sourced from the public Open Medical‑LLM Leaderboard.
Training Configuration
Full‑parameter supervised fine‑tuning was executed under Fully Sharded Data Parallel (FSDP). Hardware: 8 × H200 GPUs.
Hyperparameters:
- torch type: bfloat16
- epochs: 3
- learning rate: 2e-5
- learning rate scheduler type: cosine
- warmup ratio: 0.04
- max length: 1024
- global batch size: 128
Limitations & Responsible Use
Despite strong benchmark standing, the model can hallucinate mechanistic explanations, mis-rank differential diagnoses, or fabricate citations. It must not be used for autonomous clinical decision-making, patient triage, prescribing, or emergency guidance. Human expert verification is mandatory before any medical action.
Planned Enhancements
- Preference optimization (DPO) variants on forthcoming Llama releases.
- Expansion to JEE Advanced and NEET‑UG aligned scientific subject packs.
- Multi-turn dialogue memory and structured rationale modes.
- Integration within MedicoPlasma’s chat interface.
Citation
@misc{Neeto-1.0-8b,
author = {Sagar Verma},
title = {NEETO: A Specialized Medical LLM for NEET-PG/UKMLE/USMLE preparation},
year = {2025},
publisher = {GitHub},
journal = {GitHub repository},
note = {\url{https://huggingface.co/S4nfs/Neeto-1.0-8b}},
}
- Downloads last month
- 49
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit

Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf QuantFactory/Neeto-1.0-8b-GGUF:# Run inference directly in the terminal: llama-cli -hf QuantFactory/Neeto-1.0-8b-GGUF: