HIKARI β€” Healthcare-oriented Intelligent Knowledge Augmented Retrieval and Inference

HIKARI-Deneb-8B-SkinDx-Cascade

Healthcare-oriented Intelligent Knowledge Augmented Retrieval and Inference
Named after Deneb β€” tail star of Cygnus the Swan, tracing the cascade from group to disease


πŸ“¦ Model Type: Merged Full Model

This is a fully merged model β€” the LoRA adapter weights have been merged directly into the base model weights.

βœ… No adapter loading needed. Load and run directly with transformers, vLLM, or SGLang.

πŸ’Ύ Size: ~17 GB (4 safetensor shards)

πŸ”Œ Lightweight adapter version: E27085921/HIKARI-Deneb-8B-SkinDx-Cascade-LoRA (~1.1 GB)


Overview

HIKARI-Deneb uses 2-Stage Cascade Fine-Tuning for 10-class skin disease diagnosis. Instead of starting from the raw base model, Stage 2 disease training begins from the weights of HIKARI-Subaru-8B-SkinGroup (the 4-class group classifier). The model inherits group-level visual discrimination before learning disease-level distinctions.

Property Value
Task 10-class skin disease diagnosis (Stage 2)
Base model Qwen/Qwen3-VL-8B-Thinking
Init weights HIKARI-Subaru group classifier (cascaded)
Val accuracy (no RAG) 74.00%
Val accuracy (R2 Ξ±=0.9) 79.80% (with RAG retrieval at inference)
Model type Merged full model

Why Cascade?

Training Start Accuracy Reason
Raw base model (Altair) 74.00% Learns disease features from scratch
Group Classifier weights (Deneb) 79.80% Pre-trained on inter-group visual comparison

Starting from the group classifier teaches the model to compare visual features across skin disease groups before learning fine-grained disease distinctions β€” making it significantly more effective when used with RAG reference images.


πŸ”§ Quick Inference β€” transformers

from transformers import Qwen3VLForConditionalGeneration, AutoProcessor
import torch
from PIL import Image

model_id = "E27085921/HIKARI-Deneb-8B-SkinDx-Cascade"

processor = AutoProcessor.from_pretrained(model_id, trust_remote_code=True)
model = Qwen3VLForConditionalGeneration.from_pretrained(
    model_id, torch_dtype=torch.bfloat16, device_map="auto", trust_remote_code=True
)

image = Image.open("skin_lesion.jpg").convert("RGB")
group = "inflammatory"  # from Stage 1 (HIKARI-Subaru)

PROMPT = (
    "This skin lesion belongs to the group '{group}'. "
    "Examine the lesion morphology (papules, plaques, macules), "
    "color (red, violet, white, brown), scale/crust, border sharpness, "
    "and distribution pattern. Based on these visual features, "
    "what is the specific skin disease?"
)

messages = [{"role": "user", "content": [
    {"type": "image", "image": image},
    {"type": "text", "text": PROMPT.format(group=group)},
]}]
text = processor.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = processor(text=[text], images=[image], return_tensors="pt").to(model.device)

with torch.no_grad():
    out = model.generate(**inputs, max_new_tokens=64, temperature=0.0, do_sample=False)

print(processor.batch_decode(out[:, inputs["input_ids"].shape[1]:], skip_special_tokens=True)[0].strip())

πŸ”Œ LoRA Adapter Version

from peft import PeftModel
from transformers import Qwen3VLForConditionalGeneration
import torch

base = Qwen3VLForConditionalGeneration.from_pretrained(
    "Qwen/Qwen3-VL-8B-Thinking", torch_dtype=torch.bfloat16, device_map="auto"
)
model = PeftModel.from_pretrained(base, "E27085921/HIKARI-Deneb-8B-SkinDx-Cascade-LoRA")

β†’ E27085921/HIKARI-Deneb-8B-SkinDx-Cascade-LoRA


πŸ“„ Citation

@misc{hikari2026,
  title  = {HIKARI: RAG-in-Training for Skin Disease Diagnosis
            with Cascaded Vision-Language Models},
  author = {Watin Promfiy and Pawitra Boonprasart},
  year   = {2026},
  institution = {King Mongkut's Institute of Technology Ladkrabang,
                 Department of Information Technology, Bangkok, Thailand}
}

Made with ❀️ at King Mongkut's Institute of Technology Ladkrabang (KMITL)

Downloads last month
15
Safetensors
Model size
9B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for E27085921/HIKARI-Deneb-8B-SkinDx-Cascade

Finetuned
(50)
this model
Quantizations
1 model

Collection including E27085921/HIKARI-Deneb-8B-SkinDx-Cascade