WakeUp SigLIP-1 Base INT8 (ONNX)

ONNX INT8 exports of google/siglip-base-patch16-224 for use in the WakeUp Flutter alarm app's "Travel Mode" feature.

Files

File Size Purpose
siglip1_image_encoder_int8.onnx ~99 MB Image feature extraction (per-scan)
siglip1_text_encoder_int8.onnx ~111 MB Text feature extraction (Custom Text mode only)
model_metadata.json โ€” logit_scale, logit_bias, normalization params
tokenizer/ โ€” SigLIP-1 tokenizer files

Scoring

Both encoders L2-normalize their output. SigLIP scoring is:

logits = exp(logit_scale) * cosine(image_emb, text_emb) + logit_bias
prob   = sigmoid(logits)

Constants from model_metadata.json:

  • logit_scale = 4.765 (so exp(scale) โ‰ˆ 117.33)
  • logit_bias = -12.932

Inference

import onnxruntime as ort
import numpy as np

img_sess = ort.InferenceSession("siglip1_image_encoder_int8.onnx")
txt_sess = ort.InferenceSession("siglip1_text_encoder_int8.onnx")

# image: 1x3x224x224 normalized with mean/std [0.5, 0.5, 0.5]
image_emb = img_sess.run(None, {"pixel_values": pixel_values})[0]
# text: input_ids + attention_mask (use all-ones mask for canonical inference)
text_emb = txt_sess.run(None, {"input_ids": ids, "attention_mask": np.ones_like(ids)})[0]

cos = text_emb @ image_emb.T
prob = 1 / (1 + np.exp(-(np.exp(4.765) * cos + -12.932)))

License

Apache 2.0 (inherits from base model).

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Sentrix-Code/wakeup-siglip1-base-int8

Quantized
(3)
this model