Babelbit Miner Model

A fine-tuned utterance prediction model for the Babelbit subnet (netuid 59) on the Bittensor network.


What this model does

Given a partial utterance prefix and optional conversation context, the model predicts the most natural and complete continuation of the utterance.

The subnet evaluates predictions on:

  • Lexical similarity — exact word overlap with the ground truth
  • Semantic similarity — meaning-level match
  • Earliness — how early in the utterance the prediction was correct

Usage

Direct inference

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", torch_dtype=torch.float16)

system = (
    "You are a helpful assistant that completes the current utterance naturally and succinctly. "
    "Return only the completed utterance text without quotes or extra commentary."
)
prefix = "Hi - how"
context = ""

if context:
    user = f"Context:\n{context}\n\nContinue the utterance that begins with:\n{prefix}"
else:
    user = f"Continue the utterance that begins with:\n{prefix}"

messages = [
    {"role": "system", "content": system},
    {"role": "user", "content": user},
]

prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)

with torch.inference_mode():
    output = model.generate(**inputs, max_new_tokens=64, do_sample=False)

decoded = tokenizer.decode(output[0][inputs["input_ids"].shape[1]:], skip_special_tokens=True)
print(decoded)
# → "Hi - how are you?"

Validator API request format

Validators query the miner at POST /predict:

{
  "index": "<session-uuid>",
  "step": 3,
  "prefix": "Hi - how",
  "context": "",
  "done": false
}

Expected response:

{
  "prediction": "Hi - how are you?"
}

Limitations

  • Designed specifically for the Babelbit subnet utterance prediction task.
  • Best results on conversational English; performance varies on other languages.
  • Predictions are evaluated on partial prefixes — shorter prefixes are inherently harder.
Downloads last month
461
Safetensors
Model size
13B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support