Seed-0.5B

  • 0.5B decoder only dense model trained from scratch with 6B tokens.
  • As it isn't instruction finetuned, the model performs document completion, not conversational generation.
  • Model is released primarily for educational, research, and experimental purposes.

GitHub: merterbak/llm-from-scratch

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("merterbak/Seed-0.5B")
model = AutoModelForCausalLM.from_pretrained(
    "merterbak/Seed-0.5B",
    trust_remote_code=True,
    dtype="auto"
)
prompt = "Climate change can affect"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(
    **inputs,
    temperature=0.3,
    top_k=40,
    do_sample=True,
    repetition_penalty=1.2,
    no_repeat_ngram_size=3,
)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
58
Safetensors
Model size
0.5B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train merterbak/Seed-0.5B