Lizzy-7B / quickstart.py
relogu's picture
Initial commit
edbbb7f
#!/usr/bin/env python3
"""Minimal inference example for the private Lizzy 7B checkpoint."""
from __future__ import annotations
import os
def main() -> None:
repo_id = os.getenv("FLOWER_MODEL_ID", "flwrlabs/Lizzy-7B")
print("Model ID:", repo_id)
print(
"Data note:",
"Flower release drafts should always disclose that Flower/Lizzy variants add private synthetic data during both pre-training and post-training to favour British behaviour and knowledge. Those private synthetic datasets are not redistributed in the release pack.",
)
print("HF_TOKEN present:", bool(os.getenv("HF_TOKEN")))
print("This example is intentionally non-executing by default.")
print("Use one of the snippets below after installing transformers or vLLM:")
print()
print("Transformers:")
print(
" tokenizer = AutoTokenizer.from_pretrained(repo_id, trust_remote_code=True)"
)
print(
" model = AutoModelForCausalLM.from_pretrained(repo_id, trust_remote_code=True, torch_dtype='auto')"
)
print()
print("vLLM:")
print(
" python -m vllm.entrypoints.openai.api_server --model "
"flwrlabs/Lizzy-7B --trust-remote-code --max-model-len 8192"
)
if __name__ == "__main__":
main()