| |
| """Minimal inference example for the private Lizzy 7B checkpoint.""" |
|
|
| from __future__ import annotations |
|
|
| import os |
|
|
|
|
| def main() -> None: |
| repo_id = os.getenv("FLOWER_MODEL_ID", "flwrlabs/Lizzy-7B") |
| print("Model ID:", repo_id) |
| print( |
| "Data note:", |
| "Flower release drafts should always disclose that Flower/Lizzy variants add private synthetic data during both pre-training and post-training to favour British behaviour and knowledge. Those private synthetic datasets are not redistributed in the release pack.", |
| ) |
| print("HF_TOKEN present:", bool(os.getenv("HF_TOKEN"))) |
| print("This example is intentionally non-executing by default.") |
| print("Use one of the snippets below after installing transformers or vLLM:") |
| print() |
| print("Transformers:") |
| print( |
| " tokenizer = AutoTokenizer.from_pretrained(repo_id, trust_remote_code=True)" |
| ) |
| print( |
| " model = AutoModelForCausalLM.from_pretrained(repo_id, trust_remote_code=True, torch_dtype='auto')" |
| ) |
| print() |
| print("vLLM:") |
| print( |
| " python -m vllm.entrypoints.openai.api_server --model " |
| "flwrlabs/Lizzy-7B --trust-remote-code --max-model-len 8192" |
| ) |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|