| from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline |
|
|
| def load_model(model_name="your-username/sentinel"): |
| """ |
| Load Sentinel model and tokenizer. |
| """ |
| print(f"Loading {model_name}...") |
| tokenizer = AutoTokenizer.from_pretrained(model_name) |
| model = AutoModelForCausalLM.from_pretrained( |
| model_name, |
| device_map="auto", |
| trust_remote_code=True |
| ) |
| generator = pipeline("text-generation", model=model, tokenizer=tokenizer) |
| return generator |
|
|
| def code_with_sentinel(prompt, generator, max_new_tokens=200): |
| """ |
| Generate code from a natural language prompt. |
| """ |
| print(f"\nPrompt: {prompt}\n") |
| output = generator( |
| prompt, |
| max_new_tokens=max_new_tokens, |
| do_sample=True, |
| top_p=0.9, |
| temperature=0.7, |
| eos_token_id=generator.tokenizer.eos_token_id |
| ) |
| result = output[0]["generated_text"] |
| |
| return result[len(prompt):].strip() |
|
|
| if __name__ == "__main__": |
| |
| generator = load_model("your-username/sentinel") |
|
|
| prompt = "Write a Python function that checks if a number is prime." |
| code = code_with_sentinel(prompt, generator) |
|
|
| print("Generated Code:\n") |
| print(code) |
|
|