DatafoundryAI's picture
Initial upload of OpenVINO quantized model (INT4)
ec0c6ff verified
# Step 1: Install required package
# !pip install openvino-genai
import openvino_genai as ov_genai
# Step 2: Define local model path
model_path = "/home/anish/Desktop/Anish/Openvino/Gemma-3-1b-it-ov-sym-int4/"
# Step 3: Initialize pipeline
device = "CPU" # or "GPU" if supported
pipe = ov_genai.LLMPipeline(model_path, device)
# Step 4: Set chat template (important for Qwen)
tokenizer = pipe.get_tokenizer()
tokenizer.set_chat_template(tokenizer.chat_template)
# Step 5: Run inference
prompt = "Capital of Australia ?"
response = pipe.generate(prompt, max_length=1024, temperature=0.7, top_p=0.9)
print("\n🧾 Model Response:")
print(response)