Unable to Infer OpenVINO™ Optimized TinyLlama Model on NPU
Content Type: Troubleshooting | Article ID: 000101250 | Last Reviewed: 05/28/2025
Infer OpenVINO™ Optimized TinyLlama Model on NPU using OpenVINO™ GenAI.
import openvino_genai as ov_genai
model_path = "TinyLlama"
pipe = ov_genai.LLMPipeline(model_path, "NPU")
print(pipe.generate("What is OpenVINO?", max_length=200))