Unable to determine steps to implement OpenVINO™ Runtime inference pipeline with IR.
Resolution
Create OpenVINO™ Runtime Core import openvino.runtime as ov core = ov.Core()
Compile the Model compiled_model = core.compile_model("model.xml", "AUTO")
Create an Infer Request infer_request = compiled_model.create_infer_request()
Set Inputs # Create tensor from external memory input_tensor = ov.Tensor(array=memory, shared_memory=True) # Set input tensor for model with one input infer_request.set_input_tensor(input_tensor)
Process the Inference Results # Get output tensor for model with one output output = infer_request.get_output_tensor() output_buffer = output.data # output_buffer[] - accessing output tensor data