Intel® Edge AI Box

ID 795550
Updated 7/23/2024
Version 4.0
Public

author-image

By

Overview

AI Suite for Visual Analytics (also known as Intel Edge AI Box) Software toolkit is a modular, containerized reference architecture and a toolkit for AI pipeline prototyping and Platform Measurement to kickstart your visual AI solution development. It supports accelerated media processing and inferencing with Intel ingredients such as OneVPL, Libva (VAAPI), Intel® Distribution of OpenVINO™ toolkit and Intel® Deep Learning Streamer framework. It also includes enhanced version of OpenCV and FFMpeg to speed up your Edge AI solutions development. Configure your application end-to-end with flexible AI capacity and reference video analytics pipeline for fast development.

Select Configure & Download to download the package and the software listed below.

Configure & Download

Prerequisites

  • Programming Language: Python, C, C++
  • Available Software: 
    • Intel® Distribution of OpenVINO™ toolkit 2024.1.0 (for Base Container), 2024.0.0 (for DLStreamer container) and 2023.3.0 (for OpenCV/FFMpeg Container)
    • Intel® Deep Learning Streamer 2024.0.1
    • Intel® oneAPI Video Processing Library
    • Intel® oneAPI Deep Neural Networks Library
    • OpenCV 4.8.0 (limited features)
    • Intel® FFmpeg Cartwheel 2023Q3 (limited features)
    • XPU-Manager, Node-Exporter, Prometheus, and Grafana for CPU and GPU telemetry

Recommended Hardware

See the Recommended Hardware page for suggestions.

Target System Requirements

  • 11th, 12th, or 13th Generation Embedded Intel® Core™ processors
  • Intel® Core™ Ultra Processors
  • 12th generation Intel® Core™ Desktop processors with Intel® Arc™ A380 Graphics
  • Intel Atom® Processor X7000 Series (formerly Alder Lake-N)
  • Intel® Processor N-series (formerly Alder Lake-N)
  • 4th Gen Intel® Xeon® Scalable Processors
  • Operating System:
    • Ubuntu* Desktop 22.04 (fresh installation)
      • Preferred Ubuntu* 22.04.3 LTS (fresh installation) for Core and Core Ultra platform
      • Preferred Ubuntu* 22.04.4 LTS Server (fresh installation) for Xeon platform
  • At least 80 GB of disk space
  • At least 16 GB of memory
  • Direct Internet access

Ensure you have sudo access to the system and a stable Internet connection.

How It Works 

Intel® Edge AI Box reference architecture forms the base to create a complete video analytic system for lightweight edge devices. This package supports the 11th, 12th, and 13th generation embedded Intel® Core™ processors, Intel® Core™ Ultra Processors, 12th generation desktop Intel® Core™ processors with Intel® Arc™ A380 Graphics, Intel Atom® processor X7000-series,Intel Atom® process N-series (formerly Alder Lake-N) and 4th Gen Intel® Xeon® Scalable Processors.

An image showing the software stack of Intel Edge AI Box
Figure 1. Software Stack Diagram

 

The core of Intel® Edge AI Box reference architecture is a suite of containers. The Video Analytics (VA) Base Library container includes OpenVINO™, oneDNN, and GPU drivers for accelerated inferencing and media processing. Base-devel extends the VA Base Library container:

Container

Content

Base

VA Base Library (contains OpenVINO 2024.1.0 runtime)

Base-devel

VA Base Library and OpenVINO 2024.1.0 developer tools
OpenCV FFmpeg Ubuntu 22.04, OpenCV, FFMpeg, OpenVINO 2023.3.0

DLStreamer

Ubuntu 22.04, DLStreamer/GStreamer, OpenVINO 2024.0.0 runtime, and developer tools

Get Started

Prerequisite

  1. Perform a fresh installation of Ubuntu Desktop 22.04 on the target system.(For Core and Core Ultra platform, Ubuntu 22.04.3 LTS is preferred, for Xeon platform, Ubuntu 22.04.4 LTS Server is preferred)

Note for Xeon platform: AI Box expects a pure CPU configuration on Xeon Platform, all benchmarks will run on CPU. If your system has graphic card, please unplug the graphic card before installing AI Box package.

  1.  Install Thingsboard (Optional)

Follow these steps if you want Remote Device Management feature

  1. Use another system, install Thingsboard using steps in the link here 
  2. Launch http://<thingsboard_system_IP>:8080
  3. Enter the default tenant credential as specified in the Thingsboard installation guide.
  4. At left bar section, select Entities-> Devices
  5. Click ‘+’ button to add device
  6. Provide a device name for your Intel ® Edge AI Box 
  7. Click ‘Next: Credentials’
  8. Take note the Device Access token. Click Add

During ./edgesoftware install, you will be prompted to enter the Thingsboard system IP address, server port (default 1883) and the Device Access Token during Edge AI Box installation below.

Install the Package

  1. Select Configure & Download to download the Intel® Edge AI Box package.
    Configure & Download

A screenshot showing the available configuration and download options.
Figure 2. Configure and Download
  1. Click Download. In the next screen, accept the license agreement and copy the Product Key. 
  2. Transfer the downloaded package to the target Ubuntu* system and unzip:
unzip intel_edge_aibox.zip

NOTE: Please use the same credential that was created during Ubuntu installation to proceed with the installation for Intel® Edge AIBox

  1. Go to the intel_edge_ai_box/ directory:
cd intel_edge_aibox
  1. Change the permission of the executable edgesoftware file:
chmod 755 edgesoftware
  1. For Atom and Core platform, if your OS kernel is 6.5.0-35-generic, check if ‘echo $LC_IDENTIFICATION’ does show *UTF-8. If no, then proceed the following steps:
sudo apt-get install locales -y
sudo locale-gen en_US.UTF-8
sudo update-locale LANG=en_US.UTF-8 LC_CTYPE=en_US.UTF-8 LC_IDENTIFICATION=en_US.UTF-8
sudo reboot
  1. Install the Intel® Edge AI Box package:
./edgesoftware install
  1. When prompted, enter the Product Key. You can enter the Product Key mentioned in the email from Intel confirming your download (or the Product Key you copied in Step 2).

Note for People’s Republic of China (PRC) Network: 

  • If you are connecting from the PRC network, the following prompt will appear during bmra base installation: 

A screenshot showing a prompt to enable PRC mirror network while installing Edge AI Box on a PRC Network
Figure 3. Prompt to Enable PRC Network Mirror
  • Type No to use the default settings, or Yes to enter the local mirror URL for pip and apt package managers. 

  • Similar prompt will appear during bmra base installation:

Figure 4. Prompt to Enable PRC Network Mirror
  • Type Yes to avoid download failure in the PRC network. Below follow up prompts will appear:

A screenshot showing the installation prompting whether to access proxy server on PRC network
Figure 5. Prompt to change default GitHub and Docker hub mirror on PRC Network
  • Type No to use the default github mirror and dockerhub mirror. ** If you want to enable RDM and XPU Manager under PRC network, you need to type Yes to enter an available dockerhub mirror.

  • Next, you can choose whether to install XPU Manager, if you type Yes, please make sure you have replaced the default dockerhub mirror in previous step:

Figure 6. Prompt to Enable XPU Manager on PRC Network
  • The following prompt will appear when installing AIBox Samples, type Yes to use default GitHub mirror: 

Figure 7. Prompt to change GitHub mirror on PRC Network

  1. For Atom and Core platform, it will prompt to reboot the system if your kernel is not 6.5.0.35-generic. After reboot the system, go to step 6 again to check $LC_IDENTIFICATION. Then install again with command ./edgesoftware install

  2. If you choose to enable Remote Device Management feature, type Yes and enter the ThingsBoard server IP, port and access token from prerequisite Step 4 above. If you want to enable it under PRC network, please make sure you have replaced the default dockerhub mirror in previous step.
Figure 8. Prompt to Enable RDM
  1. When prompted for the BECOME password, enter your Linux* account password.
A screenshot showing the prompt for Linux credentials
Figure 9. Prompt for BECOME Password
  1. When prompted to reboot the machine, press Enter. Ensure to save your work before rebooting.
A screenshot showing the prompt to reboot
Figure 10. Prompt to Reboot
  1. After rebooting, resume the installation:
cd intel_edge_aibox
./edgesoftware install
  1. After the ESDQ is installed, you will be prompted for the password. Enter the password to proceed.
A screenshot showing the prompt for password at the end of ESDQ installation
Figure 11. Prompt for Password
  1. When the installation is complete, you will see the message “Installation of package complete” and the installation status for each module.
A screenshot showing the installation complete message and status of modules
Figure 12. Installation Complete Message
  1. Reboot the system:
sudo reboot 

Run Benchmarks with Device Qualification Tools

Prerequisite: We recommend connecting 2x 4K monitors to run the benchmarks. 

The Intel® Edge Software Device Qualification (Intel® ESDQ) infrastructure is used to run the test suites. All Intel® Edge AI Box tests are invoked as arguments to the Intel® ESDQ infrastructure. For the complete Intel® ESDQ CLI, refer to Intel® ESDQ CLI Overview.

Find the available Intel® Edge AI Box tests:

cd intel_edge_aibox/Intel_Edge_AI_Box_4.0
esdq --verbose module run aibox --arg "-h"

Execute all device qualification tests and generate the complete report:

cd intel_edge_aibox/Intel_Edge_AI_Box_4.0
esdq --verbose module run aibox --arg "-r all"

For more details on running the device qualification tools, refer to Intel® Edge Software Device Qualification for Intel® Edge AI Box.

Proxy Pipelines

AI Box 4.0 provides 3 Proxy pipelines, each pipeline has 2 execution modes:

  • Benchmark Execution: Run benchmark using Intel® ESDQ tools to find the max number of AI inference channels that the platform can support, as listed in the previous section.
  • Custom Execution: Run customized pipelines with config files. You can specify the number of inference streams, the models used to run pipeline and MQTT broker address and topic to send inference metadata. This section will introduce how to modify the config files and run pipelines.

Smart NVR Proxy Pipeline

Smart NVR pipeline contains the following process:

  1. Decodes 25 1080p H264 video channels
  2. Stores video stream in local storage
  3. Runs AI Inference on a subset of the video channels
  4. Sends Inference metadata to MQTT broker
  5. Composes decoded video channels into multi-view 4K video wall stream
  6. Encodes the multi-view video wall stream for remote viewing
  7. Display the multi-view video wall streams on attached monitor (4k monitor is preferred)
Figure 13. Smart NVR Proxy Pipeline

How To Run Smart NVR Pipeline

  1. Go to this directory:
cd intel_edge_aibox/Intel_Edge_AI_Box_4.0/Proxy_Pipeline_Module/smart_nvr_runner
  1. You may change pipeline configuration in this file
nano smart_nvr.cfg

 

Figure 14. Smart NVR Configuration

  

Refer to custom_models/README.md for more information

  1. To run pipeline:
./run_container.sh smart_nvr.cfg

Headed Visual AI Proxy Pipelines

Headed Visual AI Proxy Pipeline contains the following process:

  1. Decodes a specified number of 1080p video channels
  2. Runs AI Inference on all video channels
  3. Composes all video channels into a multi-view video wall stream
  4. Encodes the multi-view video wall stream for remote viewing
  5. Display the multi-view video wall stream on attached monitor (4K monitor is preferred)
Figure 15. Headed Visual AI Proxy Pipeline

How To Run Visual AI Proxy Pipeline

  1. Go to this directory
cd intel_edge_aibox/Intel_Edge_AI_Box_4.0/Proxy_Pipeline_Module/visual_ai_runner
  1. You may change pipeline configuration in this file:
nano headed_visual_ai.cfg

     

Figure 16. Headed Visual AI Configuration

 Refer to custom_models/README.md for more information

  1. To run pipeline:
./run_container.sh headed_visual_ai.cfg

VSaaS Gateway with Storage and AI Proxy Pipelines

VSaaS Gateway with Storage and AI Proxy Pipeline contains the following process:

  1. Decode a specified number of 1080p h264 video channels
  2. Stores video stream in local storage
  3. Runs AI inference on all video channel
  4. Sends inference metadata to a MQTT broker
  5. Transcodes the video channels to h265 for remote consumption
Figure 17. AI VSaaS Gateway Proxy Pipeline

 How To Run VSaaS Gateway with Storage and AI Proxy Pipeline

  1. Go to this directory
cd intel_edge_aibox/Intel_Edge_AI_Box_4.0/Proxy_Pipeline_Module/vsaas_storage_ai_runner
  1. You may change pipeline configuration in this file:
nano vsaas_storage_ai_cfg
Figure 18. AI VSaaS Gateway Configuration

 Refer to custom_models/README.md for more information

  1. To run pipeline:
./run_container.sh vsaas_storage_ai.cfg

Remote Device Management (RDM)

Note: Only available if you select Yes to install RDM feature during installation.

The Intel® In-Band Manageability Framework is software which enables an administrator to perform critical Device Management operations over-the-air remotely from the cloud. It also facilitates the publishing of telemetry and critical events and logs from an IoT device to the cloud enabling the administrator to take corrective actions if, and when necessary. The framework is designed to be modular and flexible ensuring scalability of the solution across preferred Cloud Service Providers.

AI Box support in-band remote device management with the ability to:
- Trigger reboot
- Update system APT packages
- Update container images from remote registries

Setting up RDM in Thingsboard

Setup Dashboards
  1. Open http://<thingsboard_server_ip>:8080 and login with tenant account
  2. Go to Dashboards -> Click ‘+’ and select ‘Import dashboard
Figure 19. RDM Dashboard Setup 1
  1. Click ‘Browse File’ and navigate to following path to add these 2 json files 1 at a time.
/usr/share/cloudadapter-agent/thingsboard/intel_manageability_devices_version_3.4.2.json
/usr/share/cloudadapter-agent/thingsboard/intel_manageability_batch_version_3.4.2.json
  1. You should be able to see these 2 dashboards:
Figure 20. RDM Dashboard Setup 2

Add Widgets

  1. Next is to add widgets. Go to Resources -> Widgets Library
Figure 21. RDM Dashboard Widget Library
  1. Navigate to following path to add json file
/usr/share/cloudadapter-agent/thingsboard/intel_manageability_widgets_version_3.4.2.json

Change Dashboard Profile

  1. Select Dashboard -> INB-Manageability Devices -> Edit Mode
Figure 22. RDM Dashboard - Edit Mode
  1. Entity Aliases --> Edit Aliases
Figure 23. RDM Dashboard - Edit Aliases 1
Figure 24. RDM Dashboard - Edit Aliases 2
  1. Add ‘Default’ in Device Profile and Save
Figure 25. RDM Dashboard - Edit Aliases 3
  1. Save Dashboard
Figure 26. RDM Dashboard – Save Dashboard
  1. Repeat the same steps for Dashboard ‘Intel Manageability Batch’

How To Use Thingsboard

Reboot Device
  1. Go to Dashboards -> INB-Intel Manageability Devices
  2. Click ‘Reboot Device’. The device will be reboot remotely
Figure 27. RDM Dashboard - Reboot
Update System APT Packages
  1. Go to Dashboards -> INB-Intel Manageability Devices
  2. Click ‘Trigger SOTA
  3. Command selects ‘update’. Click Send
Figure 28. RDM Dashboard – APT Update - Reboot
Add Container images from registries
  1. Go to Dashboards -> INB-Intel Manageability Device
  2. Click ‘Trigger AOTA
  3. App select ‘docker’ and Command select ‘pull
  4. For this example, container tag we enter ‘python’. Click Send
Figure 29. RDM Dashboard – Add docker images

*If you not able to see ‘Send’ button at the botton, zoom out the webpage.

  1. This action will pull python docker image and start container.

Telemetry of CPU and GPU Utilization

After installing the Intel® Edge AI Box, telemetry will automatically start during system power-up. You can visualize the CPU and GPU utilization through a dashboard.  

  1. From your web browser, go to https://127.0.0.1:3000. Log in with the default credentials:
    • Username: admin
    • Password: admin 
  2. For Xeon platform, Ubuntu Server OS does not have GUI. Follow steps below to launch the dashboard:
    • In another windows/linux system, open command prompt or terminal and run following:
ssh -L 3000:127.0.0.1:3000 <EAB_system_account>@<EAB_system_IP_address>

NOTE: If you see the Potential Security Risk Ahead warning, click Advanced and then Accept the Risk and Continue.

A screen showing potential security risk warning
Figure 30. Potential Security Risk Warning

 

  1. After logging in for the first time, when prompted, change your password.
  1. If you are connected to the device remotely, use the following command to forward port 3000:
    ssh -L 3000:127.0.0.1:3000 user@aibox_device_ip
  2. Then, log in to https://127.0.0.1:3000 from your local browser.
A screenshot showing log in screen of Telemetry Web GUI
Figure 31. Log Screen of Telemetry Web GUI

 

  1. After logging in, click the top left menu. Navigate to Home > Dashboards > General to see the available dashboards.
A screenshot showing the list of available dashboards
Figure 32. List of Available Dashboards

 

  1. Select Node Exporter Full to view the CPU and OS telemetries.
A screenshot showing CPU and OS Telemetries
Figure 33. CPU and OS Telemetries
  1. Select Intel XPU Manager Exporter from the dashboard to view the GPU telemetrics.

A screenshot showing GPU Telemetries
Figure 34. GPU Telemetries

Note: For Xeon platform, all benchmarks will run on CPU, so GPU telemetry will be "No data"
For Core Ultra platform, GPU telemetry will be "No data" due to a known issue

  1. To stop the telemetry, run the following command: 

    aibox_monitor.sh stop
  2. You can also start the telemetry manually: 

    aibox_monitor.sh start

    NOTE: The aibox_monitor.sh script is in $HOME/.local/bin/. You will be able to run the script without specifying the complete path.

 

Run Reference Application

NOTE: If you run reference applications over SSH sessions, set DISPLAY environment as following:                                                                                                                                                                                                                                                                                          export DISPLAY=:0       

This Reference Application is not supported on Xeon Platform, please use Core or Core Ultra Platform to run it.

Multi-channel Object Detection

This reference application is the containerized version of the demo of the same name available in the OpenVINO Model Zoo. The container is called multi_channel_object_detection:4.0

The following command runs the application using the default bottle-detection.mp4 video, duplicated four times:

export DEVICE=/dev/dri/renderD128
export DEVICE_GRP=$(ls -g $DEVICE | awk '{print $3}' | \
xargs getent group | awk -F: '{print $3}')

docker run -it \
--device /dev/dri --group-add=$DEVICE_GRP  \
-v /tmp/.X11-unix:/tmp/.X11-unix \
-v /run/user/1000/wayland-0:/tmp/wayland-0 \
multi_channel_object_detection:4.0 \
-i /home/aibox/bottle-detection.mp4 \
-m /home/aibox/share/models/public/yolo-v3-tf/FP16/yolo-v3-tf.xml \
-duplicate_num 4 -d GPU

To run in dGPU, change to GPU.1 and change export DEVICE to DEVICE=/dev/dri/renderD129.

Things to Try

Replace -i /home/aibox/bottle-detection.mp4 with -i rtsp://camera_url to connect to an IP camera.

Specify multiple camera sources by -i rtsp://camera1_url,rtsp://camera2_url,....

Use VA Base Library Containers

The core of Intel® Edge AI Box reference architecture is a suite of containers. The VA Base Library container includes OpenVINO™, oneDNN, and GPU drivers for accelerated inferencing and media processing.

These containers are available on the platform for your projects, and the corresponding Dockfiles are in /opt/intel/base_container/dockerfile.

REPOSITORY TAG IMAGE ID CREATED SIZE
aibox-opencv-ffmpeg 4.0 236118f0ca6d 4 days ago 4.95 GB
aibox-dlstreamer 4.0 a397a9786ee9 4 days ago 15 GB
aibox-base-devel 4.0 f6b0ee7ecddd 4 days ago 13 GB
aibox-base 4.0 724fd64df066 4 days ago 1.98 GB

 

Sample Usage 1 - Multi-Channel Object Detection Using YOLOv3

The following docker file is used to build the multi-channel object detection in the Reference Implementation package. The file uses aibox-base-devel:4.0 to download models from the OpenVINO model zoo and aibox-opencv-ffmpeg:4.0 to compile the demo.

For more details, install the Intel® Edge AI Box package and go through the intel_edge_aibox/Intel_Edge_AI_Box_4.0/Intel_Edge_AIBox_Samples/multi_channel_object_detection_yolov3 directory.

#============================================================================
# Copyright (C) 2022 Intel Corporation
#
# SPDX-License-Identifier: MIT
#============================================================================

FROM aibox-base-devel:4.0 as builder
ARG https_proxy
ARG http_proxy
ARG no_proxy

RUN mkdir -p $HOME/share/models/
RUN omz_downloader -o $HOME/share/models --name yolo-v3-tf
RUN omz_converter -o $HOME/share/models -d $HOME/share/models --name yolo-v3-tf


FROM aibox-opencv-ffmpeg:4.0
ARG https_proxy
ARG http_proxy
ARG no_proxy

USER root
RUN apt update -y && apt -y install git

USER aibox
ENV HOME=/home/aibox
WORKDIR /home/aibox

RUN git clone --branch releases/2023/3 --depth 1 --recurse-submodules https://github.com/openvinotoolkit/open_model_zoo.git /home/aibox/open_model_zoo
COPY --chown=aibox:aibox build_omz_demos.sh .
RUN chmod 755 /home/aibox/build_omz_demos.sh
RUN bash build_omz_demos.sh

COPY --from=builder /home/aibox/share /home/aibox/share

RUN curl -L -o bottle-detection.mp4 https://storage.openvinotoolkit.org/test_data/videos/bottle-detection.mp4

WORKDIR /home/aibox/omz_demos_build/intel64/Release
ENV XDG_RUNTIME_DIR=/tmp
ENV WAYLAND_DISPLAY=wayland-0

ENTRYPOINT ["/home/aibox/omz_demos_build/intel64/Release/multi_channel_object_detection_demo_yolov3"]
CMD ["-h"]

Sample Usage 2 – Video Decode and Tiled Display

The following sample shows the use of the aibox-dlstreamer: 4.0 container for streaming and media processing. This sample decodes 16 video files using VAAPI and composes them for display on a 4K resolution (3840x2160) monitor. 

Note: This sample is not supported on Xeon platform, please use Core or Core Ultra platform to run it.

  1. You can use your video or download a sample video from https://vimeo.com/644498079. The following sample uses Big_Buck_Bunny.mp4 from the /videos directory. 
  2. In your home directory, create a file decode.sh and add the following code:
#!/bin/bash
source /opt/intel/openvino_2024.0.0/setupvars.sh
source /opt/intel/dlstreamer/setupvars.sh
source /opt/intel/oneapi/setvars.sh

VIDEO_IN=videos/Big_Buck_Bunny.mp4
gst-launch-1.0 vaapioverlay name=comp0 sink_1::xpos=0 sink_1::ypos=0 sink_1::alpha=1 sink_2::xpos=960 sink_2::ypos=0 sink_2::alpha=1 sink_3::xpos=1920 sink_3::ypos=0 sink_3::alpha=1 sink_4::xpos=2880 sink_4::ypos=0 sink_4::alpha=1 sink_5::xpos=0 sink_5::ypos=540 sink_5::alpha=1 sink_6::xpos=960 sink_6::ypos=540 sink_6::alpha=1 sink_7::xpos=1920 sink_7::ypos=540 sink_7::alpha=1 sink_8::xpos=2880 sink_8::ypos=540 sink_8::alpha=1 sink_9::xpos=0 sink_9::ypos=1080 sink_9::alpha=1 sink_10::xpos=960 sink_10::ypos=1080 sink_10::alpha=1 sink_11::xpos=1920 sink_11::ypos=1080 sink_11::alpha=1 sink_12::xpos=2880 sink_12::ypos=1080 sink_12::alpha=1 sink_13::xpos=0 sink_13::ypos=1620 sink_13::alpha=1 sink_14::xpos=960 sink_14::ypos=1620 sink_14::alpha=1 sink_15::xpos=1920 sink_15::ypos=1620 sink_15::alpha=1 sink_16::xpos=2880 sink_16::ypos=1620 sink_16::alpha=1 ! vaapipostproc ! xvimagesink display=:0 sync=false \
\
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_1 \
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_2 \
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_3 \
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_4 \
\
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_5 \
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_6 \
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_7 \
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_8 \
\
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_9 \
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_10 \
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_11 \
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_12 \
\
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_13 \
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_14 \
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_15 \
filesrc location=${VIDEO_IN} ! qtdemux ! vaapih264dec ! gvafpscounter ! vaapipostproc scale-method=fast ! video/x-raw,width=960,height=540 ! comp0.sink_16
  1. Assign execute permission to the file:
    chmod 755 decode.sh
  2. Execute the decode script:
    export DEVICE=/dev/dri/renderD128
    export DEVICE_GRP=$(ls -g $DEVICE | awk '{print $3}' | xargs getent group | awk -F: '{print $3}')
    docker run -it --rm --net=host \
    -e no_proxy=$no_proxy -e https_proxy=$https_proxy \
    -e socks_proxy=$socks_proxy -e http_proxy=$http_proxy \
    -v /tmp/.X11-unix \
    --device /dev/dri --group-add ${DEVICE_GRP} \
    -e DISPLAY=$DISPLAY \
    -v $HOME/.Xauthority:/home/aibox//.Xauthority:ro \
    -v $HOME/Videos:/home/aibox/Videos:ro \
    -v $HOME/decode.sh:/home/aibox/decode.sh:ro \
    aibox-dlstreamer:4.0 /home/aibox/decode.sh
    

Here is the result:

A screenshot showing 4x4 video streaming
Figure 35. 4x4 Video Streaming

Sample Usage 3 – OpenVINO Model Benchmark 

The following sample shows how to download a model, convert the OpenVINO model, and run a benchmark using the aibox-base-devel: 4.0 container.  

  1. In the home directory, create a file openvino_benchmark.sh and add the following code. You may change the model name and target device. In this sample, the model is resnet-50-tf, and the target device is GPU.
#!/bin/bash
MODEL=resnet-50-tf
DEVICE=GPU

omz_downloader --name ${MODEL}
omz_converter --name ${MODEL}

echo "====================================="
echo "Running Benchmark FP32 with ${DEVICE}"
echo "====================================="
cd /home/aibox/public/${MODEL}/FP32
benchmark_app -m ${MODEL}.xml -d ${DEVICE}

echo "====================================="
echo "Running Benchmark FP16 with ${DEVICE}"
echo "====================================="
cd /home/aibox/public/${MODEL}/FP16
benchmark_app -m ${MODEL}.xml -d ${DEVICE}
  1. Assign execute permission to the file:
    chmod 755 openvino_benchmark.sh
  2. Run the following docker command:
    docker run -it --rm --net=host \
    -e no_proxy=$no_proxy -e https_proxy=$https_proxy \
    -e socks_proxy=$socks_proxy -e http_proxy=$http_proxy \
    -v /tmp/.X11-unix --device /dev/dri --group-add 110 \
    -e DISPLAY=:0 \
    -v $HOME/.Xauthority:/home/aibox//.Xauthority:ro \
    -v $HOME/openvino_benchmark.sh:/home/aibox/openvino_benchmark.sh:ro \
    aibox-base-devel:4.0/home/aibox/openvino_benchmark.sh
    

Here is the sample output:

Figure 36. Example Benchmark Output

Known Issues

[AIBOX-293]: Report.zip is empty when first time to generate the report
[AIBOX-420]: Permission Issues Found During Installation with new user ID
[AIBOX-623]: No Data in Intel XPU Manager Exporter on MTL Platform
[AIBOX-624]: MODULE_STDERR Occurred During EAB4.0 Installation. Workarounds refer to Troubleshooting. Workarounds refer to Run The Application in Intel® Edge Software Device Qualification for Intel® Edge AI Box.
[AIBOX-629]: GPU Telemetry data is Too Low in ESDQ Report of MTL Platform.


Troubleshooting

  1. As the system is being automatically updated in the background, you may encounter the following errors:
A screenshot showing an error message
Figure 37. Example Error

 

A screenshot showing an error message
Figure 38. Example Error

To resolve this issue, reboot and manually update the system. Then, rerun the installation:

sudo reboot
sudo apt update
./edgesoftware install
  1. If the network quality is poor, you may encounter the following error:
Figure 39. Example Error

To resolve this issue, you can configure to run sudo without entering a password before installing:

sudo visudo

Navigate to “# User privilege specification” section and add the following line to grant the current user sudo privileges without requiring a password:

USER ALL=(ALL) NOPASSWD: ALL

Please replace USER with your actual username.
After adding the line, save and exit, then reinstall the AIBox package.

If you’re unable to resolve your issues, contact the Support Forum.


Release Notes

Current Version: 4.0

  • Added INT8 models in OpenVINO benchmark
  • Added NPU support in OpenVINO benchmark
  • Replaced Video Decode Performance Benchmark with Media Performance Benchmarks
  • Replaced Video Analytic Pipeline Benchmark with domain specific proxy pipeline benchmark
  • Added Proxy Pipelines
  • Added telemetry information in benchmark report
  • Added graph results in benchmark report
  • Added Intel Core Ultra and SPR-SP platform support
  • Added Remote Device Management
  • Updated OpenVINO to 2024.1.0 (for Base containers), 2024.0.0(for DLStreamer container), and 2023.3.0(for OpenCV/FFMpeg container)
  • Updated FFMpeg to 2023q3
  • Updated DLStreamer to 2024.0.1

Version: 3.1.1

  • Fixed h264 file size issue that causing opencv-ffmpeg BKC validator failed
  • Fixed Python binding missing for OpenCV
  • Fixed sycl_meta_overlay plugin not found when using aibox_dlstreamer container as base image
  • Added locale encoding to fix bmra_base ansible-playbook
  • Fixed ESDQ hung error when prompt for user input for "parted"
  • Fixed installation failed on headless devices
  • Fixed ESDQ Test module error with "-h"
  • Fixed ESDQ Test module report generation error

Version: 3.1

  • Updated to OpenVINO 2023.1 and 2023.0 (only available in the dlstreamer container)
  • Updated to DLStreamer 2023.0
  • Updated to OpenCV 4.8.0
  • Updated to FFMPEG cartwheel 2023q2
  • Updated to GPU driver stack 20230912
  • Added CPU and GPU telemetry and observability