Customize
Offload Modeling Perspective
Offload Modeling
PerspectiveCustomize the perspective flow to better fit your goal and your application.
If you change any of the analysis settings from the
Analysis Workflow
tab, the accuracy level changes to
Custom
automatically. With this accuracy level, you can customize the perspective flow and/or analysis properties.
To change the properties of a specific analysis:
- Expand the analysis details on theAnalysis Workflowpane with
.
- Select desired settings.
- For more detailed customization, click thegear
icon. You will see the
Project Propertiesdialog box open for the selected analysis. - Select desired properties and clickOK.
For a full set of available properties, click the
icon on the left-side pane or go to
.

The following tables cover project properties applicable to the analyses in the
Offload Modeling
perspective.
Common Properties
Use This
| To Do This
|
---|---|
Inherit settings from Visual Studio project checkbox and field (Visual Studio* IDE only)
| Inherit
Intel Advisor project properties from the Visual Studio* startup project (enable).
If enabled, the
Application ,
Application parameters , and
Working directory fields are pre-filled and cannot be modified.
|
Application field and
Browse... button
| Select an analysis target executable or script.
If you specify a script in this field, consider specifying the executable in the
field (required for Dependencies analysis).
|
Application parameters field and
Modify... button
| Specify runtime arguments to use when performing analysis (equivalent to command line arguments).
|
Use application directory as working directory checkbox
| Automatically use the value in the
Application directory to pre-fill the
Working directory value (enable).
|
Working directory field and
Browse... button
| Select the working directory.
|
User-defined environment variables field and
Modify... button
| Specify environment variables to use during analysis.
|
Managed code profiling mode drop-down
|
|
Child application field
| Analyze a file that is not the starting application. For example: Analyze an executable (identified in this field) called by a script (identified in the
Application field).
Invoking these properties could decrease analysis overhead.
For the
Dependencies Analysis Type : If you specify a script file in the
Application field, you must specify the target executable in the
Child application field.
|
Modules radio buttons, field, and
Modify... button
|
Including/excluding modules could minimize analysis overhead.
|
Use MPI launcher checkbox
| Generate a command line (enable) that appears in the
Get command line field based on the following parameters:
|
Automatically stop collection after (sec) checkbox and field
| Stop collection after a specified number of seconds (enable and specify seconds).
Invoking this property could minimize analysis overhead.
|
Survey Analysis Properties
Use This
| To Do This
|
---|---|
Automatically resume collection after (sec) checkbox and field
| Start running your target application with collection paused, then resume collection after a specified number of seconds (enable and specify seconds).
Invoking this property could decrease analysis overhead.
The corresponding CLI action option is
--resume-after=<integer> , where the integer argument is in milliseconds, not seconds.
|
Sampling Interval selector
| Set the wait time between each analysis collection CPU sample while your target application is running.
Increasing the wait time could decrease analysis overhead.
|
Collection data limit, MB selector
| Set the amount of collected raw data if exceeding a size threshold could cause issues. Not available for hardware event-based analyses.
Decreasing the limit could decrease analysis overhead.
|
Callstack unwinding mode drop-down list
| Set to
After collection if:
Otherwise, set to
During Collection . This mode improves stack accuracy but increases overhead.
|
Stitch stacks checkbox
| Restore a logical call tree for
Intel® oneAPI Threading Building Blocks (oneTBB ) or OpenMP* applications by catching notifications from the runtime and attaching stacks to a point introducing a parallel workload (enable).
Disable if Survey analysis runtime overhead exceeds 1.1x.
|
Analyze MKL Loops and Functions checkbox
| Show
Intel® oneAPI Math Kernel Library (oneMKL ) loops and functions in
Intel Advisor reports (enable).
Enabling could increase analysis overhead.
|
Analyze Python loops and functions checkbox
| Show Python* loops and functions in
Intel Advisor reports (enable).
Enabling could increase analysis overhead.
|
Analyze loops that reside in non-executed code paths checkbox
| Collect a variety of data during analysis for loops that reside in non-executed code paths, including loop assembly code, instruction set architecture (ISA), and vector length (enable).
Enabling could increase analysis overhead.
Analyzing non-executed code paths in binaries that target multiple ISAs (contain multiple code paths) is available only for binaries compiled using the
-ax (Linux* OS) /
Qax (Windows* OS) option with an Intel compiler.
|
Enable registry spill/fill analysis checkbox
| Calculate the number of consecutive load/store operations in registers and related memory traffic (enable).
Enabling could increase analysis overhead.
|
Enable static instruction mix analysis checkbox
| Statically calculate the number of specific instructions present in the binary (enable).
Enabling could increase analysis overhead.
|
Source caching drop-down list
|
|
Trip Counts and FLOP Analysis Properties
Use This
| To Do This
|
---|---|
Inherit settings from the Survey Hotspots Analysis Type checkbox
| Copy similar settings from Survey analysis properties (enable).
When enabled, this option disables application parameters controls.
|
Automatically resume collection after (sec) checkbox and field
| Start running your target application with collection paused, then resume collection after a specified number of seconds (enable and specify seconds).
Invoking this property could decrease analysis overhead.
The corresponding CLI action option is
--resume-after=<integer> , where the integer argument is in milliseconds, not seconds.
|
Trip Counts /
Collect information about Loop Trip Counts checkbox
| Measure loop invocation and execution (enable).
|
FLOP /
Collect information about FLOP, L1 memory traffic, and AVX-512 mask usage checkbox
| Measure floating-point operations, integer operations, and memory traffic (enable).
|
Callstacks /
Collect callstacks checkbox
| Collect call stack information when performing analysis (enable).
Enabling could increase analysis overhead.
|
Capture metrics for dynamic loops and functions checkbox
| Collect metrics for dynamic Just-In-Time (JIT) generated code regions.
|
Capture metrics for stripped binaries checkbox
| Collect metrics for stripped binaries.
Enabling could increase analysis overhead.
|
Cache Simulation /
Enable Memory-Level Roofline with cache simulation checkbox
| Model multiple levels of cache for data, such as counts of loaded or stored bytes for each loop, to plot the Roofline chart for all memory levels (enable).
Enabling could increase analysis overhead.
|
Cache simulator configuration field
| Specify a cache hierarchy configuration to model (enable and specify hierarchy).
The hierarchy configuration template is:
[num_of_level1_caches]:[num_of_ways_level1_connected]:[level1_cache_size]:[level1_cacheline_size]/
[num_of_level2_caches]:[num_of_ways_level2_connected]:[level2_cache_size]:[level2_cacheline_size]/
[num_of_level3_caches]:[num_of_ways_level3_connected]:[level3_cache_size]:[level3_cacheline_size]
For example:
4:8w:32k:64l/4:4w:256k:64l/1:16w:6m:64l
is the hierarchy configuration for:
|
Data Transfer Simulation /
Data transfer simulation mode drop-down
| Select a level of details for data transfer simulation:
|
Dependencies Analysis Properties
Use This
| To Do This
|
---|---|
Inherit settings from the Survey Hotspots Analysis Type checkbox
| Copy similar settings from Survey analysis properties (enable).
When enabled, this option disables application parameters controls.
|
Suppression mode radio buttons
|
|
Loop call count limit selector
| Choose the maximum number of instances each marked loop is analyzed. 0 = analyze all loop instances.
Supplying a non-zero value could decrease analysis overhead.
|
Instance of interest selector
| Analyze the nth child process, where 1 = the first process of the specified name in the application process tree. 0 = analyze all processes.
Supplying a non-zero value could decrease analysis overhead.
|
Analyze stack variables checkbox
| Analyze parallel data sharing for stack variables (enable).
Enabling could increase analysis overhead.
|
Filter stack variables by scope checkbox
| Enable to report:
Enabling could increase analysis overhead.
|
Reduction Detection /
Filter reduction variables checkbox
| Mark all potential reductions by a specific diagnostic (enable).
Enabling could increase analysis overhead.
|
Markup type checkbox
| Select loops/functions by pre-defined markup algorithm. Supported algorithms are:
This option is available only from the
Analysis Workflow pane for the Offload Modeling perspective.
|
Performance Modeling Properties
Use This
| To Do This
|
---|---|
Assume Dependencies checkbox
| Assume that loops have dependencies if their dependency type is unknown.
This option is available only from the
Analysis Workflow pane.
|
Single Kernel Launch Tax checkbox
| Assume the invocation tax is paid only for the first kernel launch when estimating invocation taxes.
This option is available only from the
Analysis Workflow pane.
|
Data Reuse Analysis checkbox
| Analyze potential data reuse between code regions for the data transferred between host and target platforms.
This option is available only from the
Analysis Workflow pane.
Make sure to set the
Data Transfer Analysis to the
Full mode in the
Characterization step to analyze data reuse.
|
Target Config drop-down
| Select a pre-defined hardware configurations from a drop-down list to model application performance on.
|
Other parameters
field
| Enter a space-separated list of command-line parameters. See
Command Option Reference for available parameters.
|
Baseline Device drop-down
| Select a baseline device that your application runs on for the Intel® Advisor to collect performance data.
|
Custom Device Configuration field
| Specify the absolute path or name for a custom TOML configuration file with additional modeling parameters.
|