Intel® oneAPI Deep Neural Network Developer Guide and Reference
A newer version of this document is available. Customers should click here to go to the newest version.
Constant Tensor Cache
Overview
A set of functions that provide constant tensor cache control. More…
// global functions
void dnnl::graph::set_constant_tensor_cache(int flag);
int dnnl::graph::get_constant_tensor_cache();
dnnl_status_t DNNL_API dnnl_graph_set_constant_tensor_cache(int flag);
dnnl_status_t DNNL_API dnnl_graph_get_constant_tensor_cache(int* flag);Detailed Documentation
A set of functions that provide constant tensor cache control.
Global Functions
void dnnl::graph::set_constant_tensor_cache(int flag)Control the enabling or disabling of constant tensor cache.
This API must be called once before compilation stage. By default, constant tensor cache is disabled in the library.
Parameters:
| flag | Set to positive value to enable the cache and set to 0 to disable the cache. Negative values are invalid. | 
int dnnl::graph::get_constant_tensor_cache()Return the enabling status of constant tensor cache.
dnnl_status_t DNNL_API dnnl_graph_set_constant_tensor_cache(int flag)Control the enabling or disabling of constant tensor cache.
This API must be called once before compilation stage. By default, constant tensor cache is disabled in the library.
Parameters:
| flag | Set to positive value to enable the cache and set to 0 to disable the cache. Negative values are invalid. | 
Returns:
dnnl_invalid_arguments if the flag value is invalid, and dnnl_success on success.
dnnl_status_t DNNL_API dnnl_graph_get_constant_tensor_cache(int* flag)Return the enabling or disabling status of constant tensor cache.
Parameters:
| flag | The constant tensor cache enabling status to query. | 
Returns:
dnnl_invalid_arguments if the flag value is nullptr, and dnnl_success on success.