- Intel® Cluster Checker must be accessible by the same path on all nodes.
- A readable, writable shared directory must be available from the same path on all nodes for temporary file creation.
- as the shared directory is used by default, but you can change this option by setting the environment variable$HOMEto the shared directory.$CLCK_SHARED_TEMP_DIR
- For admin privileged users, such as root, the environment variablemust be explicitly set.$CLCK_SHARED_TEMP_DIR
- Determine if passwordless ssh access to all nodes is set up. (e.g. test if the commandssh<nodename>responds with a valid hostname, while not asking for ‘Password:’)hostname
- If passwordless ssh to all nodes is available - go ahead with Environment Setup and Running using Slurm below. By default Intel® Cluster Checker is configured to use passwordless ssh (through the command) to launch remotely on nodes of the cluster. Note: you may need to add enabling passwordless access in your local ssh configuration setup.pdsh
- If passwordless ssh is not available, use these steps to configure Intel® Cluster Checker to instead use the commandfrom the Intel® MPI Library to launch remotely on nodes:mpirun
- requires Intel MPI Library to be set up, and either
- locate and copy the <installdir>/clck/<version>/etc/clck.xml file locally and uncomment theby removing the commenting statements<extension>mpi.so</extension>before it and the<!--after it. Then add the following option when running the-->command:clck
- or locate and edit the <installdir>/clck/<version>/etc/clck.xml file to uncomment the line containing, which changes the default to use mpirun instead of pdsh/ssh for remotely launch.<extension>mpi.so</extension>
- To use scripts, follow these steps to setup the shell environment. (By default these scripts are found with the packages they are installed with; default install location is /opt/intel/<package-name>/bin/ or <installdir>/<package-name>/bin/ )
- If using the Intel® oneAPI HPC Toolkitsourcefrom oneAPI HPC Toolkit by default this would besetvars.shsourceand will analyze all software installed from oneAPI and add it to your path./opt/intel/oneapi/setvars.shIf you rather individually choose specific software packages you can still do so, i.e.:source/opt/intel/oneapi/clck/<version>/bin/clckvars.sh
- or if you are using individual package versions
- or from Intel® Parallel Studio XE Cluster Edition including all above components
- An alternative to these scripts is ‘modulefiles’ to setup your runtime environment.
- Alternatively the Intel® Cluster Checker modulefile is available using the module commandsmoduleuse<installdir>/clck/<version>/modulefilesmoduleloadclck
node1 node2 node3 node4
source /opt/intel/oneapi/setvars.sh clck
source mpivars.[sh | csh] source mklvars.[sh | csh] source compilervars.[sh | csh] source clckvars.[sh | csh] # alternatively use psxevars.[sh | csh] or setvars.sh (Intel oneAPI), or modulefiles to setup environment clck
health_extended_admin |-- health_admin | |-- health_base | | |-- cpu_user | | |-- environment_variables_uniformity | | |-- ethernet | | |-- infiniband_user | | |-- network_time_uniformity | | |-- node_process_status | | `-- opa_user | |-- basic_shells | |-- cpu_admin | |-- dgemm_cpu_performance | |-- mpi_bios | |-- infiniband_admin | |-- kernel_version_uniformity | |-- local_disk_storage | |-- memory_uniformity_admin | |-- mpi_libfabric | |-- opa_admin | |-- perl_functionality | |-- privileged_user | |-- python_functionality | |-- rpm_uniformity | |-- services_status | `-- stream_memory_bandwidth_performance |-- kernel_parameter_uniformity `-- lshw_hardware_uniformity
health_extended_user |-- health_user | |-- health_base | | |-- cpu_user | | |-- environment_variables_uniformity | | |-- ethernet | | |-- infiniband_user | | |-- network_time_uniformity | | |-- node_process_status | | `-- opa_user | |-- basic_internode_connectivity | |-- basic_shells | |-- file_system_uniformity | |-- imb_pingpong_fabric_performance | |-- kernel_version_uniformity | |-- memory_uniformity_user | |-- mpi_local_functionality | |-- mpi_multinode_functionality | |-- perl_functionality | |-- python_functionality | |-- sgemm_cpu_performance | `-- stream_memory_bandwidth_performance |-- dgemm_cpu_performance `-- hpl_cluster_performance