2
0
mirror of https://github.com/xcat2/xcat-core.git synced 2025-08-01 00:57:37 +00:00

Remove trailing spaces in file docs/source/advanced/gpu/nvidia/management.rst

This commit is contained in:
GONG Jie
2017-12-31 23:59:59 +00:00
parent 7d7476aebf
commit 2f9534902d

View File

@@ -1,7 +1,7 @@
GPU Management and Monitoring
=============================
The ``nvidia-smi`` command provided by NVIDIA can be used to manage and monitor GPUs enabled Compute Nodes. In conjunction with the xCAT``xdsh`` command, you can easily manage and monitor the entire set of GPU enabled Compute Nodes remotely from the Management Node.
The ``nvidia-smi`` command provided by NVIDIA can be used to manage and monitor GPUs enabled Compute Nodes. In conjunction with the xCAT``xdsh`` command, you can easily manage and monitor the entire set of GPU enabled Compute Nodes remotely from the Management Node.
Example: ::
@@ -14,7 +14,7 @@ Example: ::
Management
----------
Some useful ``nvidia-smi`` example commands for management.
Some useful ``nvidia-smi`` example commands for management.
* Set persistence mode, When persistence mode is enabled the NVIDIA driver remains loaded even when no active clients, DISABLED by default::
@@ -29,11 +29,11 @@ Some useful ``nvidia-smi`` example commands for management.
nvidia-smi -i 0 -p 0/1
* Set MODE for compute applications, query with --query-gpu=compute_mode::
* Set MODE for compute applications, query with --query-gpu=compute_mode::
nvidia-smi -i 0 -c 0/1/2/3
* Trigger reset of the GPU ::
* Trigger reset of the GPU ::
nvidia-smi -i 0 -r
@@ -48,7 +48,7 @@ Some useful ``nvidia-smi`` example commands for management.
Monitoring
----------
Some useful ``nvidia-smi`` example commands for monitoring.
Some useful ``nvidia-smi`` example commands for monitoring.
* The number of NVIDIA GPUs in the system ::
@@ -75,7 +75,7 @@ Some useful ``nvidia-smi`` example commands for monitoring.
nvidia-smi -i 0 --query-gpu=compute_mode --format=csv,noheader
* Percent of time over the past sample period during which one or more kernels was executing on the GPU::
nvidia-smi -i 0 --query-gpu=utilization.gpu --format=csv,noheader
* Total errors detected across entire chip. Sum of device_memory, register_file, l1_cache, l2_cache and texture_memory ::
@@ -86,14 +86,14 @@ Some useful ``nvidia-smi`` example commands for monitoring.
nvidia-smi -i 0 --query-gpu=temperature.gpu --format=csv,noheader
* The ECC mode that the GPU is currently operating under::
* The ECC mode that the GPU is currently operating under::
nvidia-smi -i 0 --query-gpu=ecc.mode.current --format=csv,noheader
* The power management status::
nvidia-smi -i 0 --query-gpu=power.management --format=csv,noheader
* The last measured power draw for the entire board, in watts::
nvidia-smi -i 0 --query-gpu=power.draw --format=csv,noheader