Reference

CUDA

Applications that need to perform massively parallel operations, like processing large arrays, may use the CUDA framework to accelerate their processing on graphics processing units (GPUs). The CUDA framework has two complementary pieces to it.

There are a set of GPU instruction set extensions that are implemented by various graphics cards. These instruction set extensions are known as the CUDA Compute Capability.

The second part of the framework is an SDK that allows developers to take advantage of the hardware’s instruction set extensions of a particular version (a specific CUDA Compute Capability version, that is).

An application will link with a version of the CUDA SDK, and the version of the CUDA SDK controls which CUDA Compute Capability versions the application will be able to work with.

The os_traits.hw.gpu.cuda module contains traits for both the CUDA compute capability version as well as the CUDA SDK version. For example, os_traits.hw.gpu.cuda.COMPUTE_CAPABILITY_V3_2 and os_traits.hw.gpu.cuda.SDK_V6_5.

The os_traits.hw.gpu.cuda module contains a utility function called compute_capabilities_supported() that accepts a trait indicating the CUDA SDK version and returns a set() containing the matching CUDA compute capability traits that that version of the CUDA SDK knows how to utilize.

Here is an example of listing the CUDA compute capability version traits that the CUDA SDK 8.0 is capable of working with:

>>> from os_traits.hw.gpu import cuda
>>> import pprint
>>>
>>> sdk8_caps = cuda.compute_capabilities_supported(cuda.SDK_V8_0)
>>> pprint.pprint(sdk8_caps)
set(['HW_GPU_CUDA_COMPUTE_CAPABILITY_V2_0',
     'HW_GPU_CUDA_COMPUTE_CAPABILITY_V2_1',
     'HW_GPU_CUDA_COMPUTE_CAPABILITY_V3_0',
     'HW_GPU_CUDA_COMPUTE_CAPABILITY_V3_2',
     'HW_GPU_CUDA_COMPUTE_CAPABILITY_V3_5',
     'HW_GPU_CUDA_COMPUTE_CAPABILITY_V3_7',
     'HW_GPU_CUDA_COMPUTE_CAPABILITY_V5_0',
     'HW_GPU_CUDA_COMPUTE_CAPABILITY_V5_2',
     'HW_GPU_CUDA_COMPUTE_CAPABILITY_V5_3',
     'HW_GPU_CUDA_COMPUTE_CAPABILITY_V6_0',
     'HW_GPU_CUDA_COMPUTE_CAPABILITY_V6_1',
     'HW_GPU_CUDA_COMPUTE_CAPABILITY_V6_2'])

For more information on CUDA, see the Wikipedia article.

AMD SEV

While data is typically encrypted today when stored on disk, it is stored in DRAM in the clear. This can leave the data vulnerable to snooping by unauthorized administrators or software, or by hardware probing. New non-volatile memory technology (NVDIMM) exacerbates this problem since an NVDIMM chip can be physically removed from a system with the data intact, similar to a hard drive. Without encryption any stored information such as sensitive data, passwords, or secret keys can be easily compromised.

AMD’s SEV (Secure Encrypted Virtualization) is a VM protection technology which transparently encrypts the memory of each VM with a unique key. It can also calculate a signature of the memory contents, which can be sent to the VM’s owner as an attestation that the memory was encrypted correctly by the firmware. SEV is particularly applicable to cloud computing since it can reduce the amount of trust VMs need to place in the hypervisor and administrator of their host system.

The os_traits.hw.cpu.amd.SEV trait is reserved in order to indicate that a compute host contains support for SEV not only on-CPU, but also in all other layers of the hypervisor stack required in order to take advantage of this feature.