Compute Shapes
A shape is a template that determines the type and amount of resources that are allocated to a compute instance. Compute Cloud@Customer offers a choice between a flexible shape for generic workloads, and dedicated shapes for GPU-accelerated workloads.
You choose the shape configuration when you create an instance. See Creating an Instance.
You can use the following shapes for instances that are created on Compute Cloud@Customer:
- Flexible Shapes: You choose the number of OCPUs and amount of memory that are allocated to an instance.
- GPU Shapes: Each shape has a fixed number of OCPUs, memory, and GPUs that are allocated to an instance when the instance is created.
Flexible Shapes
A flexible shape lets you customize the number of OCPUs and the amount of memory for an instance. When you update an instance, you can change these properties of the flexible shape. This flexibility lets you create instances that meet your workload requirements, while optimizing performance and using resources efficiently.
The VM.PCAStandard.E5.Flex shape can be selected only for Compute Cloud@Customer X10 systems.
The VM.PCAStandard.E6.Flex shape can be selected only for Compute Cloud@Customer X11 systems.
Shape |
OCPUs |
Memory (GB) |
Maximum VNICs |
Maximum Bandwidth (Gbps) |
---|---|---|---|---|
VM.PCAStandard1.Flex |
1–32 |
64 GB maximum per OCPU 512 GB maximum per instance |
1 OCPU: 2 VNICs 2 to 24 OCPUs: 1 VNIC per OCPU 25 to 32 OCPUs: 24 VNICs |
1 to 24 OCPUs: 24.6 Gbps 25 to 32 OCPUs: 1 Gbps per OCPU |
VM.PCAStandard.E5.Flex |
1–96 |
64 GB maximum per OCPU 960 GB maximum per instance |
1 OCPU: 2 VNICs 2 to 24 OCPUs: 1 VNIC per OCPU 25 to 96 OCPUs: 24 VNICs |
1 to 24 OCPUs: 24.6 Gbps 25 to 40 OCPUs: 1 Gbps per OCPU 41 to 96 OCPUs: 40.0 Gbps |
VM.PCAStandard.E6.Flex |
1–96 |
64 GB maximum per OCPU 960 GB maximum per instance |
1 OCPU: 2 VNICs 2 to 24 OCPUs: 1 VNIC per OCPU 25 to 96 OCPUs: 24 VNICs |
1 to 24 OCPUs: 24.6 Gbps 25 to 40 OCPUs: 1 Gbps per OCPU 41 to 96 OCPUs: 40.0 Gbps |
GPU Shapes
The GPU VM shapes are optimized for enterprise GPU-accelerated workloads. They can only be used if the Compute Cloud@Customer deployment includes a GPU expansion rack. Instances created with a GPU shape have direct (passthrough) access to 1-4 physical GPUs. The ratio between GPUs, OCPUs and memory is fixed.
For GPU-accelerated workloads, you have a choice between these shapes: VM.GPU.L40S.1, VM.GPU.L40S.2,VM.GPU.L40S.3, VM.GPU.L40S.4. To access these dedicated shapes, you must create an instance based on the Oracle Linux 8 or Oracle Linux 9 platform image.
No GPU drivers are included in the current Oracle Linux platform images. The instance OS detects the allocated GPUs, but to use them, you need the CUDA Toolkit from the NVIDIA developer site to install the required drivers.
The large download and local repository installation need a large amount of disk space. The default 50GB boot volume is insufficient on Oracle Linux 9 and only just large enough on Oracle Linux 8. It is highly recommended to increase the boot volume size to at least 60GB, and extend the file system accordingly.
-
From the command line of the instance, download and install the CUDA Toolkit rpm for your OS.
$ wget https://developer.download.nvidia.com/compute/cuda/12.8.0/local_installers/cuda-repo-rhel9-12-8-local-12.8.0_570.86.10-1.x86_64.rpm $ sudo rpm -i cuda-repo-rhel9-12-8-local-12.8.0_570.86.10-1.x86_64.rpm $ sudo dnf clean all $ sudo dnf install cuda-toolkit-12-8
-
Enable the Oracle Linux 9 EPEL yum repository. Install the
dkms
package.$ sudo yum-config-manager --enable ol9_developer_EPEL $ sudo dnf install dkms
-
Install the GPU drivers.
$ sudo dnf install cuda-12-8
-
Verify the installation with the NVIDIA System Management Interface.
$ nvidia-smi +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 570.86.10 Driver Version: 570.86.10 CUDA Version: 12.8 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA L40S Off | 00000000:00:05.0 Off | 0 | | N/A 26C P8 23W / 350W | 1MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | No running processes found | +-----------------------------------------------------------------------------------------+
-
From the command line of the instance, download and install the CUDA Toolkit rpm for your OS.
$ wget https://developer.download.nvidia.com/compute/cuda/12.8.0/local_installers/cuda-repo-rhel8-12-8-local-12.8.0_570.86.10-1.x86_64.rpm $ sudo rpm -i cuda-repo-rhel8-12-8-local-12.8.0_570.86.10-1.x86_64.rpm $ sudo dnf clean all $ sudo dnf install cuda-toolkit-12-8
-
Enable the Oracle Linux 8 EPEL yum repository. Install the
dkms
package.$ sudo yum-config-manager --enable ol8_developer_EPEL $ sudo dnf install dkms
-
Install the GPU drivers.
$ sudo dnf install cuda-12-8
-
Install the NVIDIA kernel module.
$ sudo scl enable gcc-toolset-13 bash # dkms install nvidia-open -v 570.86.10
If this
make
error appears while the kernel module is built, you can safely ignore it.Cleaning build area...(bad exit status: 2) Failed command: make -C /lib/modules/5.15.0-206.153.7.el8uek.x86_64/build M=/var/lib/dkms/nvidia-open/570.86.10/build clean
-
Verify the installation with the NVIDIA System Management Interface.
# nvidia-smi +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 570.86.10 Driver Version: 570.86.10 CUDA Version: 12.8 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA L40S Off | 00000000:00:05.0 Off | 0 | | N/A 26C P8 23W / 350W | 1MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | No running processes found | +-----------------------------------------------------------------------------------------+
Specification | Possible Values |
---|---|
Shape name |
|
GPUs | 1-4 – corresponding with the shape name |
OCPUs | 27 per GPU |
Memory | 200 GB per GPU |
VNICs | Up to 24 |
Bandwidth | Up to 400 Gbps |