2 Deployment Environment

This section provides information about the cloud native platform used for SCP benchmarking.

2.1 Deployed Components

Deployment Platform

Oracle Communications Cloud Native Core, Cloud Native Environment (CNE) 24.2.0 and CNE on BareMetal 24.1.0 can be used for performing benchmark tests.

Observability Services

The following table lists services that are part of CNE and used for fetching SCP metrics.

Table 2-1 Observability Services

Service Name Version
Grafana 1.26.1
Jaeger 1.52.0
Oracle OpenSearch 2.3.0
Oracle OpenSearch Dashboard 2.3.0
Prometheus 1.7.0

Cloud Native Orchestrator

Kubernetes 1.28.6 is used to manage application pods across the cluster.

cnDBTier

cnDBTier 25.2.100 is used to perform benchmark tests.

For more information about above mentioned software, see Oracle Communications Cloud Native Core, cnDBTier Installation, Upgrade, and Fault Recovery Guide.

2.2 Deployment Resources

The performance and capacity of SCP can vary based on the chosen environment and how SCP is deployed. This section provides information about CNE and cnDBTier resources used to perform benchmark tests.

2.2.1 Cluster Details

The following table provides information about the types of servers and the number of servers used in the test environment:

Table 2-2 Test Bed 1 - CNE on BareMetal

Nodes Type Count
Primary Nodes HP Gen10 RMS 3
Worker Nodes HP Gen10 Blades 29
HP Gen8 Blades 7
Top of Rack Switch Cisco Nexus9000 93180YC-EX 2
Enclosure Switch HP 6120 2

The following table provides information about the number of pods required by each CNE service.

Table 2-3 CNE Common Services Observability Resources

Service Name Number of Pods RAM Request/Limit vCPU Request/Limit PVC Size Recommendation
Prometheus Server 2 4Gi/150Gi 2m/20m 800Gi
Alert Manager 2 4Gi/32Gi 2m/6m NA
Fluentd 1 per Worker Node 2Gi/20Gi 100m/1m NA
Prom-node-exporter 1 per Worker Node 512Mi/512Mi 800m/800m NA
MetalLB speaker 1 per Worker Node 100Mi/100Mi 100m/100m NA
OpenSearch Data 5/5 32Gi/32Gi 2m/2m 400Gi
OpenSearch Master 3/3 16Gi/16Gi 1m/1m 400Gi
ISM Policy 3/3 128Mi/128Mi 100m/100m NA
OpenSearch Client 3/3 16Gi/16Gi 2m/2m NA
Grafana 1 500Mi/500Mi 200m/200m NA
kube-state-metrics 1 32Mi/1Gi 20m/20m NA
jaeger-agent 1 per Worker Node 128Mi/512Mi 256m/500m NA
jaeger-collector 1 512Mi/1Gi 500m/1250m NA
jaeger-query 1 128Mi/512Mi 256m/500m NA
promxy 1 512Mi/2Gi 100m/100m NA
prom-stack-kube-operator 1 100Mi/200Mi 100m/200m NA
rook-ceph-osd 1 for each raw disk available to OS on all Worker Node 1Gi/8Gi 500m/1m NA
rook-ceph-mgr 1 1Gi/1Gi 500m/500m NA
rook-ceph-mon 3 1Gi/1Gi 500m/500m NA
rook-ceph-operator 1 2Gi/2Gi 100m/500m NA

Table 2-4 Test Bed 2 - VMware Tanzu

Nodes Type Count
Primary Nodes VM (8 CPU and 64 GB Memory) 3
Worker Nodes VM(32 CPU and 128 GB Memory) 51
Underlying Hardware Cisco Nexus9000 93180YC-EX 19

Table 2-5 Test Bed 3 - CNE on BareMetal

Nodes Type Count
Primary Nodes X9 Server and NVME 3
Worker Nodes X9 Server and NVME 17

Table 2-6 Test Bed 4 - CNE on BareMetal

Nodes Type Count
Primary Nodes ORACLE SERVER X8-2 3
Worker Nodes ORACLE SERVER X8-2 45
Top of Rack Switch Cisco 93108tc-ex 2

Table 2-7 Test Bed 5 - vCNE on OpenStack

Nodes Type Count
Master Nodes ORACLE SERVER X8-2 3
Worker Nodes ORACLE SERVER X8-2 42
Top of Rack Switch Cisco 93108TC-ex 2

Table 2-8 Test Bed 5 - LBVM Resources used for vCNE

Resources Values
Number of LBVM

4

Note: There are two pairs of LBVMs in this setup, with 2 LBVMs in each pair. For testing purposes, only one LBVM pair was used, with one LBVM operating in active mode and the other in standby mode.

RAM 16 GB
vCPU 4
Disk Size 40 GB

The following table provides information about the number of pods required by each CNE service.

Table 2-9 CNE Common Services Observability Resources

Service Name Number of Pods RAM Request/Limit vCPU Request/Limit PVC Size Recommendation
Prometheus Server 2 4Gi/150Gi 2/20 800Gi
Alert Manager 2 4Gi/32Gi 2/6 NA
Fluentd 1 per Worker Node 2Gi/20Gi 100m/1 NA
Prom-node-exporter 1 per Worker Node 512Mi/512Mi 800m/800m NA
Metal LB speaker 1 per Worker Node 100Mi/100Mi 100m/100m NA
OpenSearch Data 5/5 32Gi/32Gi 2/2 400Gi
OpenSearch Master 3/3 16Gi/16Gi 1/1 400Gi
ISM Policy 3/3 128Mi/128Mi 100m/100m NA
OpenSearch Client 3/3 16Gi/16Gi 2/2 NA
Grafana 1 500Mi/500Mi 200m/200m NA
kube-state-metrics 1 32Mi/1Gi 20m/20m NA
jaeger-agent 1 per Worker Node 128Mi/512Mi 256m/500m NA
jaeger-collector 1 512Mi/1Gi 500m/1250m NA
jaeger-query 1 128Mi/512Mi 256m/500m NA
promxy 1 512Mi/2Gi 100m/100m NA
prom-stack-kube-operator 1 100Mi/200Mi 100m/200m NA
rook-ceph-osd 1 for each raw disk available to OS on all Worker Node 1Gi/8Gi 500m/1 NA
rook-ceph-mgr 1 1Gi/1Gi 500m/500m NA
rook-ceph-mon 3 1Gi/1Gi 500m/500m NA
rook-ceph-operator 1 2Gi/2Gi 100m/500m NA

2.2.2 cnDBTier Resources

The following table provides information about cnDBTier resources for both ASM and non-ASM setups required to perform SCP benchmark tests:

Table 2-10 cnDBTier Resources (Non-ASM)

Service Name CPU/Pod Memory/Pod (in GB) PVC Size (in GB) Ephemeral Storage (MB) Sidecar CPU/Pod Sidecar Memory/Pod (in GB) Sidecar Ephemeral Storage (MB)
Min Max Min Max PVC 1 PVC 2 Min Max Min Max Min Max Min Max
MGMT (ndbmgmd) 2 2 4 5 14 NA 90 1000 0.2 0.2 0.256 0.256 90 1000
DB(ndbmtd) 2 2 8 8 15 8 90 1000 1.2 1.2 2.256 2.256 180 3000
SQL - Replication(ndbmysqld) 4 4 10 10 25 NA 90 1000 0.3 0.3 0.512 0.512 180 2000
SQL - Access(ndbappmysqld) 4 4 8 8 20 NA 90 1000 0.3 0.3 0.512 0.512 180 2000
Monitor Service(db-monitor-svc) 4 4 4 4 0 NA 90 1000 0 0 0 0 0 0
db-connectivity-service 0 0 0 0 0 NA 0 0 NA NA NA NA NA NA
Replication Service - Leader(db-replication-svc) 2 2 12 12 190 NA 90 1000 0.2 0.2 0.5 0.5 90 1000
Replication Service - Other(db-replication-svc) 0.6 1 1 2 NA NA 90 1000 0.2 0.2 0.5 0.5 NA NA
Backup Manager Service(db-backup-manager-svc) 1 1 1 1 0 NA 90 1000 0 0 0 0 0 0

Table 2-11 cnDBTier Resources (ASM)

Service Name CPU/Pod Memory/Pod (in GB) PVC Size (in GB) Ephemeral Storage (MB) Sidecar CPU/Pod Sidecar Memory/Pod (in GB) DB Sidecar CPU/Pod DB Sidecar Memory/Pod (in GB)
Min Max Min Max PVC 1 PVC 2 Min Max Min Max Min Max Min Max Min Max
MGMT (ndbmgmd) 2 2 5 5 14 NA 1 1 1 1 1 1 1 1 1 1
DB(ndbmtd) 2 2 8 8 14 27 1 1 1 1 1 1 1 1 1 2
SQL - Replication(ndbmysqld) 4 4 10 10 25 NA 1 1 2 2 2 2 0.1 0.1 0.256 0.256
SQL - Access(ndbappmysqld) 4 4 8 8 20 NA 1 1 2 2 2 2 2 2 2 2
Monitor Service(db-monitor-svc) 4 4 4 4 NA NA 1 1 2 2 1 1 2 2 1 1
db-connectivity-service 0 0 0 0 0 NA 0 0 NA NA NA NA 0 0 0 0
Replication Service(db-replication-svc) 2 2 12 12 11 0.01 0 0 2 2 2 2 0.2 0.2 0.5 0.5
Backup Manager Service(db-backup-manager-svc) 1 1 1 1 NA NA 1 1 2 2 1 1 2 2 1 1

2.2.3 SCP Resources

The following table provides information about resource requirements to perform SCP benchmark tests:

Table 2-12 SCP Resources

Microservice Name SCP Service Pods
vCPU/Pod Memory/Pod (in Gi)
Min Max Min Max
Helm test 3 3 3 3
Helm Hook 3 3 3 3
scpc-subscription 1 1 1 1
scpc-notification 8 8 8 8
scpc-audit 3 3 4 4
scpc-configuration 2 2 2 2
scp-cache 8 8 8 8
scp-loadmanager 8 8 8 8
scp-nrfproxy 8 8 8 8
scp-worker (Profile 1) 4 4 8 8
scp-worker (Profile 2) 8 8 12 12
scp-worker (Profile 3) 12 12 24 24
scp-mediation 8 8 8 8
scp-nrfproxy-oauth 8 8 8 8
scpc-alternate-resolution 2 2 2 2