If the concurrent execution capability of the standard size is insufficient, increase the number of platform nodes and accordingly increase the resource requests and limits for the platform services to create a larger deployment.
Domino recommends that you consult your Customer Success Manager for customizations and advice before you deploy this model.
Platform infrastructure details
- Number of platform nodes
-
6
- CPU per node
-
8 cores
- Memory per node
-
32 GB
- Maximum concurrent executions
-
~600
Default compute node pool instance size
- CPU per node
-
16 cores
- Memory per node
-
64 GB
Deploy the platform
If you use this sizing model, ensure that your platform node pool can scale up to six nodes. Then, add the following resource request and limit overrides to the fleetcommand-agent configuration file:
release_overrides:
cluster-autoscaler:
chart_values:
resources:
limits:
cpu: 1
memory: 2Gi
requests:
cpu: 50m
memory: 100M
data-plane:
chart_values:
agent:
deployment:
resources:
limits:
cpu: 4000m
memory: 4Gi
requests:
cpu: 2000m
memory: 2Gi
mongodb-replicaset:
chart_values:
persistentVolume:
size: 120Gi
resources:
requests:
cpu: 2
memory: "4Gi"
elasticsearch:
chart_values:
replicas: 5
esJavaOpts: '-Xms8g -Xmx8g'
resources:
requests:
cpu: 1
memory: 8Gi
git:
chart_values:
resources:
limits:
cpu: 4
memory: 4Gi
requests:
cpu: 2
memory: 2Gi
persistence:
size: 160Gi
prometheus:
chart_values:
resources:
requests:
cpu: 1100m
memory: 7Gi
limits:
memory: 14Gi
nucleus:
chart_values:
keycloak: {}
replicaCount:
frontend: 3
config:
javaMaxHeap: 6G
resources:
limits:
cpu: 5
memory: 11Gi
requests:
cpu: 4
memory: 8Gi
newrelic-logging:
chart_values:
resources:
limits:
cpu: 1
memory: 600Mi
image-cache-api:
chart_values:
resources:
limits:
cpu: 2
memory: 2Gi
requests:
cpu: 2
memory: 2Gi
rabbitmq-ha:
chart_values:
resources:
limits:
cpu: 2
memory: 4Gi
requests:
cpu: 1
memory: 2Gi
prometheus-adapter:
chart_values:
resources:
limits:
cpu: 600m
memory: 512Mi
requests:
cpu: 250m
memory: 256Mi
newrelic-open-metrics:
chart_values:
resources:
limits:
cpu: 1200m
memory: 2Gi
requests:
cpu: 200m
memory: 512Mi
newrelic-infrastructure:
chart_values:
resources:
limits:
memory: 2Gi
requests:
cpu: 50m
memory: 30Mi
To check the status of your available platform resources and allocation, run the following:
kubectl describe nodes -l dominodatalab.com/node-pool=platform | grep -A 10 Allocated
The following example shows six platform nodes:
Allocated resources: (Total limits may be over 100 percent, i.e., overcommitted.) Resource Requests Limits -------- -------- ------ cpu 6405m (80%) 11200m (141%) memory 30601Mi (99%) 29787Mi (97%) ephemeral-storage 0 (0%) 0 (0%) hugepages-1Gi 0 (0%) 0 (0%) hugepages-2Mi 0 (0%) 0 (0%) attachable-volumes-aws-ebs 0 0 smarter-devices/fuse 0 0 -- Allocated resources: (Total limits may be over 100 percent, i.e., overcommitted.) Resource Requests Limits -------- -------- ------ cpu 6530m (82%) 13150m (166%) memory 17853Mi (58%) 20201Mi (65%) ephemeral-storage 0 (0%) 0 (0%) hugepages-1Gi 0 (0%) 0 (0%) hugepages-2Mi 0 (0%) 0 (0%) attachable-volumes-aws-ebs 0 0 smarter-devices/fuse 0 0 -- Allocated resources: (Total limits may be over 100 percent, i.e., overcommitted.) Resource Requests Limits -------- -------- ------ cpu 7805m (98%) 22300m (281%) memory 29317Mi (95%) 40653Mi (132%) ephemeral-storage 0 (0%) 0 (0%) hugepages-1Gi 0 (0%) 0 (0%) hugepages-2Mi 0 (0%) 0 (0%) attachable-volumes-aws-ebs 0 0 smarter-devices/fuse 0 0 -- Allocated resources: (Total limits may be over 100 percent, i.e., overcommitted.) Resource Requests Limits -------- -------- ------ cpu 7655m (96%) 13100m (165%) memory 24513995008 (77%) 25391Mi (83%) ephemeral-storage 0 (0%) 0 (0%) hugepages-1Gi 0 (0%) 0 (0%) hugepages-2Mi 0 (0%) 0 (0%) attachable-volumes-aws-ebs 0 0 smarter-devices/fuse 0 0 -- Allocated resources: (Total limits may be over 100 percent, i.e., overcommitted.) Resource Requests Limits -------- -------- ------ cpu 405m (5%) 2100m (26%) memory 755Mi (2%) 2863Mi (9%) ephemeral-storage 0 (0%) 0 (0%) hugepages-1Gi 0 (0%) 0 (0%) hugepages-2Mi 0 (0%) 0 (0%) attachable-volumes-aws-ebs 0 0 smarter-devices/fuse 0 0 -- Allocated resources: (Total limits may be over 100 percent, i.e., overcommitted.) Resource Requests Limits -------- -------- ------ cpu 7875m (99%) 19750m (249%) memory 25041Mi (81%) 15943Mi (52%) ephemeral-storage 0 (0%) 0 (0%) hugepages-1Gi 0 (0%) 0 (0%) hugepages-2Mi 0 (0%) 0 (0%) attachable-volumes-aws-ebs 0 0 smarter-devices/fuse 0 0