Contents
- 1 Sizing
- 1.1 1. Calculate Workbench Node/Host Disk Space
- 1.2 2. Only if/when the default Retention Period and Metric Frequency settings are changed
- 1.3 3. Determine the Workbench Node/Host Count
- 1.4 4. Workbench Node/Host Resources
- 1.5 Required Number of additional Node(s)/Host(s) at each Workbench Data-Center
- 1.6 Example 1 - Ingest from 10 Engage Hosts - 30 day Retention Period - 60 second Metric Frequency
- 1.7 Example 2 - Ingest from 30 Engage Hosts - 7 day Retention Period - 10 second Metric Frequency
- 1.8 Example 3 - Ingest from 90 Engage Hosts - 90 day Retention Period - 30 second Metric Frequency
Sizing
- It's imperative you review, plan and define the details below before installing Workbench; failure to do so could result in a Workbench re-installation
- Review and complete each sub-section below before moving onto the next
- Consider that if/when upgrading Workbench, the Workbench Host(s) "free" disk space requires at least 3 times the size of the "<WORKBENCH_INSTALL>/ElasticSearch/data" directory - else the Workbench upgrade process will fail and Workbench data integrity and operation will likely be compromised.
- Workbench can be deployed as a single-node/host or as a multi-node/host cluster.
- The Workbench multi-node cluster deployment is available to support high-availability and/or environments that have a high volume of events/metrics.
- Multiple Data-Centers are supported, where Workbench can be deployed as single-node/host or as a cluster per Data-Center.
- Workbench deployments across Data-Centers can then be connected and synced in real-time to provide holistic visibility of the Alarms, Changes, Channel Monitoring and Auditing features.
- To determine the number of Workbench nodes/hosts, and the resource requirements for each, please follow the steps below.
The Workbench 9.x Sizing steps below should be followed for each Data-Center where Workbench will be deployed.
1. Calculate Workbench Node/Host Disk Space
Based on the number of Hosts (i.e. Engage SIP, URS, FWK etc) that Workbench will ingest Metric data from, review the table below to determine the respective disk space required for each Workbench Host at a given Data-Center:
Number of Hosts
|
Total Disk Space
|
---|---|
1-50 | 300 GB |
51-100 | 600 GB |
101-150 | 900 GB |
150+ | 1.2 TB [+300 GB for every 50 hosts > 200] |
Note the Total Disk Space = ___________ (used for next steps)
- Currently Workbench 9.x is limited to a maximum of 100 Hosts (the global combined Workbench or Engage Hosts), the table above details beyond the 100 Host limit for future Workbench sizing context.
2. Only if/when the default Retention Period and Metric Frequency settings are changed
The table in section 1 above, assumes the Workbench default data Retention Period of 30 days and a Workbench Agent/Remote Metric collection frequency of every 60 seconds.
If these default Retention Period and Metric Frequency values require modification, please re-calculate the Total Disk Space, by using the scale factors below:
- Retention Scale Factor = [New Retention Period Days] / 30
- Metric Frequency Scale Factor = 60 / [New Collection Frequency Seconds]
- Re-calculated Total Disk Space = Disk Space (from the section 1 table above) * Retention Scale Factor * Metric Frequency Scale Factor
- The global Workbench Retention Period is editable via Workbench Configuration\General\Retention Period\Workbench Data Retention Period (Days)
- The Metric Frequency collection setting can be changed on each Workbench Agent and Workbench Agent Remote application via:
- Workbench Configuration\Applications\Application Name (i.e. WB_Agent_Primary)\MetricBeat Host Metrics\Host Metric Collection Frequency (seconds)
- Workbench Configuration\Applications\Application Name (i.e. WB_Agent_Primary)\MetricBeat Associated Application Metrics\Application/Process Metric Collection Frequency (seconds)
3. Determine the Workbench Node/Host Count
Using the Total Disk Space calculation from the previous step, next determine the required number of Workbench Nodes/Hosts:
Total Disk Space from Step 1 or 2 above | Number of Workbench Nodes/Hosts Required |
---|---|
is less than 2.5 TB | A single (1) Node/Host Workbench can be used |
is greater than 2.5 TB OR if Workbench High Availability is required | A 3 x Nodes/Hosts Workbench Cluster is required |
- Workbench High-Availability (HA) is resiliency of event data (via Workbench Elasticsearch) and configuration data (via Workbench ZooKeeper)
4. Workbench Node/Host Resources
This section details the per Workbench Node/Host recommended resources based on the previous steps:
Type | Specification |
---|---|
Workbench Primary Node/Host
|
|
Non Workbench Primary Nodes/Hosts
|
|
- The following Memory allocation is need for each Workbench Elasticsearch Node/Host in the deployment.
- Please review ES Heap Settings for details on configuring the RAM for each Workbench Elasticsearch instance.
Total Disk Space per Node/Host | Dedicated Workbench Elasticsearch Memory Required |
---|---|
< 100 GB | 2 GB RAM |
100 - 750 GB | 4 GB RAM |
750 - 1.5 TB | 6 GB RAM |
1.5 - 2.5 TB | 8 GB RAM |
- If/when Total Disk Space is greater than 2.5 TB per Node/Host, please raise a Genesys Customer Care Case for consultation/guidance.
Required Number of additional Node(s)/Host(s) at each Workbench Data-Center
Workbench currently supports ingesting Metric data from a maximum of 100 Hosts.
Required Number of WB additional Nodes/Hosts | Number of Hosts sending Metric data to Workbench | Frequency of Metrics being sent from each Host to Workbench |
0 (WB on Primary host) | 100 | 60 (default) |
1 (WB on Primary host and Logstash on the additional node) | 100 | 30 |
1 (WB on Primary host and Logstash on the additional node) | 100 | 10 |
Example 1 - Ingest from 10 Engage Hosts - 30 day Retention Period - 60 second Metric Frequency
A production Workbench deployment ingesting Metric data from 10 Engage Hosts:
- Number of Hosts to ingest Metric data from = 10
- Retention Period = 30 days (default)
- Metric Frequency Collection = 60 seconds (default)
- Total Disk Space = 300 GB
- 1 x Workbench Node/Host
- CPU: 10 Cores
- RAM: 24 GB
- NIC: 100 MB
- DISK: 300 GB
- DEDICATED Elasticsearch RAM: 4 GB
Example 2 - Ingest from 30 Engage Hosts - 7 day Retention Period - 10 second Metric Frequency
A production Workbench deployment ingesting Metric data from 30 Engage Hosts:
- Number of Hosts to ingest Metric data from = 30
- Retention Period = 7 days
- therefore re-calculated Retention Scale Factor is 7 (days) / 30 => 0.23
- Metric Frequency Collection = 10 seconds
- therefore re-calculated Metric Frequency Scale Factor is 60 / 10 => 6
- Re-calculated Total Disk Space is 300 GB * 0.23 * 6 => 414 GB
- 1 x Workbench Node/Host
- CPU: 10 Cores
- RAM: 24 GB
- NIC: 100 MB
- DISK: 414 GB
- DEDICATED Elasticsearch RAM: 4 GB
Example 3 - Ingest from 90 Engage Hosts - 90 day Retention Period - 30 second Metric Frequency
A production Workbench HA deployment ingesting Metric data from 90 Engage Hosts:
- Number of Hosts to ingest Metric data from = 90
- Retention Period = 90 days
- therefore re-calculated Retention Scale Factor is 90 (days) / 30 => 3
- Metric Frequency Collection = 30 seconds
- therefore re-calculated Metric Frequency Scale Factor is 60 / 30 => 2
- Re-calculated Total Disk Space is 600 GB * 3 * 2 => 3600 GB (~3.5 TB)
- 3 x Workbench Nodes/Hosts required given Total Disk Space is greater than 2.5 TB
- Workbench Primary
- CPU: 10 Cores
- RAM: 24 GB
- NIC: 100 MB
- DISK: 1200 GB (1.2 TB on each Node/Host given the Cluster architecture)
- DEDICATED Elasticsearch RAM: 8 GB
- Workbench Nodes 2 and 3
- CPU: 10 Cores
- RAM: 16 GB
- NIC: 100 MB
- DISK: 1200 GB (1.2 TB on each Node/Host given the Cluster architecture)
- DEDICATED Elasticsearch RAM: 8 GB