Dedicated master nodes in Amazon OpenSearch Service - Amazon OpenSearch Service

Dedicated master nodes in Amazon OpenSearch Service

Amazon OpenSearch Service uses dedicated master nodes to increase cluster stability. A dedicated master node performs cluster management tasks, but does not hold data or respond to data upload requests. This offloading of cluster management tasks increases the stability of your domain. Just like all other node types, you pay an hourly rate for each dedicated master node.

Dedicated master nodes perform the following cluster management tasks:

  • Track all nodes in the cluster.

  • Track the number of indexes in the cluster.

  • Track the number of shards belonging to each index.

  • Maintain routing information for nodes in the cluster.

  • Update the cluster state after state changes, such as creating an index and adding or removing nodes in the cluster.

  • Replicate changes to the cluster state across all nodes in the cluster.

  • Monitor the health of all cluster nodes by sending heartbeat signals, periodic signals that monitor the availability of the data nodes in the cluster.

The following illustration shows an OpenSearch Service domain with 10 instances. Seven of the instances are data nodes and three are dedicated master nodes. Only one of the dedicated master nodes is active. The two gray dedicated master nodes wait as backup in case the active dedicated master node fails. All data upload requests are served by the seven data nodes, and all cluster management tasks are offloaded to the active dedicated master node.

OpenSearch Service domain with data nodes and dedicated master nodes, illustrating cluster management.

Choosing the number of dedicated master nodes

We recommend that you use Multi-AZ with Standby, which adds three dedicated master nodes to each production OpenSearch Service domain. If you deploy with Multi-AZ without Standby or single-AZ, we still recommend three dedicated master nodes. Never choose an even number of dedicated master nodes. Consider the following when choosing the number of dedicated master nodes:

  • One dedicated master node is explicitly prohibited by OpenSearch Service because you have no backup in the event of a failure. You receive a validation exception if you try to create a domain with only one dedicated master node.

  • If you have two dedicated master nodes, your cluster doesn't have the necessary quorum of nodes to elect a new master node in the event of a failure.

    A quorum is the number of dedicated master nodes / 2 + 1 (rounded down to the nearest whole number). In this case, 2 / 2 + 1 = 2. Because one dedicated master node has failed and only one backup exists, the cluster doesn't have a quorum and can't elect a new master.

  • Three dedicated master nodes, the recommended number, provides two backup nodes in the event of a master node failure and the necessary quorum (2) to elect a new master.

  • Four dedicated master nodes are not better than three and can cause issues if you use multiple Availability Zones.

    • If one master node fails, you have the quorum (3) to elect a new master. If two nodes fail, you lose that quorum, just as you do with three dedicated master nodes.

    • In a three Availability Zone configuration, two AZs have one dedicated master node, and one AZ has two. If that AZ experiences a disruption, the remaining two AZs don't have the necessary quorum (3) to elect a new master.

  • Having five dedicated master nodes works as well as three and allows you to lose two nodes while maintaining a quorum. But because only one dedicated master node is active at any given time, this configuration means that you pay for four idle nodes. Many users find this level of failover protection excessive.

If a cluster has an even number of master-eligible nodes, OpenSearch and Elasticsearch versions 7.x and later ignore one node so that the voting configuration is always an odd number. In this case, four dedicated master nodes are essentially equivalent to three (and two to one).

Note

If your cluster doesn't have the necessary quorum to elect a new master node, write and read requests to the cluster both fail. This behavior differs from the OpenSearch default.

Choosing instance types for dedicated master nodes

OpenSearch Service domain and instance quotas

The following table lists the quotas related to OpenSearch Service domains

Name Default Adjustable Description
Dedicated master instances per domain Each supported region: 3 or 5 No

The maximum number of dedicated master instances in a single Amazon OpenSearch Service domain.

Domains per region Each supported region: 100 Yes

The maximum number of Amazon OpenSearch Service domains you can create in each AWS Region.

Instance per domain Each supported region: 80 Yes

The maximum number of instances in a single Amazon OpenSearch Service domain. You can request an increase of up to 1002 instances per domain.

Instance per domain (T2 instance type) Each supported region: 10 Yes

The maximum number of T2 instances in a single Amazon OpenSearch Service domain.

Warm instances per domain Each supported region: 150 No

The maximum number of warm nodes in a single Amazon OpenSearch Service domain. You can request an increase of up to 750 instances per domain.

Number of cross-cluster connections per domain 40 No
Dedicated coordinator instance per AZ Each supported region: 200 Yes

The number must be between 1 and 200. Coordinator node counts can not exceed data node count.

Total storage per domain 25 PB No

This maximum is the sum of all data nodes and warm nodes. For example, your domain might have 45 r6gd.16xlarge.search

instances and 140 ultrawarm1.large.search

instances for a total of 2.88 PiB of storage. New limits are 10 PB for data nodes and 15 PB for Warm Nodes.

Custom packages per region 25 No
Custom packages per domain 20 No

Although dedicated master nodes don't process search and query requests, their size is highly correlated with the instance size and number of instances, indexes, and shards that they can manage. For production clusters, we recommend, at a minimum, the following instance types for dedicated master nodes.

These recommendations are based on typical workloads and can vary based on your needs. Clusters with many shards or field mappings can benefit from larger instance types. Monitor the dedicated master node metrics to see if you need to use a larger instance type.

Instance count

Master node RAM size Maximum supported shard count

Recommended minimum dedicated master instance type

1–10

8 GiB 10K

m5.large.search or m6g.large.search

11–30

16 GiB 30K

c5.2xlarge.search or c6g.2xlarge.search

31–75 32 GiB 40K

r5.xlarge.search or r6g.xlarge.search

76 – 125 64 GiB 75K

r5.2xlarge.search or r6g.2xlarge.search

126 – 200

128 GiB 75K

r5.4xlarge.search or r6g.4xlarge.search