WARNING: Version 6.2 of Elasticsearch has passed its EOL date.
This documentation is no longer being maintained and may be removed. If you are running this version, we strongly advise you to upgrade. For the latest information, see the current release documentation.
Full cluster restart upgrade
editFull cluster restart upgrade
editA full cluster restart upgrade requires that you shut all nodes in the cluster down, upgrade them, and restart the cluster. A full cluster restart was required when upgrading to major versions prior to 6.x. Elasticsearch 6.x supports rolling upgrades from Elasticsearch 5.6. Upgrading to 6.x from earlier versions requires a full cluster restart. See the Upgrade paths table to verify the type of upgrade you need to perform.
To perform a full cluster restart upgrade:
-
Disable shard allocation.
When you shut down a node, the allocation process waits for
index.unassigned.node_left.delayed_timeout(by default, one minute) before starting to replicate the shards on that node to other nodes in the cluster, which can involve a lot of I/O. Since the node is shortly going to be restarted, this I/O is unnecessary. You can avoid racing the clock by disabling allocation before shutting down the node:PUT _cluster/settings { "persistent": { "cluster.routing.allocation.enable": "none" } } -
Stop indexing and perform a synced flush.
Performing a synced-flush speeds up shard recovery.
POST _flush/synced
When you perform a synced flush, check the response to make sure there are no failures. Synced flush operations that fail due to pending indexing operations are listed in the response body, although the request itself still returns a 200 OK status. If there are failures, reissue the request.
-
Shutdown all nodes.
-
If you are running Elasticsearch with
systemd:sudo systemctl stop elasticsearch.service
-
If you are running Elasticsearch with SysV
init:sudo -i service elasticsearch stop
-
If you are running Elasticsearch as a daemon:
kill $(cat pid)
-
-
Upgrade all nodes.
To upgrade using a Debian or RPM package:
-
Use
rpmordpkgto install the new package. All files are installed in the appropriate location for the operating system and Elasticsearch config files are not overwritten.
To upgrade using a zip or compressed tarball:
-
Extract the zip or tarball to a new directory. This is critical if you
are not using external
configanddatadirectories. -
Set the
ES_PATH_CONFenvironment variable to specify the location of your externalconfigdirectory andjvm.optionsfile. If you are not using an externalconfigdirectory, copy your old configuration over to the new installation. -
Set
path.datainconfig/elasticsearch.ymlto point to your external data directory. If you are not using an externaldatadirectory, copy your old data directory over to the new installation. -
Set
path.logsinconfig/elasticsearch.ymlto point to the location where you want to store your logs. If you do not specify this setting, logs are stored in the directory you extracted the archive to.
When you extract the zip or tarball packages, the
elasticsearch-n.n.ndirectory contains the Elasticsearhconfig,data,logsandpluginsdirectories.We recommend moving these directories out of the Elasticsearch directory so that there is no chance of deleting them when you upgrade Elasticsearch. To specify the new locations, use the
ES_PATH_CONFenvironment variable and thepath.dataandpath.logssettings. For more information, see Important Elasticsearch configuration.The Debian and RPM packages place these directories in the appropriate place for each operating system. In production, we recommend installing using the deb or rpm package.
-
Use
-
Upgrade any plugins.
Use the
elasticsearch-pluginscript to install the upgraded version of each installed Elasticsearch plugin. All plugins must be upgraded when you upgrade a node. -
Start each upgraded node.
If you have dedicated master nodes, start them first and wait for them to form a cluster and elect a master before proceeding with your data nodes. You can check progress by looking at the logs.
As soon as the minimum number of master-eligible nodes have discovered each other, they form a cluster and elect a master. At that point, you can use
_cat/healthand_cat/nodesto monitor nodes joining the cluster:GET _cat/health GET _cat/nodes
The
statuscolumn returned by_cat/healthshows the health of each node in the cluster:red,yellow, orgreen. -
Wait for all nodes to join the cluster and report a status of yellow.
When a node joins the cluster, it begins to recover any primary shards that are stored locally. The
_cat/healthAPI initially reports astatusofred, indicating that not all primary shards have been allocated.Once a node recovers its local shards, the cluster
statusswitches toyellow, indicating that all primary shards have been recovered, but not all replica shards are allocated. This is to be expected because you have not yet reenabled allocation. Delaying the allocation of replicas until all nodes areyellowallows the master to allocate replicas to nodes that already have local shard copies. -
Reenable allocation.
When all nodes have joined the cluster and recovered their primary shards, reenable allocation by restoring
cluster.routing.allocation.enableto its default:PUT _cluster/settings { "persistent": { "cluster.routing.allocation.enable": null } }Once allocation is reenabled, the cluster starts allocating replica shards to the data nodes. At this point it is safe to resume indexing and searching, but your cluster will recover more quickly if you can wait until all primary and replica shards have been successfully allocated and the status of all nodes is
green.You can monitor progress with the
_cat/healthand_cat/recoveryAPIs:GET _cat/health GET _cat/recovery