Configuring and operating a Raft ordering service¶
Audience: Raft ordering node admins
Note: this topic describes the process for configuring a Raft ordering service that has not been bootstrapped with a system channel genesis block. For a version of this topic that includes information about the system channel, check out Configuring and operating a Raft ordering service.
Conceptual overview¶
For a high level overview of the concept of ordering and how the supported ordering service implementations (including Raft) work at a high level, check out our conceptual documentation on the Ordering Service.
To learn about the process of setting up an ordering node, check out our documentation on Planning for an ordering service.
Configuration¶
A Raft cluster is configured in two places:
- Local configuration: Governs node specific aspects, such as TLS communication, replication behavior, and file storage.
- Channel configuration: Defines the membership of the Raft cluster for the corresponding channel, as well as protocol specific parameters such as heartbeat frequency, leader timeouts, and more.
Raft nodes identify each other using TLS pinning, so in order to impersonate a Raft node, an attacker needs to obtain the private key of its TLS certificate. As a result, it is not possible to run a Raft node without a valid TLS configuration.
Recall, each channel has its own instance of a Raft protocol running. Thus, a
Raft node must be referenced in the configuration of each channel it belongs to
by adding its server and client TLS certificates (in PEM
format) to the channel
config. This ensures that when other nodes receive a message from it, they can
securely confirm the identity of the node that sent the message.
The following section from configtx.yaml
shows three Raft nodes (also called
“consenters”) in the channel:
Consenters:
- Host: raft0.example.com
Port: 7050
ClientTLSCert: path/to/ClientTLSCert0
ServerTLSCert: path/to/ServerTLSCert0
- Host: raft1.example.com
Port: 7050
ClientTLSCert: path/to/ClientTLSCert1
ServerTLSCert: path/to/ServerTLSCert1
- Host: raft2.example.com
Port: 7050
ClientTLSCert: path/to/ClientTLSCert2
ServerTLSCert: path/to/ServerTLSCert2
When the channel config block is created, the configtxgen
tool reads the paths
to the TLS certificates, and replaces the paths with the corresponding bytes of
the certificates.
Note: it is possible to remove and add an ordering node from a channel dynamically without affecting the other nodes, a process described in the Reconfiguration section below.
Local configuration¶
The orderer.yaml
has two configuration sections that are relevant for Raft
orderers:
Cluster, which determines the TLS communication configuration. And consensus, which determines where Write Ahead Logs and Snapshots are stored.
Cluster parameters:
By default, the Raft service is running on the same gRPC server as the client facing server (which is used to send transactions or pull blocks), but it can be configured to have a separate gRPC server with a separate port.
This is useful for cases where you want TLS certificates issued by the organizational CAs, but used only by the cluster nodes to communicate among each other, and TLS certificates issued by a public TLS CA for the client facing API.
ClientCertificate
,ClientPrivateKey
: The file path of the client TLS certificate and corresponding private key.ListenPort
: The port the cluster listens on. It must be same asconsenters[i].Port
in Channel configuration. If blank, the port is the same port as the orderer general port (general.listenPort
)ListenAddress
: The address the cluster service is listening on.ServerCertificate
,ServerPrivateKey
: The TLS server certificate key pair which is used when the cluster service is running on a separate gRPC server (different port).
Note: ListenPort
, ListenAddress
, ServerCertificate
, ServerPrivateKey
must
be either set together or unset together.
If they are unset, they are inherited from the general TLS section,
in example general.tls.{privateKey, certificate}
.
When general TLS is disabled:
- Use a different
ListenPort
than the orderer general port - Properly configure TLS root CAs in the channel configuration.
There are also hidden configuration parameters for general.cluster
which can be
used to further fine tune the cluster communication or replication mechanisms:
SendBufferSize
: Regulates the number of messages in the egress buffer.DialTimeout
,RPCTimeout
: Specify the timeouts of creating connections and establishing streams.ReplicationBufferSize
: the maximum number of bytes that can be allocated for each in-memory buffer used for block replication from other cluster nodes. Each channel has its own memory buffer. Defaults to20971520
which is20MB
.PullTimeout
: the maximum duration the ordering node will wait for a block to be received before it aborts. Defaults to five seconds.ReplicationRetryTimeout
: The maximum duration the ordering node will wait between two consecutive attempts. Defaults to five seconds.ReplicationBackgroundRefreshInterval
: the time between two consecutive attempts to replicate existing channels that this node was added to, or channels that this node failed to replicate in the past. Defaults to five minutes.TLSHandshakeTimeShift
: If the TLS certificates of the ordering nodes expire and are not replaced in time (see TLS certificate rotation below), communication between them cannot be established, and it will be impossible to send new transactions to the ordering service. To recover from such a scenario, it is possible to make TLS handshakes between ordering nodes consider the time to be shifted backwards a given amount that is configured toTLSHandshakeTimeShift
. This setting only applies when a separate cluster listener is in use. If the cluster service is sharing the orderer’s main gRPC server, then instead specifyTLSHandshakeTimeShift
in theGeneral.TLS
section.
Consensus parameters:
WALDir
: the location at which Write Ahead Logs foretcd/raft
are stored. Each channel will have its own subdirectory named after the channel ID.SnapDir
: specifies the location at which snapshots foretcd/raft
are stored. Each channel will have its own subdirectory named after the channel ID.
There are also two hidden configuration parameters that can each be set by adding
them the consensus section in the orderer.yaml
:
EvictionSuspicion
: The cumulative period of time of channel eviction suspicion that triggers the node to pull blocks from other nodes and see if it has been evicted from the channel in order to confirm its suspicion. If the suspicion is confirmed (the inspected block doesn’t contain the node’s TLS certificate), the node halts its operation for that channel. A node suspects its channel eviction when it doesn’t know about any elected leader nor can be elected as leader in the channel. Defaults to 10 minutes.TickIntervalOverride
: If set, this value will be preferred over the tick interval configured in all channels where this ordering node is a consenter. This value should be set only with great care, as a mismatch in tick interval across orderers could result in a loss of quorum for one or more channels.
Channel configuration¶
Apart from the (already discussed) consenters, the Raft channel configuration has
an Options
section which relates to protocol specific knobs. It is currently
not possible to change these values dynamically while a node is running. The
node have to be reconfigured and restarted.
The only exceptions is SnapshotIntervalSize
, which can be adjusted at runtime.
Note: It is recommended to avoid changing the following values, as a misconfiguration
might lead to a state where a leader cannot be elected at all (i.e, if the
TickInterval
and ElectionTick
are extremely low). Situations where a leader
cannot be elected are impossible to resolve, as leaders are required to make
changes. Because of such dangers, we suggest not tuning these parameters for most
use cases.
TickInterval
: The time interval between twoNode.Tick
invocations.ElectionTick
: The number ofNode.Tick
invocations that must pass between elections. That is, if a follower does not receive any message from the leader of current term beforeElectionTick
has elapsed, it will become candidate and start an election.ElectionTick
must be greater thanHeartbeatTick
.HeartbeatTick
: The number ofNode.Tick
invocations that must pass between heartbeats. That is, a leader sends heartbeat messages to maintain its leadership everyHeartbeatTick
ticks.MaxInflightBlocks
: Limits the max number of in-flight append blocks during optimistic replication phase.SnapshotIntervalSize
: Defines number of bytes per which a snapshot is taken.
Reconfiguration¶
The Raft orderer supports dynamic (meaning, while the channel is being serviced) addition and removal of nodes as long as only one node is added or removed at a time. Note that your cluster must be operational and able to achieve consensus before you attempt to reconfigure it. For instance, if you have three nodes, and two nodes fail, you will not be able to reconfigure your cluster to remove those nodes. Similarly, if you have one failed node in a channel with three nodes, you should not attempt to rotate a certificate, as this would induce a second fault. As a rule, you should never attempt any configuration changes to the Raft consenters, such as adding or removing a consenter, or rotating a consenter’s certificate unless all consenters are online and healthy.
If you do decide to change these parameters, it is recommended to only attempt such a change during a maintenance cycle. Problems are most likely to occur when a configuration is attempted in clusters with only a few nodes while a node is down. For example, if you have three nodes in your consenter set and one of them is down, it means you have two out of three nodes alive. If you extend the cluster to four nodes while in this state, you will have only two out of four nodes alive, which is not a quorum. The fourth node won’t be able to onboard because nodes can only onboard to functioning clusters (unless the total size of the cluster is one or two).
So by extending a cluster of three nodes to four nodes (while only two are alive) you are effectively stuck until the original offline node is resurrected.
To add a new node to the ordering service:
- Ensure the orderer organization that owns the new node is one of the orderer organizations on the channel. If the orderer organization is not an administrator, the node will be unable to pull blocks as a follower or be joined to the consenter set.
- Start the new ordering node. For information about how to deploy an ordering node, check out Planning for an ordering service. Note that when you use the
osnadmin
CLI to create and join a channel, you do not need to point to a configuration block when starting the node. - Use the
osnadmin
CLI to add the first orderer to the channel. For more information, check out the Create a channel tutorial. - Wait for the Raft node to replicate the blocks from existing nodes for all channels its certificates have been added to. When an ordering node is added to a channel, it is added as a “follower”, a state in which it can replicate blocks but is not part of the “consenter set” actively servicing the channel. When the node finishes replicating the blocks, its status should change from “onboarding” to “active”. Note that an “active” ordering node is still not part of the consenter set.
- Add the new ordering node to the consenter set. For more information, check out the Create a channel tutorial.
It is possible to add a node that is already running (and participates in some channels already) to a channel while the node itself is running. To do this, simply add the node’s certificate to the channel config of the channel. The node will autonomously detect its addition to the new channel (the default value here is five minutes, but if you want the node to detect the new channel more quickly, reboot the node) and will pull the channel blocks from an orderer in the channel, and then start the Raft instance for that chain.
After it has successfully done so, the channel configuration can be updated to include the endpoint of the new Raft orderer.
To remove an ordering node from the consenter set of a channel, use the osnadmin channel remove
command to remove its endpoint and certificates from the channel. For more information, check out Add or remove orderers from existing channels.
Once an ordering node is removed from the channel, the other ordering nodes stop communicating with the removed orderer in the context of the removed channel. They might still be communicating on other channels.
The node that is removed from the channel automatically detects its removal either immediately or after EvictionSuspicion
time has passed (10 minutes by default) and shuts down its Raft instance on that channel.
If the intent is to delete the node entirely, remove it from all channels before shutting down the node.
TLS certificate rotation for an orderer node¶
All TLS certificates have an expiration date that is determined by the issuer. These expiration dates can range from 10 years from the date of issuance to as little as a few months, so check with your issuer. Before the expiration date, you will need to rotate these certificates on the node itself and every channel the node is joined to.
Note: In case the public key of the TLS certificate remains the same, there is no need to issue channel configuration updates.
For each channel the node participates in:
- Update the channel configuration with the new certificates.
- Replace its certificates in the file system of the node.
- Restart the node.
Because a node can only have a single TLS certificate key pair, the node will be unable to service channels its new certificates have not been added to during the update process, degrading the capacity of fault tolerance. Because of this, once the certificate rotation process has been started, it should be completed as quickly as possible.
If for some reason the rotation of the TLS certificates has started but cannot complete in all channels, it is advised to rotate TLS certificates back to what they were and attempt the rotation later.
Metrics¶
For a description of the Operations Service and how to set it up, check out our documentation on the Operations Service.
For a list at the metrics that are gathered by the Operations Service, check out our reference material on metrics.
While the metrics you prioritize will have a lot to do with your particular use case and configuration, there are two metrics in particular you might want to monitor:
consensus_etcdraft_is_leader
: identifies which node in the cluster is currently leader. If no nodes have this set, you have lost quorum.consensus_etcdraft_data_persist_duration
: indicates how long write operations to the Raft cluster’s persistent write ahead log take. For protocol safety, messages must be persisted durably, callingfsync
where appropriate, before they can be shared with the consenter set. If this value begins to climb, this node may not be able to participate in consensus (which could lead to a service interruption for this node and possibly the network).consensus_etcdraft_cluster_size
andconsensus_etcdraft_active_nodes
: these channel metrics help track the “active” nodes (which, as it sounds, are the nodes that are currently contributing to the cluster, as compared to the total number of nodes in the cluster). If the number of active nodes falls below a majority of the nodes in the cluster, quorum will be lost and the ordering service will stop processing blocks on the channel.
Troubleshooting¶
- The more stress you put on your nodes, the more you might have to change certain parameters. As with any system, computer or mechanical, stress can lead to a drag in performance. As we noted in the conceptual documentation, leader elections in Raft are triggered when follower nodes do not receive either a “heartbeat” messages or an “append” message that carries data from the leader for a certain amount of time. Because Raft nodes share the same communication layer across channels (this does not mean they share data — they do not!), if a Raft node is part of the consenter set in many channels, you might want to lengthen the amount of time it takes to trigger an election to avoid inadvertent leader elections.