NVMe Multipath

Introduction

The NVMe bdev module supports two modes: failover and multipath. In failover mode, only one active connection is maintained and alternate paths are connected only during the switch-over. This can lead to delays and failed I/O reported to upper layers, but it does reduce the number of active connections at any given time. In multipath, active connections are maintained for every path and used based on a policy of either active-passive or active-active. The multipath mode also supports Asymmetric Namespace Access (ANA) and uses that to make policy decisions.

Design

Multipath Mode

A user may establish connections on multiple independent paths to the same NVMe-oF subsystem for NVMe bdevs by calling the bdev_nvme_attach_controller RPC multiple times with the same NVMe bdev controller name. Additionally, the multipath parameter for this RPC must be set to "multipath" when connecting the second or later paths.

For each path created by the bdev_nvme_attach_controller RPC, an NVMe-oF controller is created. Then the set of namespaces presented by that controller are discovered. For each namespace found, the NVMe bdev module attempts to match it with an existing NVMe bdev. If it finds a match, it adds the given namespace as an alternate path. If it does not find a match, it creates a new NVMe bdev.

I/O and admin qpairs are necessary to access an NVMe-oF controller. A single admin qpair is created and is shared by all SPDK threads. To submit I/O without taking locks, for each SPDK thread, an I/O qpair is created as a dynamic context of an I/O channel for an NVMe-oF controller.

For each SPDK thread, the NVMe bdev module creates an I/O channel for an NVMe bdev and provides it to the upper layer. The I/O channel for the NVMe bdev has an I/O path for each namespace. I/O path is an additional abstraction to submit I/O to a namespace, and consists of an I/O qpair context and a namespace. If an NVMe bdev has multiple namespaces, an I/O channel for the NVMe bdev has a list of multiple I/O paths. The I/O channel for the NVMe bdev has a retry I/O list and has a path selection policy.

Path Error Recovery

If the NVMe driver detects an error on a qpair, it disconnects the qpair and notifies the error to the NVMe bdev module. Then the NVMe bdev module starts resetting the corresponding NVMe-oF controller. The NVMe-oF controller reset consists of the following steps: 1) disconnect and delete all I/O qpairs, 2) disconnect admin qpair, 3) connect admin qpair, 4) configure the NVMe-oF controller, and 5) create and connect all I/O qpairs.

If the step 3, 4, or 5 fails, the reset reverts to the step 3 and then it is retried after reconnect_delay_sec seconds. Then the NVMe-oF controller is deleted automatically if it is not recovered within ctrlr_loss_timeout_sec seconds. If ctrlr_loss_timeout_sec is -1, it retries indefinitely.

By default, error detection on a qpair is very slow for TCP and RDMA transports. For fast error detection, a global option, transport_ack_timeout, is useful.

Path Selection

Multipath mode supports two path selection policies, active-passive or active-active.

For both path selection policies, only ANA optimal I/O paths are used unless there are no ANA optimal I/O paths available.

For active-passive policy, each I/O channel for an NVMe bdev has a cache to store the first found I/O path which is connected and optimal from ANA and use it for I/O submission. Some users may want to specify the preferred I/O path manually. They can dynamically set the preferred I/O path using the bdev_nvme_set_preferred_path RPC. Such assignment is realized naturally by moving the I/O path to the head of the I/O path list. By default, if the preferred I/O path is restored, failback to it is done automatically. The automatic failback can be disabled by a global option disable_auto_failback. In this case, the bdev_nvme_set_preferred_path RPC can be used to do manual failback.

The active-active policy uses the round-robin algorithm and submits an I/O to each I/O path in circular order.

I/O Retry

The NVMe bdev module has a global option, bdev_retry_count, to control the number of retries when an I/O is returned with error. Each I/O has a retry count. If the retry count of an I/O is less than the bdev_retry_count, the I/O is allowed to retry and the retry count is incremented.

NOTE: The bdev_retry_count is not directly used but is required to be non-zero for the process of multipath mode failing over to a different path because the retry count is checked first always when an I/O is returned with error.

Each I/O has a timer to schedule an I/O retry at a particular time in the future. Each I/O channel for an NVMe bdev has a sorted I/O retry list. Retried I/Os are inserted into the I/O retry list.

If an I/O is returned with error, the I/O completion handler in the NVMe bdev module executes the following steps:

  1. If the DNR (Do Not Retry) bit is set or the retry count exceeds the limit, then complete the I/O with the returned error.
  2. If the error is a path error, insert the I/O to the I/O retry list with no delay.
  3. Otherwise, insert the I/O to the I/O retry list with the delay reported by the CRD (Command Retry Delay).

Then the I/O retry poller is scheduled to the closest expiration. If there is no retried I/O, the I/O retry poller is stopped.

When submitting an I/O, there may be no available I/O path. If there is any I/O path which is recovering, the I/O is inserted to the I/O retry list with one second delay. This may result in queueing many I/Os indefinitely. To avoid such indefinite queueing, per NVMe-oF controller option, fast_io_fail_timeout_sec, is added. If the corresponding NVMe-oF controller is not recovered within fast_io_fail_timeout_sec seconds, the I/O is not queued to wait the recovery but returned with an I/O error to the upper layer.

Asymmetric Namespace Accesses (ANA) Handling

If an I/O is returned with an ANA error or an ANA change notice event is received, the ANA log page may be changed. In this case, the NVMe bdev module reads the ANA log page to check the ANA state changes.

As described before, only ANA optimal I/O paths will be used unless there are no ANA optimal paths available.

If an I/O path is in ANA transition, i.e., its namespace reports the ANA inaccessible state or the ANA change state, the NVMe bdev module queues I/Os to wait until the namespace becomes accessible again. The ANA transition should end within the ANATT (ANA Transition Time) seconds. If the namespace does not report the ANA optimized state or the ANA accessible state within the ANATT seconds, I/Os are returned with an I/O error to the upper layer.

I/O Timeout

The NVMe driver supports I/O timeout for submitted I/Os. The NVMe bdev module provides three actions when an I/O timeout is notified from the NVMe driver, ABORT, RESET, or NONE. Users can choose one of the actions as a global option, action_on_timeout. Users can set different timeout values for I/O commands and admin commands by global options, timeout_us and timeout_admin_us.

For ABORT, the NVMe bdev module tries aborting the timed out I/O, and if failed, it starts the NVMe-oF controller reset. For RESET, the NVMe bdev module starts the NVMe-oF controller reset.

Usage

The following is an example to attach two NVMe-oF controllers and aggregate these into a single NVMe bdev controller Nvme0.

./scripts/rpc.py bdev_nvme_attach_controller -b Nvme0 -t rdma -a 192.168.100.8 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -l -1 -o 20
./scripts/rpc.py bdev_nvme_attach_controller -b Nvme0 -t rdma -a 192.168.100.9 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -l -1 -o 20 -x multipath

In this example, if these two NVMe-oF controllers have a shared namespace whose namespace ID is 1, a single NVMe bdev Nvme0n1 is created. For the NVMe bdev module, the default value of bdev_retry_count is 3 and I/O retry is enabled by default. ctrlr_loss_timeout_sec is set to -1 and reconnect_delay_sec is set to 20. Hence, NVMe-oF controller reconnect will be retried once per 20 seconds until it succeeds.

To confirm if multipath is configured correctly, two RPCs, bdev_get_bdevs and bdev_nvme_get_controllers are available.

./scripts/rpc.py bdev_get_bdevs -b Nvme0n1
./scripts/rpc.py bdev_nvme_get_controllers -n Nvme0

To monitor the current multipath state, a RPC bdev_nvme_get_io_paths are available.

./scripts/rpc.py bdev_nvme_get_io_paths -n Nvme0n1

Limitations

SPDK NVMe multipath is transport protocol independent. Heterogeneous multipath configuration (e.g., TCP and RDMA) is supported. However, in this type of configuration, memory domain is not available yet because memory domain is supported only by the RDMA transport now.

The RPCs, bdev_get_iostat and bdev_nvme_get_transport_statistics display I/O statistics but both are not aware of multipath.