The Complete Magazine on Open Source

Using the iSCSI Protocol to Provide Remote Block Storage

1.09K 0

The iSCSI protocol functions in a familiar client-server configuration. Client systems configure initiator software to send SCSI commands to remote server storage targets. The accessed iSCSI target appears on the client system as local, unformatted SCSI block devices, identical to devices connected with SCSI cabling, FC direct attached or FC switched fabric. Let’s take a look at how the iSCSI protocol can be used to provide remote block storage.

The Internet Small Computer System Interface (iSCSI) is a TCP/IP based protocol for emulating a SCSI high-performance local storage bus over IP networks, providing data transfer and management to remote block storage devices. As a storage area network (SAN) protocol, iSCSI extends SANs across local and wide area networks (LANs, WANs and the Internet) providing location-independent data storage retrieval with distributor servers and arrays.

The SCSI protocol suite provides the Command Descriptor Block (CDB) command set over a device bus communication protocol. The original SCSI topology used physical cabling with a 20 metre limitation for all devices per channel (cabled bus). Devices used unique numeric target IDs (0-7 or 0-15, with dual channels). Physical SCSI disks were made obsolete by popular implementation of fibre channel (FC), which retained the SCSI CDB command set but replaced the disk and bus communication with protocols for longer and faster optical cabling.

The iSCSI protocols also retain the CDB command set, performing bus communication between iSCSI systems that is encapsulated over standard TCP/IP. iSCSI servers emulate SCSI devices using files, logical volumes, or disks of any type as the underlying storage (backstore) presented as targets. An iSCSI service is typically implemented in software above either an operating system TCP/IP stack or a TCP offload engine (TOE)—a specialised Ethernet network interface card (NIC) that includes the TCP/IP network layers to increase performance. iSCSI can also be hardware implemented as a host bus adaptor (HBA) for better performance.

Enterprise-grade SANs require dedicated traffic infrastructure. The FC’s independent optical cabling and switches guarantee isolation. iSCSI should also be implemented on cabling that is independent of standard LAN traffic, since performance can fall due to bandwidth congestion on shared networks. Both the Internet and FC now offer copper and optical cabling options, allowing network consolidation combined with traffic classification.

Storage area network traffic is typically unencrypted, since physical server-to-storage cabling is normally enclosed within secure data centres. For WAN security, iSCSI and fibre channel over Ethernet (FCoE) can utilise Internet Protocol Security (IPSec), which is a protocol suite for securing IP network traffic. Select networking hardware (preferred NICs, TOEs and HBAs) can provide encryption. iSCSI offers the Challenge-Handshake Authentication Protocol (CHAP) user names and passwords as an authentication mechanism to limit connectivity between chosen initiators and targets.

Until recently, iSCSI was not considered an enterprise-grade storage option, primarily due to the use of slower 100Mbps and 1000Mbps Ethernet, compared to 4Gbps optical infrastructure. With the current 10Gbps or 40Gbps FC, bandwidth availability is now similar for both.

The use of iSCSI extends a SAN beyond the limit of local cabling, facilitating storage consolidation in local or remote data centres. Because iSCSI structures are logical, new storage allocations are made using only software configurations, without the need for additional cable or physical disks. iSCSI also simplifies data replication, migration and disaster recovery using multiple remote data centres.

iSCSI fundamentals

The iSCSI protocol functions in a familiar client-server configuration. Client systems configure initiator software to send SCSI commands to remote server storage targets. The accessed iSCSI target appears on the client system as local, unformatted SCSI block devices, identical to devices connected with SCSI cabling, FC direct attached, or FC switched fabric.

Figure 1: Targetcli interactive mode shown with ls command

Figure 2: A block has been created for the logical volume /dev/myvg/mylvm

Figure 3: IQN has been created for the target

iSCSI component terminology

1. Initiator: This is an iSCSI client, typically available as software but also implemented as iSCSI HBAs. Initiators must be given unique names.

2. Target: This is an iSCSI storage resource, configured for connection from an iSCSI server. Targets must be given unique names. A target provides one or more numbered block devices called logical units. An iSCSI server can provide many targets concurrently.

3. ACL: An Access Control List (entry) is an access restriction using the node IQN (commonly the iSCSI initiator name) to validate access permissions for an initiator.

4. Discovery: This involves querying a target server to list configured targets. Using targets requires additional access steps.

5. IQN: An iSCSI Qualified Name, this is a worldwide unique name used to identify both initiators and targets, in the following mandated naming format:[:optional_string]

IQN signifies that this name will use a domain as its identifier. YYYY-MM is the first month in which the domain name was owned. com.reversed.domain is the reversed domain name of the organisation creating the iSCSI name. optional_string is an optional, colon-prefixed string assigned by the domain owner as desired while remaining unique, worldwide. It may include colons to separate the organisation boundaries.

6. Login: This authenticates to a target or to a LUN to begin the client’s block device use.

7. LUN: A Logical Unit Number, it numbers block devices attached to and available through a target. One or more LUNs may be attached to a single target although, typically, a target provides only one LUN.

8. Node: This is an iSCSI initiator or iSCSI target identified by its IQN.

9. Portal: This is an IP address and port on a target or initiator used to establish connections. Some iSCSI implementations use the terms portal and node interchangeably.

10. TPG or Target Portal Group: This is the set of interface IP addresses and TCP ports to which a specific iSCSI target will listen. Target configuration can be added to TPG to coordinate settings for multiple LUNs.

iSCSI uses ACLs to perform LUN masking, managing the accessibility of appropriate targets and LUNs to initiators. Access to targets may also be limited with CHAP authentication. iSCSI ACLs are similar to FC’s use of a device’s worldwide numbers (WWNs) for soft zoning management restrictions, although FC switch level compulsory port restrictions (hard zoning) have no comparable iSCSI mechanisms. Ethernet VLANs could provide similar isolation security.

Unlike local block devices, iSCSI network-accessed block devices are discoverable from many remote initiators. Typical local file systems (ext4, XFS) do not support concurrent multi-system mounting, which can result in significant file system corruption. Clustered file systems resolve multiple-system access by the use of the Global File System (GFS2), which is designed to provide distributed file locking and concurrent multi-mode file system mounting. An attached iSCSI block device appears as a local SCSI block device (sdx) for use underneath a local file system, swap space or raw database installation.

Figure 4: Access Control List has been created for the client

Figure 5: LUN created for backstore

Figure 6: Portal has been created for listening IP address and port

An overview of the iSCSI target

In original SCSI protocol terminology, a target is a single connectible storage or output device uniquely identified on a SCSI bus. In iSCSI, in which the SCSI bus is emulated across an IP network, a target can be a dedicated physical device in a network attached storage (NAS) enclosure or an iSCSI software configured logical device on a networked storage server. Targets like HBAs and initiators are end points in SCSI bus communication, passing command descriptor blocks to request or provide storage transactions.

To provide access to the storage or output device, a target is configured with one or more logical unit numbers (LUNs). In iSCSI, LUNs appear as the target’s sequentially numbered disk drives, although targets typically have only one LUN. An initiator performs SCSI negotiations with a target to establish a connection to the LUN, which responds as an emulated SCSI disk block device, which can be used in raw form or formatted with a client supported file system.

Warning: Do not mount single system file systems to more than one system at a time. iSCSI allows shared target and LUN access from multiple initiator nodes, requiring the use of a cluster capable file system such as GFS2. Mounting standard file systems designed for local single system access (e.g., ext3, ext4, FAT32, XFS and ZFS) from more than one system concurrently will cause file system corruption.

iSCSI provides for LUN masking by using ACLs to restrict LUN accessibility to specific initiators, except when shared access is intended. ACLs can ensure that only a designated client node can log in to a specific target. On the target server, ACLs can be set at the TPG level to secure groups of LUNs, or can be set individually per LUN.

iSCSI target configuration

Let’s go through a demonstration of target server configuration.

targetcli is both a command line utility and an interactive shell in which to create, delete and configure iSCSI target components. Target stacks are grouped into a hierarchical tree of objects, allowing easy navigation and contextual configuration. Familiar Linux commands are used in this shell. These include cd, ls, pwd and set.

The targetcli shell also supports TAB completion. An administrator can use TAB to either complete partially-typed commands or view a list of acceptable keywords at the current location in a command.

Figure 7: Final configuration

Figure 8: Client IQN has been updated

Figure 9: The target has been discovered and the successful login has been given

Configuring a target server

Install targetcli by using the following command:

# yum –y install targetcli.

Run targetcli with no options to enter interactive mode:

# targetcli

Create backing storage (backstores). There are several types of backing storage, as listed below:

  • Block: A block device is defined on the server. It is a disk drive, disk partition, a logical volume, multi-path device, or any device files defined on the server that are of Type b.
  • fileio: This creates a file of a specified size in the file system of the server. This method is similar to using image files to be storage for virtual machine disk images.
  • pscsi: Physical ISCSI, this permits pass through to a physical SCSI device connected to the server. This backstore type is not typically used.
  • Ramdisk: This creates a ramdisk device of a specified size, in memory on the server. This type of storage will not store data persistently. When the server is rebooted, the ramdisk definition will return when the target is instantiated but all data would have been lost.

Note: For such cases, a logical volume has been created: /dev/myvg/mylvm, although a simple partition can also be created.

Examples include using an existing logical volume, a disk partition, and a new file at a specified size. The backstores get displayed as deactivated.

Next, create an IQN for the target. This step will also create a default TPG underneath the IQN.

An administrator can use create without specifying the IQN to create. targetcli will generate an IQN similar to the following:

Specifying the IQN value provides the administrator the ability to use a meaningful namespace for their IQNs.

Next, in the TPG, create an ACL for the client node to be used later. Because the global auto_add_mapped_luns parameter is set to true (default), any existing LUNs in the TPG are mapped to each ACL as it is created.

The ACL configures the target to only accept initiator connections from a client presenting as its initiator IQN, also known as the initiator name.

Next, in the TPG, create a LUN for each existing backstore. This step also activates each backstore. Because ACLs exist for the TPG, they will automatically be assigned to each LUN created.

A number of LUNs assigned to a target means that when the initiator connects to the target, it will receive n number of SCSI devices.

Now, still inside the TPG, create a portal configuration to designate the listening IP address and ports. Create a portal using the system’s public networks interface. Without specifying the TCP ports to be used, the portal creation will default to the standard iSCSI port 3260.

Next, view the entire configuration, then exit targetcli, which will automatically save upon exit. The resulting persistent configuration file is stored in JavaScript Object Notation (JSON) format.

Now add a port exemption to the default firewall for port 3260, the standard iSCSI port, as follows:

# firewall-cmd --add-port=3260/tcp

# firewall-cmd --add-port=3260/tcp --permanent.

In the final step, enable the target.service systemd unit. The target.service will recreate the target configuration from the json fill at boot. If this step is skipped, any configuration target will work until the machine is rebooted. However, after a reboot, no target will be offered by the server.

# systemctl enable target.


In addition to ACL node verification, password based authentication can be implemented. Authentication may be required during the iSCSI discovery phase, and it can be unidirectional or bi-directional.

CHAP authentication does not use strong encryption for the passing of credentials. While CHAP does offer an additional factor of authentication besides having a correctly configured initiator name configured in an ACL, it should not be considered as secure enough, if security of iSCSI data is of concern. Controlling the network side of the protocol is a better method to ensure security. Providing a dedicated, isolated network or vlans to pass the iSCSI traffic will be a more secure implementation of the protocol.

Figure 10: Newly available target

An introduction to iSCSI initiators

In Red Hat Enterprise Linux, an iSCSI initiator is typically implemented in software and functions similar to a hardware iSCSI HBA to access targets from a remote storage server. Using software based iSCSI initiators requires connecting to an existing Ethernet network of sufficient bandwidth to carry the expected storage traffic.

iSCSI can also be implemented using a hardware initiator that includes the required protocols in a dedicated host bus adaptor. iSCSI HBAs and TCP offload engines, which includes the TCP network stack on an Ethernet NIC, and move the processing of iSCSI or TCP overhead and Ethernet interrupts to hardware, easing the load on system CPUs.

Accessing iSCSI storage

1.) Prepare the client system to become an iSCSI initiator’s node by installing the initiator utilities, setting the unique iSCSI client name and starting the iSCSI client service.

Install the iscsi-initiators-utils RPM, if not already installed, as follows:

# yum –y install iscsi-initiators-utils
  • Create a unique iSCSI qualified name for the client initiators by modifying the InitiatorName setting in /etc/iscsi/initiatorname.iscsi. Use the client system as the optional string after the colon.
  • Enable and start the iSCSI client service, as follows:
# systemctl enable iscsi

# systemctl start iscsi

2.) Discover and log into the configured target from the iSCSI target server.

  • Discover and configure the iSCSI target provided by the iSCSI target server portal, as follows:
# iscsiadm –m discovery –t st –p 172.25.X.XX
  • Log into the presented iSCSI target, as follows:
# iscsiadm –m node –T –p 172.25.X.XX
  • Identify the newly available block created by the iSCSI target login, as follows:
# lsblk


Your email address will not be published.