Configuring EMC Unity Replication

Following on from the EMC Unity Setup Guide and EMC Unity Configuration Guide, we will walk through setting up replication between 2 Unity arrays. For Remote Office and Branch Offices replication can be configured between the Unity VSA and a physical Unity array in the datacentre.

Replication between storage devices provides data redundancy and protects against storage system failures. EMC Unity provides synchronous or asynchronous replication; synchronous replication is only available to physical arrays and can protect LUNs, Consistency Groups, and VMware VMFS datastores. Asynchronous replication is applicable to all products in the Unity range and can protect the storage resources listed previously, as well as File Systems, NAS Servers, and VMware NFS datastores. Replication can be configured within the same system or to a different system locally, or in a remote location. All EMC Unity systems are licensed for replication as standard. For more information on EMC Unity replication technologies see the Unity Replication White Paper.

Establishing Replication Connections

Before configuring replication a secure link between Unity systems must be established. All tasks are carried out using the HTML5 Unisphere web client. Browse to the IP or FQDN of either Unity system.

replication1

Under Data Protection on the left hand navigation pane select Replication. First we need to configure the interfaces to use for replication traffic, so click the Interfaces tab.

interfaces

Click the plus symbol (Create Replication Interface). Select the interface to use on each storage processor, if you have created link aggregation groups these are also listed. For assistance with creating link aggregation groups see the EMC Unity Configuration Guide. Enter the IP address to use for replication traffic for each storage processor, configure a VLAN ID if required, and click Ok.

replication2

The created replication interfaces will now be listed, you can also edit and delete replication interfaces from this tab. You need to configure the replication interfaces on both the source and destination system.

Next we setup a remote connection between the storage systems. From either device select the Connections tab.

connections

Enter the management IP address and credentials of the remote system, and the admin password for the local system. Select the replication type; in this example we will use Asynchronous. Click Ok.

replication3

A replication connection will now be established on both the local and remote systems. Once complete click Close.

Configuring Replication

Select the storage resource to configure replication, in this post we will replicate a file system. The NAS Server hosting the file system must be configured for replication. Browse to File under the Storage menu, open the NAS Servers tab and select the NAS server to replicate. Click Edit and select the Replication tab.

replication7

Select the replication mode and RPO (Recovery Point Objective) time. The replication destination is the remote system connection we established earlier. Click Next.

nas2

Select the storage pool and storage processor for the destination storage system, the NAS Server name will auto-populate. Any existing file systems stored on the NAS Server will be listed for replication. Click Next.

nas3

Review the summary page and click Finish and Close.

nas4

When creating new file systems they can be configured for replication on the Replication page of the create a file system wizard.

replication4

A destination file system is automatically created on the destination storage system.

replication5

Alternatively we can configure replication at a later date. To do this open the File Systems tab and select the file system to replicate, click Edit.

replication6

Select the Replication tab. Click Configure Replication.

replication7

The replication wizard will open. The replication session inherits the configuration from the NAS Server. Click Next.

replication8

A file system on the destination file system will automatically be created. Select the storage pool to use on the destination storage system, click Next.

replication9

Review the summary page and click Finish. The replication session will be established, once complete click Close.

replication10

We can confirm the replication status by going back into the properties of the file system and the Replication tab where the replication status will be displayed. The replication role of the file system on the source storage system is Source, on the remote system the file system role is Destination. We can also go back to the Replication page and open Sessions. The replication sessions will be listed.

properties

EMC VNXe Setup Guide

The VNXe is the most affordable hybrid and all-flash array across the EMC product range. Although the future potentially sits with the newly released Unity line, the VNXe remains a popular, flexible, and efficient storage solution for SMBs and ROBOs. This post will walk through the setup of an EMC VNXe device.

vnxe

Architecture

The VNXe 3200 is powered by dual Intel Xeon E5-2407 4-Core processors, providing up to 3x the performance of its 3150 predecessor. The Disk Processor Enclosure (DPE) leverages dual controllers and 6-Gb SAS back-end connectivity to deliver high levels of availability and efficiency, whilst lowering storage costs per I/O. Disk Array Enclosures (DAE) are added to scale out capacity up to 500 TB top end. There is concurrent support for NAS and SAN, with CIFS, SMB3, NFS, iSCSI and Fibre Channel (up to 8Gb) protocols, whilst the unit itself has a small datacentre footprint. For more information see the EMC VNXe Data Sheet.

Some considerations when creating storage pools; typically we want to configure less storage pools to reduce complexity and increase flexibility, however configuring multiple storage pools may be required if you want to separate workloads. Storage pools must maintain free capacity to operate, EMC recommend at least 10%. You will need to make design decisions based on your environment around storage pool capacities and configured RAID protection. The VNXe range offers multicore RAID 1/10/5/6 configured at storage pool level. EMC generally recommends smaller RAID widths as providing the best performance and availability, at the cost of slightly less usable capacity, e.g. for RAID 6 use 4+2 or 6+2 instead of 10+2 or 14+2.

VNXe arrays use the first 4 drives to store configuration information and critical system data, these are known as the system or vault drives and run from DPE Disk 0 through to DPE Disk 3. The system drives can be added to storage pools however usable capacity of system drives is reduced, therefore storage pools utilising system drives should use a smaller RAID width. For larger configurations with high drive counts EMC does not recommend using the system drives as heavy client workload may slow down management operations (does not apply to all-flash).

Requirements

In addition to the boxed system components you will need:

  • Cabinet vertical space of 2U for the DPE, and 2U for each optional 25-drive DAE.
  • 2 x Cat 5e or better GbE management connections.
  • Between 2 and 8 Cat 5e or better GbE or 10GbE data connections, or, between 2 and 8 Gb FC connections, depending on your chosen connection protocol.
  • A Windows based computer to run the initialisation and setup.
  • If you are unable to connect the Windows computer to the same subnet as the EMC VNXe then you will need a USB drive to configure the array with a management IP address.
  • Phillips screwdriver for installation.

Unboxing

The VNXe base comes with the following:

  • Disk Processor Enclosure (DPE) 2U component consisting of 12 x 3.5″ bays or 25 x 2.5″ bays.
  • Rail kit consisting of 2 adjustable rails and 10 screws, or 2 snap-in rails and 6 screws.
  • Accessory kit consisting of an mini-USB adaptor, cable ties, stickers, etc.
  • Front bezel for DPE.
  • Power cords.

Any additional disk shelves contain:

  • Disk Array Enclosure (DAE) 2U component consisting of 12 x 3.5″ bays or 25 x 2.5″ bays.
  • Rail kit consisting of 2 adjustable rails and 10 screws, or 2 snap-in rails and 6 screws.
  • Front bezel for DAE.
  • Power cords.
  • Mini-SAS and mini-SAS HD to mini-SAS cables.

Racking

EMC recommend installing the DPE at the bottom of the cabinet and installing any additional DAE’s above. The snap-in method is the most commonly used rail set and the one we will use here. For assistance with racking the adjustable rails see page 16 of the EMC VNXe Install Guide.

Locate the left and right markings on each rail. Align the 2U key tabs with the U-space in the rear rack channel. Push the key tabs and adaptors into the rear mounting holes until the spring clips snap into place. Round the front push in the spring clip and release once the rail is lined up with the mounting holes. Secure the rear of the rail using 1 x M5 screw on each side.

Slide the DPE into the rails until they click into the rear tabs on each rail. The tabs secure and support the rear of the enclosure, the front is secured using 2 x M5 screws on each side. Repeat the process for any additional DAEs.

 

racking

Cabling

First connect the 2 management ports to the switch, management ports have a LAN/management symbol above them. Do not use the service ports, service ports have a wrench/spanner symbol above them. Next plug in the cables for your chosen front end connectivity, i.e. Fibre Channel or Ethernet. Front end ports need to be connected and configured symmetrically across both storage processors to facilitate high availability. Furthermore you should use all front-end ports that are installed in the system, so that workload is spread across as many resources as possible. NAS ports should also be configured with LACP grouped per storage processor, to provide path redundancy and performance improvements.

If you have purchased additional DAEs then these need to be connected using the included SAS cables. There are 2 on-board 6Gb SAS ports in each storage processor in the DPE. When cabling DAEs to the DPE, balance them as evenly as possible across all available buses. The drives in the DPE are serviced by SAS Bus 0; therefore, the first DAE should be cabled to SAS Bus 1. Connect SP A SAS Port 1 to DAE 1 Link Controller Card (LCC) A (cable 1 in the image below). Connect SP B SAS Port 1 to DAE 1 LCC B (cable 2).

sas

The mini-SAS HD connectors are used for the DPE ports, the mini-SAS connectors are used for DAE ports. Mini-SAS to mini-SAS cables are used for cabling DAEs together. If you are attaching additional DAE’s see page 28 of the EMC VNXe Install Guide.

The power cables included with the array are colour coded with an intended use of: grey for Power Distribution Unit (PDU) A, black for PDU B. Once the array has power it will take approximately 10 – 15 minutes to power up. Finally, clip the front bezels into place and secure with the key included.

power

Setup

To access the web UI for setup we have a couple of options for automatic or manual IP addressing.

Automatic – if the array has access to network DHCP and DNS servers (with dynamic DNS enabled) then it will automatically be assigned an IP address. After power up if the SP Fault LED is solid blue then a management address has been assigned. This IP is dynamically added to DNS in the format of serialnumber.domain. If the SP Fault LED alternates between solid blue and flashing amber then a management address has not been assigned as the DHCP or DNS server could not be reached.

Manual – download and install the Connection Utility from EMC Downloads. The Connection Utility gives you two options; automatically detect unconfigured storage systems in the same subnet as your Windows client, or manually configure an IP in a configuration file for use with a USB flash drive which the array automatically reads.

Depending on how IP addressing has been assigned open a browser and enter the IP address manually configured, or the DNS entry (serialnumber.dnszone). Log in to Unisphere using the default credentials admin Password123#.

unisphere1

The Initial Configuration Wizard launches the first time you login. This self explanatory wizard guides you through the basic setup of the array, any settings you skip here can be configured later through the appropriate menus.

unisphere2

Once the configuration wizard is complete you will be returned to the home dashboard. It is recommended that the operating system is updated straight away. This can be achieved from the Settings drop down menu, and selecting Update software. Software can either be obtained online direct from the VNXe, or downloaded from EMC Downloads and then uploaded to the array. If you skipped the configuration wizard there are some basic configuration settings below to get you started.

First browse to the Management Settings page of the Settings drop down menu. Under the General tab we can configure the system name and management network settings. The Network tab features DNS settings, NTP settings, and remote logging.

network

To apply a license (.lic file provided by EMC) go to Settings, Manage Licenses; upload and install the license file. Also under Settings select Configure alerts, connect to EMC and configure SMTP and alert settings here.

alerts.png

It is recommended that physical network interfaces are  pooled together. To configure link aggregation browse to Settings, More Configuration, Advanced Configuration. Tick the Aggregation box.

linkaggregation

Storage pools are configured under System, Storage Pools. You will see 2 default pools; Hot Spare Pool and Unconfigured Disks. To configure the number of hot spares, or configure a storage pool and RAID group, select the appropriate pool and click Configure Disks. Follow the Disk Configuration Wizard.

diskconfig

To change the admin password at any time go to Settings, User administration. To enable SSH (optional) navigate to Settings, Service System and enter the service password. Select Enable SSH and click Execute service action.

You can now move on to configure the chosen protocol for the array, whether that be creating CIFS/NFS servers and shares through Settings, Manage Shared Folder Server Settings, or presenting iSCSI or FC storage through Hosts or Settings, iSCSI Server Settings. For further assistance with the VNXe GUI see EMC Unisphere for VNXe.

EMC Unity Configuration Guide

Following on from the EMC Unity Setup Guide this post will walk through the configuration of an EMC Unity array with iSCSI connectivity using the management web interface. Before beginning, ensure your Unity device is up to date by following the EMC Unity Update Guide. The EMC Unity is also available as a Virtual Storage Appliance.

unityhybrid

Architecture

The EMC Unity hybrid and all flash storage range implements an integrated architecture for block, file, and VMware VVOLs powered by the Intel E5-2600 processors. The Disk Processor Enclosure (DPE) leverages dual storage processors and full 12-Gb SAS back-end connectivity to deliver high levels of performance and efficiency. Disk Array Enclosures (DAE) are added to scale out capacity up to 3 PB top end. There is concurrent support for native NAS, iSCSI, and Fibre Channel protocols whilst the unit itself takes up less rack space than it’s competitors. Unity arrays can be managed from the HTML5 web client, or through the CloudIQ service, and offer a full range of enterprise storage features. For more information see the Unity platform white paper.

Some considerations when creating storage pools; typically we want to configure less storage pools to reduce complexity and increase flexibility. However configuring multiple storage pools may be required if you want to separate workloads for different I/O profiles or use FAST Cache. When sizing a storage pool remember that all data written to LUNs, file systems, and datastores is stored in the pool, as well as configuration information, change tracking, and snapshots. Storage pools must maintain free capacity to operate, EMC recommend at least 10%.

You will need to make design decisions based on your environment around storage pool capacities and configured RAID protection. The Unity range offers RAID 1/0, RAID 5, or RAID 6 configured at storage pool level. EMC generally recommends smaller RAID widths as providing the best performance and availability, at the cost of slightly less useable capacity, e.g. for RAID 6 use 4+2 or 6+2 instead of 10+2 or 14+2. Unity automatically reserves 1 out of every 30 drives of the same type for use as a hot spare, you can reduce the number of hot spare drives by decreasing the number of individual drive types.

Unity arrays use the first 4 drives to store configuration information and critical system data, these are known as the system drives and run from DPE Disk 0 through to DPE Disk 3. The system drives cannot be used as hot spares but can be added to storage pools in smaller configurations, if no other disks are available. The usable capacity of system drives is reduced by around 100 GB, therefore storage pools utilising system drives should use a smaller RAID width. For larger configurations with high drive counts EMC does not recommend using the system drives as heavy client workload may slow down management operations. This restriction does not apply to all-flash.

Configuration Settings

Browse to the management IP address of the Unity array configured during installation. If you have not changed the admin password the default login is admin Password123#.

The welcome dashboard gives an overview of health and capacity. Note the icons in the top right hand corner. The first symbol shows the overall system state, if there are no issues this will be a green tick. The second icon lists active jobs and the third any active alarms. Next is the settings menu, logged in user menu, and help.

icons

Let’s start by opening the settings menu using the gear icon. The Software and Licenses page lists the licensed enabled features. To install a license click Install License and upload the .lic file provided by EMC. You can also view system limits, install language packs, software updates, and disk firmware.

settings1

The Users and Groups page can be used to add local users or an LDAP identity source.

settings2

Use the Management page to configure NTP servers and DNS. The host name and management address can also be changed here if required as well as optional services such as Unisphere Central (centralised management), remote logging, and encryption.

settings3.png

The Storage Configuration page allows for configuration of FAST cache; FAST cache extends existing cache using enterprise flash drives to provide instant access to frequently used data. You can also view the spare disks in the system, but it’s best to come back to this after we’ve configured our storage pool.

settings4

Configure auto-support on the Support Configuration page by entering the support credentials and contact details. Make sure you use the EMC support account where the support contract is associated.

settings5

The Access page lists the iSCSI (Ethernet) and FC ports. Double click the port to view further details, all ports should be connected and green.

For Ethernet ports it is good practise to create link aggregation where more than one port is used for the same traffic, e.g. iSCSI data, or replication. Aggregating ports together pools the resources to create a highly available configuration, iSCSI or other services then use the port aggregation group to distribute I/O and provide redundancy. Select the first port for the group and click Link Aggregation, Create Link Aggregation. You can add or remove additional ports by selecting the port and clicking Link Aggregation, and Add to Link Aggregation or Remove from Link Aggregation.

settings6

Configure email alerts, and SNMP traps if required, using the Alerts page.

settings7

Next we’ll go through the menu options in the left hand navigation pane.

System

The System View page lists basic system information such as the model, serial number, and software version. If any hardware issues are detected they will be listed here.

unisphere1

The Performance page shows IOPS and bandwidth , you can also create I/O limits.

The Service page shows a number of service related tasks and logs, as well as any technical advisories issued by EMC. Auto-support functionality should already be enabled as we configured it earlier using the Support Configuration page of the Settings menu. The support contract will auto-populate once refreshed providing the correct support settings have been entered.

unisphere2

Access

The Hosts page allows for configuration of network hosts, such as Windows or Linux machines, for storage access. An individual host can be added, or a subnet or netgroup; to allow access to multiple hosts or network segments. The VMware page provides a single workflow for adding vCenter servers and ESXi host discovery. Virtual machine and VMDK information can also be imported.

For block storage resources you must register initiators using the Initiators tab. Initiators are servers initiating Fibre Channel or iSCSI sessions, and are identified by a unique World Wide Name (WWN) or iSCSI Qualified Name (IQN). The link between the initiator and the port on the storage system is called the initiator path; an initiator can be associated with multiple initiator paths. At this point for iSCSI paths to show up iSCSI interfaces must be configured on the Block page, see the Storage section above for further details. For FC paths the appropriate zoning on the FC switch must be complete for the initiator paths to be seen by the storage system.

Data Protection

The Data Protection section gives you two ways of protecting data on the array. The first is Snapshots; snapshots are used to create point in time copies of your data. There are 3 built in snapshot policies with different retention periods, or you can create your own by clicking the add symbol.

The second option is Replication, replication allows data to be copied to a different Unity array or Virtual Storage Appliance, on or off-site. To facilitate replication you must first create an interface by clicking the Interfaces tab and the add symbol. Chose an Ethernet interface, or link aggregation group, to use and configure the network settings. Next click the Connections tab and the add symbol. Enter the details of the remote Unity system to be a replication target and the connection mode; asynchronous replication, which takes an initial copy and then only updates with incremental (changed) data (recommended for most use cases) or synchronous replication, which takes full copies of the data at each replication interval. Finally configure replication on the storage resource you wish to replicate, as outlined under the Storage section below.

To configure replication see the Configuring EMC Unity Replication post.

data

Storage

Before using any disks in the system they must be allocated to a storage pool. When creating storage pools take into consideration the notes in the Architecture section above. To create a storage pool click Pools and the add symbol. Assign disks to the storage pool and select a RAID configuration, a storage pool can be made up of 2 performance tiers (types of disks) with different RAID types.

The Unity array is able to provide both block level and file level storage. For block level resources click Block and iSCSI Interfaces. Use the add button to add iSCSI interfaces for use with block level storage, chose the interface(s), storage pool, and configure the networking settings. LUNs can be created and mapped to a host, subnet, or netgroup using the LUNs tab.

For file level resources click File and NAS Servers, click the add symbol to create a NAS server, chose the interface(s), storage pool, configure the networking settings, and select the sharing protocols to use. It is good practise to create at least one NAS server each on SPA and SPB, and distribute resources evenly. Once your NAS servers are ready you can create File Systems, and then SMB shares or NFS Shares using the appropriate tabs.

During the creating of storage objects such as LUNs or file systems, you have the option to configure snapshots and replication. These features can also be configured at a later date by selecting the storage object and clicking the edit icon. Snapshots can be configured using one of the built in policies or creating your own under the Data Protection section above. When creating replication sessions you need to specify a replication schedule and target.

The VMware page can be used to configure VVOLs, read more about this at Configuring VVOLs with EMC Unity.

pools

Events and Support

The Events page lists all alerts from information to critical, as well as a record of all jobs that have been initiated on the device. The Support page provides links to documentation, training, and support.

support