N5 NETWORKING BEST PRACTICES
Transcription
N5 NETWORKING BEST PRACTICES
N5 NETWORKING BEST PRACTICES Table of Contents Nexgen N5 Networking ................................................................................................................................................................... 2 Overview of Storage Networking Best Practices ............................................................................................................. 2 Recommended Switch features for an iSCSI Network .................................................................................................. 2 Setting up the iSCSI Network for High Performance and High Availability ......................................................... 3 iSCSI SAN Topologies .................................................................................................................................................................. 7 Networking Best Practices Summary................................................................................................................................... 9 iSCSI Network Best Practices Summary ...........................................................................................................................10 General Application Setup for iSCSI Volume Access ....................................................................................................11 Overview of NexGen N5 Networking Options ................................................................................................................12 NexGen N5 Network Cabling Options ................................................................................................................................12 Network Cabling: 10GBT & 10GbE SFP+ ..........................................................................................................................13 Appendix A: NexGen N5 TCP/UDP Port Numbers ........................................................................................................14 NexGen N5 Networking Best Practices | 1 NEXGEN N5 NETWORKING Overview of Storage Networking Best Practices A high performance, high availability, storage network can be done in many ways. In the case of the NexGen N5 Hybrid Array, we recommend the following settings which are explained in further detail throughout this guide. Implement a fault-tolerant switch environment with multiple redundant switches. Implement MPIO at the host for high availability and performance. Implement a high performing network for data (10GbE SFP+ or 10GBT RJ45) Implement separate data and management subnets. Implement separate subnets or VLAN for dedicated data bandwidth. Set/verify the individual ports on the switch, host and storage to full-duplex mode. Enable Jumbo frames on all ports to maximize throughput. Flow Control should be enabled on all ports. Recommended Switch features for an iSCSI Network NexGen does not recommend a particular switch. However, the following illustrates the minimum set of switch capabilities to optimize the operation and performance of the N5. Item Description 10Gb Ethernet Support Full duplex 10Gb Ethernet operation ensures the minimum network latency and highest throughput. In order to ensure the network is not the bottleneck for application server performance, implement end-to-end 10Gb from application server to storage Non-blocking Backplane Optimal iSCSI data communications require switches with a backplane that has enough bandwidth to support fullduplex connectivity and full utilization, line rate for all ports at the same time. NexGen N5 Networking Best Practices | 2 Buffer Cache High performing iSCSI data communications require switches with at least 512KB of buffer space for each port. Therefore, a 48 port switch needs at least 24 MB of buffer cache. Jumbo Frames Support for Jumbo Frames ensures maximum performance for sequential read write workloads. Flow Control Flow Control ensures graceful communication between initiator and target. Switch Trunking Use link aggregation of two or more 10Gb or 40Gb links to connect multiple switches together. Switch Features Managed switch Layer 2 switching VLAN Spanning Tree Protocol Setting up the iSCSI Network for High Performance and High Availability The NexGen N5 Hybrid Flash Array is equipped with four 1GbE management ports; two on each storage processor. By default, the Management Port-1 is enabled for DHCP. Management Port-2 is configured from the factory with a static IP address (details below). For data ports, each N5 is set up with four 10GbE data ports; two on each processor. The following is summary of out-of-factory network configuration. Network Interface Mgmt. Port-2 Mgmt. Port-1 Data Ports 1 – 2 Lights Out Management Ports (Accessed via Mgmt. Port-1) Storage Processor A Enabled Static IP: 192.168.100.100 Mask: 255.255.255.0 Gateway: None Enabled DHCP Disabled No IP Configuration Disabled No IP Configuration Storage Processor B Enabled Static IP: 192.168.100.200 Mask: 255.255.255.0 Gateway: None Enabled DHCP Disabled No IP Configuration Disabled No IP Configuration Figure 1 – Default Network Settings Reference Appendix A – TCP/UDP Inbound Ports Used for Normal SAN Operations for additional information. NexGen N5 Networking Best Practices | 3 To setup the network interfaces on the NexGen N5 Hybrid Flash Array; navigate to the ‘Settings’ window and select the ‘Networking Addressing’ tab. Both the management and data ports on each storage processor in the system are listed in a dialog that allows you to select each port and configure it individually by clicking on the ‘Edit’ button. Figure 2 – Network Addressing screen After clicking on the ‘Edit’ button, the Edit Network Interface Configuration window appears. Set the Mode (DHCP, Disabled or Static), Address, Mask, Gateway, and Frame Size (1500 or 9000). Click ‘Save Config Changes’ to save the information. The Validate Configuration region of the window can be used to test network connectivity: NexGen N5 Networking Best Practices | 4 Figure 3 – Configuring a Management Port Click on the ‘Edit’ button next to the specific data port that you want to configure. Set the Mode (DHCP, Disabled or Static), Address, Mask, Gateway, and Frame Size (1500 or 9000). Click ‘Save Config Changes’ to save the information. NexGen N5 Networking Best Practices | 5 Figure 4 – Configuring a Data Port The Validate Configuration region of the window can be used to test network connectivity by clicking on the ‘Ping Address’ option. If successful, the RTT (round trip time) and Count (how many successful trips) are displayed: Figure 5 – Configuration Success Screen NexGen N5 Networking Best Practices | 6 iSCSI SAN Topologies There are two types of recommended iSCSI SAN topologies for use with the NexGen N5 Storage System: Single IP SAN network (single subnet or single vlan) Dual IP SAN networks (two subnets or two vlans) For the Single IP SAN network topology configuration, all Data Ports on both storage processors will be configured with IP addresses on the same IP SAN network. Application servers are also connected to the single IP SAN network and volumes are connected via iSCSI with either a single session or multiple sessions (MPIO). The following logical network diagrams depict how to setup single and dual IP SAN network configurations with the NexGen N5 Storage System. Figure 6 – Single IP SAN Network Configuration NexGen N5 Networking Best Practices | 7 Figure 7 – Dual IP SAN Network Configuration For optimal performance and availability in the application servers, it is recommended that MPIO be used. The best MPIO option is to configure two or more paths from the application server to the NexGen N5 Hybrid Flash Array in an Active-Active MPIO mode. The minimum number of MPIO paths should be two paths across two NICS in the application server connected to both storage processors. The recommended physical network topology for the IP SAN network(s) is to have redundant physical paths for the volume connections made from the application servers to the storage. This is easily done with multiple switches in the environment connected to multiple NICs in the application server and storage. Below is an example of a two physical switch configuration. The switches are trunked together so that the Single IP SAN network can span both switches. If a Dual IP SAN network is implemented, the trunk links between the switches are not necessary unless other VLANs are being spanned across switches. Figure 8 – Two Switch Physical Topology in Single IP SAN Network Configuration NexGen N5 Networking Best Practices | 8 Networking Best Practices Summary Item Logical Network Topologies Physical Network Topologies MPIO Jumbo Frames Flow Control Link Aggregation Technology (LACP, MC-LAG, Virtual Port Channels, etc.) NexGen N5 Networking Best Practices | 9 Best Practice Implement a Single IP SAN or Dual IP SAN network topology. Implement redundant switches. In a Single IP SAN implementation, the two switches must be trunked together. In a Dual IP SAN implementation, the logically separated networks should also be on physically different switches for redundancy. Utilize core switch topologies that utilize multiple high bandwidth low-latency trunk links without requiring use of Spanning Tree. Use host-based MPIO. Setup at least two paths to a volume. Use MPIO ALUA with a Round-Robin path selection policy. Use caution when implementing Jumbo Frames. Configure Jumbo Frame support on all switches between the application servers and the NexGen N5. Enable Jumbo Frames on all application servers and storage network interfaces that are connected to the IP SAN network(s). There is no need to configure Jumbo Frames on the Management Ports on the NexGen N5 unless you are using replication. Proper configuration of Jumbo Frames should yield anywhere from 0-20% performance benefit depending on the workload. Misconfiguration of Jumbo Frames can result in a negative performance impact. Enable Flow Control on all switches and switch ports connected to the IP SAN network(s). Flow Control is good practice for optimal iSCSI performance on 10 Gigabit Ethernet networks. Enable tx/rx flow control in the application server and VM environment NIC configurations if not on by default. Flow Control is enabled by default on the NexGen N5 NICs. There is no need to configure this. Link aggregation technologies cannot be used with the NexGen N5 network ports for management or data. Use MPIO at the host to provide path redundancy and improved performance. Link aggregation technologies (LACP, MC-LAG, VTP) should be used for trunking switches together in order to achieve connection reliability and higher performance (bandwidth). Consult the switch vendor documentation on proper setup. iSCSI Network Best Practices Summary Item Data Ports and Management Ports Flow Control Jumbo Frames NexGen N5 Networking Best Practices | 10 Best Practice Management and Data ports should be configured on separate networks. Data Ports should use static IP addresses on a dedicated IP SAN network, ideally isolated from all other traffic. Network ports which are not being used should be set to Disabled. Flow Control should be enabled on all switches and ports that will carry iSCSI traffic when using 10 Gigabit Ethernet hosts and storage. Enable tx/rx flow control in the application server NIC configurations if not on by default. Flow Control is enabled by default on the NexGen N5 NICs. There is no need to configure this. Data Ports should be enabled for Jumbo Frames (9000 Frame Size) only if all switches, switch ports and application servers connected to the IP SAN network are configured for Jumbo Frames. Misconfiguration of Jumbo Frames can have a negative performance impact. Configure Jumbo Frame support on the switches first, followed by the application server iSCSI network interfaces, etc. There is no need to configure Jumbo Frames on the Management Ports on the NexGen N5. General Application Setup for iSCSI Volume Access The NexGen N5 Hybrid Flash Array presents data volumes for access by application servers via the iSCSI protocol with Asymmetric Logical Unit Access (ALUA)-enabled. There are several general setup instructions for connecting volumes to all iSCSI initiators available for the most common operating systems. Best Practice Use multiple iSCSI discovery addresses (portals) for high availability Use MPIO for volume connectivity high availability Use Round-Robin MPIO policy for optimal host connectivity performance Use iqn (iSCSI Qualified Name) based iSCSI Security for volume access. NexGen N5 Networking Best Practices | 11 Details Configure at least two discovery addresses in the application server’s iSCSI initiator for accessing volumes on the N5 Storage System. For high availability of iSCSI discovery on the iSCSI initiator, specify at least two (preferably four) iSCSI discovery IP addresses that correspond to data port IP addresses on each Storage Processor in the NexGen N5. All volumes on the NexGen N5 are advertised for discovery on all Data Ports. Volumes are NOT advertised for discovery on the Management Ports. If the host operating system and iSCSI initiator support MPIO, configure at least two iSCSI sessions per volume. Create one session connected to Storage Processor-A and the other session connected to Storage ProcessorB. If the host operating system and iSCSI initiator support MPIO, use MPIO ALUA with a RoundRobin path selection policy. Each iSCSI initiator will have a unique iqn on the iSCSI network. Specify the iqn(s) of the application servers in the Host Access Group and assign volumes to that group. The ‘Allow-all Access’ Host Access Group is not recommended for use on production server volumes. Overview of NexGen N5 Networking Options Model Management Ports Redundant Data Port NICs N5-200 4x 1Gb RJ45 4x 1/10GbE SFP+ -or- 4x 1/10GbaseT RJ45 N5-300 4x 1Gb RJ45 4x 1/10GbE SFP+ -or- 4x 1/10GbaseT RJ45 N5-500 4x 1Gb RJ45 4x 1/10GbE SFP+ -or- 4x 1/10GbaseT RJ45 N5-1000 4x 1Gb RJ45 4x 1/10GbE SFP+ -or- 4x 1/10GbaseT RJ45 NexGen N5 Network Cabling Options Port # of Cable Type NIC type What to buy with N5 Ports 1GbE Management 4 Cat6 or better (RJ45) N/A Cables only 1GbE Data 4 Cat6 or better (RJ45) SFP+ GBIC SKU plus cables 1GBT Data 4 Cat6 or better (RJ45) GBase-T Cables only 10GbE Data 4 SFP+ Twinax SFP+ Cables only 10GbE Data 4 SFP+ Optical SFP+ Cables with optic (OM3 or better) 10GBT Data 4 Cat6a or better (RJ45) NexGen N5 Networking Best Practices | 12 adapters/modules GBase-T Cables only Network Cabling: 10GBT & 10GbE SFP+ Type 10G Base-T 10Gb SFP+ SR Optical 10Gb SFP+ SR Twinax Passive 10Gb SFP+ SR Twinax Active Speed 10 Gb/sec 10 Gb/sec 10 Gb/sec 10 Gb/sec Figure 9 – 10GBT Cat7 Cable Figure 10 – 10GbE SFP+ Twinax Cable NexGen N5 Networking Best Practices | 13 Max Distance 100m 300m 10m 25m Latency per Link 2.6µs 0.3µs 0.3µs 0.3µs Figure 11 – 10GbE SFP+ Optical Cable Figure 12 – 10GbE SR Optical Gbic Power 2.7W 0.7W 0.7W 0.7W Appendix A: NexGen N5 TCP/UDP Port Numbers IP Protocol Port(s) Name Description TCP 22 SSH TCP 80 HTTP UDP 123 NTP UDP 161 SNMP TCP/UDP 162 SNMPTRAP TCP 443 HTTPS TCP TCP 860 3260 iSCSI iSCSI target Secure Shell access for SAN support only. Not meant for normal day-today operations. Permit intermediate network elements to improve or enable communications between clients and servers. Network Time Protocol used for time synchronization. Simple Network Management Protocol Simple Network Management Protocol Trap HTTPS (Hypertext Transfer Protocol over SSL/TLS) iSCSI System port iSCSI port NexGen N5 Networking Best Practices | 14