Overlay Transport Virtualization
Transcription
Overlay Transport Virtualization
Overlay Transport Virtualization Brian Farnham Technical Marketing Engineer BRKDCT-2049 Agenda • Introduction • Distributed Data Centers: Goals and Challenges • OTV Architecture Principles • Principles of Interconnecting Networks at Layer-2 • OTV New Features OTV – Overlay Transport Virtualization Simplifying Data Center Interconnect Any Workload Anytime Anywhere Session Objectives • The main goals of this session are: • This session features a detailed analysis of the architectural aspects and deployment benefits behind OTV • The attendees will learn how OTV is aimed at providing Layer 2 connectivity beyond the Layer 3 boundary while maintaining the failure containment and operational simplicity that the Layer 3 boundary provides • The attendees will get a deep knowledge of how the OTV control-plane and data-plane work to provide the VLAN extension Session Non-objectives • This session does not include: • In depth discussion of Path Optimization technologies (DNS, LISP, etc.) • Storage extension considerations associated to DCI deployments • Workload mobility application specific deployment considerations Related Cisco Live Events Session-ID BRKDCT-2131 BRKDCT-3060 BRKDCT-3103 Session Name Mobility and Virtualization in the Data Center with LISP and OTV Deployment Considerations with Interconnecting Data Centers Advanced OTV – Configure, Verify and Troubleshoot OTV Agenda • Introduction • Distributed Data Centers: Goals and Challenges • OTV Architecture Principles • Principles of Interconnecting Networks at Layer-2 • OTV New Features Distributed Data Centers Goals • Ensure business continuity • Distributed applications • Seamless workload mobility • Maximize compute resources Data Center Interconnect Traditional Layer 2 Extensions EoMPLS • VSS & vPC or FabricPath • Applies easily for dual site interconnection • Over dark fiber or protected D-WDM • Easy crypto using end-to-end 802.1AE Ethernet IP • OTV – Overlay Transport Virtualization VPLS • MAC in IP Dark Fiber • MPLS EoMPLS & VPLS & A-VPLS & H-VPLS • PE style • Multi-tenants • Most deployed today Challenges in Traditional Layer 2 VPNs Flooding Behavior Pseudo-wire Maintenance Multi-Homing - Unknown Unicast for MAC propagation - Unicast Flooding reaches all sites - Full mesh of Pseudo-wire is complex - Head-End replication is a common problem - Requires additional Protocols & extends STP - Malfunctions impacts multiple sites Technology Pillars No Pseudo-Wire State Maintenance Optimal Multicast Replication Dynamic Encapsulation Multipoint Connectivity Point-to-Cloud Model Preserve Failure Boundary Built-in Loop Prevention Protocol Learning Automated Multi-Homing Site Independence OTV – Overlay Transport Virtualization Simplifying Data Center Interconnect Any Workload Anytime • Nexus 7000 First platform to support OTV (since 5.0 NXOS Release) • ASR 1000 Now also supporting OTV (since 3.5 XE Release) Anywhere Agenda • Introduction • Distributed Data Centers: Goals and Challenges • • • • • • • Control Plane and Data Plane Failure Isolation Multi-homing Mobility L2 Multicast Forwarding QoS and Scalability Path Optimization • OTV Architecture Principles • Principles of Interconnecting Networks at Layer-2 • OTV New Features Terminology OTV Devices and Interfaces • OTV Edge Device Edge Device • Performs all OTV functionality • Usually located at the Aggregation Layer or at the Core Layer • Support for multiple OTV Edge Devices (multi-homing) in the same site • OTV Edge Device Internal Interface • • • • • Site facing Interfaces of the Edge Devices Carry VLANs extended through OTV Regular Layer 2 interfaces No OTV configuration required Supports IPv4 & IPv6 Core Device Aggregation Device OTV Internal Interfaces OTV Internal Interface OTV Join Interface OTV Overlay Interface Terminology OTV Devices and Interfaces • Join Interface • • • • • • OTV Join Interface Overlay Interface One of the uplink of the Edge Device Point-to-point routed interface (physical interface, sub-interface or port-channel supported) Used to physically “join” the Overlay network No OTV specific configuration required IPv4 only OTV Edge Device Core Device Aggregation Device Overlay Interface • Virtual interface with most of the OTV configuration • Logical multi-access multicast-capable interface • Encapsulates Layer 2 frames in IP unicast or multicast OTV Internal Interface OTV Join Interface OTV Overlay Interface OTV Control Plane Building the MAC Tables No unknown unicast flooding (selective unicast flooding in 6.2) Control Plane Learning with proactive MAC advertisement Background process with no specific configuration IS-IS used between OTV Edge Devices MAC Addresses Advertisements OTV IP A OTV IP B East West IP C OTV South OTV Control Plane Neighbor Discovery and Adjacency Formation Before any MAC address can be advertised the OTV Edge Devices must: ‒ ‒ Discover each other Build a neighbor relationship with each other Neighbor Relationship built over a transport infrastructure: ‒ ‒ Multicast-enabled (all shipping releases) Unicast-only (from NX-OS release 5.2 & IOS-XE 3.9) OTV Control Plane Neighbor Discovery (over Multicast Transport) Multicast-enable Transport OTV Control Plane OTV OTV OTV Control Plane IP A IP B East West End Result Mechanism • Edge Devices (EDs) join an multicast group in the transport, as they were hosts (no PIM on EDs) • OTV hellos and updates are encapsulated in the multicast group • Adjacencies are maintained over the multicast group • A single update reaches all neighbors OTV Control Plane (Multicast Transport) Neighbor IP Addr West IP A Neighbor IP Addr 3 OTV Hello OTV Control Plane 4 OTV Hello IP A G 7 OTV OTV OTV Hello OTV Control Plane Multicast-enabled Transport IP A West IGMP Join G IP B OTV Hello OTV Hello IP A G East 6 Encap Decap IGMP Join G 1 OTV Hello OTV Hello All edge devices join OTV control-group G IP A G IP A G 5 Transport natively replicates multicast to all OIFs 2 IGMP Join G Multicast state for group G established throughout transport Decap 6 IP C OTV OTVHello Hello OTV IP A G OTV Control Plane 7 OTV Hello Neighbor IP Addr West IP A South OTV Control Plane (Multicast Transport) Neighbor IP Addr South IP C OTV Hello Bidirectional adjacency formed 5 OTV Control Plane 5 OTV OTV IP C G OTV OTV Hello Hello Neighbor IP Addr West IP A South IP C OTV Control Plane Multicast-enabled Transport IP A IP B OTV OTV Hello Hello East West 4 OTV Hello Decap Decap 3 OTV Hello OTV Hello IP C G IP C G Encap The South Site creates its hello with West’s address in the TLV 2 IP C OTV OTV Hello IP C G OTV Control Plane 1 OTV Hello Neighbor IP Addr West IP A South 4 IP C G OTV Control Plane MAC Advertisements (over Multicast Transport) Craft OTV 2 update with new MACs VLAN 100 Update A OTV West MAC Table VLAN 100 100 101 100 102 MAC MAC A MAC B MAC C Multicast-enabled Transport IP A G Update A East MAC Table 5 Encap Decap 4 Update A Update A IP A G IP A G VLAN 100 101 102 MAC MAC A MAC B MAC C Add MACs learned through OTV 1 New MACs learned in VLANs that are OTV extended Decap 7 5 OTV Update UpdateAA 6 VLAN 100 MAC IF MAC A 100 MACA B Update 100 MAC C IP A IP A IP A IP A G MAC Table South IP A IP A IP A IP A G Update UpdateAA 3 IF e1/1 e1/1 e1/1 MAC IF MAC A 100 A MAC B Update 100 MAC C OTV VLAN 100 100 101 100 102 MAC MAC A MAC B MAC C IF IP A IP A IP A 7 Add MACs learned through OTV IF IP A IP A IP A 6 Multicast Transport OTV Control and Data Plane over Multicast Transport • Use a High-Available Multicast RendezVous Point (RP) configuration ‒ • Requirements to Control Plane ‒ • PIM Anycast (RFC4610) or MSDP (Multicast Source Discovery Protocol) PIM Any-Source-Multicast (ASM) Sparse-Mode Requirements to Data Plane ‒ PIM Source-Specific-Multicast (SSM) or BiDir Example: Multicast for OTV on Nexus 7000 feature pim ! interface loopback 0 ip pim spare-mode ip address 192.168.1.100/32 ! interface loopback 1 ip pim sparse-mode ip address 10.254.254.n1-x/32 ! ip pim rp-address 192.168.1.100 ip pim anycast-rp 192.168.1.100 ip pim anycast-rp 192.168.1.100 ip pim ssm range 232.239.1.0/24 ! interface port-channel1 # This Interface peers with the ip igmp version3 group-list 239.1.1.1 10.254.254.n1 10.254.254.n2 OTV Join Interface * “n” in the last Octet reflects a unique IP address per Router joining the PIM Anycast Group Release 5.2 and above OTV Control Plane Neighbor Discovery (Unicast-only Transport) • Ideal for connecting a small number of sites • With a higher number of sites a multicast transport is the best choice Unicast-only Transport OTV Control Plane OTV OTV OTV Control Plane IP A IP B East West Mechanism End Result • Edge Devices (EDs) register with an “Adjacency Server” ED • Neighbor Discovery is automated by the “Adjacency Server” • EDs receive a full list of Neighbors (oNL) from the AS • All signaling must be replicated for each neighbor • OTV hellos and updates are encapsulated in IP and unicast to each neighbor • Data traffic must also be replicated at the head-end OTV Control Plane CLI Verification Establishment of control plane adjacencies between OTV Edge Devices (multicast or unicast transport): dc1-agg-7k1# show otv adjacency Overlay Adjacency database Overlay-Interface Overlay100 Hostname System-ID dc2-agg-7k1 001b.54c2.efc2 dc1-agg-7k2 001b.54c2.e1c3 dc2-agg-7k2 001b.54c2.e142 : Dest Addr 20.11.23.2 20.12.23.2 20.22.23.2 Up Time 15:08:53 15:43:27 14:49:11 Adj-State UP UP UP Unicast MAC reachability information: dc1-agg-7k1# show otv route OTV Unicast MAC Routing Table For Overlay100 VLAN MAC-Address Metric Uptime Owner ---- -------------- ------ -------- --------2001 0000.0c07.ac01 1 3d15h site 2001 0000.1641.d70e 1 3d15h site 2001 0000.49f3.88ff 42 2d22h overlay 2001 0000.49f3.8900 42 2d22h overlay Next-hop(s) ----------Ethernet1/1 Ethernet1/2 dc2-agg-7k1 dc2-agg-7k2 Local Site MAC Remote Site MAC OTV Data Plane Inter-Site Packet Flow 4 Transport Infrastructure MAC TABLE VLAN 100 2 Layer 2 Lookup MAC IF MAC 1 Eth 2 100 OTV MAC 2 Eth 1 100 MAC 3 IP B 100 MAC 4 IP B MAC 1 MAC 3 1 Server 1 IP A Decap 5 IP B 3 OTV MAC TABLE West Site IP A IP B MAC IF 100 MAC 1 IP A OTV Encap MAC 1 MAC 3 VLAN MAC MAC11 MAC MAC33 East Site IP A IP B OTV 100 MAC 2 IP A 100 MAC 3 Eth 3 100 MAC 4 Eth 4 MAC 1 MAC 3 Server 3 7 6 Layer 2 Lookup OTV Data Plane Encapsulation • 42 Bytes overhead to the packet IP MTU size (IPv4 packet) • Outer IP + OTV Shim - Original L2 Header (w/out the .1Q header) • 802.1Q header is removed and the VLAN field copied over to the OTV shim header • Outer OTV shim header contains VLAN, overlay number, etc. 802.1Q header removed • Consider Jumbo MTU Sizing 802.1Q DMAC SMAC Ether Type 6B 6B 2B * The 4 Bytes of .1Q header have already been removed IP Header 20B 802.1Q DMAC SMAC OTV Shim 8B L2 Header 14B* Ether Type CRC Payload Original L2 Frame 20B + 8B + 14B* = 42 Bytes of total overhead 4B Agenda • Introduction • Distributed Data Centers: Goals and Challenges • • • • • • • Control Plane and Data Plane Failure Isolation Multi-homing Mobility L2 Multicast Forwarding QoS and Scalability Path Optimization • OTV Architecture Principles • Principles of Interconnecting Networks at Layer-2 • OTV New Features Spanning-Tree and OTV Site Independence • Site transparency: no changes to the STP topology • Total isolation of the STP domain • Default behavior: no configuration is required OTV • BPDUs sent and received ONLY on Internal Interfaces The BPDUs stop here OTV L3 The BPDUs L2 stop here Unknown Unicast and OTV No Longer Unknown Unicast Storms Across the DCI MAC TABLE VLAN OTV MAC IF 100 MAC 1 Eth1 L3 100 MAC 2 IP B L2 - - - OTV MAC 1 MAC 3 • No requirements to forward unknown unicast frames • Assumption: end-host are not silent or uni-directional • Default behavior: no configuration is required No MAC 3 in the MAC Table New Release 6.2 Unknown Unicast and OTV Selective Unicast Flooding • Some Application requirement to forward unknown unicast frames • Selective Unicast Flooding can be enabled per mac address • Default behavior: no unknown unicast forwarding Enable Flooding for MAC .0101 OTV-a # conf Enter configuration commands, one per line. End with CNTL/Z OTV-a(config)# otv flood mac 0000.2102.1111 vlan 172 Unknown Unicast OTV OTV MAC State IF .0000 Blk Overlay1 L3 .0101 Blk Overlay1 L2 .1111 Fwd Overlay1 MAC 1 MAC 3 VLAN 100 MAC 6 MAC 7 VLAN 102 Controlling ARP Traffic New Release 6.1 ARP Neighbor-Discovery (ND) Cache • ARP cache maintained in Edge Device by snooping ARP replies • First ARP request is broadcasted to all sites. Subsequent ARP requests are replied by local Edge Device • Timeout can be adjusted (as per NX-OS 6.1(1)) • Drastic reduction of ARP traffic on DCI • ARP spoofing can be disabled • IPv4 only feature • Default behavior: no configuration is required OTV-a(config)# interface overlay 1 OTV-a(config-if-overlay)# no otv surpress-arp-nd # Allows ARP requests over an overlay network and disables ARP caching on edge devices. This command does not support IPv6. OTV-a(config)# interface overlay 1 OTV-a(config-if-overlay)# otv arp-nd timeout 70 # Configures the time, in seconds, that an entry remains in the ARP-ND cache. The time is in seconds varying from 60 to 86400. The default timeout value is 480 seconds. Agenda • Introduction • Distributed Data Centers: Goals and Challenges • • • • • • • Control Plane and Data Plane Failure Isolation Multi-homing Mobility L2 Multicast Forwarding QoS and Scalability Path Optimization • OTV Architecture Principles • Principles of Interconnecting Networks at Layer-2 • OTV New Features OTV Multi-homing Fully Automated Multi-homing • No additional protocols required (i.e. BGP) • OTV site-vlan used to discover OTV neighbor in the same site • Authoritative Edge Device (AED) Election takes place • Extended VLANs are split across the AEDs • The AED is responsible for: ‒ MAC address advertisement for its VLANs ‒ Forwarding its VLANs’ traffic inside and outside the site AED OTV OTV Site Adjacency AED L3 L2 Site Adjacency used for AED election Release 5.2 and above Hardened Multi-homing Introducing OTV Site-identifier • Same site devices must use common site-identifier • Site-id information is included in the control plane • Makes OTV multi-homing more robust and resilient ‒ • Site Adjacency and Overlay Adjacency are now both leveraged for AED election An overlay will not come up until a site-id is configured ‒ Site and Overlay Adjacency are both leveraged for AED election Overlay Adjacency AED OTV OTV Site Adjacency AED L3 L2 feature otv otv site-identifier 0x1 otv site-vlan 99 OTV Multi-homing VLANs Split across AEDs Remote OTV Device MAC Table • Automated and deterministic algorithm • In a dual-homed site: VLAN MAC IF 100 MAC 1 IP A 101 MAC 2 IP B • Lower IS-IS System-ID (Ordinal 0) = EVEN VLANs • Higher IS-IS System-ID (Ordinal 1) = ODD VLANs OTV-a# show otv vlan OTV Extended VLANs and Edge Device State Information (* - AED) VLAN ---100 101* 102 Auth. Edge Device -----------------East-b East-a East-b Vlan State ---------inactive(Non AED) active inactive(Non AED) Overlay ------Overlay100 Overlay100 Overlay100 AED IP A ODD VLANs OTV Overlay Adjacency OTV AED IP B EVEN VLANs Site Adjacency OTV-a OTV-b# show otv vlan OTV Extended VLANs and Edge Device State Information (* - AED) VLAN ---100* 101 102* Auth. Edge Device -----------------East-b East-a East-b Vlan State ---------active inactive(Non AED) active Overlay ------Overlay100 Overlay100 Overlay100 OTV-b OTV Multi-homing AED and Broadcast Handling Broadcast reaches all the Edge Devices within the site 2. Only the AED forwards the traffic to the Overlay 3. All the Edge Devices at the other sites receive the broadcast 4. At the remote sites only the AEDs forward it into the site 1. OTV Broadcast stops here OTV Broadcast stops here OTV Bcast pkt OTV Core AED AED Agenda • Introduction • Distributed Data Centers: Goals and Challenges • • • • • • • Control Plane and Data Plane Failure Isolation Multi-homing Mobility L2 Multicast Forwarding QoS and Scalability Path Optimization • OTV Architecture Principles • Principles of Interconnecting Networks at Layer-2 • OTV New Features OTV and MAC Mobility MAC Moving and OTV Updates (1) 1. Workload moved between Data Center sites VM Moves OTV OTV MAC X MAC X MAC X OTV ESX OTV MAC X ESX Core MAC X MAC X AED AED OTV and MAC Mobility MAC Moving and OTV Updates (2) Workload moved between Data Center sites 2. Workload is detected in East DC and OTV control plane is triggered 1. 2.3) AED advertises MAC X with a metric of zero OTV OTV MAC X MAC X MAC X MAC X MAC X OTV ESX Core OTV MAC X MAC X MAC X MAC X MAC X MAC X AED AED 2.4) EDs in site West see MAC X advertisement with a better metric from site East and change them to remote MAC address. ESX 2.2) AED detects MAC X is now local 2.1) Server originates an ARP (RARP) frame OTV and MAC Mobility MAC Moving and OTV Updates (3) Workload moved between Data Center sites 2. Workload is detected in East DC and OTV control plane is triggered 3. East to West OTV data plane traffic allows to update the MAC tables of the L2 devices in West Site 1. 3.2) AED in site West forwards the RARP into the site and the L2 switches update their CAM tables OTV OTV MAC X MAC X MAC X MAC X OTV ESX MAC X Core OTV MAC X MAC X MAC X AED 3.1) AED in site East forwards the RARP broadcast frame across the overlay AED ESX Agenda • Introduction • Distributed Data Centers: Goals and Challenges • • • • • • • Control Plane and Data Plane Failure Isolation Multi-homing Mobility L2 Multicast Forwarding QoS and Scalability Path Optimization • OTV Architecture Principles • Principles of Interconnecting Networks at Layer-2 • OTV New Features L2 Multicast Traffic between Sites Multicast Enabled Transport • OTV can leverage the multicast support available in the transport network to optimize the delivery of the multicast traffic for the VLANs stretched across sites • Three steps: Automated mapping of the sites’ multicast groups to a range of multicast groups in the transport network 2. Creation of the Multicast state information at the OTV Edge Devices 3. Sites’ Multicast traffic delivered over the Overlay 1. L2 Multicast with Multicast Transport Step 1 – Mapping of the Site Multicast Group • The site multicast groups are mapped to a SSM group range in the core • Each (S1,Gs1) maps to a different SSM group in round-robin fashion 3) The West ED communicates the mapping information (including the source VLAN) to the other EDs Mcast Group Mapping 1) The Mcast source starts sending traffic to the group Gs1 Site Group Core Group Gs1 Gd1 Gs2 Gd2 1 Mcast Stream 2 OTV 3 The Mapping is communicated to the other EDs Multicast-enabled Transport Mapping to a Delivery Group OTV S1 Gs1 S1 West IP B IP A East OTV S2 Gs2 S2 4 2) The West ED maps (S1,Gs1) to a delivery group Gd1 4) Same process happens once source S2 is enabled (sending to a different group Gs2) IP C South L2 Multicast with Multicast Transport Step 2 – Multicast State Creation 3.1) ED Announces the receivers in a GroupMembership Update (GMUpdate) to all other EDs 4) The source ED adds the Overlay interface to the Outbound Interfaces (OIF) OIF-List Group IF Gs1 Gd1 Overlay OTV 2) The OTV ED snoops the IGMP join (without forwarding it) Multicast-enabled Transport 4 Receive GM-Update Update OIL 2 Client IGMP snoop 3.1 GM-Update OTV 1 1) A receiver in the East site sends an IGMP join for Gs1 Client IGMP report to join Gs1 S1 Gs1 S1 West IP A SSM Tree for Gd1 5) The SSM tree for Gd1 (rooted at the source ED) is IP built in the core) 3.2 OTV C IP B IGMPv3 report to join (IP A, Gd1) , the SSM group in the Core. Receiver (for Gs1) East 3.2) ED Sends an IGMPv3 South report to join the (IP A, Gd1) SSM group in the core It is important to clarify that the edge devices join the core multicast groups as hosts, not as routers! L2 Multicast with Multicast Transport Step 3 – Multicast Packet Flow OIF-List 1 Group IF Lookup Gs1 Gd1 Overlay 3 Multicast-enabled Transport Transport Replication OTV OTV S1 Gs1 S1 Gs1 S1 S1 Gs1 IP A Gd1 East OTV 2 Encap S1 Gs1 Receiver (for Gs1) IP 4 B IP A West IP A Gd1 4 IP C S1 Gs1 IP A Gd1 South Decap 5 Decap S1 Gs1 Receiver (for Gs1) 5 L2 Multicast with Multicast Transport Multicast Groups in the Core OTV can leverage the benefits of a multicast-enabled transport for both control and data planes. The following summarizes the requirements for a multicast transport: • Control group – Single PIM-SM or PIM-Bidir group used to form adjacencies and exchange MAC reachability information • Data groups – Range of SSM groups used to carry multicast data traffic generated by the sites interface Overlay100 otv otv otv otv join-interface e1/1 control-group 239.1.1.1 data-group 232.192.1.0/24 extend-vlan 100-150 The right number of SSM groups to be used depends on a tradeoff between the amount of multicast state to be maintained in the core and the optimization of Layer 2 multicast traffic delivery Agenda • Introduction • Distributed Data Centers: Goals and Challenges • • • • • • • Control Plane and Data Plane Failure Isolation Multi-homing Mobility L2 Multicast Forwarding QoS and Scalability Path Optimization • OTV Architecture Principles • Principles of Interconnecting Networks at Layer-2 • OTV New Features Release 5.2 and above QoS and OTV Marking on Encapsulation • On Encapsulation • CoS bits (802.1p) copied to the OTV shim header • If IP traffic: The original (inner) DSCP value is also copied to “outer” DSCP DMAC SMAC 802.1Q ETHERTYPE CoS 802.1p IP (optional) Inner DSCP OTV 1 OTV 802.1Q West IP A 2 Encap IP (optional) Outer DSCP OTV OTV shim Original Frame IP B East Release 5.2 and above QoS and OTV Marking on De-capsulation • On De-capsulation • CoS value is recovered from the OTV shim and added to the 802.1Q header • Original CoS and DSCP are both preserved • OTV Control Traffic is statically marked at CoS = 6/DSCP = 48 OTV West Decap IP A IP (optional) Outer DSCP OTV OTV shim 1 OTV 2 Original Frame 802.1Q IP B East DMAC SMAC 802.1Q CoS 802.1p ETHERTYPE IP (optional) Inner DSCP Release 6.2 OTV Scalability 1500 NX-OS 6.2 8* NX-OS 5.2 6* Sites * two ED per Site 256 32k 4000 16k 2000 OTV extended MAC addresses Multicast VLANs across all the Data Groups extended VLANs Agenda • Introduction • Distributed Data Centers: Goals and Challenges • • • • • • • Control Plane and Data Plane Failure Isolation Multi-homing Mobility L2 Multicast Forwarding QoS and Scalability Path Optimization • OTV Architecture Principles • Principles of Interconnecting Networks at Layer-2 • OTV New Features Path Optimization Egress Routing Optimization Hot Potato Routing Path Optimization Egress Routing with LAN Extension • Extended VLANs typically have associated HSRP groups • By default, only one HSRP router elected active, with all servers pointing to HSRP VIP as default gateway Packet from • HSRP Hellos Result: sub-optimal routing Vlan for 10 to Vlan 20 ARP DMAC HSRP VIP = DGW ARP reply Routing HSRP Active HSRP Standby HSRP Listen HSRP Listen Packet from Vlan 10 to Vlan 20 DMAC = Host Vlan 20 VLAN 20 VLAN 10 Egress Routing Localization FHRP Filtering Solution • Filter FHRP with combination of VACL and MAC route filter • Result: Still have one HSRP group with one VIP, but now have active router at each site for optimal first-hop routing HSRP Hellos ✗✗ ✗✗ HSRP Hellos HSRP Filter HSRP Active HSRP Standby HSRP Active Listen HSRP Listen Standby ARP for HSRP VIP ARP reply VLAN 20 VLAN 10 Path Optimization Optimal Routing Challenges • Layer 2 extensions represent a challenge for optimal routing • Challenging placement of gateway and advertisement of routing prefix/subnet WAN Ingress: Ingress: North-South / Client-Server North-South / Client-Server HSRP Filter HSRP Active HSRP Active HSRP Standby HSRP Standby East-West / Server-Server Egress: South-North / Server-Client Egress: South-North / Server-Client Path Optimization Is it relevant to my Data Center model? • Logical Data Center or Physical Data Center? • High Availability or Disaster Recovery? Ingress: North-South / Client-Server WAN Ingress: North-South / Client-Server Is this ONE Logical Data Center ? Or do I have TWO (High Availability) … separated Data Physical & Logical Center? East-West / … Server-Server Egress: South-North / Server-Client Egress: South-North / Server-Client Release 5.2 and above Specific Use-Case IPv6 and OTV • IPv6 Unicast Forwarding and Multicast Flooding supported across OTV - • Requires to disable optimized multicast forwarding (OMF) in IGMP snooping on OTV ED IPv6 Transport Network (Join Interface & Source Interface, not yet supported) OTV DC West OTV DC OTV Edge Device (VDC) East Global (all VLAN): no ip igmp snooping optimise-multicast-flood OTV vPC/vPC+ Domain Per VLAN with IPv6 Traffic OTV vlan vlan-id vlan configuration no ip igmp snooping optimise-multicast-flood Ingress Routing Localization Possible Solutions Challenge • Subnets are spread across locations • Subnet information in the routing tables is not specific enough • Routing doesn’t know if a server has moved between locations • Traffic may be sent to the location where the application is not available Options • DNS Based • Route Injection • LISP – Locator/ID Separation Protocol For more details on LISP and OTV Deployment see: BRKDCT-2131 OTV – Overlay Transport Virtualization Simplifying Data Center Interconnect Any Workload Anytime Anywhere Agenda • Introduction • Distributed Data Centers: Goals and Challenges • • • • • • • Control Plane and Data Plane Failure Isolation Multi-homing Mobility L2 Multicast Forwarding QoS and Scalability Path Optimization • OTV Architecture Principles • Principles of Interconnecting Networks at Layer-2 • OTV New Features OTV Support ASR1000 • OTV has been introduced in IOS XE 3.5 (Nov 2011) • To use OTV on ASR1000, you require: • • ASR1k <-> N7k Inter-Site Interoperability has been tested • • Advance Enterprise Image or Advance IP Service + OTV feature license No ASR1k <-> N7k Multihoming Support (Intra-Site Interoperability) OTV on ASR1000 Use Cases are: Legacy Deployments – where DC may still be Catalyst based • New Small Data Center and/or Disaster Recovery Sites – where Main DC is equipped with Nexus 7000 • OTV with Layer-3 Encryption – where MACSec is no option for Inter-DC Encryption • OTV Support ASR 1000 • New Features for IOS-XE 3.9 • OTV Adjacency Server (unicast) • OTV with LISP ESM • RPVST STP Support • New Features for IOS-XE 3.10 • Portchannel for join interface • VRF Aware • Subinterface for join interface • Layer 2 portchannel Agenda • Introduction • Distributed Data Centers: Goals and Challenges • • • • • • • Control Plane and Data Plane Failure Isolation Multi-homing Mobility L2 Multicast Forwarding QoS and Scalability Path Optimization • OTV Architecture Principles • Principles of Interconnecting Networks at Layer-2 • OTV New Features Principals of Interconnecting Networks at Layer-2 • Control-Plane • • • Using redundant Path Providing Loop protection V V V V Fault Containment • • • Automated Multi-Homing for Resiliency Loop Prevention • • OTV Multi-Homing • • Core (Layer-3) Learn and Distribute MAC information (no Flood&Learn) Separate Control-Plane information Limit Flood (ARP caching) Transport Agnostic • Can leverage literally any Transport Technology Principals for Interconnecting Networks Do Apply for Ethernet, FabricPath and VXLAN Switch# show nve peers Interface Peer-IP ---------- ----------nve1 10.10.10.1 nve1 10.10.10.2 nve1 10.10.10.3 nve1 20.20.20.1 nve1 20.20.20.3 End to End VXLAN A Very Bad Idea • One common Control Plane – One failure can affect all sites Switch# show nve peers Interface Peer-IP – No site concept • Manual Multihoming – BGP and/or vPC config • Multicast ---------nve1 nve1 nve1 nve1 nve1 ----------- VNI ------ Up Time --------- 10.10.10.1 10.10.10.3 20.20.20.1 20.20.20.2 20.20.20.3 30000 30000 30000 30000 30000 03:18:06 05:44:24 02:17:03 03:08:44 02:58:21 VNI ------ Up Time --------- 30000 30000 30000 30000 30000 03:18:06 08:06:22 05:44:24 02:17:03 02:58:21 Core (Layer-3) – Multiple multicast groups required • Reduced Scale – Every VTEP learns all MACs • Flooding Across Sites V V V V – BUM Traffic is flooded V V Principals of Interconnecting Networks at Layer-2 Inter-Pod Connectivity • Simplified Transport Requirement • • OTV Multicast Optimization • Offers optimized Multicast Forwarding • Path Diversity • • Core (Layer-3) Multicast dependent and independent Forwarding of BUM* Traffic (no hairpin) Flow based Entropy VXLAN or VXLAN+EVPN V V Multi-Site • Provides Site to Multi-Site connectivity V VXLAN or VXLAN+EVPN V V V V V V V Interconnecting VXLAN Networks (Layer-3) Inter-Pod Connectivity • Interconnecting VXLAN/EVPN Pods with VXLAN/EVPN is possible • • Control-Plane Domains (EVPN) can be separated (iBGP/eBGP) VXLAN/EVPN VNI 99000 With Layer-3 interconnect, Data-Plane Encapsulation is separated • • • Core (Layer-3) Routing decision at DC-Edge results in Decapsulation Requires a Transit VNI between Sites No Layer-2 Interconnect! V V V VXLAN or VXLAN/EVPN VXLAN or VXLAN/EVPN V VNI 30000 VNI 31000 V V V V V Not All Principles Satisfied “Good Enough” Solution V Principals of Interconnecting Networks at Layer-2 Control-Plane Multi-Homing Loop Prevention Fault Containment Transport Agnostic Multicast Optimization Path Diversity Multi-Site Inter-Pod Connectivity FabricPath ✖ ✔1 ✔✔ ✖ ✖ ✖ ✔ ✖ VXLAN (Flood&Learn) ✖ ✔1 ✔2 ✖ ✔ ✔ ✔✔ ✖ VXLAN-EVPN ✔ ✔1 ✔2 ✔✔ ✔✔ ✔ ✔✔ ✖ VPLS ✖ ✔1 ✔✔ ✖ ✖ ✖ ✔ ✔ OTV ✔✔ ✔✔ ✔✔ ✔✔ ✔✔ ✔✔ ✔✔ ✔✔ Good Better Best 1) Only with Multi-Chassis Link Aggregation (MC-LAG / VPC) 2) Limited Overlay Loop Prevention OTV 2.5 VXLAN Encapsulation UDP Header Utilizes VXLAN Encapsulation Original IETF draft of OTV F3 only feature Use in conjunction with Tunnel Depolarization Source Port 16 VXLAN Port 16 UDP Length 16 Checksum 0x0000 16 8 Bytes VXLAN Header Original Layer-2 Frame Overlay 50 Bytes of Overhead Outer IP Header Underlay Outer MAC Header • • • • UDP 4789 VXLAN Flags RRRRIRRR 8 Reserved 24 VNI 24 Reserved 8 8 Bytes Agenda • Introduction • Distributed Data Centers: Goals and Challenges • • • • • • • Control Plane and Data Plane Failure Isolation Multi-homing Mobility L2 Multicast Forwarding QoS and Scalability Path Optimization • OTV Architecture Principles • Principles of Interconnecting Networks at Layer-2 • OTV New Features New Feature for OTV in NX-OS 6.2 Nexus 7000 Hardware Support F3 Support for OTV in 6.2(6) Routed Uplinks to Core – Enable OTV on Nexus 7700 Series OTV Join Interface – Utilize port-level VLAN Translation on F3 OTV VDC Join-Interface Internal Interface M1 M1 M2 F1 F2e F3 • M2 F3 Aggregation VDC L3 (M-only, M1-F1 or F2/F2e) L2 Interfaces to Access (Classic-Ethernet or FabricPath) F1 and F2e support for OTV internal Interface • OTV Internal Interface (CE) F1 and F2e linecards have the ability to be internal interfaces when M series linecard is used for OTV M-Series interface F/M-Series interface New Features for OTV Tunnel Depolarization & Secondary IP • Secondary IP command introduced • • Configured within interface, not OTV interface OTV VDC Introduction of multiple IPs results in tunnel depolarization OTV-a (config-if)# sh otv OTV-a(config-if)# ip address 2.100.11.1/24 secondary OTV Overlay Information Disabling IP Redirects on port-channel11 :secondary address Site Identifier 0000.0000.0011 configured. OTV-a(config-if)# sh run int po11 Overlay interface Overlay1 !Command: show running-config interface port-channel11 VPN name : Overlay1 !Time: Wed Mar 27 23:05:21 2013 VPN state : UP Extended vlans : 25-50 72-227 (Total:182) version 6.2(2) Control group : 224.1.1.0 Data group range(s) : 232.1.0.0/24 interface port-channel11 Broadcast group : 224.1.1.0 no ip redirects Join interface(s) : Po11 (2.100.11.100) ip address 2.100.11.100/24 Secondary IP Addresses: : 2.100.11.1 ip address 2.100.11.1/24 secondary Site vlan : 1 (up) ip ospf network point-to-point AED-Capable : Yes1 ip router ospf 1 area 0.0.0.0 Capability : Multicast-Reachable ip igmp version 3 Release 6.2 New Features for OTV VLAN Translation: Translation through transit VLAN • When a different VLAN is used at multiple sites • Usually for 3 or more sites VLAN OTV400 VLAN 200 OTV VLAN 100 DC West DC East OTV OTV Release 6.2 New Features for OTV VLAN Translation: Translation through transit VLAN OTV-a(config)# int overlay1 OTV-a(config-if-overlay)# otv vlan mapping 100 to 400 OTV-B(config)# int overlay1 OTV-B(config-if-overlay)# otv vlan mapping 200 to 400 OTV-B(config-if-overlay)# sh run int overlay1 OTV-a(config-if-overlay)# sh run int overlay1 !Command: show running-config interface Overlay1 !Time: Fri Mar 29 19:01:04 2013 !Command: show running-config interface Overlay1 !Time: Fri Mar 29 19:02:29 2013 version 6.2(2) version 6.2(2) interface Overlay1 otv isis hello-multiplier 9 otv join-interface port-channel11 otv control-group 224.1.1.0 otv data-group 232.1.0.0/24 otv extend-vlan 25-50, 72-497 otv vlan mapping 100 to 400 no shutdown OTV-a(config-if-overlay)# sh otv vlan-mapping Original VLAN -> Translated VLAN -------------------------------100 -> 400 interface Overlay1 otv isis hello-multiplier 9 otv join-interface port-channel21 otv control-group 224.1.1.0 otv data-group 232.1.0.0/24 otv extend-vlan 25-50, 72-497 otv vlan mapping 200 to 400 no shutdown OTV-B(config-if-overlay)# sh otv vlan-mapping Original VLAN -> Translated VLAN -------------------------------200 -> 400 New Release 6.2 OTV Convergence Small and Large Scale Targets (Extreme Failures) Large Scale • • Small Scale <30sec <10sec < 10sec <5sec Remember to place join-interface into a dynamic routing protocol (OSPF, EIGRP, etc) Configure BFD in site-vlan Challenges in Traditional Layer 2 VPNs Solved by OTV Flooding Behavior ✔ Pseudo-wire Maintenance ✔ Multi-Homing Control-Plane Based - Unknown Unicast for MACLearning propagation - Unicast Flooding reaches all sites Dynamic - Full mesh Encapsulation of Pseudo-wire is complex - Head-End replication is a common problem Native additional Automated - Requires Multi-Homing Protocols & extends STP - Malfunctions impacts multiple sites ✔ OTV – Overlay Transport Virtualization Simplifying Data Center Interconnect Any Workload Anytime Anywhere Thank you