This is a place holder page for my latest Cisco ACI/HPE Synergy InterOp decks.
I’d like to share one Synergy OneView CLI I have been working on.
The tool is about 10M bytes and can be used on windows/Linux/MAC platform.
If you are IT admin and just want to run the tool without worrying about install any programming dependencies, you can download .exe file from the web directly. Another option is to run the tool with the docker image format
In any case, if you know Go programming language and wants to compile from the source, you can do “go get” from github repo directly.
The CLI tool intends to complement the current OneView GUI, Powershell cmdlets and other devops tools like Python/Ansible libraries.
Some of the examples are shown below like show mac, show server profile, show logical interconnect. You can do other commands like create/delete/edit Synergy resources.
When connecting Cisco ACI fabric with HPE blade servers through HPE Virtual Connect Modules, users should pay additional attention when working with VC tunnel networks. Both ACI and VC tunnel mode has some unique internal traffic forwarding mechanism when comparing with traditional L2 MAC forwarding method.
This blog intends to highlight one scenario users may see unexpected traffic behavior when connecting ACI with VC tunnel mode. Some potential design alternatives to resolve the issue are then discussed.
Virtual Connect tunnel network has the benefit of scalability and simplicity of management. Instead of configuring individual vlans matching upstream switch and downlink blade servers, a single Virtual Connect network can transit up to 4K vlans transparently. Users only need to configure vlans at upstream switch and blade servers OS/hypervisor level.
When using Virtual Connect tunnel networks, internally VC maintains a consolidated MAC forwarding table for the defined VC internal tunnel network instead of separated MAC table per user VLAN defined at switch and server side.
This unique operation characteristics of VC tunnel network needs to be taken into consideration when working with ACI Fabric. ACI uses Bridge Domain(BD) as layer 2 broadcast boundary and each BD can include multiple End Point Groups(EPG). At access layer, users can bind encapsulation vlan to the desired EPG to carry user traffic. As you can see, in some ACI design scenarios, flooding can cross different user vlans(EPGs) when the EPGs are in the same BD.
The following diagram shows one of the problematic scenarios of connecting VC tunnel network and ACI .
(Please note: The diagram only demos one use case to see this interop issue. Other cases may exist with the same issue so the key is to understand the nature of the problem.)
In the above topology, Virtual Connect has one single tunnel network defined and uses one uplink to connect with ACI leaf node. Over this link, two user vlans are carried through, vlan 160 and vlan 161. On ACI side, vlan-160 is used as encap-vlan for EPG Mgmt and vlan-161 is used as encap-vlan for EGP vmotion.
ACI BD domain is set as flooding mode as blade servers’ gateway are outside ACI cloud.
In this setup, user will have connectivity issue for the blade servers. This is how the problem will arise.
1) The blade server sends one ARP broadcast request over vlan 160 network. The ARP packet will travel through VC tunnel network. VC will record blade server source MAC address learned from the server downlink and forward the packet out to its local uplink to ACI. so far so good.
2) ACI fabric will see the ARP broadcast packet coming in on access port vlan 160 and will map it into EPG Mgmt. Because BD is set to flood ARP packets. The packet will be flooded inside the bridge domain and will be in turn to all ports under EPG vMotion as the two EPGs are in the same BD.
3) The same ARP broadcast packet will come back over the same uplink. VC will see the original blade server source MAC address from this uplink. At this point, VC will have the same MAC learned from both downlink server port and the uplink port within its single tunnel network MAC forwarding table. Remember VC tunnel mode will consolidate MAC table instead of maintaining per-vlan MAC table. This will in turn cause traffic interruption for this server.
There are several resolutions to the above scenario.
1) Set up ACI fabric EPG and BD as one to one mapping to simulate traditional layer 2 flooding behavior. So the flooding traffic in one EPG will not be able to reach other EPGs.
2) Use VC mapped mode networks. VC mapped mode networks are mapping to individual user vlans so will not have this conflicting MAC learning issue due to the fact that VC will maintain separate MAC forwarding table separation per user VLAN.
ACI release 1.1(1j) introduced the new option for BD flooding behavior called “Flood in Encapsulation”. This is to contain the flooding traffic inside the BD to only flood the traffic within the same EGP encapsulation vlan. This is a a promising solution as EPG Mgmt flooding traffic will not reach EPG vMotion with this option turned on even if these two EPGs share the same BD.
However, my test shows that even after setting this option, I still see flooding across EPG encap-vlans. Cisco ACI Fundamental also specifically pointed out the unsupported protocol for this option. This include ARP traffic, many common routing protocol like OSPF/BGP/EIGRP, multicast protocols like PIM/IGMP. So until this option can support ARP packets, it won’t be a viable solution for this case.
For readers want to dig deeper into VC tunnel mode MAC table, you can view VC internal tunnel network MAC forwarding table by the following method.
If VC is managed by HPE OneView software, you can make RestAPI call to OneView appliance like the following captures. You can see when the problem was seen, the blade server MAC address was learned from uplink LAG 25, which is VC module port X7. In working scenario, the same MAC address should be learned from server downlink like “downlink 3-a”.
HP OneView PowerShell cmdlet “show-HPOVLogicalInterconnectMacTable” can also retrieve VC MAC table.
The capture below shows incorrect MAC address entry after VC received the original packet from uplink. You can see LAG(Link Aggregation Group) 25 is on uplink X7.
If Virtual connect is managed by classic Virtual Connect Manager, you can use “show interconnect-mac-table” to retrieve MAC table information. The following capture just gives you an idea how the output looks like. It’s not for the scenario and configuration described above.
From time to time, customers will want to configure Virtual Connect PVLAN in order to work with their Cisco or VMware environment.
Virtual Connect has a checkbox for private network when you configure a “vnet” (internal vlan). However, it’s not the same thing as the PVLAN you find in Cisco Catalyst, Nexus switches or VMware VDS. I developed this deck some time ago to help to clarify some confusions, feel free to download here.
I also created Youtube video demos to show some lab captures when I tested the configuration.
With my original HP Virtual Connect Initial Configuration Youtube video getting over ten thousand views and growing, I have been thinking to make an equivalent initial configuration tutorial video for HP OneView for some time.
Whenever I learn any new product or technology, usually I will try to search if there are some good video tutorials out there instead of reading official user guide from page one. A good tutorial video can help me a long way and I want to help my readers using the same way as well because there are still many users just trying to get a good starting point.
Because of the above reasons, I made two separate videos but they are related to each other.
The first one is the step-by-step configuration from the scratch using HP OneView simulator. This will help users to get familiar with HP OneView workflow and some important terms like Uplink Set, Logical Interconnect, Network Set and etc.
The second video will cover the next step after initial configuration. Assuming you already finish setting up OneView initial configuration and want to verify your LAN/SAN connectivity along with looking for some advanced features like OneView with HP 3PAR StoreServ storage array and Brocade Network Advisor Integration, this video will use the live system to demo these items for you.
Hope the above two videos help you with getting to know HP OneView.
Hyper-converge era is coming up. Coupled with VMware latest VSAN technology, HP Proliant SL2500 scale out chassis can be a very powerful platform to deliver compute/storage/networking in data center. We just published the latest SL2500 VSAN RA white paper, please check it out.
This presentation highlights how to configure Nexus1000v with Virtual Connect. The presentation can be downloaded here.
Nexus1000v VSM deployment design for Virtual Connect is the same as general Cisco recommendation. However, two points need to be pointed out related with Virtual Connect environment. (It’d be much better to review presentation first before watching videos below to understand configuration context.)
1) Adding VEM on ESXi host connecting with Virtual Connect
2) Configuring MAC-pinning host port-channel for VEM uplinks