Professional Documents
Culture Documents
Sfs Deployment With Vxrail 7.0 PDF
Sfs Deployment With Vxrail 7.0 PDF
Deployment Guide
Abstract
In this guide, SmartFabric Services (SFS) is used to deploy a new leaf-
spine fabric for a new VxRail cluster. SFS automatically reconfigures
the fabric with user-specified VLANs during VxRail cluster deployment.
The SFS-enabled leaf-spine topology is connected to the data center's
existing network using Layer 2 or Layer 3 uplinks.
Chapter 1 Introduction 6
Purpose of this guide ................................................................................................7
Dell Technologies .....................................................................................................7
VxRail ........................................................................................................................7
SmartFabric Services ...............................................................................................7
SmartFabric Services with VxRail ............................................................................9
OpenManage Network Integration............................................................................9
Typographical conventions .....................................................................................10
Chapter 3 Topology 14
Overview .................................................................................................................15
Production topology with SmartFabric Services.....................................................15
Production topology connection details ..................................................................16
OOB management topology ...................................................................................18
OOB management connection details....................................................................19
VxRail .......................................................................................................................7
Dell Technologies
Our vision at Dell Technologies is to be the essential technology company for the data
era. Dell ensures modernization for today’s applications and for the emerging cloud-native
world.
Dell is committed to disrupting the fundamental economics of the market with an open
strategy that gives you the freedom of choice for networking operating systems and top-
tier merchant silicon. Our strategy enables business transformations that maximize the
benefits of collaborative software and standards-based hardware, including lowered costs,
flexibility, freedom, and security. Dell provides further customer enablement through
validated deployment guides that demonstrate these benefits while maintaining a high
standard of quality, consistency, and support.
VxRail
VxRail is at the forefront of a fundamental shift in IT infrastructure consumption – away
from application-specific, “build-your-own” infrastructure and toward virtualized, general-
purpose, engineered systems. Dell Technologies and VMware have embraced this shift
with the VxRail hyperconverged appliance. VxRail has a simple, scale-out architecture
that uses VMware vSphere and VMware vSAN to provide server virtualization and
software-defined storage.
SmartFabric Services
Dell EMC SmartFabric OS10 includes SmartFabric Services (SFS). With SFS, customers
can quickly and easily deploy and automate data center networking fabrics.
The VxRail L2 Single Rack personality is the original SFS personality. It is enabled by
running a Python script in the OS10 Linux shell.
Currently, this personality is limited to a single rack and cannot be expanded to a
multirack deployment. If switches with this personality enabled are upgraded, they will
continue to operate with the VxRail L2 Single Rack personality.
Note: The VxRail L2 Single Rack personality is not covered in this deployment guide. It is covered
in the VMware Integration for VxRail Fabric Automation SmartFabric User Guide, Release 1.1.
See the Dell EMC VxRail Support Matrix (account required), and the SmartFabric OS10 Solutions
(HCI, Storage, MX) Support Matrix for supported software versions.
L3 Fabric personality
Note: Dell requires using the L3 Fabric personality for new SFS deployments. All examples in this
guide use this personality. Unless otherwise specified, statements in this guide regarding
SmartFabric behavior and features are applicable to the L3 Fabric personality only.
Note: The L3 personality provides the option of deploying a VxRail cluster in a single rack but also
provides the option to extend to multi-rack in the future.
The new L3 Fabric Personality was introduced in OS10.5.0.5. It allows users to deploy
SmartFabric Services in a single rack and expand to multirack as business needs evolve.
The SFS L3 Fabric personality automatically builds an L3 leaf-spine fabric. This enables
faster time to production for hyperconverged and private cloud environments while being
fully interoperable with existing data center infrastructure.
Spine 1 Spine 2
Leaf 1A Leaf 1B
Leaf1B Leaf 2A Leaf 2B Leaf nA Leaf nB
Typographical conventions
The CLI and GUI examples in this document use the following conventions:
To use the SFS features detailed in this guide, switches must be running SmartFabric
OS10.5.0.7P3 or a later version specified in the SmartFabric OS10 Solutions (HCI,
Storage, MX) Support Matrix.
VxRail node
Dell EMC PowerSwitch model Typical role
connectivity options
Note: The roles shown are recommended, with the exception that Z9264F-ON is supported as a
spine only. S5232F-ON may be used as a leaf with ports connected to VxRail nodes broken out to
10 GbE or 25 GbE. VxRail nodes do not currently support 100 GbE NICs for VxRail system traffic.
Any combination of the leaf and spine switches listed in Table 1 may be used with the
exception that leaf switches must be deployed in pairs. Each leaf switch in the pair must
be the same model due to VLT requirements.
SFS supports up to 20 switches in the fabric. This scales to nine racks if two spines and
18 leafs are used.
Dell EMC The Dell EMC PowerSwitch S3048-ON is a 1U switch with 48x 1 GbE BASE-T ports and
PowerSwitch 4x 10 GbE SFP+ ports. This guide uses one S3048-ON switch in each rack for out-of-
S3048-ON band (OOB) management traffic. This includes connections to VxRail node iDRAC ports
and dedicated switch management ports.
Dell EMC VxRail Dell EMC VxRail P, V, S, E, and G Series nodes are built on 14th generation (14G)
14G nodes PowerEdge servers. VxRail P, V, and S Series nodes have a 2U form factor, as shown in
Figure 6.
Note: The deployment example in this guide builds a VxRail cluster with four VxRail nodes. VxRail
supports cluster sizes up to 64 nodes. With SFS, VxRail clusters must have a minimum of three
nodes. Two-node VxRail clusters are not currently supported.
VxRail node VxRail nodes support various combinations of network adapters. See the Dell EMC VxRail
network Network Planning Guide for network connectivity options by node type and the Dell EMC
adapters VxRail Support Matrix (account required) for supported network adapters.
Note: For VxRail node connections to the leaf switches listed in Table 1, use supported 10 GbE or
25 GbE network adapters only. 1 GbE network adapters are not supported for VxRail node to
SFS-enabled leaf switch connections.
Each VxRail node also includes an integrated Dell Remote Access Card (iDRAC) for out-
of-band management.
Overview ................................................................................................................15
• Production
• Out-of-band (OOB) management
The production topology contains redundant components and is used for all mission-
critical and end-user network traffic. The OOB management network is an isolated
network for remote management of hardware.
SmartFabric
VxRail node 2
VxRail node 3
Rack 1 Rack 2
Layer 3 connections
Layer 2 connections
VLTi connections
Layer 2 or Layer 3 uplink connections
Figure 7. SmartFabric topology with connections to VxRail nodes and external network
Note: The deployment examples in this guide use two network adapter ports per VxRail node, as
shown in Figure 7. See the Dell EMC VxRail Network Planning Guide for VxRail node connectivity
options.
SFS uses BGP-EVPN to stretch L2 networks across the L3 leaf-spine fabric. This allows
for the scalability of L3 networks with the VM mobility benefits of an L2 network. For
example, a VM can be migrated from one rack to another without the need to change its
IP address and gateway information.
The example in this guide builds the SmartFabric shown in Figure 7 in two stages:
1. The first stage is a single rack deployment. Leaf switches 1A and 1B are deployed
in Rack 1 without spine switches, and a two-leaf fabric is created using SFS. The
fabric is connected to the external network using either L2 or L3 uplinks. The
external network is typically a preexisting network in the data center. Three VxRail
nodes are connected to the two leaf switches, and a three-node VxRail cluster is
deployed.
2. In the second stage, two spine switches are added and connected to leaf
switches 1A and 1B. Leaf switches 2A and 2B are added in Rack 2 and are also
connected to the spine switches. The fabric is expanded to include the two spines
and two additional leafs using SFS. A fourth VxRail node is added in Rack 2 and
joined to the existing VxRail cluster.
Note: Single and multirack deployment options are discussed in Chapter 4 - Deployment
Planning.
Note: If S4100 series leaf switches are used (not shown), connections to VxRail nodes are
10 GbE. S4100 series switches have 100 GbE ports available for connections to spines.
Z9264F-ON Reset
Spine 1
Stack ID
Z9264F-ON Reset
Spine 2
S5248F-ON Stack ID
Leaf 1A
VLTi
S5248F-ON
Stack ID
Leaf 1B
VxRail P570
node
GRN=10G
GRN=10G
ACT/ LNK B
ACT/ LNK A
Note: In this example, the two QSFP28-DD double density ports (2x 100 GbE interfaces per
physical port), available on S5248F-ON switches, are used to create a 400 GbE VLTi. This
requires QSFP28-DD DAC cables or optics. On switches without QSFP28-DD ports, QSFP28
(100 GbE) or QSFP+ (40 GbE) ports are typically used for VLTi connections. The VLTi
synchronizes L2 and L3 control-plane information across the two nodes. The VLTi is used for data
traffic only when there is a link failure that requires the VLTi to reach the destination. Dell
Technologies recommends using at least two physical ports on each switch for the VLTi for
redundancy and to provide additional bandwidth if there is a failure.
For OOB management network connections, one S3048-ON switch is installed in each
rack, as shown in Figure 9.
OOB
Management
Network Core
The OOB management network also enables connections to the PowerSwitch SFS GUI. It
also enables switch console access using SSH, and VxRail node console access using
the iDRAC. This network is also used to carry heartbeat messages between switches
configured as VLT peers, and for OpenManage Network Integration (OMNI) to
communicate with the SFS master switch.
Note: This guide covers the equipment shown in Racks 1 and 2. Other devices and racks shown
in Figure 9 are for demonstration purposes only.
Four 10 GbE SFP+ ports are available on each S3048-ON for use as uplinks to the OOB
management network core.
Note: At a minimum, Dell Technologies recommends changing the admin password to a complex
password during the first login.
Note: For reference, devices on the OOB Management network in this guide use the
100.67.0.0/16 IP address block. These addresses are examples only. Use IP addresses suitable
for your environment.
OOB Management
network core
S3048-ON 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52
OOB management
switch
S5248F-ON
Leaf 1A
S5248F-ON
Leaf 1B
Stack ID
Z9264F-ON Reset
Spine 1
Stack ID
Z9264F-ON Reset
Spine 2
VxRail P570
node
GRN=10G
GRN=10G
ACT/ LNK B
ACT/ LNK A
iDRAC
1GbE link
10GbE link
External switches..................................................................................................23
• Three VxRail nodes running VxRail appliance software version 7.0 or a later as
specified in the SmartFabric OS10 Solutions (HCI, Storage, MX) Support Matrix.
• VxRail nodes must meet the hardware and software requirements listed in the
Dell EMC VxRail Support Matrix.
• Onboard NICs in VxRail nodes must be 10 GbE or 25 GbE.
• Two Dell EMC PowerSwitch units as listed in Table 1 deployed as leaf switches.
Each leaf switch in the pair must be the same model due to VLT requirements.
• Dell EMC PowerSwitch units must be running SmartFabric OS10.5.0.7P3 or a
later version as specified in the SmartFabric OS10 Solutions (HCI, Storage, MX)
Support Matrix.
• One 1 GbE BASE-T, also referred to as 1000BASE-T, switch for OOB
management connections. Dell Technologies recommends using one
PowerSwitch S3048-ON per rack.
• One DNS server which can be an existing DNS server that is reachable on the
network with host records added for this deployment. The example DNS host
records used in this guide are shown in Table 5.
Unsupported environments
SFS does not currently support the following environments:
Unsupported features
SFS does not currently support the following features:
Uplink options
SFS uplink options to external network switches include:
Note: Dell Technologies recommends using uplinks from a leaf pair as a best practice. The leafs
are referred to as border leafs. VxRail nodes and other servers in the rack may be connected to
border leafs in the same manner as other leafs in the SmartFabric.
L2 uplink If an L2 uplink is used, determine the VLAN ID to use for VxRail external management,
planning and if ports in the uplink will be tagged or untagged. Typically, this will be the same VLAN
used for DNS and NTP services on the existing network, as shown in the example in this
guide. Optionally, traffic may be routed from the external switch to the DNS/NTP servers.
The L2 uplink may be an LACP or static LAG. If L2 uplinks connect to a pair of Dell EMC
PowerSwitch systems, Dell Technologies recommends using LACP with VLT per the
example in this guide.
Note: With L2 uplinks, all routing into and out of the SmartFabric is done on external switches.
Point-to-point IP networks and addresses must be planned for each physical link in the L3
uplink.
Each leaf switch in the SmartFabric needs an IP address on the External Management
VLAN. An anycast gateway address on the same VLAN is also specified. This is the
virtual router/anycast gateway address shared by all leafs in the SmartFabric.
SmartFabric supports routing using eBGP or static routes. eBGP and static routing
examples are both provided in this guide.
If eBGP is used, ASNs and router IDs must be determined for the external switches.
These are automatically configured on all switches in the SmartFabric.
Note: SFS uses ASNs 65011 for leafs, and 65012 for spines. If these ASNs conflict with your
environment, they may be changed in the SFS GUI under 5. Edit Default Fabric Settings.
External switches
External switches must have available ports for connections from the existing network to
the SFS border leafs (or spines if applicable). For redundancy, Dell Technologies
recommends two external switches with at least two links per switch to the SmartFabric.
Use enough connections to provide sufficient bandwidth for the traffic anticipated across
these links. If using Dell EMC PowerSwitch systems as external switches, Dell
Technologies recommends configuring them as VLT peers, as shown in the examples in
this guide.
Note: This guide provides external switch configuration examples for Dell EMC PowerSwitch
systems. Cisco Nexus switch configuration examples are provided in Appendix C.
VLAN Purpose
VxRail Cluster Build Used to build the VxRail cluster. SFS automatically creates this VLAN
and names it SFS Client Management.
VxRail Internal Used for VxRail node discovery. SFS automatically creates this VLAN
Management and names it SFS Client Control.
VxRail External User-specified VLAN for VxRail Manager, ESXi, vCenter Server, NTP,
Management DNS, and vRealize Log Insight traffic
Note: All VLANs in Table 2 share the physical connections shown in Figure 8 in this deployment.
VLAN IDs and network addresses planned for this deployment example are shown in
Table 3.
Note: SFS automatically creates VLANs 4091 and 3939. VLANs 1811 through 1815 and their
network IP addresses are user-defined and are examples only. In SmartFabric mode, VLANs 2
through 3999, excluding 3939, are available for use. VLANs 4091 and 3939 may be changed from
their defaults in the SFS GUI under 5. Edit Default Fabric Settings. VLAN 3939 is also a VxRail
default VLAN. If VLAN 3939 is changed in the SFS GUI, you must also change it to match in
VxRail per the VxRail documentation.
Note: VLANs 4000 through 4094 are reserved for SFS. For more information about the reserved
VLANs, see the SmartFabric Services for OpenManage Network Integration User Guide, Release
1.3. The guide is available on the Dell EMC OpenManage Network Integration for VMware
vCenter website.
Note: SFS uses the 172.16.0.0/16, and 172.30.0.0/16 IP address blocks internally for the leaf-
spine network configuration. If these networks conflict with your environment, these default IP
addresses blocks may be changed in the SFS GUI under 5. Edit Default Fabric Settings.
The show virtual network command is used to view virtual networks, VLANs, and
port-VLAN assignments. This command is covered in more detail later in this guide.
Separator None
Iterator Num 0x
Offset 1
Suffix
Cluster name
VLAN ID 1812
VLAN ID 1813
Solutions Logging None, vRealize Log Insight, Syslog server vRealize Log Insight
Accounts Virtual appliance Use the same credentials for all virtual checked
accounts accounts
ESXi Accounts Use the same credentials for all hosts checked
**In the L2 uplink example in this guide, the DNS/NTP servers on the existing network are
on the same External Management VLAN, 1811, as the VxRail nodes. IP addresses on
this network use the 172.18.11.0/24 address block. In the L3 uplink example, the
This deployment example uses four VxRail nodes. However, host IP address pools are
configured for the Management, vMotion, and vSAN VLANs using addresses in the .101
through .164 range, as shown in Table 4. This is optional but removes the need to
configure new IP addresses if new hosts are added to the cluster.
Add forward and reverse lookup records on the DNS server using the hostnames and IP
addresses used in your deployment. The DNS entries for the deployment examples in this
guide are listed in Table 5.
Hostname IP address
vxrail01.dell.lab 172.18.11.101
vxrail02.dell.lab 172.18.11.102
vxrail03.dell.lab 172.18.11.103
vxrail04.dell.lab 172.18.11.104
vcenter01.dell.lab 172.18.11.62
vxmgr01.dell.lab 172.18.11.72
vxinsight.dell.lab 172.18.11.55
omni.dell.lab 172.18.11.56
In the L2 uplink example in this guide, the DNS server address is 172.18.11.50. In the L3
uplink example, the DNS server address is 172.19.11.50.
Cabling ...................................................................................................................29
Leaf1A Leaf1B
VLTi
VxRail node 1
VxRail node 2
VxRail node 3
Rack 1
Figure 11. VxRail node and leaf switch connections
For connection details, see Figure 8. Also, make OOB management connections, as
shown in Figure 10.
Note: Configure a unique OOB management IP address on each switch. The IP addresses shown
are examples only. Use IP addresses suitable for your environment. The management route
should not be 0.0.0.0/0, or this may interfere with the data network’s default route. Use a specific
destination prefix, as shown.
Other global settings may also be configured here, such as ip name-server and ntp
server if used by the switch. These settings are not required for the deployment example
in this guide. The hostname of the switch may be configured at the CLI or in the SFS GUI.
In this guide, the SFS GUI is used.
Enable SmartFabric
SmartFabric
Leaf1A Leaf1B
VLTi
Figure 12. First pair of leaf switches in SmartFabric mode
Caution: The following commands delete the existing switch configuration. Switch management
settings such as management IP address, management route, hostname, NTP server, and IP
name server are retained.
Ensure the physical VLTi connections are made between leaf pairs before proceeding.
Note: This example uses the two QSFP28 2x100Gb DD ports, interfaces 1/1/49-1/1/52, for the
VLTi connections on each S5248F-ON leaf.
To put the first pair of leaf switches in SmartFabric mode and configure them as VLT
peers, run the following commands on each switch:
OS10# configure terminal
OS10(config)# smartfabric l3fabric enable role LEAF vlti ethernet
1/1/49-1/1/52
Note: For more information, see SmartFabric Services for OpenManage Network Integration User
Guide, Release 1.3. The guide is available on the Dell EMC OpenManage Network Integration for
VMware vCenter web site. For additional SmartFabric CLI commands, see the SmartFabric
Services chapter of the Dell EMC SmartFabric OS10 User Guide Release 10.5.0.
Note: After reloading the switches, it takes about 2 minutes after the login prompt displays at the
switch CLI for SFS to come up and for the GUI to be fully functional.
Note: The SFS GUI supports Chrome, Firefox, and Edge browsers. Languages other than English
are not currently supported.
All GUI configuration is done on the SFS master switch. If you connect to an SFS switch
that is not the master, a link to the master is provided. This is outlined in red in Figure 13.
3. If applicable, click the link provided to go to the master switch, and log in as
admin.
Note: The IPv4 address of the SFS master may also be determined by running show
smartfabric cluster from the CLI of any switch in the SmartFabric. The master is always a
leaf switch, never a spine. Only one leaf switch in the SmartFabric will have ROLE set to MASTER.
The remaining leafs will have ROLE set to BACKUP.
When connected to the SFS master switch, the GUI appears, as shown in Figure 14.
Note: The Network Fabric ID is automatically set to 100 and cannot be changed. All directly
connected switches in SmartFabric mode join this fabric.
4. Click NEXT.
5. On the Switches page, update the Name (recommended, if not previously
configured from the CLI) and Description (optional) of the switches. Hostnames
are set to S5248F-Leaf1A and S5248F-Leaf1B in this example.
Note: If L3 uplinks are used, proceed to the Configure L3 routed uplinks to the external network
section. If there is no preference for the uplink type, consider using an L2 uplink configuration as it
has fewer steps than L3 uplink configuration.
The switches are cabled as shown in Figure 19. When L2 uplink configuration is complete,
Leaf1A and Leaf1B will connect with a VLT port channel to a switch pair named External-
A and External-B. In this example, an existing DNS/NTP server also connects to the
external switches using a VLT port channel. All VLT port channels use LACP in this guide.
Note: DNS and NTP servers do not have to connect in this manner as long as they are reachable
on the network.
External-A External-B
33 21 22 33 21 22
VLT
SmartFabric
VLT
53 54 53 54
Existing Leaf1A Leaf1B
3 2 1 1 2 3
DNS/NTP server
VLTi
L2 Connections
(All in VLAN 1811) VxRail node 1
VLTi connections
VxRail node 2
VxRail node 3
Rack 1
Figure 19. L2 uplinks to the external network
Configure L2 Note: Any ports available on the leaf switches may be used as uplinks, provided they are
uplinks in SFS compatible with the corresponding ports on the external switches. If leaf switch uplink ports will not
use their native speeds, the interfaces must be first broken out to the correct speed before the
uplinks are created. This is done using the Breakout Switch Ports option on the SFS GUI home
page. A breakout example is shown in the Change the port-group speed in the SFS GUI section of
this guide.
3. Click NEXT.
4. On the Port Configuration page, select the uplink ports used on each leaf switch
and set the LAG Mode to LACP or Static. In this example, 100 GbE ports 1/1/53-
1/1/54 are used on each switch, and the LAG mode is set to LACP.
Note: Be sure to configure the corresponding ports on the external switches with the same LAG
mode. External switch configuration examples using LACP are provided in the Configure external
switches for L2 connections section of this guide.
5. Click NEXT.
VxRail Manager must be able to contact a DNS server to resolve hostnames during
deployment. The External Management VLAN is created to enable this, and the uplinks
are added to it as follows:
6. On the Network Configuration page, click ADD NETWORK.
7. In the dialog box that opens, provide a Name, Description (optional), and a
VLAN ID for the External Management network. In this example, VLAN ID 1811
from Table 3 is used.
8. Click OK.
9. Next to Tagged Networks, select the External Management VLAN created
above, ExtMgmt-1811. Use the arrow button to move it to the box on the right, as
outlined in red in Figure 24. This makes the uplinks tagged members of the
External Management VLAN.
Note: Networks created through vCenter integration include the External Management, VSAN,
vMotion, and VM Networks created during VxRail deployment. It also includes networks added
through OMNI post-deployment.
The show smartfabric uplinks command may be used at the leaf switch CLI to
view configured interfaces and networks on the uplink.
Configure This section shows example configurations for both external switches for L2 connections
external to the SmartFabric.
switches for L2
connections Note: The external switches used in this example are Dell EMC PowerSwitch systems. If the
external switches are Cisco Nexus, see Appendix C.
Note: This is only an example. Modify your external switch configuration as needed for your
network.
General settings
Configure the hostname, OOB management IP address, and OOB management route as
shown.
External-A External-B
configure terminal configure terminal
Configure VLANs
Create the External Management VLAN. If traffic will be routed from the external switches
to other external networks, assign a unique IP address on each switch and configure
VRRP to provide gateway redundancy. Set the VRRP priority. The switch with the highest
External-A External-B
interface vlan1811 interface vlan1811
description External_Mgmt description External_Mgmt
ip address 172.18.11.252/24 ip address 172.18.11.253/24
vrrp-group 11 vrrp-group 11
priority 150 priority 100
virtual-address 172.18.11.254 virtual-address 172.18.11.254
no shutdown no shutdown
Configure interfaces
Configure the interfaces for connections to the SFS leaf switches. Interfaces 1/1/21 and
1/1/22 are configured in VLT port channel 100 in this example. Port-channel 100 is set as
an LACP port channel with the channel-group 100 mode active command.
Use the switchport mode trunk command to enable the port channel to carry traffic
for multiple VLANs. Configure the port channel as tagged on VLAN 1811 (the External
Management VLAN).
Optionally, allow the forwarding of jumbo frames with the MTU 9216 command.
In this example, interface 1/1/33 on each external switch is configured in VLT port channel
1 for connections to the DNS/NTP server. Port-channel 1 is set as an LACP port channel
with the channel-group 1 mode active command.
Configure ports directly connected to nodes, servers, or other endpoints as STP edge
ports. As a best practice, flow control settings remain at their factory defaults as shown.
External-A External-B
interface port-channel100 interface port-channel100
description "To Leaf1A/1B" description "To Leaf1A/1B"
no shutdown no shutdown
switchport mode trunk switchport mode trunk
switchport trunk allowed vlan switchport trunk allowed vlan
1811 1811
mtu 9216 mtu 9216
vlt-port-channel 100 vlt-port-channel 100
Configure VLT
This example uses interfaces 1/1/29 and 1/1/31 for the VLTi. Remove each interface from
L2 mode with the no switchport command.
Create the VLT domain. The backup destination is the OOB management IP address of
the VLT peer switch. Configure the interfaces used as the VLTi with the discovery-
interface command.
As a best practice, use the vlt-mac command to manually configure the same VLT MAC
address on both the VLT peer switches. This improves VLT convergence time when a
switch is reloaded.
Caution: Be sure the VLT MAC address is the same on both switches to avoid any unpredictable
behavior.
If you do not configure a VLT MAC address, the MAC address of the primary peer is used
as the VLT MAC address on both switches.
Note: For more information about VLT, see the Dell EMC SmartFabric OS10 User Guide on the
Dell EMC Networking OS10 Info Hub.
When the configuration is complete, exit configuration mode and save the configuration
with the end and write memory commands.
end end
write memory write memory
Validation Once the uplink interfaces have been configured on the external switches and in the SFS
GUI, additional validation is done using the switch CLI.
Run the show vlan command to verify ports are correctly assigned to the External
Management VLAN. Port channel 100 connects to the SFS leaf switches and is a tagged
member of the same VLAN configured on the SmartFabric uplinks (VLAN 1811). It is
tagged because it is also tagged on the SmartFabric leaf switches. The DNS/NTP server
is connected on port channel 1, which is an access member of VLAN 1811 in this
example.
External-A# show vlan
Codes: * - Default VLAN, M - Management VLAN, R - Remote Port Mirroring VLANs,
@ – Attached to Virtual Network, P - Primary, C - Community, I - Isolated
Q: A - Access (Untagged), T - Tagged
NUM Status Description Q Ports
* 1 Active A Eth1/1/1-1/1/20,1/1/23-/1/28,1/1/30,1/1/34
A Po100,1000
1811 Active External_Mgmt T Po100,1000
A Po1
4094 Active T Po1000
The show port channel summary command confirms port channel 100 connected to
the leaf switches is up and active. Port channel 1000 is the VLTi, and port channel 1 is
connected to the DNS/NTP server.
With SFS, port channel numbers are automatically assigned as they are created. Port
channel 1 is the uplink connected to the external switches and is up and active. Port
channel 1000 is reserved for the VLTi.
The L2 uplink, port channel 1 in this example, is added as a tagged member of VLAN
1811. This is verified at the CLI using the show virtual-network command as
follows:
S5248F-Leaf1A# show virtual-network
Codes: DP - MAC-learn Dataplane, CP - MAC-learn Controlplane, UUD - Unknown-
Unicast-Drop
Un-tagged VLAN: 4080
Virtual Network: 1811
VLTi-VLAN: 1811
Members:
VLAN 1811: port-channel1, port-channel1000
VxLAN Virtual Network Identifier: 1811
Source Interface: loopback2(172.30.0.0)
Remote-VTEPs (flood-list):
Note: If L2 uplinks were configured in the preceding section, skip this section and go to Configure
a jump host port.
Connections, port numbers, and networks used for external management are shown in
Figure 26. The External Management VLAN is VLAN 1911 on the external switches and is
VLAN 1811 on the SmartFabric switches.
External-A External-B
33 21 22 33 21 22
VLT po1
SmartFabric
53 54 53 54
Existing DNS & Leaf1A Leaf1B
3 2 1 1 2 3
NTP server
VLTi
VxRail node 3
Rack 1
VxRail nodes in SmartFabric VLAN 1811,
172.18.11.0/24 network
Point-to-point IP The point-to-point links used in this deployment are labeled A-E in Figure 27.
networks
E
External-A External-B
33 21 22 33 21 22
VLTi
A B C D
VLT po1
SmartFabric
53 54 53 54
Existing DNS & Leaf1A Leaf1B
NTP services VLTi
Rack 1
L3 uplink connections
L2 connections
VLTi connections
Figure 27. Point to Point connections
Each L3 uplink is a separate, point-to-point IP network. Table 6 details the links labeled in
Figure 27. The IP addresses in the table are used in the switch configuration examples.
BGP Example This section covers the L3 routed uplink configuration with BGP.
AS 65101
External-A External-B
10.0.2.1/32 10.0.2.2/32
33 21 22 33 21 22
VLTi
VLT po1
SmartFabric
53 54 53 54
Existing DNS & Leaf1A Leaf1B
NTP services
AS 65011 VLTi
Rack 1
L3 uplink connections
L2 connections
VLTi connections
Figure 28. BGP ASNs and router IDs
Note: Using private ASNs in the data center is a best practice. Private, 2-byte ASNs range from
64512 through 65534.
In this example, ASN 65101 is used on both external switches. SFS leaf switches use
ASN 65011 by default for all leafs in the fabric.
Note: If L3 uplinks are connected from SFS spine switches, the spine switches use ASN 65012 by
default.
Note: SFS default ASNs and IP address blocks may be changed by going to 5. Edit Default
Fabric Settings in the SFS GUI.
Prefix Length 31 31 31 31
Note: Any ports available on the leaf switches may be used as uplinks, provided they are
compatible with the corresponding ports on the external switches. If leaf switch uplink ports will not
use their native speeds, the interfaces must be first broken out to the correct speed before the
uplinks are created. This is done using the 1. Breakout Switch Ports option on the SFS GUI
home page. A breakout example is shown in the Change the port-group speed in the SFS GUI
section of this guide.
1. In the SFS GUI, select 2. Create Uplink for External Network Connectivity.
3. Click NEXT.
5. Click NEXT.
7. Click FINISH.
Repeat the steps in this section for the remaining three uplinks using the data from Table
7.
After uplink configuration, the SFS GUI Home page appears, as shown in Figure 33.
Static route This section shows L3 routed uplink configuration with a static route.
example
Note: If BGP is used instead of a static route, continue to Configure external switches for L3
connections.
Note: Currently, only one static route per L3 uplink is allowed. If multiple routes are needed, use a
default route, 0.0.0.0/0, as the destination network, or add additional uplinks for specific networks.
Support for multiple static routes per L3 uplink is planned for a future release.
Prefix Length 31 31 31 31
Routing Protocol Static Route Static Route Static Route Static Route
Prefix Length 24 24 24 24
Note: Any ports available on the leaf switches may be used as uplinks, provided they are
compatible with the corresponding ports on the external switches. If leaf switch uplink ports will not
use their native speeds, the interfaces must be first broken out to the correct speed before the
uplinks are created. This is done using the 1. Breakout Switch Ports option on the SFS GUI
home page. A breakout example is shown in the Change the port-group speed in the SFS GUI
section of this guide.
1. In the SFS GUI, select 2. Create Uplink for External Network Connectivity.
5. Click NEXT.
7. Click FINISH.
Repeat the steps in this section for the remaining three uplinks using the data from Table
8.
After uplink configuration, the SFS GUI Home page appears as shown in Figure 39.
Configure This section shows example configurations for both external switches for L3 routed
external connections to the SmartFabric.
switches for L3
connections Note: The external switches used in this example are Dell EMC PowerSwitch systems. If the
external switches are Cisco Nexus, see Appendix C.
Note: This is only an example. Modify your external switch configuration as needed for your
network.
General settings
Configure the hostname, OOB management IP address, and management route.
External-A External-B
configure terminal configure terminal
Configure VLANs
VLAN 1911 represents a preexisting management VLAN on the external network. DNS
and NTP services are located on this VLAN. Assign a unique IP address to the VLAN on
each switch.
Configure VRRP to provide gateway redundancy. Set the VRRP priority. The switch with
the highest priority value becomes the master VRRP router. Assign the same virtual
address to both switches.
vrrp-group 19 vrrp-group 19
priority 150 priority 100
virtual-address 172.19.11.254 virtual-address 172.19.11.254
Configure interfaces
Configure the interfaces for connections to the SFS switches. Ports 1/1/21 and 1/1/22 are
configured as L3 interfaces. The IP addresses used are from Table 6. Optionally, allow
the forwarding of jumbo frames with the MTU 9216 command. As a best practice, flow
control settings remain at their factory defaults as shown.
In this example, VLT port channel 1 connects to the DNS/NTP server. It is on VLAN 1911,
which represents the preexisting management VLAN, and the port channel is configured
as a spanning tree edge port.
Interface 1/1/33 on each external switch is configured in VLT port channel 1 for
connections to the DNS/NTP server. Port-channel 1 is set as an LACP port channel with
the channel-group 1 mode active command.
Configure VLT
This example uses interfaces 1/1/29 and 1/1/31 for the VLTi. Remove each interface from
L2 mode with the no switchport command. As a best practice, flow control settings
remain at their factory defaults, as shown.
Create the VLT domain. The backup destination is the OOB management IP address of
the VLT peer switch. Configure the interfaces used as the VLTi with the discovery-
interface command.
As a best practice, use the vlt-mac command to manually configure the same VLT MAC
address on both the VLT peer switches. This improves VLT convergence time when a
switch is reloaded.
Caution: Be sure the VLT MAC address is the same on both switches to avoid any unpredictable
behavior.
Note: For more information about VLT, see the Dell EMC SmartFabric OS10 User Guide on the
Dell EMC Networking OS10 Info Hub.
External-A External-B
interface range interface range
ethernet1/1/29,1/1/31 ethernet1/1/29,1/1/31
description VLTi description VLTi
no switchport no shutdown
no shutdown no switchport
flowcontrol receive on flowcontrol receive on
flowcontrol transmit off flowcontrol transmit off
Configure BGP
Note: If BGP is not used, go to the Configure static routes section.
Configure the BGP ASN with the router bgp command. The external switches share
the same ASN. Use the address that was set for interface loopback0 as the router ID.
VLAN 4000 is used for the iBGP connection between the external switches. VLAN4000 IP
addresses are configured per Table 6.
When the configuration is complete, exit configuration mode and save the configuration
with the end and write memory commands.
end end
write memory write memory
Configure two routes to the external management network. This is 172.18.11.0/24, one to
the connected IP address of Leaf1A, and one to Leaf1B.
When the configuration is complete, exit configuration mode and save the configuration
with the end and write memory commands.
end end
write memory write memory
Validate BGP Note: This section shows validation commands for the BGP example. If static routes are used,
example skip this section and go to Validate static route example.
Now that the uplink interfaces are configured on the external switches and on the SFS leaf
switches, connectivity can be verified using the switch CLI.
Run the show ip bgp summary command to verify that BGP is up for each neighbor.
When BGP is up, uptime is shown in the Up/Down column. The neighbors for External-A
shown in the output below are Leaf1A, Leaf1B, and External-B.
Run the show ip interface brief command to verify connected interfaces are up,
and IP addresses are configured correctly. VLAN 1911 is the external management VLAN
that contains the DNS/NTP server. VLAN 4094 and port channel 1000 are automatically
configured for the VLTi.
Note: Unused interfaces have been removed from the output for brevity.
The show ip route command output for the External-A switch appears as shown. No
BGP routes from the SFS fabric are learned at this stage of deployment. Interfaces 1/1/21
and 1/1/22 are connected to the SFS leaf switches.
Run the show ip bgp summary command to verify that BGP is up for each neighbor.
When BGP is up, uptime is shown in the Up/Down column. The neighbors for Leaf1A
shown in the output below are Leaf1B, External-A, and External-B.
Note: Unused interfaces have been removed from the output for brevity.
Run the show ip route command to verify routes to the external management VLAN,
172.19.11.0/24, have been learned using BGP from the external switches. In this
example, two routes to 172.19.11.0/24 are learned, one using each external switch.
Validate static Note: This section shows validation commands for the static route example. If BGP was used,
route example skip this section and go to Configure a jump host port.
Once the uplink interfaces have been configured on the external switches and in the SFS
GUI, connectivity can be verified using the switch CLI.
Run the show ip interface brief command to verify connected interfaces are up,
and IP addresses are configured correctly. In the output below, interface 1/1/33 and port
channel 1 connect to the DNS/NTP server. 1/1/21-1/1/22 are the links to the SFS leaf
switches, and 1/1/29 and 1/1/31 are the VLTi links. VLAN 4094 and port channel 1000 are
automatically configured for the VLTi.
Note: Unused interfaces have been removed from the output for brevity.
Run the show ip route command to verify static routes to the external management
VLAN, 172.18.11.0/24, are properly configured.
Run the show ip interface brief command to verify connected interfaces are up,
and IP addresses are configured correctly.
In the output below, interfaces 1/1/1-1/1/3 are connected to the VxRail nodes, 1/1/49-
1/1/52 are the VLTi links, and 1/1/53-1/1/54 are the uplinks to the external switches.
Note: Unused interfaces have been removed from the output for brevity.
Note: Since BGP is used by SFS to exchange routes within the fabric, some BGP routes appear
in the output.
Note: Before VxRail deployment, the VxRail Manager is accessible on an untagged port on the
SFS Client Management VLAN (VLAN 4091 by default). The default IP address is
192.168.10.200.
Jump host or
Laptop VxRail node 1
VxRail node 2
VxRail node 3
Rack 1
Figure 41. Jump host connected leaf switch for VxRail deployment
This section covers the configuration of a leaf switch port for connection to a jump host or
laptop computer (referred to only as a jump host for the remainder of this guide).
Change native If the jump host has a 1 GbE or 10 GbE NIC, and it is connected to a 25 GbE port on an
port speed on S5200 series switch, the switch port used must be changed from its native 25 GbE speed
S5200 series to 10 GbE for the port to come up.
switches
If the jump host has a 1 GbE or 10 GbE NIC, and is connected to a 10 GbE port on an
S4100 series leaf switch, or has a 25 GbE NIC and connects to an S5200 series leaf
switch, leave the port at its native speed, skip this section, and go to Configure the jump
host interface.
Note: When in 10 GbE mode, an S5200 series switch port will autonegotiate to 1 GbE when
connected to a 1 GbE NIC. To connect a jump host with a 1 GbE BASE-T or 10 GbE BASE-T port
to an SFP port on the leaf switch, use a Dell EMC supported SFP-1G-T or SFP-10G-T adapter.
Non-native
Port Non-
Native physical Native logical
group native Applicable switches
interface name speed interface
number speed
name
Note: Changing the speed is done for all ports in the port group. In this example, setting port
group 1/1/3 to 10g-4x changes ports 1/1/9-1/1/12 to 10 GbE, and the ports are renamed 1/1/9:1-
1/1/12:1.
Note: Port 1/1/9 was automatically renamed to 1/1/9:1 when its port group was changed from its
native setting of 4x 25 GbE to 4x 10 GbE.
Note: The output below is with an L3 uplink. If an L2 uplink to the external network is configured,
the External Management Network, VLAN 1811 in this guide, also appears in the command output
along with the L2 uplink port channel as a member. An output is shown in the Show command
output on Leaf1A section of the L2 uplink configuration chapter.
VLAN 3939 contains port channel 1000 (the VLTi) and the three interfaces connected to
the VxRail nodes, 1/1/1-1/1/3. VLAN 4091 contains port channel 1000, the three
interfaces connected to the VxRail nodes, and the jump host port, ethernet 1/1/9:1.
Note: The jump host port is only configured on one of the leaf switches.
Configure the By default, the initial VxRail Manager IP address is 192.168.10.200/24, and it is in VLAN
jump host IP 4091. After the initial configuration, the VxRail Manager address changes to its new
addresses address on the External Management VLAN (VLAN 1811 in this example). The new
VxRail Manager address used in this guide is 172.18.11.72/24 per the planning data in
Table 4.
During installation, the jump host must be able to reach both the initial and new VxRail
Manager addresses, so two addresses are configured on its network adapter, one for
each network.
The IP addresses are configured on the jump host NIC in this example as follows:
Note: Both addresses may be configured simultaneously if the network adapter supports it, or in
sequence if required. During VxRail deployment, the jump host port on the switch is automatically
moved from VLAN 4091 to the External Management VLAN, along with VxRail Manager. Since
the jump host port on the switch is untagged in VLAN 4091, and will also be untagged in the
External Management VLAN, no VLAN information is configured on the jump host NIC.
Once the jump host has been configured with an IP address on the 192.168.10.0/24
network, verify the jump host can communicate with VxRail Manager by pinging
192.168.10.200 from the jump host.
Prerequisites .........................................................................................................71
5. Check the box next to I confirm I want to configure the listed Dell EMC switch
clusters and click NEXT.
6. In the Dell EMC Switch Password Change window, set the REST_USER
password.
Note: The REST_USER is an account used by VxRail and OMNI to configure the switches.
8. Click NEXT.
9. The VxRail cluster type page displays. Select VxRail cluster.
Note: 2-node VxRail clusters are not currently supported with SFS.
11. When all nodes in the rack have been discovered, check the box next to I
confirm I want to configure the listed VxRail nodes and click NEXT.
12. On the How would you like to configure VxRail screen, select your preferred
configuration method, Step-by-step, or Configuration file. Either configuration
method may be used.
Note: A JSON-formatted configuration file may be used if you have saved one from a previous
installation using the same versions of VxRail and SmartFabric OS10, or if you have been
provided one from your sales representative. If you do not have a configuration file, select Step-
by-step.
Note: Step-by-step VxRail configuration screens are not in this guide, but are provided in the
VxRail Appliance Installation Procedures Guide that is available on Dell Technologies Solve
Online (account required). The values used in the VxRail configuration for this example are
entered using the planning data provided in Table 4.
14. When done with the configuration screens, you have the option to click REVIEW
FIRST to review your selections, as shown in Figure 52.
Note: VxRail Manager uses multicast DNS (mDNS) on the SFS Client Control network to discover
the SFS master switch virtual IP address.
Note: Before proceeding, be sure to keep the browser window that is displaying the screen shown
in Figure 53 open on the jump host.
Optionally, switch configuration can be verified at the CLI with the show virtual-
network command.
The command output below shows the External Management, vMotion, vSAN, and VM
network VLANs (VLANs 1811 through 1815 from Table 4) specified on the VxRail
deployment screens have been automatically configured on the leaf switches. The VxRail
node ports, interfaces 1/1/1-1/1/3, and the VLTi port channel, port channel 1000, are
members of all VLANs. The jump host port, interface 1/1/9:1, is still a member of VLAN
4091.
Note: The output below is with an L3 uplink. If an L2 uplink is configured, the uplink port channel
also appears as a member of VLANs 1811 through 1815.
Traffic on the External Management network, VLAN 1811, must be able to reach the DNS
server on the external network during VxRail deployment. To accomplish this with L3
uplinks, an IP address is assigned to each leaf switch on virtual network 1811. An anycast
gateway address shared by all leafs is also configured on the same network.
Table 10. Leaf switch External Management network IP addresses and anycast gateway
Item IP address/prefix
Note: If present, additional leaf switches in the fabric will also need one IP address per leaf on this
network.
Note: When IPv4 is selected, additional fields appear, as shown in Figure 55.
Note: If you plan to expand the fabric, additional leaf switches will also need IP addresses on this
network, with one IP address per leaf. This is covered in Expand SmartFabric and VxRail cluster
to multi-rack.
d. Enter the Prefix Length for the IP addresses and a Gateway IP Address.
These values are from Table 10.
When complete, the Update Network Configuration window appears, as shown in Figure
55.
3. Click OK.
BGP Validation Note: If static routes are used, proceed to the Validate and build VxRail cluster section. (Static
route validation was done earlier in the Validate static route example section of this guide).
If BGP is used on the uplinks, ensure the external switches have learned the routes to the
VxRail External Management network, 172.18.11.0/24 in this example, to reach the VxRail
nodes and VxRail Manager. This is done with the show ip route command. The BGP-
discovered route to 172.18.11.0/24 is shown in bold in the output below.
Note: Command output from a Cisco Nexus switch is shown in Appendix C: BGP validation on
N9K-External-A during VxRail deployment.
2. Click VALIDATE.
When validation is complete, a message indicating whether the configuration has passed
or failed displays, as shown in Figure 57.
Note: If validation failed, address the items that failed and validate again.
Note: Once validation passes, Dell Technologies recommends clicking the DOWNLOAD JSON
button to save an offline copy of your VxRail settings.
The switch port connected to the jump host has been automatically moved from VLAN
4091 to the External Management VLAN, VLAN 1811, on the leaf switch to enable it to
reach VxRail Manager on the new network.
Optionally, you can run the show virtual network command on the leaf switch that
the jump host is connected to verify the change. In the output below, the jump host port
1/1/9:1 is now untagged in VLAN 1811.
Note: Ensure the jump host NIC has an IP address on the new network, 172.18.11.0/24 in this
example, before proceeding with the next step. The jump host port on the leaf switch is untagged,
so do not configure a VLAN ID on the jump host NIC.
4. Click the START CONFIGURATION button shown in Figure 58. You are
redirected to the new VxRail Manager IP address in the browser, and VxRail
deployment starts, as shown in Figure 59.
Deployment takes about 1 hour. When VxRail is successfully deployed, the Hooray!
message displays, as shown in Figure 60.
Note: The jump host or workstation used must be able to reach the DNS server to resolve the
hostname of the vCenter server for the connection to succeed.
The Hosts and Clusters page of the vSphere Client appears, as shown in Figure 61.
Caution: Review any warnings that may appear in the vSphere Client.
In this chapter, the SmartFabric is expanded to a second rack with the addition of Spine1,
Spine2, Leaf2A, and Leaf2B. After the fabric is expanded, VxRail node 4 in Rack 2 is
added to the existing VxRail cluster.
Note: SFS supports up to 20 switches in the fabric. This scales to nine racks if two spines and 18
leafs are used. Any combination of leafs and spines may be used with the exception that leaf
switches must be deployed in pairs, and at least two spines are used. VxRail does not support
single spine deployments.
SmartFabric
VxRail node 2
VxRail node 3
Rack 1 Rack 2
Layer 3 connections
Layer 2 connections
VLTi connections
Layer 2 or Layer 3 connections
When you create an uplink to the external network using the SFS GUI or OMNI, the
preferred master is automatically set on all leaf switches in the fabric at that time.
Note: Spine switches are never elected SmartFabric master or preferred master switches.
Check preferred Before connecting additional leaf switches to the fabric, verify the preferred master is set
master status on on at least one pair of leaf switches currently in the SmartFabric. This is done by running
existing leafs the show smartfabric cluster command on each leaf in the existing fabric.
The output for at least one pair of leaf switches in the SmartFabric must show
PREFERRED-MASTER is set to true. The following commands and output are from
Leaf1A and Leaf1B:
For the example in this guide, there are only two leaf switches in the SmartFabric at this
stage of deployment. However, if there were additional leaf switches in the SmartFabric
when the uplink was created, they would also show PREFERRED-MASTER is set to true.
If the fabric was previously expanded after the uplink was created, the added leafs would
not have PREFERRED-MASTER set to true. This is allowed as long as PREFERRED-
MASTER is set to true on at least one pair of leaf switches currently in the SmartFabric.
If leaf switch pairs in the existing SmartFabric do not show PREFERRED-MASTER set to
true, create an uplink by following the instructions in the Configure L2 uplinks to the
external network or Configure L3 routed uplinks to the external network sections. After the
uplink is created, return to the preceding section and check the preferred master setting
again.
If you are using a demo or lab environment without the need for an uplink, create a
temporary uplink to set all of the leaf switches that are currently in the SmartFabric, to be
the preferred master.
Note: Physical port connections are not required to create this temporary uplink.
1. On the SFS GUI Home page, select 2. Create Uplink for External Network
Connectivity.
2. On the Uplink Details page:
a. Next to Uplink Connectivity, leave Layer 2 selected.
b. Enter a Name, such as temp.
3. Click NEXT.
4. On the Port Configuration page:
a. Next to Racks, select any rack.
b. Next to Configured Interfaces, select an available interface on either
switch.
Note: You cannot use this interface for other purposes until you delete the uplink.
To make the interface used in the temporary uplink available for other purposes, you can
delete the uplink without affecting the preferred master setting by performing the following
steps:
1. On the SFS GUI Uplinks page, select the uplink by name, temp in this example.
2. Click DELETE > OK.
The port used for the temporary uplink is now available.
----------------------------------------------------------
CLUSTER DOMAIN ID :
VIP : unknown
ROLE : unknown
SERVICE-TAG : unknown
MASTER-IPV4 :
PREFERRED-MASTER :
----------------------------------------------------------
After the switch reloads, run show smartfabric cluster again on each affected leaf
switch to confirm PREFERRED-MASTER is no longer set to true.
Note: New switches will be placed in SmartFabric mode in the Add switches to SmartFabric
section of this chapter.
Note: Configure a unique OOB management IP address for each switch. The management route
should not be 0.0.0.0/0, as it may interfere with the data network’s default route. Use a specific
destination prefix, as shown in the example below.
Other global settings may also be configured here, such as ip name-server and ntp
server if used by the switch. These settings are not required for the deployment example
in this guide. The hostname of the switch may be configured at the CLI or in the SFS GUI.
In this guide, the SFS GUI is used.
In this section, the two spine switches and two leaf switches are added to the
SmartFabric. These are Spine1, Spine2, Leaf2A, and Leaf2B, as shown in Figure 62.
Cable the switches as shown in Figure 62. Connection details are shown in Figure 8. Also,
make OOB management connections, as shown in Figure 10.
Caution: The following commands delete the existing switch configuration. Switch management
settings such as management IP address, management route, hostname, NTP server, and IP
name server are retained.
Spines The following commands are run on Spine1 and Spine2. This puts the switches in
SmartFabric mode as spines.
The configuration is applied, and the switch reloads. Repeat on the second spine switch.
Leafs The following commands are run on Leaf2A and Leaf2B. This puts the switches in
SmartFabric mode as leafs and configures them as VLT peers.
Note: This example uses the two QSFP28 2x100Gb DD ports, Ethernet 1/1/49-1/1/52, for the
VLTi connections on each leaf.
(Optional) To verify a leaf or spine switch is in SmartFabric mode, run the following
command:
Note: After reloading a switch in SmartFabric mode, it takes about 2 minutes after the login
prompt displays at the switch CLI for SFS to come up and for the GUI to be fully functional. If you
connect to a switch in the fabric that is not the master, a link to the master is provided in the GUI.
The IPv4 address of the master may also be determined by running show smartfabric
cluster from the CLI of any switch in the SmartFabric.
Once logged into the master, the SFS GUI Home page shows the leaf-spine topology and
configured uplinks. For the example used in this guide, it appears as shown in Figure 63.
Note: Since hostnames have not been configured on the four additional switches, each appears
with its default hostname, OS10. Hostnames for the additional switches are configured in the next
section.
1. On the Network Fabric page, update the Name (optional) and Description
(optional) of the fabric and click NEXT.
Note: The Network Fabric ID is automatically set to 100 and cannot be changed. All directly
connected switches in SmartFabric mode join this fabric.
2. On the Racks page, the second rack appears. Update the Name (recommended)
and Description (optional) of the second rack, as shown in Figure 64.
3. Click NEXT.
4. On the Switches page, the additional switches appear along with their service
tags, roles, and models. Update the Names (recommended) and Descriptions
(optional) of the newly added switches. The four additional switches with updated
names are outlined in red in Figure 65.
IP addresses are configured for the new leafs added to the SmartFabric, Leaf2A and
Leaf2B. The examples used in this guide are shown in Table 11.
Note: Existing leaf IP addresses and the gateway IP address were configured during VxRail
cluster deployment in the Additional configuration steps for L3 uplinks section of this guide.
Table 11. Leaf switch External Management network IP addresses and anycast gateway
Note: When Management Network 1811, is selected, additional fields appear, as shown in
Figure 67.
Ensure forward and reverse lookup records have been added to the DNS server for each
new node to be added to the cluster. In this example, the node FQDN is vxrail04.dell.lab,
and its IP address is 172.18.11.104. Figure 68 shows validation commands run from a
Microsoft Windows-based system with connectivity to the DNS server.
6. Click NEXT.
8. Click NEXT.
16. A Success message displays when validation has passed, as shown in Figure
76.
18. When complete, the fourth VxRail node appears in the cluster, as shown in Figure
78.
CAUTION: Review any warnings that may appear in the vSphere Client.
Optionally, verify the interface connected to the new VxRail node has been automatically
added to the VxRail networks on the leaf switches. To verify the connection, run the show
virtual network command. In this example, the new VxRail node is connected to
interface 1/1/1 on Leaf2A and Leaf2B.
Note: The command output shown in the following command is for Leaf2A. The output for Leaf2B
is the same.
The output below confirms the interface connected to the new VxRail node, ethernet1/1/1,
has been automatically placed in all VxRail virtual networks/VLANs.
100 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Untagged: ethernet1/1/1
VLAN 4091: port-channel1000
VxLAN Virtual Network Identifier: 4091
Source Interface: loopback2(172.30.0.1)
Remote-VTEPs (flood-list): 172.30.0.0(CP)
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 101
Deployment Guide
Chapter 8 Deploy and Configure OMNI
Configure OMNI...................................................................................................107
102 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Deploy OMNI VM
The OMNI VM is available for download from the Dell EMC OpenManage Network
Integration for VMware vCenter website. Download OMNI-version#.zip and extract the
OMNI-version#.ova file to a location accessible from the vSphere client.
Note: VxRail 7.0 supports OMNI 1.3.24 or a later version specified in the SmartFabric OS10
Solutions (HCI, Storage, MX) Support Matrix.
1. To deploy the OMNI VM, launch the vSphere Client and go to Hosts and
Clusters.
2. Right-click the VxRail cluster and select Deploy OVF Template.
3. On the Select an OVF template page, point to the location of the OMNI .ova file
and click NEXT.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 103
Deployment Guide
4. Enter a Virtual machine name and select a location for the OMNI VM. The
default location is used in this example, as shown in Figure 80.
5. Click NEXT.
6. On the Select a compute resource page, ensure the VxRail cluster is selected,
and Compatibility checks succeeded displays at the bottom of the page.
7. Click NEXT.
104 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
8. The Review details page displays, as shown in Figure 82.
9. Click NEXT.
10. On the License agreements page, review the terms provided. If you agree,
check the I accept all license agreements box and click NEXT.
11. On the Select storage page, select the vSAN datastore. Ensure Compatibility
checks succeeded displays at the bottom of the page.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 105
Deployment Guide
13. On the Select networks page, change the two destination networks to match the
two source networks, as shown in Figure 84.
When complete, the OMNI VM appears under the VxRail cluster, as shown in Figure 86.
106 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Figure 86. OMNI VM deployed
Configure OMNI
Note: Before proceeding, determine an IP address on the External Management VLAN and a
hostname for the OMNI VM. Add this information to the DNS server. In this example, the
hostname is omni.dell.lab, and the address is 172.18.11.56/24.
1. In the vSphere Client, power on the OMNI VM and launch the web console for the
VM.
2. Log in with the default OMNI VM username, admin, and password, admin. The
first time you log in, follow the prompts to change the admin password.
3. The OMNI Menu displays, as shown in Figure 87.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 107
Deployment Guide
4. Select 0. Full setup. The Network Manager terminal user interface (TUI)
displays.
Note: In the TUI, use the Tab and Arrow keys to navigate and the Enter key to select.
Note: The Device field is automatically populated with a MAC address and (ens160).
108 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Note: The example in Figure 90 is with L2 uplinks, so the DNS server has an IP address on the
same network as the OMNI VM, 172.18.11.50 in this example. If L3 uplinks are used, the DNS
server will be on a different network. For the L3 uplink examples in this guide, the DNS server IP
address is 172.19.11.50.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 109
Deployment Guide
8. Select Wired connection 2.
Note: The Device field is automatically populated with a MAC address and (ens192).
110 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Figure 92. Internal management connection settings
Note: Only part of the Destination/Prefix field is visible on the screen. Be sure it is set to
fde1:53ba:e9a0:cccc::/64.
10. Select OK > OK > Back to return to the Network Manager TUI menu.
11. On the Network Manager TUI menu, select Activate a connection. The
connection activation window displays, as shown in Figure 93.
Note: When active, connection names have an asterisk (*) next to them.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 111
Deployment Guide
12. Deactivate both connections as follows:
Select external management > Deactivate.
Select internal management > Deactivate.
13. Activate both connections as follows:
Select external management > Activate.
Select internal management > Activate.
14. Select Back to return to the Network Manager TUI menu.
15. On the Network Manager TUI menu, select Set system hostname, as shown in
Figure 94.
16. Change the Hostname from its default setting to the new hostname,
omni.dell.lab, in this example.
112 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Note: OMNI certificate installation is outside the scope of this guide. Certificates can be imported
later by selecting 5. Password/SSL configuration menu from the OMNI Menu. Follow the
instructions provided in SmartFabric Services for OpenManage Network Integration User Guide,
Release 1.3. The guide is available on the Dell EMC OpenManage Network Integration for
VMware vCenter website.
23. Ensure no error messages are displayed, and press Enter when prompted to
return to the OMNI menu.
24. At the OMNI menu, select 9. Logout.
This completes the OMNI VM console configuration.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 113
Deployment Guide
OMNI plugin
Note: You do not have to log out and log back into the vSphere Client for the plugin notification to
appear.
In the vSphere Client, an OMNI plugin banner appears at the top of the screen, outlined
in red in the figure below. Click the REFRESH BROWSER button that appears in the
banner.
Note: If there are other messages present, such as a license warning, the message shown in the
figure above may be located behind the other messages. When there are multiple messages,
there are "<" and ">" icons present to the left of the banner to cycle through the messages.
114 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
3. Click Services on the left menu.
Note: The Services menu displays the list of OMNI management and SmartFabric vCenter
automation services running on the OMNI appliance.
4. To start the SmartFabric vCenter automation service, select the service as shown
and click START.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 115
Deployment Guide
5. The system displays a start service success message.
116 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
2. In the left pane, click Home. In the right pane, click the radio button next to the
service instance as shown.
3. Click EDIT.
4. In the Edit a Service Instance window, enter the REST_USER Password that is
configured on the SmartFabric switches.
Note: The REST_USER password was set during VxRail deployment, as shown in Figure 47.
5. Click SUBMIT.
A Success message displays in green, as shown in Figure 106.
Note: After OMNI is deployed, use OMNI for switch configuration instead of the SFS GUI. The
SFS GUI is intended for initial deployment only.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 117
Deployment Guide
Chapter 9 Use OMNI for Post-deployment
Configuration
118 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
This chapter demonstrates different post-deployment switch configuration tasks that are
done with OMNI as needed.
SmartFabric
VxRail node 3
Rack 1 Rack 2
Layer 3 connections
Layer 2 connections
VLTi connections
Layer 2 or Layer 3 connections
Add a server to The PowerEdge server to be added is connected to Leaf2A and Leaf2B in this example,
the fabric as shown in Figure 107.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 119
Deployment Guide
Figure 108. Server Interfaces
3. Choose the first leaf switch, S5248F-Leaf2A. Click the drop-down icon to expand
the contents. Click the drop-down icon next to Breakout Ports & Jump Port.
4. Choose the port group to break out. In this example, the PowerEdge server is
connected to port 5, therefore, port-group1/1/2 is selected.
6. In the Mode drop down box, select the breakout mode. In this example,
4x10GEFixedFormFactor is selected.
7. Click SUBMIT.
120 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Figure 110. Server Interfaces
Repeat the above steps to breakout port group 1/1/2 on switch S5248F-Leaf2B. After the
wanted switch port is broken out to 10 GbE, you can continue to add the server to the
fabric.
In this example, the server's two connected ports are configured as an LACP NIC team in
the operating system running on the PowerEdge server. An IP address is assigned to the
NIC team on VM_Network_A, 172.18.14.0/24.
Note: NIC teaming and IP address configuration procedures on the PowerEdge server depend on
the operating system used and are not covered in this guide.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 121
Deployment Guide
Figure 111. Server Interfaces
4. Click CREATE. The Create Server Interface Profile window appears, as shown
in Figure 112.
5. In the Create Server Interface Profile window that opens, do the following
6. Enter a name for the Server Interface ID, for example, LACP-NIC-Team-1.
Next to Server Profile, select New Server Profile.
Enter a Server Profile ID, for example, PowerEdge-Server-1.
Next to Server Profile Bonding Type, select AutoDetect from the drop-
down list.
Next to Untagged Network, select a network for this connection from the
drop-down list. In this example, it is VM_Network_A 1814.
Next to Static Onboarding Option, select Yes.
Next to NIC Bonding, select Enable.
Select the Leaf Node and the Interface that is connected to the server. In
this example, Leaf2A and ethernet 1/1/5:1 are selected.
Leave the Routing Protocol option set to None.
122 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Figure 112. Create a server interface profile
Click CREATE.
Repeat steps 4 through 6 above for the second connection to the PowerEdge server using
the values in Table 12.
Field Value
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 123
Deployment Guide
When complete, the Server Interface Profile page appears, as shown in Figure 113.
Validation Validation may be done at a leaf switch CLI with the following commands:
Note: In the output below, port channels 96 through 97 are SFS-configured uplinks to the spines,
and port channel 1000 is the VLTi.
show virtual-network
The show virtual-network output shows that port channel 1, containing the
interface to the PowerEdge server, has been added as an untagged member of virtual
network 1814.
Note: Other virtual networks have been removed from the output for brevity.
124 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Codes: DP - MAC-learn Dataplane, CP - MAC-learn Controlplane, UUD - Unknown-
Unicast-Drop
Virtual Network: 1814
VLTi-VLAN: 1814
Members:
Untagged: port-channel1
VLAN 1814: port-channel1000, ethernet1/1/1
VxLAN Virtual Network Identifier: 1814
Source Interface: loopback2(172.30.0.1)
Remote-VTEPs (flood-list): 172.30.0.0(CP)
In this example, a port group named Application is created in the vSphere Client as
follows:
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 125
Deployment Guide
Figure 115. Application port group created on VLAN 30
The port group created in the vSphere client is propagated to the SFS switches as a
virtual network, and the VxRail node interfaces are automatically added to the virtual
network.
Note: Interfaces for statically onboarded devices, such as the non-VxRail PowerEdge server
added in the previous section, are not automatically added to new virtual networks.
Validation This is validated with the show virtual-network command on the leaf switches.
Note: In the two command output examples that follow, other virtual networks have been removed
from the output for brevity. See Figure 107 for the VxRail node locations in the topology.
The output below from Leaf1A shows that virtual network 30 was created, and the three
VxRail node interfaces, ethernet1/1/1-1/1/3, are members of the virtual network. The
output from Leaf1B, not shown, is the same.
Note: The output above is from a topology using an L2 uplink from Leaf1A and Leaf1B to the
external network. Port-channel1 is the L2 uplink in this example. Port-channel 1000 is always the
VLTi.
The output below from Leaf2A shows that virtual network 30 was created, and the single
VxRail node interface, ethernet1/1/1, is a member of the virtual network. The output from
Leaf2B, not shown, is the same.
126 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
S5248F-Leaf2A# show virtual-network
Codes: DP - MAC-learn Dataplane, CP - MAC-learn Controlplane, UUD - Unknown-
Unicast-Drop
Un-tagged VLAN: 4080
Virtual Network: 30
VLTi-VLAN: 30
Members:
VLAN 30: port-channel1000, ethernet1/1/1
VxLAN Virtual Network Identifier: 30
Source Interface: loopback2(172.30.0.1)
Remote-VTEPs (flood-list): 172.30.0.0(CP)
L3 uplinks to the external network can be moved from the leaf layer to the spine layer if
desired. This section covers the migration process.
Figure 116 shows the existing L3 uplinks from the leafs to the external switches in purple.
The new connections from the spines to the external network are shown in yellow. In this
example, 40 GbE connections are used from the external switches to the SmartFabric
switches.
Note: In SmartFabric mode, the auto breakout feature will automatically break out the 100 GbE
interfaces on the Spine switches to 40 GbE if connected to 40 GbE ports on the external switches
and Dell EMC supported cables and optics are used.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 127
Deployment Guide
VLTi
External-A External-B
21 22 27 28 21 22 27 28
27 28 27 28
Spine1 Spine2
L3 Leaf existing
connections
L3 Spine new
connections
SFS Connections
VLTi connections
Existing connections
BGP ASNs and Figure 117 shows the autonomous system numbers (ASNs) and router IDs used for the
router IDs external switches. The external switches share a common ASN. SmartFabric leafs share
a common ASN, and SmartFabric spines share a common ASN as shown.
128 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
ASN 65101
VLTi
External-A External-B
10.0.2.1 10.0.2.2
ASN 65012
Spine1 Spine2
Existing L3 uplinks
New L3 uplinks
SFS Connections
VLTi connections
Existing connections
Figure 117. BGP ASNs and router IDs
In this example, ASN 65101 is used on both external switches. By default, all SmartFabric
leaf switches use 65011, and all SmartFabric Spine switches use 65012.
The IP addresses shown in Figure 117 are loopback addresses used as BGP router IDs
on the external network switches. Router IDs on the SFS switches are automatically
configured from the SFS default private subnet address block, 172.16.0.0/16.
Point-to-point IP The existing point-to-point links used in this deployment are labeled A-D, and the new
networks point-to-point links are labeled E-H in Figure 118.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 129
Deployment Guide
VLTi
External-A External-B
E F G H
A B C D Spine1 Spine2
L3 Leaf existing
connections
L3 Spine new
connections
SFS Connections
VLTi connections
Existing connections
Each L3 uplink is a separate, point-to-point IP network. Table 13 details the links labeled
in Figure 118. The IP addresses in the table are used in the switch configuration
examples.
Configure L3 Table 14 shows the values entered in the OMNI plug-in in vCenter for this example. The
uplinks in OMNI following steps are run once for each uplink using the values from the table.
130 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Table 14. Uplink configuration details
Link label E F G H
Prefix Length 31 31 31 31
In the OMNI plugin in vCenter, select the Service Instance in the left pane. In the right
pane, click the Uplink tab. The existing uplinks from the leaf switches are displayed, as
shown.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 131
Deployment Guide
Figure 119. OMNI Uplink tab
1. Click CREATE.
2. On the Uplink Details page:
Set the Uplink Port Type to L3.
Set the L3 Type to L3 Routed Interface.
Enter a unique Name and, optionally, a Description.
3. Click NEXT.
4. On the Port Configuration page:
Set the Switch Group to Spine.
The Domain is set to the Fabric ID by default.
132 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Select the Node, Z9264F-Spine1.
Select the appropriate interface, ethernet 1/1/27, in this example.
5. Click NEXT.
6. On the Network Configuration page:
Set the Network Profile Name, SPN01-to-ExtA.
Set the Prefix Length, 31, and IP Address, 192.168.1.5.
Select eBGP as the Routing Protocol.
Set the Policy ID and Routing Policy Name, eBGP-SPN01-to-ExtA.
Note: Policy ID and Routing Policy Name are text fields and do not have to match. However, the
fields must be unique for each link, as shown in Table 14.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 133
Deployment Guide
Figure 123. Network Configuration page continued
7. Click FINISH.
Repeat the steps in this section for the remaining three uplinks using the data from Table
14.
Once all four uplinks are configured, they are visible on the Uplink tab of the OMNI plugin.
The new uplinks are outlined in red in Figure 124.
134 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Configure The external switches start at their configuration shown in Chapter 5 - Configure external
external switches for L3 connections. This section covers the configuration changes for both
switches external switches.
Configure interfaces
Configure the interfaces for L3 connections to the SmartFabric spine switches. Interfaces
1/1/27 and 1/1/28 are used in this example.
For each interface, add a description and take the port out of L2 mode with the no
switchport command. Assign an IP address to each interface per Table 13.
Optionally, allow the forwarding of jumbo frames with the MTU 9216 command. As a best
practice, flow control settings remain at their factory defaults as shown.
External-A External-B
configure terminal configure terminal
Configure BGP
Enter BGP configuration mode with the router bgp 65101 command. The external
switches share the same ASN. For the router ID, use the address that was set for
interface loopback0 in Chapter 5 - Configure external switches for L3 connections.
Configure the neighbor ASN. By default, all SmartFabric spine switches use 65012.
When the configuration is complete, exit configuration mode and save the configuration
with the end and write memory commands.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 135
Deployment Guide
External-A External-B
router bgp 65101 router bgp 65101
router-id 10.0.2.1 router-id 10.0.2.2
end end
write memory write memory
Validation Once the uplink interfaces are configured on the external switches and in the SFS GUI,
connectivity can be verified using the switch CLI.
Run the show ip bgp summary command to verify that BGP is up for each neighbor.
When BGP is up, uptime is shown in the Up/Down column. The neighbors for External-A
shown in the output below are Leaf1A, Leaf1B, Spine1, Spine2, and External-B.
Note: Leaf1A and Leaf1B can be disconnected from the external switches after validation is
complete.
Run the show ip interface brief command to verify connected interfaces are up
and IP addresses are configured correctly.
In the output below, interfaces 1/1/27 and 1/1/28 are connected to the spine switches.
Note: Unused interfaces have been removed from the output for brevity.
136 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Ethernet 1/1/27 192.168.1.4/31 YES manual up up
Ethernet 1/1/28 192.168.1.6/31 YES manual up up
Ethernet 1/1/29 unassigned YES unset up up
Ethernet 1/1/31 unassigned YES unset up up
Ethernet 1/1/33 unassigned YES unset up up
Management 1/1/1 100.67.127.37/24 YES manual up up
Vlan 1 unassigned YES unset up up
Vlan 1911 172.19.11.252/24 YES manual up up
Vlan 4000 192.168.3.20/31 YES manual up up
Vlan 4094 unassigned YES unset up up
Port-channel 1 unassigned YES unset up up
Port-channel 1000 unassigned YES unset up up
Loopback 0 10.0.2.1/32 YES manual up up
Run the show ip bgp summary command to verify that BGP is up for each neighbor.
When BGP is up, uptime is shown in the Up/Down column. The neighbors for Spine1
shown in the output below are the four leaf switches (with automatically configured IP
addresses in the 172.16.0.0/16 address range), and External-A and External-B.
Run the show ip interface brief command to verify connected interfaces are up,
and IP addresses are configured correctly.
In the output below, interfaces 1/1/27 and 1/1/28 are connected to the external switches.
Note: Unused interfaces have been removed from the output for brevity.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 137
Deployment Guide
Vlan 4001 172.16.0.11/31 YES manual up up
Vlan 4002 172.16.0.13/31 YES manual up up
Vlan 4003 172.16.0.9/31 YES manual up up
Vlan 4004 172.16.0.17/31 YES manual up up
Port-channel 96 unassigned YES unset up up
Port-channel 97 unassigned YES unset up up
Port-channel 98 unassigned YES unset up up
Port-channel 99 unassigned YES unset up up
Loopback 1 172.16.128.4/32 YES manual up up
Remove uplinks After the new uplinks from the spines are validated, the original uplinks from the leafs may
from leafs be removed as follows:
1. In the OMNI plugin in vCenter, select the Service Instance in the left pane. In the
right pane, click the Uplink tab. The existing uplinks are displayed.
2. Remove the first uplink from a leaf switch to an external switch by selecting the
link followed by DELETE. Repeat for the remaining three leaf switch uplinks.
3. Remove the physical connections from the leaf switches to the external switches.
138 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Appendix A Validated Components
General .................................................................................................................140
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 139
Deployment Guide
General
The following tables include hardware, software, and firmware that was used to configure
and validate the examples in this guide.
Note: For more information about supported components and versions, see the Dell EMC VxRail
Support Matrix (account required).
Note: Switches validated for the Cisco Nexus examples are in Appendix C.
Operating system
Qty Item
version
- BIOS 2.6.4
- CPLD 1.1.3
140 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
VxRail appliance software
The examples in this guide were validated using VxRail 7.0 appliance software. The
software consists of the component versions shown in Table 17.
Item Version
OMNI software
OMNI software used in this guide is shown in Table 18.
Item Version
OMNI 1.3.24
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 141
Deployment Guide
Appendix B CLI Commands
142 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Switch CLI validation commands
This section provides a list of the most common commands and their output for the
examples used in this guide.
Note: See the SmartFabric OS10 release notes for upgrade instructions.
Note: If SmartFabric OS10 was factory installed, a perpetual license is already on the switch.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 143
Deployment Guide
show interface status
The show interface status | grep up command is used to verify required
interfaces are up, and links are established at their appropriate speeds. In this example,
ports 1/1/1-1/1/3 are VxRail nodes, port 1/1/9:1 is a jump host, ports 1/1/49-1/1/52 are the
VLTi, ports 1/1/53-1/1/54 are uplinks to the external network, and ports 1/1/55-1/1/56 are
uplinks to the spines.
Note: The jump host port was automatically moved from VLAN 4091 to VLAN 1811 during VxRail
deployment.
144 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
S5248F-Leaf1A# show virtual-network
Codes: DP - MAC-learn Dataplane, CP - MAC-learn Controlplane, UUD - Unknown-
Unicast-Drop
Un-tagged VLAN: 4080
Virtual Network: 1811
VLTi-VLAN: 1811
Members:
Untagged: ethernet1/1/9:1
VLAN 1811: port-channel1, port-channel1000, ethernet1/1/1, ethernet1/1/2,
ethernet1/1/3
VxLAN Virtual Network Identifier: 1811
Source Interface: loopback2(172.30.0.0)
Remote-VTEPs (flood-list):
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 145
Deployment Guide
Virtual Network: 3939
Description: In-band SmartFabric Services discovery network
VLTi-VLAN: 3939
Members:
VLAN 3939: port-channel1000, ethernet1/1/1, ethernet1/1/2, ethernet1/1/3
VxLAN Virtual Network Identifier: 3939
Source Interface: loopback2(172.30.0.0)
Remote-VTEPs (flood-list):
Note: If an entry is not shown for the jump host, port 1/1/9:1 in the above example, the jump host
may be connected to the other leaf switch, or the NIC on the jump host does not have LLDP
enabled. This is not required and may be ignored.
146 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
show ip interface brief
Run the show ip interface brief command to verify connected interfaces are up,
and IP addresses are configured correctly.
In the output below, interfaces 1/1/1-1/1/3 are connected to the VxRail nodes, 1/1/49-
1/1/52 are the VLTi, and 1/1/53-1/1/54 are L3 uplinks to the external switches. VLAN
4090, Loopback 1, and Loopback 2 are automatically configured by SFS. VLAN 4090 is
used for iBGP, Loopback 1 is the router ID, and Loopback 2 is the VTEP IP address.
Note: Unused interfaces have been removed from the output for brevity.
show ip route
With L3 uplinks, the show ip route command is used to ensure the leaf switches have
routes to the external network, 172.19.11.0/24 in this example, to reach the DNS server.
This BGP-discovered route is shown in bold in the output below.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 147
Deployment Guide
B - BGP, IN - internal BGP, EX - external BGP
O - OSPF, IA - OSPF inter area, N1 - OSPF NSSA external type 1,
N2 - OSPF NSSA external type 2, E1 - OSPF external type 1,
E2 - OSPF external type 2, * - candidate default,
+ - summary route, > - non-active route
Gateway of last resort is not set
Destination Gateway Dist/Metric Change
----------------------------------------------------------------------------------
B EX 10.0.2.1/32 via 192.168.1.0 20/0 21:07:08
via 192.168.2.0
B EX 10.0.2.2/32 via 192.168.1.0 20/0 21:07:08
via 192.168.2.0
C 172.16.0.0/31 via 172.16.0.1 vlan4090 0/0 21:10:36
C 172.16.128.0/32 via 172.16.128.0 loopback1 0/0 21:10:39
B IN 172.16.128.1/32 via 172.16.0.0 200/0 21:09:53
B EX 172.19.11.0/24 via 192.168.1.0 20/0 21:07:08
via 192.168.2.0
C 172.18.11.0/24 via 172.18.11.250 virtual-network1811 0/0 16:02:54
C 172.30.0.0/32 via 172.30.0.0 loopback2 0/0 21:10:39
C 192.168.1.0/31 via 192.168.1.1 ethernet1/1/53 0/0 21:10:41
B IN 192.168.1.2/31 via 172.16.0.0 200/0 21:09:53
C 192.168.2.0/31 via 192.168.2.1 ethernet1/1/54 0/0 21:07:14
B IN 192.168.2.2/31 via 172.16.0.0 200/0 21:09:53
B EX 192.168.3.20/31 via 192.168.1.0 20/0 21:07:08
via 192.168.2.0
148 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
show vlt domain_id backup-link
This command is used to verify VLT peers are communicating on the backup link over the
OOB management network. The Destination is the management IP address of the
peer. The Peer Heartbeat status must be Up.
Peer-routing mismatch:
No mismatch
VLAN mismatch:
No mismatch
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 149
Deployment Guide
Virtual Network Remote-VTEP Mismatch:
No mismatch
EVPN Mismatch:
EVPN Mode Mismatch:
No mismatch
NVE Mismatch:
No mismatch
To delete the existing switch configuration and go from SmartFabric mode back to Full
Switch mode, run the following commands on each switch:
150 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
S5248F-Leaf1A# configure terminal
S5248F-Leaf1A(config)# no smartfabric l3fabric
Reboot to change the personality? [yes/no]:y
The switch reboots into Full Switch mode. The mode can be verified with the following
command:
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 151
Deployment Guide
Appendix C Cisco Nexus External Switch
Configuration Example
152 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Configure external Nexus switches for L3 routed connections
SmartFabric uplinks may be connected to external Cisco Nexus switches. This appendix
includes a Cisco Nexus 9000 switch configuration example for L3 routed connections to
SmartFabric leaf switches.
Note: L3 routed uplinks on the SmartFabric leaf switches are configured per the Configure L3
routed uplinks with BGP in SFS section of this guide.
Connections, port numbers, and networks used for external management in this example
are shown in Figure 125. The External Management VLAN is VLAN 1911 on the external
Nexus switches, and is VLAN 1811 on the SmartFabric switches.
Peer Link
N9K-External-A N9K-External-B
1 49 50 1 49 50
vPC
SmartFabric
53 54 53 54
Existing Leaf1A Leaf1B
3 2 1 1 2 3
DNS/NTP server
VLTi
VxRail node 3
Rack 1
VxRail nodes in SmartFabric VLAN 1811,
172.18.11.0/24 network
Figure 125. L3 routed uplinks from SmartFabric to external Nexus switches
In this example, an existing DNS/NTP server connects to the Nexus switches using a vPC
270B
in VLAN 1911.
Note: DNS and NTP servers do not have to connect in the manner shown if they are reachable on
381B
the network.
Point-to-point IP The L3 point-to-point links used in this example are labeled A-D in Figure 126.
271B
networks
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 153
Deployment Guide
Pee r Li nk
N9K-External-A N9K-External-B
1 49 50 1 49 50
A B C D
vPC
SmartFabric
53 54 53 54
Existing Leaf1A Leaf1B
DNS/NTP server VLTi
Rack 1
L3 uplink connections
L2 connections
VLTi connections
Figure 126. Point to Point connections
310B
Each L3 uplink is a separate, point-to-point IP network. Table 19 details the links labeled
27B
in Figure 126. The IP addresses in the table are used in the switch configuration
examples.
A
103B N9K-External-A
104B 192.168.1.0
105B Leaf1A
106B 192.168.1.1
107B 192.168.1.0/31
108B
B
109B N9K-External-A
10B 192.168.1.2
1B Leaf1B
12B 192.168.1.3
13B 192.168.1.2/31
14B
C
15B N9K-External-B
16B 192.168.2.0
17B Leaf1A
18B 192.168.2.1
19B 192.168.2.0/31
120B
D
12B N9K-External-B
12B 192.168.2.2
123B Leaf1B
124B 192.168.2.3
125B 192.168.2.2/31
126B
154 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
BGP ASNs and Figure 127 shows the ASNs and router IDs used for the external Nexus switches and SFS
router IDs leaf switches in this example. External switches share a common ASN, and all SFS leaf
switches share a common ASN.
AS 65101
Pee r Li nk
N9K-External-A N9K-External-B
10.0.2.1/32 10.0.2.2/32
1 49 50 1 49 50
vPC
SmartFabric
53 54 53 54
Existing DNS & Leaf1A Leaf1B
NTP services
AS 65011 VLTi
Rack 1
L3 uplink connections
L2 connections
VLTi connections
Figure 127. BGP ASNs and router IDs
31B
In this example, ASN 65101 is used on both Nexus external switches. SFS leaf switches
274B
Note: If L3 uplinks are connected from SFS spine switches, the spine switches use ASN 65012 by
382B
default.
The IP addresses shown on the external network switches in Figure 127 are loopback
275B
addresses used as BGP router IDs. On the SmartFabric switches, BGP router IDs are
automatically configured from the SFS default private subnet address block,
172.16.0.0/16.
Note: SFS default ASNs and IP address blocks may be changed by going to 5. Edit Default
38B
Note: All of the Nexus switch configuration commands used to validate this topology are shown in
384B
the sections that follow. The Nexus switches were reset to their default configuration settings
using the write erase command before running the configuration commands below. This is
only an example. Modify your external switch configuration as needed for your environment.
General settings Enable the following features: interface-vlan, lacp, vrrp, vpc, bgp, lldp.
276B
Configure the hostname, OOB management IP address on VRF management, and the
VRF management route as shown.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 155
Deployment Guide
N9K-External-A
120B N9K-External-B
12B
configure terminal
168B configure terminal
170B
feature
1687B interface-vlan feature
170B interface-vlan
feature
168B lacp feature
1702B lacp
feature
1689B vrrp feature
1703B vrrp
feature
1690B vpc feature
1704B vpc
feature
169B bgp feature
1705B bgp
feature
1692B lldp feature
1706B lldp
hostname N9K-External-A
1693B hostname N9K-External-B
170B
interface mgmt 0
1694B interface mgmt 0
1708B
ip address 100.67.127.30/24
1695B ip address 100.67.127.29/24
1709B
no shutdown
1697B no shutdown
17B
ip route 100.67.0.0/16
169B ip route 100.67.0.0/16
173B
100.67.127.254 100.67.127.254
Configure the VLAN 1911 represents a preexisting management VLAN on the external network. DNS
27B
External and NTP services are located on this VLAN. Optionally, enable jumbo frames with the mtu
Management 9216 command. Assign a unique IP address to the VLAN on each switch.
VLAN
Configure VRRP to provide gateway redundancy and assign the same virtual address to
278B
both switches.
N9K-External-A
12B N9K-External-B
123B
vlan 1911
174B vlan 1911
1725B
name ExtMgmt
175B name ExtMgmt
1726B
no shutdown
176B no shutdown
172B
interface Vlan1911
17B interface Vlan1911
1728B
description ExtMgmt
178B description ExtMgmt
1729B
no shutdown
179B no shutdown
1730B
mtu 9216
1720B mtu 9216
173B
ip address 172.19.11.252/24
172B ip address 172.19.11.253/24
1732B
vrrp 11
172B vrrp 11
173B
address 172.19.11.254
1723B address 172.19.11.254
1734B
no shutdown
1724B no shutdown
1735B
Configure the Create the vPC domain. The peer-keepalive destination is the OOB management IP
279B
and allow the default and External Management VLANs, 1 and 1911 respectively.
156 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Configure the interfaces to use in the vPC peer link. Put the interfaces in trunk mode and
281B
allow the default and External Management VLANs, 1 and 1911 respectively. Add the
interfaces to the peer link port channel.
N9K-External-A
124B N9K-External-B
125B
role priority 1
173B role priority 65535
1754B
peer-keepalive destination
1738B peer-keepalive destination
175B
100.67.127.29 100.67.127.30
description "Peer-Link to
1740B description "Peer-Link to
175B
External-B" External-A"
switchport
174B switchport
1758B
1,1911 1,1911
vpc peer-link
174B vpc peer-link
176B
no shutdown
1745B no shutdown
1762B
switchport
1748B switchport
1765B
1,1911 1,1911
channel-group 1000 mode active
175B channel-group 1000 mode active
1768B
no shutdown
1752B no shutdown
1769B
Configure Configure the interfaces for connections to the SFS switches. Ports 1/49 and 1/50 are
28B
interfaces configured as L3 interfaces. The IP addresses used are from Table 19. Optionally, allow
the forwarding of jumbo frames using the mtu 9216 command.
Create port channel 1. In this example, port channel 1 connects to the DNS/NTP server. It
283B
is on VLAN 1911, which represents the preexisting management VLAN. Add the port
channel to vPC 1.
Interface 1/1 on each external switch is connected to the DNS/NTP server. Each interface
284B
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 157
Deployment Guide
N9K-External-A
126B N9K-External-B
127B
description Leaf1A
17B description Leaf1A
179B
no shutdown
172B no shutdown
1798B
no switchport
173B no switchport
179B
mtu 9216
174B mtu 9216
180B
ip address 192.168.1.0/31
175B ip address 192.168.2.0/31
180B
description Leaf1B
17B description Leaf1B
1803B
no shutdown
178B no shutdown
1804B
no switchport
179B no switchport
1805B
mtu 9216
1780B mtu 9216
1806B
ip address 192.168.1.2/31
178B ip address 192.168.2.2/31
1807B
interface port-channel 1
1782B interface port-channel 1
180B
switchport
1784B switchport
180B
vpc 1
178B vpc 1
183B
no shutdown
178B no shutdown
184B
switchport
179B switchport
187B
no shutdown
1795B no shutdown
182B
Configure BGP Configure a loopback interface to use for the BGP router ID.
285B
Allow BGP to distribute routes with the route-map allow permit command.
286B
Configure the BGP ASN with the router bgp command. The external switches share
287B
the same ASN. Use the address that was set for interface loopback0 as the router ID.
BGP adds to the route table for equal-cost multipath load balancing.
Specify the neighbor IP addresses and ASNs. Configure an address family for each
289B
neighbor.
When the configuration is complete, exit configuration mode and save the configuration
290B
158 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
External-A
128B External-B
129B
interface loopback0
182B interface loopback0
1840B
description router_ID
1823B description router_ID
184B
no shutdown
1824B no shutdown
1842B
ip address 10.0.2.1/32
1825B ip address 10.0.2.2/32
1843B
router-id 10.0.2.1
182B router-id 10.0.2.2
1846B
allow allow
maximum-paths 2
183B maximum-paths 2
1849B
65011 65011
address-family ipv4 unicast
183B address-family ipv4 unicast
185B
no shutdown
1834B no shutdown
1852B
65011 65011
address-family ipv4 unicast
1836B address-family ipv4 unicast
1854B
no shutdown
1837B no shutdown
185B
end
183B end
1856B
config config
Validate L3 After the uplink interfaces are configured on the Nexus external switches and on the SFS
291B
connections to leaf switches, connectivity can be verified using the switch CLI.
Cisco Nexus
switches Show command output on N9K-External-A
Note: The command output shown in the following commands is for the N9K-External-A switch.
The output for N9K-External-B is similar.
Run the show ip bgp summary command to verify that BGP is up for each neighbor.
29B
When BGP is up, uptime is shown in the Up/Down column. The neighbors for N9K-
External-A shown in the output below are Leaf1A and Leaf1B.
BGP summary information for VRF default, address family IPv4 Unicast
269B
BGP table version is 15, IPv4 Unicast config peers 2, capable peers 2
2694B
Neighbor
2698B V AS MsgRcvd MsgSent TblVer InQ OutQ Up/Down State/PfxRcd
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 159
Deployment Guide
192.168.1.1
269B 4 65011 2912 2529 15 0 0 1d18h 5
192.168.1.3
270B 4 65011 2907 2529 15 0 0 1d18h 5
Run the show ip interface brief command to verify IP addresses are configured
380B
correctly. VLAN 1911 is the external management VLAN that contains the DNS/NTP
server. Loopback 0 is the router ID, and interfaces 1/49-1/50 are connected to the SFS
leaf switches.
N9K-External-A# show ip interface brief
2701B
Interface
2703B IP Address Interface Status
Vlan1911
2704B 172.19.11.252 protocol-up/link-up/admin-up
Lo0
2705B 10.0.2.1 protocol-up/link-up/admin-up
Eth1/49
2706B 192.168.1.0 protocol-up/link-up/admin-up
Eth1/50
270B 192.168.1.2 protocol-up/link-up/admin-up
The show ip route command output for the N9K-External-A switch appears as shown.
293B
Note: The 172.18.11.0/24 External Management network has not yet been configured on the SFS
fabric, so it is not learned using BGP at this stage of deployment.
160 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Show command output on Leaf1A
Note: The command output shown in the following commands is for Leaf1A. The output for
386B
Leaf1B is similar.
Run the show ip bgp summary command to verify that BGP is up for each neighbor.
294B
When BGP is up, uptime is shown in the Up/Down column. The neighbors for Leaf1A
shown in the output below are Leaf1B, N9K-External-A, and N9K-External-B.
Neighbor
2739B AS MsgRcvd MsgSent Up/Down State/Pfx
172.16.0.1
2740B 65011 3222 3240 1d:22:14:58 8
192.168.1.0
2741B 65101 2794 3231 1d:18:29:11 4
192.168.2.0
274B 65101 2795 3226 1d:18:26:04 4
Run the show ip interface brief command to verify connected interfaces are up,
381B
1/1/52 are the VLTi, and 1/1/53-1/1/54 are the uplinks to the external switches. VLAN
4090, Loopback 1, and Loopback 2 are used internally by SFS. VLAN 4094 and port
channel 1000 are automatically configured for the VLTi.
Note: Unused interfaces have been removed from the output for brevity.
387B
Interface Name
274B IP-Address OK Method Status Protocol
================================================================================
2745B
Ethernet 1/1/1
2746B unassigned YES unset up up
Ethernet 1/1/2
274B unassigned YES unset up up
Ethernet 1/1/3
2748B unassigned YES unset up up
Ethernet 1/1/49
2749B unassigned YES unset up up
Ethernet 1/1/50
2750B unassigned YES unset up up
Ethernet 1/1/51
2751B unassigned YES unset up up
Ethernet 1/1/52
275B unassigned YES unset up up
Ethernet 1/1/53
2753B 192.168.1.1/31 YES manual up up
Ethernet 1/1/54
2754B 192.168.2.1/31 YES manual up up
Management 1/1/1
275B 100.67.127.28/24 YES manual up up
Vlan 4000
2756B unassigned YES unset up up
Vlan 4090
275B 172.16.0.1/31 YES manual up up
Vlan 4094
2758B unassigned YES unset up up
Port-channel 1000
2759B unassigned YES unset up up
Loopback 1
2760B 172.16.128.0/32 YES manual up up
Loopback 2
2761B 172.30.0.0/32 YES manual up up
Virtual-network 3939
276B unassigned YES unset up up
Run the show ip route command to verify routes to the External Management VLAN,
295B
172.19.11.0/24, have been learned using BGP from the Nexus switches. In this example,
two routes to 172.19.11.0/24 are learned, one using each Nexus switch. The routs are
shown in bold in the output below.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 161
Deployment Guide
S5248F-Leaf1A# show ip route
2763B
Codes: C - connected
2764B
S - static
2765B
Destination
27B Gateway Dist Last Change
----------------------------------------------------------------------------------
273B
B EX 10.0.2.1/32
274B via 192.168.1.0 20/0 00:43:16
via 192.168.2.0
275B
B EX 10.0.2.2/32
276B via 192.168.1.0 20/0 00:43:16
via 192.168.2.0
27B
C 172.16.0.0/31
278B via 172.16.0.1 vlan4090 0/0 02:19:46
C 172.16.128.0/32
279B via 172.16.128.0 loopback1 0/0 02:20:07
B IN 172.16.128.1/32
2780B via 172.16.0.0 200/0 02:19:44
B EX 172.19.11.0/24
2781B via 192.168.1.0 20/0 00:43:32
via 192.168.2.0
278B
C 172.30.0.0/32
2783B via 172.30.0.0 loopback2 0/0 02:20:07
C 192.168.1.0/31
2784B via 192.168.1.1 ethernet1/1/53 0/0 01:12:49
B IN 192.168.1.2/31
2785B via 172.16.0.0 200/0 01:09:12
C 192.168.2.0/31
2786B via 192.168.2.1 ethernet1/1/54 0/0 01:10:18
B IN 192.168.2.2/31
278B via 172.16.0.0 200/0 01:07:51
To continue deployment, go to the Configure a jump host port section of this guide.
296B
SmartFabric leaf switches. IP addresses are then manually assigned to each leaf switch
on the External Management network, 172.18.11.0/24 in this guide, as shown in the
Additional configuration steps for L3 uplinks section.
Once the items above are done, run the show ip route command on the external
298B
Note: The following command output is for the N9K-External-A switch. The output for N9K-
38B
External-B is similar.
278B
162 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
N9K-External-A# show ip route
IP Route Table for VRF "default"
2789B
To continue deployment, go to the Validate and build VxRail cluster section of this guide.
29B
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 163
Deployment Guide
Configure external Nexus switches for L2 connections
The external Nexus and SmartFabric leaf switches are cabled as shown in Figure 128 and
38B
are powered on. When L2 uplink configuration is complete, Leaf1A and Leaf1B connect
with a VLT port channel to a virtual PortChannel (vPC) on the external Nexus switches. In
this example, an existing DNS/NTP server also connects to the Nexus switches using a
vPC.
Pee r Li nk
N9K-External-A N9K-External-B
1 49 50 1 49 50
VLT/vPC
SmartFabric
vPC
53 54 53 54
Existing Leaf1A Leaf1B
3 2 1 1 2 3
DNS/NTP server
VLTi
L2 Connections
(All in VLAN 1811) VxRail node 1
VLTi/Peer Link
connections
VxRail node 2
VxRail node 3
Rack 1
Figure 128. L2 uplinks to external Nexus 9000 switches
312B
Note: DNS and NTP servers do not have to connect in this manner if they are reachable on the
389B
network.
All ports on the four switches shown in Figure 128 are in the External Management VLAN,
30B
Note: All Nexus switch configuration commands used to validate this topology are shown in the
3840B
sections that follow. These are only examples. Modify your Nexus external switch configuration as
needed for your environment.
General settings Enable the following features: interface-vlan, lacp, vrrp, vpc, lldp.
301B
Configure the hostname, OOB management IP address on VRF management, and the
VRF management route as shown.
Note: Nexus spanning tree settings are at their factory defaults in this example. You may
3841B
configure spanning tree on the Nexus switches as needed for your environment. On Dell leaf
switches in SmartFabric mode, spanning tree is disabled on L2 uplinks. See Dell EMC Networking
SmartFabric Services Deployment with VxRail for more information.
164 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
N9K-External-A
1230B N9K-External-B
123B
configure terminal
185B configure terminal
187B
feature
1859B interface-vlan feature
1872B interface-vlan
feature
1860B lacp feature
1873B lacp
feature
186B vrrp feature
1874B vrrp
feature
1862B vpc feature
1875B vpc
feature
1863B lldp feature
1876B lldp
hostname N9K-External-A
1864B hostname N9K-External-B
187B
interface mgmt 0
1865B interface mgmt 0
187B
ip address 100.67.127.30/24
186B ip address 100.67.127.29/24
1879B
no shutdown
186B no shutdown
18B
ip route 100.67.0.0/16
1870B ip route 100.67.0.0/16
183B
100.67.127.254 100.67.127.254
Configure the VLAN 1811 represents a preexisting management VLAN on the external network. DNS
302B
External and NTP services are located on this VLAN. Optionally, enable jumbo frames with the mtu
Management 9216 command.
VLAN
If traffic will be routed from the external switches to other external networks, assign a
30B
unique IP address on each switch and configure VRRP to provide gateway redundancy.
Assign the same virtual address to both switches.
N9K-External-A
123B N9K-External-B
123B
vlan 1811
184B vlan 1811
1895B
name ExtMgmt
185B name ExtMgmt
1896B
no shutdown
186B no shutdown
1897B
interface Vlan1811
187B interface Vlan1811
189B
description ExtMgmt
18B description ExtMgmt
189B
no shutdown
189B no shutdown
190B
mtu 9216
1890B mtu 9216
190B
ip address 172.18.11.252/24
189B ip address 172.18.11.253/24
1902B
vrrp 11
1892B vrrp 11
1903B
address 172.18.11.254
1893B address 172.18.11.254
1904B
no shutdown
1894B no shutdown
1905B
Configure the Create the vPC domain. The peer-keepalive destination is the OOB management IP
304B
and allow the default and External Management VLANs, 1 and 1811 respectively.
Configure the interfaces to use in the vPC peer link. Put the interfaces in trunk mode and
306B
allow the default and External Management VLANs, 1 and 1811 respectively. Add the
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 165
Deployment Guide
interfaces to the peer link port channel. Port-channel 1000 is set as an LACP port-channel
with the channel-group 1000 mode active command.
N9K-External-A
1234B N9K-External-B
1235B
role priority 1
1907B role priority 65535
1924B
peer-keepalive destination
1908B peer-keepalive destination
1925B
100.67.127.29 100.67.127.30
description "Peer-Link to
190B description "Peer-Link to
1927B
External-B" External-A"
switchport
19B switchport
1928B
1,1811 1,1811
vpc peer-link
194B vpc peer-link
193B
no shutdown
195B no shutdown
1932B
switchport
198B switchport
1935B
1,1811 1,1811
channel-group 1000 mode active
192B channel-group 1000 mode active
1938B
no shutdown
192B no shutdown
193B
Configure
329B
Configure the interfaces for connections to the SFS leaf switches. Interfaces 1/49 and
307B
interfaces 1/50 are configured in vPC 100 in this example. Port-channel 100 is set as an LACP port-
channel with the channel-group 100 mode active command.
Use the switchport mode trunk command to enable the port-channel to carry traffic
308B
for multiple VLANs. Allow VLAN 1811 (the External Management VLAN).
Optionally, allow the forwarding of jumbo frames with the mtu 9216 command.
309B
In this example, interface 1/1 on each external switch is configured in vPC 1 for
310B
When the configuration is complete, exit configuration mode and save the configuration
31B
166 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
N9K-External-A
1236B N9K-External-B
1237B
switchport
1942B switchport
1974B
1811 1811
vpc 100
1945B vpc 100
197B
mtu 9216
1946B mtu 9216
1978B
no shutdown
1947B no shutdown
197B
switchport
1950B switchport
1982B
1811 1811
mtu 9216
1953B mtu 9216
1985B
no shutdown
195B no shutdown
1987B
interface port-channel 1
1956B interface port-channel 1
198B
switchport
1958B switchport
190B
vpc 1
196B vpc 1
193B
no shutdown
1962B no shutdown
194B
switchport
1965B switchport
197B
no shutdown
196B no shutdown
201B
end
1970B end
20B
config config
Validation Once the uplink interfaces have been configured in the SFS GUI and on the external
312B
Show command Note: The command output shown in the following commands is for Leaf1A. The output for
3842B
With SFS, port channel numbers are automatically assigned as they are created. In this
31B
example, port channel 1 is the uplink connected to the Nexus switches. It has two
members that are both up and active. Port channel 1000 is reserved for the VLTi.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 167
Deployment Guide
S5248F-Leaf1A# show port-channel summary
280B
Flags:
281B D - Down I - member up but inactive P - member up and active
U - Up (port-channel)
28B F - Fallback Activated
--------------------------------------------------------------------------------
283B
Group Port-Channel
284B Type Protocol Member Ports
--------------------------------------------------------------------------------
285B
1
286B port-channel1 (U) Eth DYNAMIC 1/1/53(P) 1/1/54(P)
1000 port-channel1000 (U)
287B Eth STATIC 1/1/49(P) 1/1/50(P) 1/1/51(P)
1/1/52(P)
28B
The L2 uplink, port channel 1 in this example, is a tagged member of VLAN 1811. This is
384B
Unicast-Drop
Un-tagged VLAN: 4080
2831B
VLTi-VLAN: 1811
283B
Members:
2834B
Untagged: ethernet1/1/9:1
2835B
ethernet1/1/3
VxLAN Virtual Network Identifier: 1811
2837B
Remote-VTEPs (flood-list):
2839B
VLTi-VLAN: 1812
2841B
Members:
284B
ethernet1/1/3
VxLAN Virtual Network Identifier: 1812
284B
Remote-VTEPs (flood-list):
2846B
VLTi-VLAN: 1813
284B
Members:
2849B
ethernet1/1/3
VxLAN Virtual Network Identifier: 1813
2851B
Remote-VTEPs (flood-list):
2853B
VLTi-VLAN: 1814
285B
Members:
2856B
ethernet1/1/3
168 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
VxLAN Virtual Network Identifier: 1814
285B
Remote-VTEPs (flood-list):
2860B
VLTi-VLAN: 1815
286B
Members: 2863B
ethernet1/1/3
VxLAN Virtual Network Identifier: 1815
2865B
Remote-VTEPs (flood-list):
2867B
VLTi-VLAN: 3939
2870B
Members: 2871B
Remote-VTEPs (flood-list):
2875B
VLTi-VLAN: 4091
287B
Members: 2879B
Remote-VTEPs (flood-list):
284B
Use the show vlt 255 vlt-port-detail to verify the status of VLT ports. Port
385B
channel 1 is the L2 uplink to the Nexus switches. The output shows information for both
VLT peer switches. An asterisk (*) denotes the local switch. In this case, Leaf1A is VLT
unit 1, and Leaf1B is VLT unit 2.
S5248F-Leaf1A# show vlt 255 vlt-port-detail
285B
vlt-port-channel ID : 1
286B
VLT Unit ID
287B Port-Channel Status Configured ports Active ports
-------------------------------------------------------------------------------
28B
* 1
289B port-channel1 up 2 2
2 2890B port-channel1 up 2 2
Show command Note: The command output shown in the following commands is for the N9K-External-A switch.
384B
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 169
Deployment Guide
The show port-channel summary command confirms port channels are up. Po1
386B
connects to the DNS/NTP server, Po100 connects to the SFS leaf switches, and Po1000
is the peer link.
N9K-External-A# show port-channel summary
2891B
Flags: D - Down
289B P - Up in port-channel (members)
I - Individual H - Hot-standby (LACP only)
2893B
s - Suspended
2894B r - Module-removed
S - Switched
2895B R - Routed
U - Up (port-channel)
2896B
--------------------------------------------------------------------------------
289B
Group Port-
290B Type Protocol Member Ports
Channel
2901B
--------------------------------------------------------------------------------
290B
1
2903B Po1(SU) Eth LACP Eth1/1(P)
100
2904B Po100(SU) Eth LACP Eth1/49(P) Eth1/50(P)
1000 Po1000(SU) Eth
2905B LACP Eth1/51(P) Eth1/52(P)
Run the show vlan command to verify ports are correctly assigned to the External
387B
Management VLAN (VLAN 1811). Po1 connects to the DNS/NTP server, Po100 connects
to the SFS leaf switches, and Po1000 is the peer link.
N9K-External-A# show vlan
2906B
VLAN
2907B Name Status Ports
----
2908B -------------------------------- --------- -------------------------------
1
290B default active Po1000, Eth1/51, Eth1/52
1811
2910B ExtMgmt active Po1, Po100, Po1000, Eth1/49
Eth1/50, Eth1/51, Eth1/52
291B
VLAN
291B Type Vlan-mode
----
2913B ----- ----------
1
2914B enet CE
1811
2915B enet CE
-------------------------------------------------------------------------------
2917B
Primary
2918B Secondary Type Ports
-------
291B --------- --------------- -------------------------------------------
Run the show vpc command to verify all vpc connections are up. In this example,
314B
Po1000 is the peer link, Po1 connects to the DNS/NTP server, and Po100 connects to the
SFS leaf switches.
Legend:
291B
vPC domain id
293B : 129
170 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Peer status
294B : peer adjacency formed ok
vPC keep-alive status
295B : peer is alive
Configuration consistency status
296B : success
Per-vlan consistency status
297B : success
Type-2 consistency status
298B : success
vPC role
29B : primary
Number of vPCs configured
2930B : 2
Peer Gateway
2931B : Disabled
Dual-active excluded VLANs
293B : -
Graceful Consistency Check
293B : Enabled
Auto-recovery status
2934B : Disabled
Delay-restore status
2935B : Timer is off.(timeout = 30s)
Delay-restore SVI status
2936B : Timer is off.(timeout = 10s)
---------------------------------------------------------------------
2938B
id
293B Port Status Active vlans
--
2940B---- ------ --------------------------------------------------
1
2941BPo1000 up 1,1811
vPC status
294B
----------------------------------------------------------------------
2943B
id
294B Port Status Consistency Reason Active vlans
--
2945B---- ------ ----------- ------ ------------
1
2946BPo1 up success success 1811
100
2947B Po100 up success success 1811
Qty
1238B Item
1239B Operating system version
1240B
2
127B Cisco Nexus 93180YC-EX switches
128B 7.0(3)I4(2)
129B
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 171
Deployment Guide
Appendix D Support and Feedback
172 Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0
Deployment Guide
Technical resources
Dell EMC Networking Info Hub
316B
Dell EMC OpenManage Network Integration for VMware vCenter – This site includes
32B
OMNI software and the SmartFabric Services for OpenManage Network Integration User
Guide, Release 1.3
the planning, design, and deployment of network fabrics that power Dell EMC compute,
storage, and hyperconverged infrastructure solutions. The FDC is ideal for turnkey
solutions and automation based on validated deployment guides.
guides. For additional information, go to the Dell EMC Fabric Design Center.
Dell EMC Networking SmartFabric Services Deployment with VxRail 7.0 173
Deployment Guide