Professional Documents
Culture Documents
VxRail 7.0.XXX Administration Participant Guide
VxRail 7.0.XXX Administration Participant Guide
VxRail 7.0.XXX Administration Participant Guide
7.0.XXX
ADMINISTRATION
PARTICIPANT GUIDE
PARTICIPANT GUIDE
ESCPXD05344 ~ VxRail 7.0.XXX Administration
Perform Capacity Expansion for an Existing VxRail Node (Add Disks) 184
VxRail Node - Storage Capacity Expansion 184
Add Disk Wizard - Disk Group Add 185
Interaction: Add Disks to a VxRail Node 191
Glossary 315
VMware Technologies
VxRail Manager
VxRail clusters are deployed with one of two vCenter options. The first
option is to use the VxRail-managed vCenter Server.
Considerations:
Includes a nontransferable standard vCenter license for VxRail nodes
only
Uses VxRail Lifecycle Management for vCenter software upgrades
Does not support vCenter High Availability (HA)
Deployed with a noncustomizable vSphere Single Sign-on (SSO)
domain called vsphere.local
Supports internal or external DNS server options
Considerations:
Includes a nontransferable standard vCenter license for VxRail nodes
only
Uses VxRail Lifecycle Management for software upgrades
Does not support vCenter HA
Uses existing noncustomizable vSphere SSO domain called
vsphere.local
A new VxRail cluster is deployed with an existing vCenter server that runs
on infrastructure that is external to a VxRail cluster. This option is used for
single or multiple cluster deployments.
Considerations:
Requires a separate vCenter license
Uses customer-defined procedures for vCenter LCM
Supports vCenter HA
Supports a custom vSphere SSO domain
Requires an external DNS server
Supports a new VxRail-deployed or existing Virtual Distributed
Switches
The physical switch infrastructure supports all the network traffic for the
VxRail cluster. VxRail logical networking depends on the ToR switches for
physical connectivity between the nodes, and to the core production
network. VxRail network traffic is configured for Layer 2 networking. For
supported VxRail network speeds, interfaces, and supported network card
options, see the Dell VxRail Spec Sheet.
VxRail nodes connect to one or more ToR network switches and are
configured as part of the VxRail cluster.
VxRail nodes with two ToR switches and optional management switch
Each VxRail node has a 1 GbE RJ45 port for Integrated Dell Remote
Access Controller (iDRAC) connectivity.
VxRail networking is grouped into six types of traffic. Each traffic type is
associated with a distributed port group that exists on one or two VDSs.
Discovery
Usage: Internal network for VxRail node discovery during initial
implementation and cluster expansion
VLAN ID: Configurable, default 3939
Management
Usage: External network for ESXi host management
VLAN ID: Configurable, default untagged
vCenter Server
Usage: External network for vCenter Server and VxRail Manager
management
VLAN ID: Configurable, default untagged
vMotion
Usage: Internal network for cluster vMotion traffic
VLAN ID: Configurable
vSAN
The predefined port group uplink configuration for the VxRail deployed
VDS depends on the number and type of node ports that are used for
VxRail traffic. Two uplinks are configured for each port group, one active
and one standby. Predefined profiles require specific cabling. To learn
more about the cabling and port group uplink usage for the predefined
network profiles, select each tab.
The cabling for 4x25 GbE (2x25 GbE Integrated NIC + 2x25 GbE PCIe)
profile is different from cabling for the 4x10 GbE (All Integrated NIC ports)
profile. Integrated NIC P0 is cabled to Switch 1, while PCIe Port P0 is
cabled to Switch 2. These two ports form a HA failover pair. In a similar
manner, Integrated NIC P1 is cabled to Switch 2, while PCIe P1 is cabled
to Switch 1. These two ports form an HA failover pair as well.
The port groups in a VxRail deployed VDS with a predefined profile are
configured with the teaming and failover settings that are shown in the
table.
Post-Deployment Considerations
Consider a VxRail system that has been deployed with the predefined
4x10 GbE (all Integrated NIC ports) network profile.
The Integrated NIC on each node is a single point of failure.
Since each port group is configured with one active and one standby
uplink, load balancing is not available.
Only one VDS is configured for all network traffic.
The VxRail nodes must have available PCIe ports that are cabled to the
Top of Rack (ToR) switches. The ToR switch ports must be configured to
support VxRail traffic.
Example 2 - Reconfigured to support NIC redundancy - 2x10 GbE Integrated NIC and
2x10 GbE PCIe Ports
By default, the port groups for VxRail networks with predefined profiles are
configured with one active and one standby uplink; load balancing is not
available. However, the vSAN and vMotion networks could benefit from
load balancing across the NICs.
vmnic0 and vmnic3 carry the VxRail management traffic, they are
configured as an active/standby pair with no link aggregation.
The default VxRail configuration is a single VDS for all VxRail networks.
VxRail can be deployed with two VDS to support network segmentation
requirements. The two virtual switches are used to separate the VxRail
networks.
Either two or four uplinks can be assigned to each VDS. In the example,
VDS 1 carries the VxRail Management traffic, while VDS 2 carries the
vSAN and vMotion traffic.
VxRail Manager is used for the hardware and software life cycle
management and serviceability of VxRail. VxRail Manager is not used for
virtualization management.
1: Deploy
Initial configuration
Automation used for deployment tasks
2: Update
3: Monitor
Health
Events
Operational state
4: Maintain
The vSphere Client is the vCenter Server user interface for managing the
virtual environment. The vCenter Server managing a VxRail vSAN cluster
contains components that are configured during the VxRail deployment.
Components include the ESXi hosts, vSAN cluster, vSAN datastore,
VxRail Manager VM, and the VxRail Plugin.
VxRail Plugin
The VxRail Plugin for vCenter Server provides VxRail cluster management
from the vSphere Client. VxRail actions are available both at the VxRail
cluster and VxRail node levels.
VxRail Licensing
The VxRail system is shipped with 60-day evaluation licenses for vSphere
and vSAN. The vSphere evaluation period starts when the VxRail nodes
are first powered on. New licenses must be applied before the end of the
evaluation period.
1: The licenses tab shows all the available licenses that are applied to the
system. If new vSphere and vSAN licenses are needed, they can be
2: The Products tab displays the available licenses that are organized by
product.
3: The Assets tab displays the license status of various assets like
vCenter Server Systems, Hosts, vSAN Clusters, Supervisor Clusters, and
Solutions. The Assets tab is also used to assign available license to the
various assets. To assign an available license, select the asset and click
Assign License.
Additional information:
Software Licensing Options for VxRail
VMware vSphere and vSAN Editions Feature
Comparison
About ESXi Evaluation and Licensed Modes
Considerations about the vSAN License
How to activate Partner Activation Codes (PAC) for
vSphere License
License Management
In this example, vSphere Client is used to add and assign a vSAN license
to a VxRail cluster. To learn more about the process, click each tab.
View License
Add License
There are multiple ways to add a license. One of the ways is to go to the
Licenses tab and select ADD. The Licenses tab lists all the available
licenses. The ADD option allows adding multiple licenses for multiple ESXi
hosts, CLUSTERS, and SOLUTIONS. Clicking ADD launches the New
Licenses wizard.
Assign License
After the license keys are added, go to the Assets > VSAN CLUSTERS
tab to assign the licenses to the applicable asset. Select the VxRail
Cluster asset and click ASSIGN LICENSE to launch the Assign License
wizard. In this example, the newly added vSAN license has being
assigned to the VxRail Cluster asset.
VxRail Plugin on the Client Plugins page showing deployed and enabled
VxRail. This section is not fully populated until a support account is set up
and the VxRail Manager has Internet connectivity.
At the cluster level, the Configure > VxRail > System page displays
system information about VxRail, and provides links to product
documentation, the privacy statement, and software updates. The System
page also provides information about the cluster. Clicking the question
mark in the System Information section opens VxRail Help in a new
browser tab.
The Updates page displays the installed components along with their
version and facilitates software upgrades. On this page, administrators
can generate a compliance report. The report compares the expected
software and firmware versions for VxRail components to the versions
installed on the system.
VxRail Plugin Updates page showing installed components and their version
The Certificate page displays the current VxRail Manager TLS Certificate.
During deployment, a self-signed certificate is generated for VxRail
Manager. Administrators can use this page to manually update the
customer-supplied certificate or configure automated renewals.
VxRail Plugin Certificate page showing that the cluster is using a valid TLS Certificate
The Hosts page displays details of the VxRail hosts in the cluster.
Administrators can use this page to add new VxRail nodes or update the
hostname and management IP address of the ESXi hosts.
The Networking page displays settings to manage Internet access for the
VxRail Manager VM, configure proxy settings and configure traffic
throttling. The proxy settings allow VxRail to connect to an external
network. Traffic throttling limits the communications between VxRail
Manager and vCenter.
VxRail Plugin Health Monitoring page showing that health monitoring is enabled
At the host level, the Configure > VxRail > iDRAC Configuration page
displays the iDRAC settings for the selected node. Administrators can use
this page to add users and change network settings.
VxRail Plugin iDRAC Configuration page showing IPv4 and VLAN Settings
At the cluster level, the Configure > VxRail > Support page allows
administrators to connect VxRail Manger to Dell Support resources. By
linking to support resources, administrators can access the VxRail
Community and KB articles from the VxRail Dashboard. Administrators
can also access the VxRail Market and perform Internet software
upgrades using the VxRail Plugin.
To learn how to link the support account to VxRail Manager, select each
tab.
Select the EDIT button to open the Edit Dell Technologies Support
Account wizard.
VxRail Plugin Support page with the Support Account Edit button identified
Enter the Support Account credentials to link the account with the VxRail
system and allow access to support services.
VxRail Plugin Support page with the Connectivity Enable button identified.
Enable Connectivity wizard - Network Connection step with the Connect Directly
configuration options identified
Network Connection -
Connect using Gateway
Enable Connectivity wizard - Network Connection step with the Connect using Gateway
Server configuration options identified.
Enable Connectivity wizard - Customer Improvement Program step with the data
collection options identified
At the cluster level, the Monitor > VxRail > Physical view page shows
the physical view of all VxRail nodes in the VxRail cluster. The top of the
page shows details about the cluster. To view information about an
individual VxRail node, select the node. The information panel on the right
presents the node-specific information.
Click the ACTIONS drop-down menu to view the available options for the
node.
1:
At the host level, the Monitor > VxRail > Physical view page shows the
front and back view of the selected VxRail node. The top of the page
shows details about the node. The information panel on the right presents
additional details of the node.
To learn about the details that are presented for the node and
maintenance activities available, select each hotspot.
1: The BOOT DEVICE tab shows information about the boot devices in
the node.
2: The ALERTS tab shows alerts that are related to the node. The alerts
are displayed with a link to the related Dell Technologies Knowledge Base
(KB) article.
The ALERTS tab showing no new alerts found for the node
3: The actions available on this page are the same as the actions
available on the Physical View page of the cluster.
4: The OVERVIEW tab shows details about the node, such as the power
status, the health, and the firmware versions.
To view the details of a disk, select the disk in the Front View of the node.
To learn about the details that are presented for the selected disk and
maintenance activities available, select each tab.
Information
The INFORMATION tab shows details such as the health, disk type, and
the remaining write endurance.
Alerts
The ALERTS section shows alerts that are related to the disk. Alerts are
generated when the remaining write endurance reaches certain levels:
Warning: 30%
Error: 20%
Critical: 5%
The ALERTS tab showing no new alerts found for the selected disk
Actions
To view the NIC details of a node, select the network interface ports in the
Back View.
To learn about the details presented for the selected NIC, select each tab.
Information
The INFORMATION tab shows details such as the MAC address of the
NIC port, link speed, and the link status.
Driver Version
DRIVER VERSION tab showing the current driver version for the selected NIC
To view the details of a power supply, click the power supply in the Back
View of the Node.
To learn about the details presented for the selected power supply, select
each tab.
Information
The INFORMATION tab shows details such as the health, serial number,
and part number of the power supply.
Alerts
The ALERTS section shows alerts that are related to the power supply.
The alerts have links for the related Dell Technologies KB articles.
The ALERTS tab showing no new alerts found for the selected power supply.
vSAN Datastore
A vSAN cluster consists of two or more physical ESXi hosts. Each host
either contains all-flash devices or a combination of flash and HDDs.
These devices contribute cache and capacity to the vSAN datastore. The
cache device is always a flash device, and the capacity devices could be
either all flash or all hard drive. For example, in an all-flash system, both
the cache and capacity drives are flash. In a hybrid system, all the
capacity disks are HDDs. Each host can have one or more disk groups. A
disk group contains one cache device and one or more capacity devices.
The disk groups of all ESXi hosts in a vSAN cluster are combined to
create a vSAN datastore. The datastore capacity depends on the number
and the size of capacity disks on each ESXi host, and the number of ESXi
hosts in the cluster.
1: This VMDK object uses a vSAN storage policy with the Failures to
tolerate set to 1 failure - RAID 1 (Mirroring).
VM Storage Policies
RAID 1 Mirroring 1, 2, or 3
To manage VM Storage Policies from within the vSphere Client, select the
Menu icon in the upper-left corner. In the navigation bar, select Policies
and Profiles. In the Policies and Profiles window, select VM Storage
Policies. For more information about the vSAN Default Storage Policy,
see the VMware About the vSAN Default Storage Policy document.
To learn more about changing the default storage policy for a vSAN
Datastore, see the VMware Change the Default Storage Policy for vSAN
Datastores document.
VM Storage Policies
1: The VM Compliance tab shows which VMs are using the selected
storage policy and their compliance status. The VxRail Manager and the
VMware vCenter Server Appliance VMs are examples of VxRail system
VMs that use the VXRAIL-SYSTEM-STORAGE-PROFILE. vSAN
monitors and reports on the policy compliance status of the VMs.
VM Template tab showing there are not any VMs associated with this storage policy
Image showing the Storage Compatibility showing the policy is compatible with the
VxRail Virtual vSAN Datastore
4: The Rules tab shows general and placement information about the
selected storage policy. In this example, a standard VxRail cluster is
deployed with VxRail-managed vCenter Server. The number of failures to
tolerate is set to 1. Object space reservation is set to Thick provisioning.
A VM with this policy applied is guaranteed storage and is RAID 1
protected.
1: The vSAN Skyline Health check runs every 60 minutes by default. The
date and time of the last check is displayed here.
Online Health
5: The vSAN Disk Balance test checks each healthy disk in the vSAN
cluster, and gets a summary of the disk usage. If the imbalance exceeds a
system-determined threshold, vSAN Disk Balance shows the warning.
6: Move the slider to the right to view vSAN health history. vSAN stores up
to 30 days of health data.
7: The overview shows the health check count on the health status,
Primary issues with the nested issues, and Other health issues. The
nested issues are the result of the primary issue and resolving the primary
issue might solve the related nested issues.
The Virtual Objects page displays the Object State and the Storage
Policy for each virtual object. To monitor vSAN virtual objects, select the
Monitor tab for the cluster and then Virtual Objects under the vSAN
heading.
1:
To monitor vSAN resyncing, select the Monitor tab and then Resyncing
Objects under the vSAN heading.
The Object repair timer field is at the top of the window. This field
displays the amount of time vSAN waits before repairing an object after a
host failure or during host maintenance. The repair timer only applies to
Absent vSAN objects as opposed to Degraded objects which are repaired
immediately (if possible). The page that is displayed below shows the
Total resyncing objects, Bytes left to resync, and the Total
resynching ETA. If there are any scheduled resync operations, they are
also displayed next to the Scheduled resyncing field.
The vSAN Capacity view displays the capacity of the vSAN datastore,
deduplication and compression efficiency, and the breakdown of capacity
usage.
To monitor the capacity of a vSAN cluster, select the Monitor tab for the
cluster and then Capacity under the vSAN heading.
2: The What if analysis displays the estimated Effective free space with
the selected storage policy. The storage policy selection options are
dependent upon the policies that exist. In the example, the storage policy
that is selected is vSAN Default Storage Policy.
In case the space efficiency options were enabled, the What if analysis
can report Effective free space due to Compression or Deduplication and
Compression.
The VM tab at the top of the performance page is selected. vSAN displays
performance charts for clients running on the cluster, including IOPS,
throughput, latency, congestions, and outstanding I/O. The statistics on
these charts are aggregated from the hosts within the cluster.
vSAN cluster level performance page, displaying metrics about clusters in the
perspective of VM consumption
vSAN performance charts are also available at the host level. Select a
host under the VxRail cluster > Monitor > vSAN > Performance.
The Frontend data is any I/O activity before vSAN processing. The
Backend is any I/O activity including vSAN processing such as
encryption, deduplication, compression, parity, mirroring, striping,
witnesses, checksums, and metadata.
VM Level Performance
The two techniques are unrelated but work together to ensure space
efficiency. Deduplication removes redundant data blocks, while
compression removes additional redundant data within each data block.
Determining whether to enable compression or deduplication/compression
is based on if a given dataset benefit enough to justify enabling one
setting or the other. The decision to implement is workload-dependent.
Space efficiency options are disabled by default. The options are enabled
at the vSAN cluster level.
1
Capacity savings are workload-dependent and not guaranteed.
VMware recommends the following space efficiency settings for the best
balance of capacity savings and performance impact:2
2
Table source: VMware vSAN Virtual Blocks blog -
https://blogs.vmware.com/virtualblocks/2020/09/22/space-efficiency-using-the-new-
compression-only-option-in-vsan-7-u1/.
vSAN space efficiency options are enabled from the vSAN Services
page.
To learn more about enabling vSAN space efficiency, select each hotspot.
1: The Space efficiency radio buttons are used to select the required
space efficiency option.
2: The Allow reduced redundancy option is used when the vSAN cluster
might not have enough resources for the disk groups to be fully
evacuated. For example, a three-node cluster with FTT set to 1 does not
have the resources to evacuate data for the disk group reformat. The
Allow reduced redundancy option keeps the VMs running. However the
VMs might be unable to tolerate the full level of failure defined in the
storage policy. As a result, during the format change, the VMs might be
temporarily at risk of experiencing data loss. vSAN restores full
compliance and redundancy after conversion is complete.
The recent tasks page shows all the tasks that are associated with the
disk reformatting.
The storage savings depend on different factors such as the datatype and
the number of duplicate blocks. A larger disk group potentially provides a
higher deduplication ratio. Space efficiency capacity savings are
monitored from the vSAN Capacity page.
compression are applied. You can further expand each usage category.
The Overhead for compression is shown in the graphic.
vSAN Encryption
Additional information:
Using Encryption in a vSAN Cluster
3
The data for this table was sourced from the Using Encryption in a vSAN Cluster
document - https://docs.vmware.com/en/VMware-
vSphere/7.0/com.vmware.vsphere.vsan.doc/GUID-F3B2714F-3406-48E7-AC2D-
3677355C94D3.html
Data and metadata is encrypted Data is encrypted using XTS AES 256
using AES 256-bit encryption. cipher, in both the cache and capacity
tiers of the vSAN datastore.
4
A Witness host applies only to vSAN 2-node and vSAN stretched cluster
architectures.
To add a key provider, in the left pane, select the vCenter Server. Then
select the Configure tab. In the middle pane, scroll down to Security,
select Key Providers, and click ADD.
Establish Trust
A trust must be established between the KMS and the vCenter Server
after adding a key provider. Establishing trust is only required when
configuring a Standard (External) Key Provider.
Image showing the location of the Establish Trust option required when configuring an
external key provider
Additional information:
Set up the Standard Key Provider
Set up a Native Key Provider
VMware Certified Key Management Servers list
vSAN Services page with the vSAN Services options window open
The Wipe residual data check box erases residual data from devices
before enabling encryption. Keep this box cleared unless you want to wipe
existing data from the storage devices when encryption a cluster
containing VM data.
The Key provider defaults to the key provider that has been added to the
vCenter Server.
3: vSAN initiates the rekey process to generate new keys at the scheduled
intervals. The default rekey interval is set to one day. The Rekey interval
can be changed to match corporate compliance requirements.
You can also set a CUSTOM Rekey interval by manually entering the
values between 30 to 10080 minutes.
Slack space is not a hard limit that vSAN enforces. The ratio was a
conservative estimation that is based on an approximation for all
environments. However, often the reserve capacity was underutilized.
The second is a 24-node VxRail cluster with total capacity of 670.68 TB.
vSAN reserves 7.86% of the capacity. This percentage is well below the
slack space recommendation.
vSAN Capacity Reserve is enabled either from the vSAN Capacity page,
or the vSAN Services page.
Either method opens the Reservation and Alerts dialog. On the dialog
page, Host rebuild reserve can only be enabled after enabling
Operations reserve. It is optional for administrators to enable Operations
Reserve and Host rebuild reserve.
percentages between the highest and lowest consumed disk to less than
half the rebalancing threshold.
vSAN Disk Balance Overview tab shows the disk usage statistics for the cluster.
990749567
The web version of this content contains an interactive activity.
VM Storage Policies are created from the VM Storage Policies page of the
vSphere client. Go to Menu > Policies and Profiles. In the left pane,
select VM Storage Policies.
On the Name and description page, create a unique name and add an
optional description for the new policy.
Policy structure
On the Policy structure page, check Enable rules for "vSAN" storage.
A vSAN page is added to the wizard after enabling the rules for vSAN
storage.
vSAN Availability
vSAN Availability With Site disaster tolerance and Failures to tolerate identified
Mirroring supports one, two, or three Fault Domain failures. For N failures
tolerated, N+1 copies of the object are created.
Erasure coding supports one Fault Domain failure with RAID 5 or two
Fault Domain failures with RAID 6. Erasure coding is supported on All-
flash systems only.
The example below places VMs on datastores which are on an all flash
storage tier, with no preference for encryption services, or space
efficiency.
For additional vSAN policy rules, see the vSAN Advanced Policy Rules.
Storage compatibility
The Storage compatibility page lists the datastores compatible with the
defined vSAN rules. In this example, there is only one vSAN datastore that
meets the requirements of the configured storage policy.
On the Review and finish page, Review the defined policy. Clicking
FINISH creates the policy.
2: Clicking EDIT opens the Edit VM Storage Policy wizard. This wizard is
similar to the Create VM Storage Policy wizard.
3: Clicking CLONE opens the Clone VM Storage Policy wizard. Use this
wizard to create a storage policy using the existing policy as a template.
VM Compliance tab
The default storage policy for a vSAN datastore is the vSAN Default
Storage Policy. If a storage policy is not explicitly selected during the
deployment of a new VM, the default policy is used.
The default storage policy for the vSAN datastore can be changed. For
space efficiency, administrators with all-flash VxRail systems may choose
to implement a storage policy with RAID 5 erasure coding as the default
policy. In the example, a storage policy that is named RAID5 Policy has
been created. This policy can be made the default for the vSAN datastore.
Image of the Change Default Storage Policy selection window for the VxRail vSAN
Datastore Default Storage Policy
New VM
OVF template
VM from template
Clone VM
The Policies page is located under the Configure tab for the Virtual
Machine. The VM Policies page is used to monitor the storage policy
compliance and to change the assigned storage policy.
The Policies page displays all the vSAN objects that belong to the VM,
the associated storage policy, and the compliance status.
Best practices:
Ensure that sufficient vSAN storage capacity exists for the rebuilds.
Apply new policies one object at a time and inspect the impact.
Ensure that there are enough fault domains to meet the new policies
requirements.
Changing the storage policy of this object to the vSAN Default Storage
Policy would result in the creation of new vSAN components.
The default vSAN policy is a RAID 1 policy, the change in the RAID layout
requires the creation of two data and one witness component.
The Physical disk placement view of RAID 1 object - two data and one witness
components
Until the policy change is complete, two complete sets of components co-
exist and consume about 333 GB (133 GB + 200 GB) of storage.
The vSAN HCI Mesh client clusters5 can be standard vSAN clusters or
vSAN HCI Mesh compute clusters6.
5
Client cluster -The vSAN cluster to which a remote vSAN datastore has
been mounted.
6
vSAN HCI Mesh compute clusters do not have local vSAN datastores.
These clusters do not require a vSAN license.
vSAN HCI Mesh configuration - VxRail Cluster 1 is the server cluster and VxRail Cluster
2 is a client cluster
Requirements:
Clusters in the same virtual data center in vCenter
vSAN Enterprise license or higher for the server cluster7
vSAN network connectivity between clusters8
Layer 2 or Layer 3
Latency below 5 milliseconds
Scalability:
A server cluster can export its vSAN datastore to a maximum of five
client clusters.
A client cluster can mount a maximum of five remote vSAN datastores.
A vSAN datastore can be mounted on a maximum of 128 ESXi hosts
(includes both server and client hosts).
Supported features:
vSphere vMotion and Storage vMotion
vSphere HA
vSphere DRS
vSAN Storage Policy Based Management
Unsupported features:
Data-In-Transit encryption
7
Server cluster - The vSAN cluster where the storage is locally hosted.
This vSAN cluster provides storage resources to the client clusters.
8
VMware recommends 25 GbE networks for vSAN HCI Mesh configurations.
VxRail Cluster Shut Down wizard - Example of a validation failure due to VMs which are
still running on a client cluster
A client cluster enters the Datastore with All Paths Down (APD)
condition 60 seconds after detecting a cross-cluster communication issue.
VMware recommends setting the vSphere HA response for a datastore
with APD to Power off and restart VMs. VMs are restarted when the
datastore becomes available.
In the example, the vCenter Server manages two VxRail clusters, VxRail-
DEV and VxRail-MGT. Both the clusters reside in the data center VxRail-
DC.
Local datastores are identified with the (Local) prefix. For the VxRail-DEV
cluster, only the local vSAN datastore DEV-vSAN-Datastore is mounted.
The mount wizard displays all the available remote vSAN datastores.
Select datastore step of the Mount Remote Datastore wizard - One remote vSAN
datastore is available
Compatibility Check
VM Deployment
New VMs can be deployed with remote vSAN storage. The remote vSAN
datastore is listed as a compatible datastore as long it satisfies the
required storage policy. In the example, MGT-vSAN-Datastore is a
remote vSAN datastore.
Select storage page of the Deploy OVF Template wizard - Compatible remote vSAN
datastore
Storage vMotion
Storage vMotion can be used to migrate VMs to and from remote vSAN
storage. In the example, VM2 is hosted on the VxRail-DEV cluster and
uses storage from the local vSAN datastore DEV-vSAN-Datastore. Based
on the existing storage policy, remote vSAN datastore MGT-vSAN-
Datastore is compatible.
Select storage page of the Migrate VM wizard - Compatible remote vSAN datastore
To learn more about the health of vSAN Objects and components, select
each tab.
Virtual Objects
Go to the Monitor tab of the VxRail cluster. Under vSAN select Virtual
Objects. On the below page, VMs listed in the Name column have objects
using a storage policy that cannot fulfill the expected number of fault
domains.
To view, the component Physical Placement, select any object under the
Virtual Objects page and click VIEW PLACEMENT DETAILS.
The amount of minutes vSAN waits before repairing an object after a host
is either in a failed state (absent failures) or in maintenance mode.
Some environments may not want their data unprotected for the 60 minute
default or may even have requirements for thresholds to be less than 60
minutes.
To edit the Object repair timer, from the Configure tab of the vSAN
cluster, choose EDIT under the Advanced Options of vSAN Services.
Configure Object repair timer page - 60 minutes is a default Object repair timer
The page for vSAN object health includes a button for rebuilding
components immediately, rather than waiting as specified by the Object
repair timer.
Before the Object repair timer expiry, the administrator can immediately
repair the objects, if the resources are not coming back in a reasonable
time.
To repair objects immediately, go to Cluster > Monitor > vSAN > Skyline
Health > vSAN object health and select REPAIR OBJECTS
IMMEDIATELY. Similar functionality is also under Cluster > Monitor tab >
Resyncing Objects > RESYNC NOW.
Resynchronizing Components
Absent and Degraded are the two failure states of vSAN components.
View the table below for more information.
If the vSAN cluster experiences more failures than the policy for a virtual
machine object tolerates, a virtual machine object may become
inaccessible.
Durability Components
The durability components contain the new writes. Once the original
component is back online, its data is synchronized with the durability
component.
Active Components
Consider a four-node VxRail cluster and all the nodes are in a healthy
state. Place one of the nodes in maintenance mode with ensure
accessibility option, so that there is no impact on the running VMs.
Durability Components
In the example, a vSAN cluster has eight hosts across four racks, four
fault domains have been configured for rack awareness. A storage policy
with RAID 1 protection has been applied to the VMs in the cluster. vSAN
ensures that the components of a VM object are placed in different fault
domains.
Example of a vSAN cluster with four fault domains showing the components of a RAID 1
object placed in different fault domains
For a balanced storage load and fault tolerance when using fault domains,
consider the following guidelines:
The tables show the minimum number of fault domains that are required
for various protection options.
1 2 3
2 3 5
3 4 7
1 RAID 5 4
2 RAID 6 6
After the initial VxRail deployment, each node acts as its own fault
domain. If necessary, fault domains can be configured. To manage vSAN
fault domains, select the VxRail cluster, and then go to Configure > vSAN
> Fault Domains.
In the example, the VxRail cluster has eight nodes. To create a fault
domain, click the plus (+) icon, the New Fault Domain window opens.
Name the domain, select nodes, and then click CREATE.
vSAN Fault Domains page showing the creation of a new fault domain.
In the example, four fault domains have been configured for the VxRail
cluster. Each fault domain has two nodes.
FD1 vcluster730-esx01.edu.local
vcluster730-esx02.edu.local
FD2 vcluster730-esx03.edu.local
vcluster730-esx04.edu.local
FD3 vcluster530-esx01.edu.local
vcluster530-esx02.edu.local
FD4 vcluster530-esx03.edu.local
vcluster530-esx04.edu.local
In the example, VM1 is hosted on a VxRail cluster with four fault domains
configured. A storage policy with RAID 5 protection has been assigned to
the VM. The components of each of the RAID 5 objects are placed in a
different fault domain.
VM Physical disk placement page showing the object component placement for VM1
The vSAN storage resources for a VM are distributed across the cluster.
Maintenance mode uses vMotion to move the active memory of VMs to
other cluster nodes. vSAN storage components on the node are moved as
directed by the selected vSAN data migration option. The impacted vSAN
components can be moved only if the remaining cluster nodes have
enough free space to meet the storage policy requirements.
To place a host into maintenance mode, in the left pane, select the node.
On the node page, select ACTIONS > Maintenance Mode > Enter
Maintenance Mode.
There are three vSAN data migration options when placing the node in
maintenance mode:
Ensure accessibility (default) - vSAN only migrates the components
that are essential for running VMs. This option is typically used while
performing upgrades.
Full data migration - vSAN transfers all the data from this host to the
other hosts in the cluster. This option is typically used when removing
a host permanently.
No data migration - vSAN does not evacuate any data from the host.
This option is typically used when shutting down an entire cluster.
Start PRE-CHECK
On the node page, select ACTIONS > Maintenance Mode > Enter
Maintenance Mode > GO TO PRE-CHECK. Select the "vSAN data
migration" type from the drop-down menu. Success for one type of
migration might be a failure for another since they have different
requirements. Click PRE-CHECK.
Data Migration PRE-CHECK page for selecting the node and vSAN data migration option
PRE-CHECK
Results
The Ensure accessibility PRE-CHECK passed and the host can enter maintenance mode
Object state
The Objects state tab lists the vSAN objects that are affected. In this
example, 16 vSAN objects become non-compliant when the node moves
to maintenance mode. vSAN waits for the delay timer to expire before
Cluster Capacity
The Cluster Capacity tab shows the impact to the vSAN cluster capacity.
In this example, the cluster has adequate capacity.
Predicted Health
The Predicted Health tab shows the impact of the vSAN health checks. In
this example, 16 vSAN objects would have Reduced availability with no
rebuild - delay timer alarm. vSAN waits for the delay timer to expire
before rebuilding the objects.
The Ensure accessibility option ensures that all running VMs, including
VMs with no data redundancy remain accessible. Components of
unprotected vSAN objects are migrated to another host.
Before placing the node in maintenance mode, the component residing on vcluster730-
esx01
After placing the node in maintenance mode with Ensure accessibility, the component
migrated to vcluster730-esx03
Hard disk 1 is RAID 1 protected and has two components on two different
nodes. Placing the vcluster730-esx01 node into maintenance mode with
Ensure accessibility does not trigger a migration of the component.
Before placing the node in maintenance mode, the components residing on vcluster730-
esx01 are in Active state
The components on the node vCluster730-esx01 are marked Absent, when the node
vCluster730-esx01 is placed in maintenance mode with Ensure accessibility
9
Dell Employees can impersonate all access levels.
VxRail Firmware Upgrade Procedures - Select VxRail Appliance Model and Component
to Upgrade Firmware
VxRail Software Upgrade Procedures - Select the Installed VxRail Software Version
VxRail Software Upgrade Procedures - Select the Target VxRail Software Version
2: The available 'How To' Procedures are different if the VxRail system is
part of a VMware Cloud Foundation (VCF) on VxRail environment.
VxRail Power Control Procedures - Select the activity and VxRail Software Version
The VxRail version and model of the node are required for generating the
SolVe procedure. In the vSphere Client, select the cluster in question and
then select Configure tab, and under the VxRail heading select System.
Collect the VxRail software version and model details from the System
tab.
VxRail Model
Select the VxRail model. Once the model is selected, scroll down to select
the component to be replaced.
Replacement Component
Answer the questions about the installed VxRail version. The capacity
SSD replacement procedure is different for older VxRail versions.
VxRail Version
Usage Information
Generate Procedure
While a VxRail cluster can be extended beyond a single physical rack, the
typical use case is for cluster expansion within the same rack. When
adding nodes within the same Layer 2 network segment, the IP addresses
must be on the same subnets as the existing cluster.
The add node process checks for compatibility after the new nodes are
discovered. Hardware compatibility checks ensure that the VxRail models
are compatible. Software compatibility checks ensure that the ESXi,
drivers, firmware, and the vSphere Installation Bundle (VIB) versions on
the new node are at the supported version levels. Compatible nodes at an
older version are automatically upgraded to the cluster version during the
node add process.
Excerpt of a customer facing SolVe procedure showing the selections for adding a node
to the same Layer 2 segment.
Follow the procedure to complete the node add process. High-level steps:
Mount the node in the rack.
Connect the node to AC power.
Connect the node to the ToR switches.
Power on the node.
Configure iDRAC for the node.
Verify date and time settings on the node.
Adjust the VLAN IDs of the port groups on the node if necessary.
Add the node to the VxRail cluster.
For VxRail clusters configured with automatic node discovery, the VxRail
Dashboard displays a notification when new nodes are discovered.
Selecting ADD on the VxRail Hosts page launches the Add VxRail Hosts
wizard. To learn how to use the Add VxRail Hosts wizard, select each
tab.
Discovered Hosts
Select the nodes that should be added to the cluster. In the example, one
compatible node has been discovered.
The vCenter user could be the vCenter administrator or a user with the
VMware HCI Management role. Provide the vCenter user credentials.
NIC Configuration
Map the physical NICs to the uplinks on the VDS. The NIC to uplink
mapping defaults to that of the first node in the VxRail cluster.
Host Settings
Host Location
Network Settings
Validate
Validate the input settings before adding the node to the cluster. The
validation process is similar to the validation performed by the VxRail
Deployment Wizard during the initial configuration. The validation process
could be used to determine the readiness of the environment for node
addition without adding the nodes.
Finish
The host can be left in Maintenance Mode after the node add process
completes. The default setting is No. In the example, the Yes option is
selected.
The node resources are available to the cluster when the node is taken
out of maintenance mode. Taking the node out of maintenance mode
could lead to DRS or vSAN disk rebalance activities. Take the node out of
maintenance mode when the cluster activity is low.
A VxRail cluster can extend beyond a single physical rack. All the network
addresses applied to the VxRail nodes within a single rack must be within
the same subnet. In a multirack VxRail system, the VxRail System VMs
run on the nodes in the first rack.
To learn about the available options for extending VxRail beyond a single
rack, select each tab.
The same subnet, gateway, and VLANs are assigned to all the VxRail
networks on the ToR switches in both racks. Physical connectivity is
required between the ToR switches in both racks. In the example, a new
All VxRail networks in the expansion rack are assigned a different subnet,
gateway, and VLAN ID. VLANs for first rack are configured on its ToR
switches. VLANs for the expansion rack are configured on its ToR
switches. Routing services for the different VxRail networks is enabled
using the upstream network. In the example, a new expansion rack (Rack
2) is being added to the existing VxRail cluster in Rack 1 over a Layer 3
network.
10
The internal management network is also known as the Discovery
network or Private Management network. The default VLAN ID assigned
to the internal management network is 3939 and can be changed.
Example presenting a Layer 3 Multirack VxRail cluster expansion setup with a Proxy
node.
The disk group configuration and disk slot placement rules are VxRail
model dependent. See appendixes A through D of the VxRail 7.x Support
Matrix for details. SolVe procedures are available for Capacity Drive
(HDD/SDD) Expansion and Disk Group Add. The procedures are VxRail
model and software version specific.
The Add Disk wizard handles the addition of capacity disks and disk
groups. To launch the wizard, click Add Disk under the ACTIONS menu
in the VxRail node Physical View.
To learn about adding a disk group using the Add Disk wizard, select
each tab.
To add a disk group, select No, I want suggestions about disk slots for
the new disks.
Set the Disk Type to Cache Disk, and then select NEXT.
Insert New Disk step - Request disk slot suggestion for Cache Disk
Disk Suggestions
The wizard scans the system and recommends disk slots for the cache
and capacity disks. Review the recommended disk slot locations. Insert
the cache disk, and at least one capacity disk into the recommended slots.
In the example, the recommended cache disk slot is 9, capacity disks
should start at slot 4.
Cache and capacity disks are not interchangeable and have different
model numbers. Ensure that the new disks are supplied by Dell and have
not been used in any system previously. If a disk has any signature or a
previously configured file system, the add disk procedure fails.
After the disks have been inserted, select I have inserted new disks, and
then select NEXT.
Disk Suggestions step - Recommended disk slot locations for the new disks
Verify Disk
The wizard discovers the inserted disks and presents a table with the disk
details. In this example, the cache disk is in slot 9 and one capacity disk is
in slot 4.
Confirm that the disks are in the expected slots, and then select NEXT.
If space efficiency features are enabled on the cluster, the Add Disk
wizard displays the Select vSAN Services step. In this example,
compression is enabled on the cluster. Select Yes, and then select NEXT.
The compression service is applied to the newly added disk group.
Validate
The wizard performs a validation. If all the validation tasks pass, select
NEXT.
Complete Addition
Follow the progress of the disk group addition. Select FINISH after the
process is complete. The new vSAN disk group becomes part of the fault
domain that the VxRail node belongs to.
For a VxRail cluster where vSAN is the primary storage, the backup
archives are stored in a folder on the VxRail vSAN datastore. The backup
can be run manually or scheduled.
Customers must engage Dell Support for assistance with the restore
process.
The VMware vCenter Server and the vSAN cluster must be healthy for a
successful restore.
Manual Backup
To learn about the arguments used, and the steps the script performs,
select each hotspot.
3: The script creates the backup archive locally on the VxRail Manager
VM and then uploads the archive to the VxRail vSAN datastore.
4: Once the backup archive has been successfully uploaded to the VxRail
vSAN datastore, the local archive on the VxRail Manager VM is deleted.
Scheduled Backup
3: The script updates the Cron job on the VxRail Manager VM with the
requested schedule. The schedule information is updated in the
vxmbackup.json file. The updated vxmbackup.json file is uploaded to
the VxRail vSAN datastore.
Use the scp command to copy the recovery bundle from the VxRail
Manager VM to the vSAN datastore. The bundle is in the
/data/store2/recovery directory.
Before running the scp command, enable SSH on the VxRail node that is
used for the copy process. In the example shown, VxRail node
vcluster730-esx01.edu.local is used for the copy process. The scp
command challenges for the root password.
The VxRail Manager restore process starts with the deployment of a new
VxRail Manager VM, a recovery bundle is not included. The VxRail
Manager file-based restore script uses a VxRail backup archive and the
recovery bundle on the vSAN datastore to complete the restoration.
11
The recovery bundle contains code that is used during the VxRail node
add process. If the node is at a lower version than the VxRail cluster, the
bundle is used to upgrade the node to the version of the cluster.
The backup archives, the backup catalog, and the recovery bundle are all
stored in the VxRail vSAN datastore in a folder named
VxRail_backup_folder. The example shows the recovery bundle, the
vxmbackup.json file, and two backup archives.
Use the CLI Simulator to log in to the VxRail Manager VM and perform
VxRail Manager file-based backup.
The CLI Simulator has limited capabilities. Enter the commands in the
order described in the instruction panel.
The simulator does not hide passwords, nor does it support tab
completion of commands.
The simulator supports copying and pasting, and using the up arrow to
retrieve previous commands.
Use the CLI simulator to view the help for the vxm_backup_restore.py
script. After reviewing the help, answer the questions.
The VxRail software upgrade can be initiated from the VxRail Plugin. To
initiate a software upgrade, select the Configure tab for a cluster. In the
middle pane, select VxRail > Updates.
COMPLIANCE
This tab also displays the components that are installed on the cluster and
their versions.
INTERNET UPDATES
To use the Internet updates option, VxRail Manager must have Internet
connectivity and be configured with a Dell support account.
LOCAL UPDATES
VxRail Plugin LOCAL UPDATES tab showing multiple files that are uploaded to VxRail
SETTINGS
Enabling vLCM
vSphere Client Lifecycle Manager Patch Setup page showing VxRail Manager as the
source depot
Review the How does vSphere LCM compare with VxRail LCM?
Blog for more information about vLCM.
The graphic shows the VxRail Updates page for a system at version
7.0.372.
The release notes also provide details of the fixes and features in each
release compared to previous releases, these details help drive the
decision to upgrade.
1: To create a drift report, select Create New Report, and select Create.
Access the new report from the View the report link in the Compliance
tab.
2: The View the report link displays the Compliance Drift Report. The
report shows which components are compliant, and which components
are not compliant with the expected version.
The Installer Metadata file is one of the files in the multipart upgrade
bundle. It contains a baseline file that describes each supported VxRail
component and the supported version for that specific release.
The installer metadata file is downloaded from the Dell Support site >
VxRail Drivers and Downloads tab.
VxRail 7.0.410 upgrade packages on the Dell Support site with the installer metadata
package identified
The advisory report only requires the metadata input file – not the entire
upgrade bundle. This small file allows administrators to compare their
current version of VxRail to various upgrade candidates to see the
differences that each potential upgrade would entail.
VxRail > Updates > LOCAL UPDATES page with the SELECT REPORT FILE and
UPLOAD options identified
VxRail > Updates > LOCAL UPDATES page with the VIEW REPORT option identified
View Report
The advisory report identifies the components that must be upgraded and
the version they will be upgraded to.
Sample advisory report showing the components that must be upgraded and their target
version.
1: Change VLAN ID
Dell Customers must contact support for assistance with some IP address
change procedures as shown in the graphic. The rest of the procedures
can be run without the assistance of Dell support.
The procedures to repoint the DNS and NTP servers are also located
here.
Change other VxRail Cluster settings - 'How To' Procedures, with procedures to modify
switch settings identified
High-level steps:
1. Cable VxRail PCIe ports to ToR switches.
A VxRail cluster with a VDS that was initially configured with two
Integrated NIC ports for all VxRail network traffic has been reconfigured.
The final configuration uses two Integrated NIC ports and two PCIe ports.
To view the changes that were made to the configuration, select each tab.
The initial configuration had two uplinks. Integrated NIC ports (vmnic0 and
vmnic1) were mapped to the uplinks. The final configuration has four
uplinks. Two uplinks were added and the PCIe ports (vmnic4 and vmnic5)
were mapped to these uplinks.
In the initial configuration, the management port group used uplink1 and
uplink2. In the final configuration, the management port group has been
reconfigured to use uplink1 and uplink3. This port group now has NIC
redundancy across an integrated NIC port (vmnic0 mapped to uplink1)
and a PCIe port (vmnic4 mapped to uplink3).
Management port group Policies view - Management port group Policies view -
Initial configuration with uplinks identified Final configuration with uplinks identified
In the initial configuration, the vSAN port group used uplink2 and uplink1.
In the final configuration, the vSAN port group has been reconfigured to
use uplink4 and uplink2. This port group now has NIC redundancy across
an integrated NIC port (vmnic1 mapped to uplink2) and a PCIe port
(vmnic5 mapped to uplink4).
vSAN port group Policies view - Initial vSAN port group Policies view - Final
configuration with uplinks identified configuration with uplinks identified
NIC teaming with link aggregation can be configured for VxRail non-
management traffic - vSAN, vMotion, and production networks.
12
Route based on originating virtual port is the default setting for VxRail port groups that
are deployed with the predefined network profiles.
Initial state
The port group has been modified. Load balancing is set to Route based
on physical NIC load, both uplinks are active.
For VxRail systems deployed with predefined network profiles, the default
configuration is a single VDS for all VxRail networks. If the network
requirements change after deployment, create another VDS and migrate
the selected VxRail networks to the new VDS. An additional VDS can also
be deployed for production VM traffic.
High-level steps:
1. Cable VxRail NIC ports to ToR switches.
2. Configure ToR switches to support VxRail network traffic.
3. Create a VDS with the required number of uplinks.
4. Create port groups for the VxRail networks that must be migrated to
the new VDS.
5. Configure new VDS:
Initial configuration
The VxRail system has been configured with one VDS and two uplinks for
all VxRail and production traffic. Each node in the cluster has an
integrated NIC with four ports. The first two ports of the NIC are mapped
to the uplinks on the VDS, two ports are unused. VMkernel adapter vmk3
on each node carries the vSAN traffic.
The VxRail administrator wants to move the vSAN traffic to the unused
NIC ports on a separate VDS.
VxRail system with one VDS - Topology view of VDS with the vSAN port group and
uplinks identified
Final configuration
A second VDS with two uplinks and a port group for the vSAN traffic has
been created. Physical adapters vmnic2 and vmnic3 of each node have
been assigned to the uplinks. The VMkernel adapters that carry the vSAN
traffic have been moved to the port group in the new VDS.
VxRail system with two VDS - Topology view of new VDS with the vSAN port group and
uplinks identified
vSAN stretched cluster configuration showing RAID 1 protection across sites and RAID 6
protection at each site
vCenter Server
No permissions assigned
No roles assigned
Witness Host
In the example, the VxRail vSAN stretched cluster has four nodes in each
data site (fault domain). The node names starting with vcluster730 are in
the Preferred fault domain, the node names starting with vcluster530 are
in the Secondary fault domain. The witness host is witness1.edu.local.
Create VM Storage Policy wizard - vSAN Availability page showing the Site disaster
tolerance options for a stretched cluster
In this example, a storage policy with site mirroring and RAID 5 protection
at each site is applied to a VM. To learn about the policy rules, VM
compliance status, and VM object placement, select each tab.
Policy Rules
Storage policy rules page showing the Site disaster tolerance and Failures to tolerate
settings
VM Compliance Status
Storage policy RAID5 Local + Site Mirroring has been applied to VM1,
all the VM objects are compliant.
VM Object Placement
Hard disk 1 has RAID 1 protection across the data sites, and RAID 5
protection in each of the data sites (Preferred and Secondary). The
witness components for the RAID 1 protection across the data sites are
stored on the witness host witness1.edu.local.
VM Physical disk placement page showing the object component placement for VM1
Hard disk 1
Policy Rules
Storage policy rules page showing the Site disaster tolerance and Failures to tolerate
settings
VM Compliance Status
VM Object Placement
Hard disk 1 has RAID 5 protection only at the Preferred site. There is no
site mirroring.
VM Physical disk placement page showing the object component placement for VM2
Hard disk 1
vSphere DRS manages the initial placement and on-going load balancing
of VMs. DRS is unaware of vSAN stretched cluster sites and without the
use of VM-Host affinity rules could vMotion a VM to any host in the cluster.
Consider a VM which only has local protection at the Preferred site. DRS
may vMotion the VM to the Secondary site. In this case, the VM uses the
CPU and memory resources at the Secondary site while its storage is still
at the Preferred site.
Using VM-Affinity rules along with the VM storage policy ensures that VMs
run on the appropriate site.
In the example, VMs in group Preferred Site VMs should run on hosts in
group Preferred Site Hosts.
The table shows the recommended VM-Host affinity rules for vSAN
stretched clusters.
To manage DRS groups in the vSphere Client, select the cluster, and then
go to Configure > VM/Host Groups. To create a VM or host group, click
ADD, the Create VM/Host Group dialog opens.
Enter a name and select the group type. For a new VM group, set the type
to VM Group. To add VMs, click ADD, the Add Group Member dialog
opens.
Group Members
Select the required VMs and click OK, the dialog closes. Click OK again to
close the VM/Host Group dialog.
Create VM/Host Group - Add Group Member dialog showing one selected VM
Group Created
The group has been created. To manage a group, select the group.
Members can be added or removed as necessary. The last member of a
group cannot be removed. Groups can be deleted. A warning is displayed
when trying to delete a group that is in use.
To manage VM-Host affinity rules in the vSphere Client, select the cluster,
and then go to Configure > VM/Host Rules. To create a VM-Host affinity
rule, click Add, the Create VM/Host Rule dialog opens.
Enter a name and select the rule type. For a VM-Host affinity rule, set the
type to Virtual Machines to Hosts.
Select the VM group, the appropriate rule, the host group, and then click
OK.
For VM groups with site mirroring, set the rule to Should run on hosts in
group. In this example, the VMs in the group Preferred Site VMs, should
run on the hosts in group Preferred Site Hosts. If there is a site failure at
the Preferred site, vSphere HA attempts to restart the VMs at the
Secondary site.
Rule Created
The affinity rule has been created. To manage a rule, select the rule.
Rules can be edited or deleted. Members can be added or removed from
either the VM or the host group. The last member of a group cannot be
removed.
SolVe procedures, VxRail Manager, and the Dell Support site contain links
to VxRail KB articles.
To learn how to access VxRail KB articles from the Dell Support site,
select each tab.
Log in to VxRail Support page and then select the Documentation tab.
Select VxRail Appliance Series and model if the issue is specific to the
model.
KB Articles List
To search for a KB article from the VxRail Support page, select the
Documentation tab to get a Search VxRail Appliance Series Support
Information box. In the Search VxRail Appliance Series Support
Information box, type the error message or the Article ID and click enter.
Select the related KB article.
VxRail Advisories
Dell issues advisories for Dell products in order to notify about critical
issues and potential impacts on the VxRail environment. Dell Security
Advisories (DSAs) and Dell Technical Advisories (DTAs) are two types of
advisories.
A DTA is a Dell proactive notification for different issues that may affect IT
operations. Issues such as data loss, data unavailability, loss of system
functionality, or anything that could result in a significant safety risk. A
DTA provides the steps to eliminate or minimize the impact of such issues.
Select the VxRail Appliance Series and then select the node model for
the selected VxRail Appliance Series.
VxRail Support home page - Select VxRail Appliance Series and Node Model
Advisories
Log in to the VxRail Support page. Hover over the account icon and click
Preferences.
Preferences
Notification Settings
Click the Notifications & Alerts tab and click the Product
Advisories tab. Update the Email and make the selection for types of
The VxRail SolVe Procedure - How to Collect Log Bundle, details the log
collection steps with the VxRail Plugin and with the VxRail Manager VM
CLI scripts.
Platform: -p --platform
Witness in a 2-Node VxRail cluster: -w, --witness
Short option names can be stacked, for example: "-vc" compared to long
names for example: "--vxm --vcenter". Both of these examples collect
VxRail Manager and vCenter logs. Both short and long options exist for
flexibility.
For downloading the full log Bundle, run the script without options. The
witness log bundle is not in the full log bundle collection option. The
witness log bundle collection must be performed separately.
Custom scripts and applications can be built with the VxRail API to
automatically generate and download the required logs. The VxRail API
base URL is: https://<VxRail Manager IP Address or
FQDN>/rest/vxm/.
The table lists the relevant VxRail APIs for gathering and downloading
logs.
Each cmdlet from the VxRail API PowerShell Modules maps to a VxRail
API which is used in the background by the cmdlets. The VxRail API
PowerShell Modules complement VMware PowerCLI.
Dell Support or VMware Support may request logs that are related to the
vCenter Server or the ESXi hosts to diagnose issues. The vSphere Client
can be used to export the logs for the ESXi hosts, vCenter Server, and
vSphere UI Client. Performance data from the ESXi hosts can be
optionally included.
To learn how to perform log collection using the vSphere Client, select
each tab.
Select vCenter
To export system logs from the vSphere Client, right-click the vCenter
server in the inventory list and select Export System Logs.
Select Components
In the Export System Logs dialog, select the relevant ESXi hosts.
Optionally, select the Include vCenter Server and vSphere UI Client
logs box. Click NEXT.
Select Logs
In the Select logs page, select the specific system logs. Typically, the
support representative specifies the logs that are required. To generate
and download the log bundle, click EXPORT LOGS.
For VxRail, operations are broken into categories. Day 1 operations are
cluster initialization tasks. Day 2 operations are tasks that are performed
after initialization, like expansion and upgrades.
The table lists VxRail Manager VM log files that are useful when
troubleshooting both categories. These files are in the log bundles
collected using the VxRail Plugin or the VxRail Manager VM log collection
script.
The All Issues page in the vSphere Client is a good place to start the
troubleshooting process for a VxRail vSAN Cluster. This page shows all
triggered alarms relevant to the cluster. vSAN alarm definitions contain the
term "vSAN'' in the alarm name and VxRail alarm definitions contain the
term "VXR" in the alarm name.
Skyline Health page can be used to monitor and remediate some of the
vSAN health issues. Each Skyline Health test has an Info tab with details
of the test and the link to the relevant VMware knowledge base article.
In the Partition list below, observe that the last host is in partition 2, while
the other hosts are in partition 1. This cluster is partitioned and must be
investigated further.
In the failed ping tests below, observe that the esx04 host appears in each
of the rows. In all probability, there are network issues specific to this host.
Observe that the network connectivity issues in this vSAN Cluster have
resulted in vSAN objects with reduced availability.
It is the VxRail alert code that begins with "VXR" which is in the Issue. See
the Dell VxRail Event Code Reference document for more details.
VXR event codes are documented in the VxRail Event Code Reference
and in the Alarm Definitions within vCenter Server. To learn more about
VXR alarms, select each tab.
The VxRail Event Code Reference lists all the VXR alarms, severity level,
message, and a link to the related Dell KB article.
This example from the VxRail Event Code Reference shows the
information for the VXR014020 event:
VXR Alarm Definitions are shown on the Alarm Definitions page at the
VxRail cluster level in the vSphere Client. To view VXR alarm definitions,
filter the Alarm Name column by VXR or vxr as the filter is case
insensitive.
vSAN includes two proactive tests for troubleshooting the cluster: the VM
Creation Test, and the Network Performance Test. The vSAN proactive
tests are available under the Monitor tab of the VxRail cluster. To learn
more about vSAN Proactive Tests, select each tab.
VM Creation Test
The VM Creation Test creates a VM on every host and then deletes it.
This active test only takes a few seconds to run and can find issues that
cannot be found with passive tests. The test can be used to diagnose
node isolation, cluster segmentation, and other configuration issues.
Go to the Monitor tab of the VxRail cluster. Select Proactive Tests > VM
Creation Test > RUN TEST > RUN.
If the creation and deletion tasks succeed, it can be concluded that many
aspects of vSAN are operational.
structure; every host sends packets to its next host while receiving from
the previous host. The test examines the network bandwidth that can be
achieved between hosts and reports warnings when the bandwidth is less
than 850 Mbps. Low network bandwidth can negatively impact vSAN
performance.
Go to the Monitor tab of the VxRail cluster. Select Proactive Tests >
Network Performance Test > RUN TEST > RUN. After the network
performance test is complete, be sure to disable network diagnostic mode.
This command lists out the status of each of the health test names for the
vSAN cluster. The below output shows the vSAN object health and Stats
DB object status as red means that the object does not exist.
This command checks vSAN hosts communication. The below output with
Partition is unknown means that the host in the VSAN cluster cannot
communicate with each other.
Consider the below prerequisites for removing a node from the VxRail
Cluster:
VxRail Cluster must be in a healthy state.
VxRail Cluster must have enough nodes remaining after the node
removal to support the current Failures to Tolerate (FTT).
For detailed information about the node removal procedure, generate and
download the SolVe Procedure.
Maintenance
Planned power outage
Excerpt of a customer facing SolVe procedure showing the selections to Power Down a
Running VxRail Cluster
To shut down a VxRail cluster, go to Hosts and Clusters view and select
the VxRail cluster. Click Actions in the right pane and select VxRail >
Shut Down.
To learn more about shutting down the VxRail cluster, select each tab.
Welcome
Validate
On the Validate page, ensure that the validation is successful and click
NEXT. The validation process shows an error if the client VMs are
powered on.
Confirm Operation
On the Shut Down Cluster page, ensure that the shutdown operation is
initiated successfully and click FINISH.
After the successful cluster shutdown, all the nodes go into a powered off
state.
Excerpt of a customer facing SolVe procedure showing the selections for Power-up a
VxRail cluster after a scheduled shutdown.
If either vCenter or DNS is external to the cluster, ensure that they are
running before powering on the cluster.
If the VxRail cluster is using internal DNS, perform the following steps:
The Create the VxRail Manager Key and Certificate Files video shows an
example of the above steps.
Open the vSphere Client, select the Configure tab for the cluster, and
select VxRail > Certificate.
Enter Content
Paste the contents of the certificate file, private key file, and certificate
chain into the wizard.
Certificate Update wizard with certificate, key, and chain fields entered
Enter Password
The VxRail Manger private key and certificates are updated, and trusts are
configured with other VMware components. The password is used to
secure the VxRail Manager key file.
After replacing the vCenter certificate, use the VxRail: How to manually
import vCenter SSL certificate on VxRail Manager KB to address
certificate issues.
After replacing the ESXi host certificates, use the Solve procedure to
Import ESXi Host Certificates to VxRail Manager.
Manage Passwords
The VxRail mystic and root accounts should have different passwords as
described when using the first run wizard to deploy a VxRail. Use the
following steps to change passwords using the CLI:
1. SSH to the VxRail Manager system or launch the Web Console from
the vSphere Client for the VxRail Manager VM. If an SSH session is
opened, root login is disabled as default. Log in with the mystic
account and then SU to root.
2. Log in with the user account requiring the password change.
3. Type passwd. This is the standard command that is found in UNIX
and UNIX like operating systems such as Linux.
4. Enter and verify the new password.
Use the following steps to change the mystic password using the CLI:
1. Launch the web console of the VxRail Manager through the vSphere
Client.
2. Log in to the interface with the root user credentials.
3. Type pam_tally2 --user=mystic --reset
Once the account is unlocked, change the password by using the passwd
command.
Note: If the root account is locked, use Dell VxRail: VxRail Manager root
password is lost knowledge base article to reset the password.
Click the Save Progress and Exit button in the course menu or
below to record this content as complete.
Go to the next learning or assessment, if applicable.
For more information about VMware vSphere licensing, see the VMware
vSphere Compute Virtualization - Licensing, pricing and packaging guide
Deduplication
and
Compression
RAID 5/6
Erasure
Coding
vRealize
Operations
within
vCenter
Data-at-Rest
Encryption
Stretched
Cluster with
Local Failure
File Services
VMware HCI
Mesh
vROPS 8
Advanced
Services
For more information about VMware vSAN licensing, see the VMware
vSAN - licensing, pricing, and packaging guide.
vSphere Encryption
vSphere Encryption
Applied on a VM basis
Supported with vSAN deduplication and compression (minimal space
savings for encrypted VMs)
1 2 3
2 3 5
3 4 7
1 RAID 5 4
2 RAID 6 6
13
Clusters running at 1 GbE base network speed are limited to a maximum of eight
nodes. 1 GbE base network speed is only supported with hybrid single processor nodes.
VMware Loudmouth is a service that runs on VxRail nodes and the VxRail
Manager VM. The service enables automatic discovery of nodes. The
service is used during Initial cluster configuration, cluster expansion, and
failed node replacement. Loudmouth is an implementation of Zero
Configuration Networking which enables networking without having to
configure a DHCP, DNS, and static network settings. Loudmouth uses
IPv6 multicast network protocol
When powered on, VxRail Manager uses IPv6 multicast network protocols
to discover VxRail nodes. The IPv6 multicast traffic is sent over an internal
management network that is used for discovery only. The default VLAN ID
assigned to the internal management network is 3939 and can be
changed.
The IP address discovery option is used when the customer has concerns
about using either IPv6 or multicast network protocols. Since nodes are
available over the external management network, the default ESXi root
password can be changed to address security concerns.
The disk slots are numbered 0 through 23 starting from left to right. Slots 0
through 19 are capacity disk slots. Slots 20 through 23 are cache disk
slots. In the example slots 4, 9, 13 through 19, and 23 are available for
expansion.
The disk slots are numbered 0 through 9 starting from upper left to bottom
right. Slots 0 through 7 are capacity disk slots. Slots 8 and 9 are cache
disk slots. In the example, slots 3 and 7 are available for expansion.
VxVerify
VxVerify is a tool for checking the health of the nodes and the service
VMs. It is designed to detect issues that could cause complications or
failures during VxRail LCM upgrades. VxVerify is run from the VxRail
Manager. It automatically uploads and runs scripts against each ESXi
node and system VM and then analyzes the collected data. Since the tool
is updated regularly, ensure you have the latest version before using it.
Acquire the latest version of the tool before using it. Follow KB 21527 for
instructions on how to run VxVerify.
Proxy Node
A Proxy node is a VxRail node with a manually assigned static ESXi
management IP address on the new Layer 3 network segment. It is used
to discover the other VxRail nodes in that network segment.
RAID 1 Mirroring
Mirroring can accommodate an FTT setting of 1, 2, or 3. If FTM is set to
mirroring, for the N number of failures that are tolerated, N+1 copies of the
object are created. Mirroring requires witness components. The number of
witnesses is equal to the FTT setting. RAID 1 with FTT setting of 1
requires a minimum of 3 hosts. RAID 1 with FTT setting of 2 requires a
minimum of 5 hosts and, RAID 1 with FTT setting of 3 requires a minimum
of 7 hosts.
VxRail-managed vCenter
A vCenter that is deployed during an initial VxRail cluster build and is
upgraded using VxRail Manager.
Witness
The witness host is an appliance or a physical ESXi host which does not
store VM data. However, it does store the witness components for each
VM object.