Field Installation Guide v4 - 5
Field Installation Guide v4 - 5
2. Foundation Considerations..................................................................................5
Foundation Use Case Matrix.................................................................................................................................5
CVM vCPU and vRAM Allocation....................................................................................................................... 6
7. Post-Installation Steps.........................................................................................36
Configuring a New Cluster in Prism................................................................................................................36
ii
Verify Hypervisor Support...................................................................................................................................39
Updating an iso_whitelist.json File on Foundation VM...........................................................................40
9. Network Requirements........................................................................................41
12. Troubleshooting................................................................................................... 49
Fixing IPMI Configuration Issues...................................................................................................................... 49
Fixing Imaging Issues............................................................................................................................................50
Frequently Asked Questions (FAQ)..................................................................................................................51
Copyright...................................................................................................................59
License......................................................................................................................................................................... 59
Conventions............................................................................................................................................................... 59
Default Cluster Credentials................................................................................................................................. 59
Version.........................................................................................................................................................................60
iii
1
FIELD INSTALLATION OVERVIEW
For a node to join a Nutanix cluster, it must have a hypervisor and AOS combination that Nutanix supports. AOS
is the operating system of the Nutanix Controller VM, which is a VM that must be running in the hypervisor to
provide Nutanix-specific functionality. Find the complete list of supported hypervisor/AOS combinations at https://
portal.nutanix.com/page/documents/compatibility-matrix.
Foundation is the official deployment software of Nutanix. Foundation allows you to configure a pre-imaged node, or
image a node with a hypervisor and an AOS of your choice. Foundation also allows you to form a cluster out of nodes
whose hypervisor and AOS versions are the same, with or without re-imaging. Foundation is available for download
at https://portal.nutanix.com/#/page/Foundation.
If you already have a running cluster and want to add nodes to it, you must use the Expand Cluster option in Prism,
instead of using Foundation. Expand Cluster allows you to directly re-image a node whose hypervisor/AOS version
does not match the cluster's version, or a node that is only running DiscoveryOS. More details on DiscoveryOS are
provided below.
Nutanix and its OEM partners install some software on a node at the factory, before shipping it to the customer. For
shipments inside the USA, this software is a hypervisor and an AOS. For Nutanix factory nodes, the hypervisor is
AHV. In case of the OEM factories, it is up to the vendor to decide what hypervisor to ship to the customer. However,
they always install AOS, regardless of the hypervisor.
For shipments outside the USA, Nutanix installs a light-weight software called DiscoveryOS, which allows the node
to be discovered in Foundation or in the Expand Cluster option of Prism.
Since a node with DiscoveryOS is not pre-imaged with a hypervisor and an AOS, it must go through imaging
first before joining a cluster. Both Foundation and Expand Cluster allow you to directly image it with the correct
hypervisor and AOS.
Vendors who do not have an OEM agreement with Nutanix ship a node without any software (not even DiscoveryOS)
installed on it. Foundation supports bare-metal imaging of such nodes. In contrast, Expand Cluster does not support
direct bare-metal imaging. Therefore, if you want to add a software-less node to an existing cluster, you must first
image it using Foundation, then use Expand Cluster.
This document only explains procedures that apply to NX and OEM nodes. For non-OEM nodes, you must perform
the bare-metal imaging procedures specifically adapted for those nodes. For those procedures, see the vendor-specific
field installation guides, available on the Nutanix Support portal at https://portal.nutanix.com/page/documents/list?
type=compatibilityList.
Note: Mixed-vendor cluster is not supported. For more information, see the Product Mixing Restrictions in the NX and
SX Series Hardware Administration Guide.
• See Prepare Factory-Imaged Nodes for Imaging on page 9 to re-image factory-prepared nodes, or create a
cluster from these nodes, or both.
• See Prepare Bare-Metal Nodes for Imaging on page 15 to image bare-metal nodes and optionally configure
them into a cluster.
2
FOUNDATION CONSIDERATIONS
This section provides the various guidelines, compatibility list, limitations, and capabilities of Foundation.
Function
• Factory-imaged nodes • Factory-imaged nodes • Factory-imaged nodes
• Bare-metal nodes • Bare-metal nodes
If IPv6 is disabled Cannot image nodes. IPMI IPv4 required on IPMI IPv4 required on
the nodes the nodes
Can configure the VLAN No. Manually configure No. Manually configure Yes
of Foundation in the vSwitch of the in Windows or macOS.
host.
Platform vCPU
AMD Naples, AMD Rome, or PowerPC
• Fixed to 8 for AMD Naples
• Fixed to 12 for AMD Rome
• Fixed to 6 for PowerPC
• 32+ TB of HDD
• OR 48+ TB of SSD (excluding NVMe)
If the platform is generic, that is, 8+ physical cores Between 8 and 12 (inclusive)
in the CVM NUMA node OR both of the following:
For low end platforms Equal to the number of physical cores in the CVM
NUMA node
Note:
• Foundation GUI and Prism do not provide an option to override vCPU defaults.
Platform vRAM
“storage_heavy” or “minimal_compute_node” 28 GB
“high_perf” 32 GB
• 32+ TB of HDD
• OR 48+ TB of SSD and NVMe combined
If AOS is older than 5.1, or total memory is <62GB Subtract 4 GB from the previous value, only if the
value is < 32 GB
If the determined value leaves < 6 GB of remaining As much memory that leaves 6 GB of remaining
memory, all of the above are ignored. memory
Features like dedupe and compression require vRAM beyond the default values. For such cases, you can override the
default values by performing one of the following tasks:
• Manually changing the value in the Foundation GUI. If this value leaves <6 GB of remaining memory, then it is
ignored and instead the default values mentioned earlier is used.
• The allocation can also be changed from Prism after installation, by going to Configure CVM from the gear
icon in the web console.
Platform vRAM
If AOS is older than 5.1, or total memory is <62 GB Subtract 4 GB from the previous value, only if the
value is <32GB.
If the determined value leaves < 6 GB of remaining As much memory that leaves 6 GB of remaining
memory, all of the above are ignored. memory.
NUMA Pinning
Pinning is enabled only when the following conditions are met:
Note:
• “CVM NUMA node” means the NUMA node that connects to the SSD storage controller.
• When pinning is enabled, all vCPUs is pinned to this single NUMA node. Pinning vCPUs to a NUMA
node enables the CVM to maximize its I/O performance under heavy load.
• vRAM allocation is not pinned to any NUMA node to ensure that enough memory is available for the
CVM when it starts after shutdowns like maintenance mode.
3
PREPARE FACTORY-IMAGED NODES
FOR IMAGING
This procedure describes how to install a selected hypervisor and the Nutanix Controller VM on discovered
nodes and how to configure the nodes into a cluster. "Discovered nodes" are factory prepared nodes that
are not currently part of any cluster, and are reachable within the same subnet. This procedure runs the
Foundation tool through the Nutanix Controller VM (Controller VM–based Foundation).
• Make sure that the nodes that you want to image are factory-prepared nodes that have not been configured in any
way and are not part of a cluster.
• Physically install the Nutanix nodes at your site. For general installation instructions, see "Mounting the Block"
in the Getting Started Guide. For installation instructions specific to your model type, see "Rack Mounting" in the
NX Series Hardware Administration Guide.
Your workstation must be connected to the network on the same subnet as the nodes you want to image.
Foundation does not require an IPMI connection or any special network port configuration to image discovered
nodes. See Network Requirements for general information about the network topology and port access required
for a cluster.
• Determine the appropriate network (gateway and DNS server IP addresses), cluster (name, virtual IP address), and
node (Controller VM, hypervisor, and IPMI IP address ranges) parameter values needed for installation.
Note: The use of a DHCP server is not supported for Controller VMs, so make sure to assign static IP addresses to
Controller VMs.
Note: Nutanix uses an internal virtual switch to manage network communications between the Controller VM
and the hypervisor host. This switch is associated with a private network on the default VLAN and uses the
192.168.5.0/24 address space. For the hypervisor, IPMI interface, and other devices on the network (including
the guest VMs that you create on the cluster), do not use a subnet that overlaps with the 192.168.5.0/24 subnet on
the default VLAN. If you want to use an overlapping subnet for such devices, make sure that you use a different
VLAN.
• AOS installer named nutanix_installer_package-version#.tar.gz from the AOS (NOS) download page.
• Hypervisor ISO if you want to instal Hyper-V or ESXi. The user must provide the supported Hyper-V or ESXi
ISO (see Hypervisor ISO Images on page 38); Hyper-V and ESXi ISOs are not available on the support
portal.
It is not necessary to download AHV because the AOS bundle includes an AHV installation bundle. However,
you can download an AHV installation bundle if you want to install a non-default version.
• Make sure that IPv6 is enabled on the network to which the nodes are connected and IPv6 multicast is supported.
• If the nodes contain self-encrypting drives (SEDs), disable encryption on the SEDs before imaging the nodes.
If the nodes contain only SEDs, you can enable encryption after you image the nodes. If the nodes contain both
Note: This method can image discovered nodes or create a single cluster from discovered nodes or both. This method
is limited to factory prepared nodes running AOS 4.5 or later. If you want to image factory prepared nodes running an
earlier AOS (NOS) version or image bare-metal nodes, see Prepare Bare-Metal Nodes for Imaging on page 15.
Procedure
1. Run discovery and launch Foundation (see Node Discovery and Foundation Launch on page 10).
2. Update Foundation to the latest version (see Upgrade CVM Foundation by Using the Foundation Java Applet).
Note:
3. Run CVM Foundation (see Configuring Foundation VM by Using the Foundation GUI on page 31).
4. After the cluster is created successfully, begin configuring the cluster (see Configuring a New Cluster in Prism on
page 36).
Procedure
Note: A security warning message may appear to indicate that the list of nodes is from an unknown source. To run
the application, click Accept and Run.
Note: Use the network configuration tool only on factory-prepared nodes that are not part of a cluster. Using the tool
on a node that is part of a cluster makes the node inaccessible to the other nodes in the cluster. If there are issues, the
only way to resolve an issue is to reconfigure the node to the previous IP addresses by using the network configuration
tool again.
1. Connect a console to one of the nodes and log on to the Acropolis host by using the root credentials.
2. Change your working directory to /root/nutanix-network-crashcart/, and then start the network
configuration utility.
root@ahv# ./network_configuration
a. Review the network card details to ascertain interface properties and identify connected interfaces.
b. Use the arrow keys to go to the interface that you want to configure, and then use the Spacebar key to select
the interface.
Repeat this step for each interface that you want to configure.
c. Use the arrow keys to navigate through the user interface and specify values for the following parameters:
Launching Foundation
Launching Foundation depends on whether you used the Foundation Applet to discover nodes in the same
broadcast domain or the crash cart user interface to discover nodes in a VLAN-segmented network.
Procedure
• If you used the Foundation Applet to discover nodes in the same broadcast domain, do the following:
Note: A warning message stating that the value provided is not the highest available version of Foundation
found in the discovered nodes may appear. If you select a node using an earlier Foundation version (one that
does not recognize one or more of the node models), installation may fail when Foundation attempts to image
a node of an unknown model. Therefore, select the node with the highest Foundation version among the nodes
b. (Optional but recommended) Upgrade Foundation on the selected node to the latest version. See Upgrade
CVM Foundation by Using the Foundation Java Applet.
c. With the node having the latest Foundation version selected, click the Launch Foundation button.
Foundation searches the network subnet for unconfigured Nutanix nodes (factory prepared nodes that are
not part of a cluster) and then displays information about the discovered blocks and nodes in the Discovered
Nodes screen. (It does not display information about nodes that are powered off or in a different subnet.) The
discovery process normally takes just a few seconds.
Note: If you want Foundation to image nodes from an existing cluster, you must first either remove the target
nodes from the cluster or destroy the cluster.
• If you used the crash cart user interface to discover nodes in a VLAN-segmented network, in a browser on your
workstation, enter the following URL: http://CVM_IP_address:8000
Replace CVM_IP_address with the IP address that you assigned to the Controller VM when using the network
configuration tool.
Note: Ensure that you use the minimum version of Foundation required by your hardware platform. To determine
whether Foundation needs an upgrade for a hardware platform, see the respective System Specifications guide. If
the nodes you want to include in the cluster are of different models, determine which of their minimum Foundation
versions is the most recent version, and then upgrade Foundation on all the nodes to that version.
Procedure
Click the version link in the Foundation GUI.
Note: Update only the foundation-platforms submodule. This update enables Foundation to support the latest hardware
models or components qualified after the release of installed Foundation version.
Procedure
1. In the Foundation Java applet, select the node on which you want to upgrade the CVM Foundation.
3. Browse to the folder where you downloaded the Foundation .tar file and double-click the .tar file.
The upgrade process begins. After the upgrade completes, Genesis restarts on the node, and that in turn restarts the
Foundation service. After the Foundation service becomes available, the upgrade process reports the status of the
upgrade.
4
PREPARE BARE-METAL NODES FOR
IMAGING
Prepare Bare-Metal Nodes for Imaging
You can perform bare-metal (also referred to as standalone) imaging from a workstation with access to
the IPMI interfaces of the nodes. Imaging a cluster in the field requires installing tools (such as Oracle VM
VirtualBox, VMware Fusion ) on the workstation and setting up the environment to run these tools. This
chapter describes how to install a selected hypervisor and the Nutanix Controller VM on bare-metal nodes
and configure the nodes into one or more clusters. "Bare metal" nodes are not factory-prepared and cannot
be detected through discovery. However, you can also use this method to image factory-prepared nodes.
• Physically install the nodes at your site. For installing Nutanix hardware platforms, see the NX and SX
Series Hardware Administration and Reference for your model type. For installing hardware from any other
manufacturer, see that manufacturer's documentation.
• Set up the installation environment (see Preparing the Workstation on page 16).
• Ensure that you have the appropriate global, node, and cluster parameter values needed for installation. The use of
a DHCP server is not supported for Controller VMs, so make sure to assign static IP addresses to Controller VMs.
Note: If the Foundation VM is configured with an IP address that is different from other clusters that require
imaging in a network (for example Foundation VM is configured with a public IP address while the cluster resides
in a private network), repeat step 8 in Installing the Foundation VM on page 17 to configure a new static IP
address for the Foundation VM.
• If the nodes contain self-encrypting drives (SEDs), disable encryption on the SEDs before imaging the nodes. If
the nodes contain only SEDs, enable encryption after you image the nodes. If the nodes contain both regular hard
disk drives (HDDs) and SEDs, do not enable encryption on the SEDs at any time during the lifetime of the cluster.
For information about enabling and disabling encryption, see the "Data-at-Rest Encryption" chapter in the Prism
Web Console Guide.
Note: After you prepare the bare-metal nodes for Foundation, configure the Foundation VM by using the GUI. For
details, see Node Configuration and Foundation Launch on page 29.
• If you change the boot device order in the BIOS to boot from a USB flash drive, you will get a Foundation
timeout error unless the first boot device is set to CD-ROM in BIOS (under boot order menu).
• If Spanning Tree Protocol (STP) is enabled on the ports that are connected to the Nutanix host, Foundation might
time out during the imaging process. Therefore, ensure to disable STP by using PortFast or an equivalent feature
on the ports that are connected to the Nutanix host before starting Foundation.
Recommendations
• Nutanix recommends that imaging and configuration of bare-metal nodes is performed or supervised by Nutanix
sales engineers or partners. If a Nutanix sales engineer or a partner is unavailable, contact Nutanix Support for
assistance.
Procedure
1. Go to the Nutanix Support portal and download the following files to a temporary directory on the workstation:
• Foundation_VM-version#-
disk1.vmdk. This file is the
Foundation VM VMDK file
for the version# release. For
example, Foundation_VM-3.1-
disk1.vmdk.
Note:
• For post installation performance validation execute the Four Corners Microbenchmark
using X-ray. For more information, refer https://portal.nutanix.com/page/documents/list?
type=software&filterKey=software&filterVal=X-Ray.
• To install a hypervisor other than AHV, you must provide the ISO image of the hypervisor (see
Hypervisor ISO Images on page 38). Make sure that the hypervisor ISO image is available on
the workstation.
• Verify the support for the hypervisor and the corresponding version of the hypervisor. For details,
see Verify Hypervisor Support on page 39.
2. Download the installer for Oracle VM VirtualBox (a free open source tool used to create a virtualized
environment on the workstation) and install it with the default options. For installation and start-up instructions
(https://www.virtualbox.org/wiki/Documentation), see the Oracle VM VirtualBox User Manual.
Note: You can also use any other virtualization environment (VMware ESXi, AHV, and so on) instead of Oracle
VM VirtualBox.
3. Go to the location where you downloaded the Foundation .tar file and extract the contents.
$ tar -xf Foundation_VM_OVF-version#.tar
Note: If the tar utility is not available, use the appropriate utility for your environment.
4. Copy the extracted files to the VirtualBox VMs folder that you created.
Procedure
2. Click the File menu and select Import Appliance... from the pull-down list.
3. In the Import Virtual Appliance dialog box, browse to the location of the Foundation .ovf file, and select the
Foundation_VM-version#.ovf file.
4. Click Next.
5. Click Import.
8. (Optional) If you want to enable the file drag-and-drop functionality between your workstation and the Foundation
VM, install Oracle Additions as follows:
a. On the VirtualBox window for the Foundation VM, select Devices > Insert Guest Additions CD
Image... from the menu.
A VBOXADDITIONS CD entry appears on the Foundation VM desktop.
b. When the Open Autorun Prompt dialog box appears, click OK, and then click Run.
d. After the installation completes, press the return key to close the VirtualBox Guest Additions installation
window.
e. Right-click the VBOXADDITIONS CD entry on the desktop and select Eject.
f. Restart the Foundation VM by selecting System > Shutdown... > Restart from the Linux GUI.
g. After the Foundation VM restarts, select Devices > Drag 'n' Drop > Bidirectional from the menu.
Note: The Foundation VM must be on a public network to copy the selected ISO files to the Foundation VM. This
requirement requires setting a static IP address and setting it again when the workstation is on a different (typically
private) network for the installation.
Note: Use the indication keys only to perform selections in the terminal window. Mouse clicks do not work.
• 1. Remove the asterisk in the Use DHCP field (which is set by default).
2. Enter appropriate addresses in the Static IP, Netmask, and Default gateway IP fields.
3. Click the OK.
Procedure
2. If you downloaded the diagnostics files for one or more hypervisors, copy them to the appropriate directories on
the Foundation VM. The directories for the diagnostic files are as follows:
Note: You can connect to either a managed switch (routing tables) or a flat switch (no routing tables). Nutanix
recommends a flat switch to protect the production environment against configuration errors. Foundation includes a
multi-homing feature that allows you to image nodes by using the production IP addresses even when connected to a
flat switch. See Network Requirements on page 41 for information about the network topology and port access
required for a cluster.
Procedure
1. Enable IPv6 on the network to which the nodes are connected and ensure that IPv6 multicast is supported.
Note: If you use a shared IPMI port to reinstall the hypervisor, follow the instructions in KB 3834.
• Nutanix NX Series: Connect the dedicated IPMI port and any one of the data ports to the switch. Ensure that
you use the dedicated IPMI port instead of the shared IPMI port. In addition to the dedicated IPMI port, we
highly recommend that you use a 10G data port. You can use a 1G port instead of a 10G data port at the cost
of increased imaging time or imaging failure. If you use SFP+ 10G NICs and a 1G RJ45 switch for imaging,
connect the 10G data port to the switch by using one of our approved GBICs. If the BMC is configured to use
it, you can also use the shared IPMI/1G data port in place of the dedicated port. However, the shared IPMI/1G
data port is less reliable than the dedicated port. The IPMI LAN interfaces of the nodes must be in failover
mode (factory default setting).
If you choose to use the shared IPMI port on G4 and later platforms, make sure that the connected switch can
auto-negotiate to 100 Mbps. This auto-negotiation capability is required because the shared IPMI port can
support 1 Gbps throughput only when the host is online. If the switch cannot auto-negotiate to 100 Mbps when
What to do next
Update the Foundation VM service, if necessary. After you prepare the bare-metal nodes for Foundation,
configure the Foundation VM by using the GUI. For details, see Node Configuration and Foundation Launch
on page 29.
Foundation VM Upgrade
You can upgrade the Foundation VM either by using the GUI or CLI.
Note: Ensure that you use the minimum version of Foundation required by your hardware platform. To determine
whether Foundation needs an upgrade for a hardware platform, see the respective System Specifications guide. If
the nodes you want to include in the cluster are of different models, determine which of their minimum Foundation
versions is the most recent version, and then upgrade Foundation on all the nodes to that version.
Procedure
» (Do not use with Lenovo platforms) To perform a one-click over-the-air update, click Update.
» (For Lenovo platforms; optional for other platforms) To update Foundation by using an installer that you
downloaded to the workstation, click Browse, browse and select the .tar file, and then click Install.
Procedure
1. Download the Foundation upgrade bundle (foundation-version#.tar.gz) from the Nutanix Support portal to
the /home/nutanix/ directory.
3. Upgrade Foundation.
$ ./foundation/bin/foundation_upgrade -t foundation-version#.tar.gz
Procedure
4. To install the Foundation app, drag the Foundation app to the Application folder.
The Foundation app is installed.
Note: To upgrade the app, download and install a higher version of the app from the Nutanix Support portal.
The Foundation GUI is launched in the default browser or you can browse to http://localhost:8000.
6. Allow the app to accept incoming connections when prompted by your Mac computer.
7. To close the app, right-click on the Foundation icon in the launcher and click Force Quit.
Note: The installation stops any running Foundation process. If you have initiated Foundation with a previously
installed app, ensure that it is complete before launching the installation again.
Procedure
1. Ensure that IPv6 is enabled on the network interface that connects the Windows PC to the switch.
2. Download the Foundation .msi installer file from the Nutanix Support portal.
Note: To upgrade the app, download a higher version of the app from the Nutanix portal and perform the
installation again. The new installation stops any running Foundation operation and updates the older version to the
higher version. If initiated Foundation with the older app, ensure that it is complete before doing a new installation
of the higher version.
The Foundation GUI is launched in the default browser or you can browse to http://localhost:8000/gui/index.html.
Procedure
1. If the Foundation app is running, right-click on the Foundation icon in launcher and click Force Quit.
Note: Foundation app does not remove the log and configuration files generated during uninstallation. For a clean
installation, Nutanix recommends that you delete these files manually.
Procedure
Do one of the following:
• Serves as a reusable baseline file that helps skip repeat manual entry of configuration details.
• Plan the configuration details in advance.
• Invite others to review and edit your planned configuration.
• Import NX nodes from a Salesforce order to avoid manually adding NX nodes.
Procedure
2. In a web browser inside the Foundation VM, browse to the http://localhost:8000/gui/index.html URL.
3. After you assign an IP address to the Foundation VM, browse to the http://<foundation-vm-ip-
address>:8000/gui/index.html URL from a web browser outside the VM .
4. On the Start page, you can import a Foundation configuration file created on the https://install.nutanix.com
portal.
To create or edit this configuration file, log into https://install.nutanix.com with your Nutanix Support portal
credentials. You can populate this file with either partial or complete Foundation GUI configuration details.
Note: The configuration file only stores configuration details and not AOS or hypervisor images. You can upload
images only in the Foundation GUI.
a. Click Next.
• BLOCK SERIAL
• NODE
• VLAN
• IPMI MAC
• IPMI IP
• HOST IP
• CVM IP
• HOSTNAME OFHOST
To configure values to the table, select the Tools drop-down list, and select one of the following options:
• Add Nodes Manually: To manually add nodes to the list if they are not populated automatically.
• Add Compute-Only Nodes: To add compute-only nodes. For more information about compute only
nodes, see Compute-Only Node Configuration (AHV Only) in the Prism Web Console Guide.
• Range Autofill: To bulk-assign the IP addresses and hostnames for each node.
Note: Unlike CVM Foundation, standalone Foundation does not validate these IP addresses by checking for
its uniqueness. Hence, manually cross-check and ensure that the IP addresses are unique and valid.
• Reorder Blocks: To match the order of IP addresses and hypervisor hostnames that you want to assign.
• Select Only Failed Nodes: To select all the failed nodes.
• Remove Unselected Rows: To remove a node from the Foundation process, de-select a node, and click
the Remove Unselected Rows option from the Tools drop-down.
a. Click Next.
6. On the Cluster page, provide the cluster details, configure cluster formation, or just image the nodes without
forming a cluster. You can also enable network segmentation to separate CVM network traffic from guest VMs
and hypervisors' network traffic.
Note:
• The Cluster Virtual IP field is essential for Hyper-V clusters but optional for ESXi and AHV
clusters.
• To provide multiple DNS or NTP servers, enter a comma-separated list of IP addresses. For best
practices in configuring NTP servers, see the Recommendations for Time Synchronization section
in the Prism Web Console Guide.
Note:
• You can select one or more nodes to be storage-only nodes that host AHV only. However, image
the remaining nodes with another hypervisor and form a multi-hypervisor cluster.
• For discovered nodes, if you skip updating CVMs with AOS, you can still reimage the hypervisors.
Hypervisor-only imaging is supported. However, do not image CVMs with AOS before imaging
the hypervisors.
• [Hyper-V only] If you choose Hyper-V from the Choose Hyper-V SKU list, select the SKU
that you want to use.
The following four Hyper-V SKUs are supported: Standard, Datacenter, Standard with
GUI, and Datacenter with GUI.
a. Click Next.
9. (For standalone Foundation only) On the IPMI page, provide the IPMI access credentials for each node.
Standalone Foundation needs IPMI remote access to all the nodes.
To provide credentials for all the nodes, select the Tools drop-down list to either use the Range Autofill
option or assign a vendor's default IPMI credentials to all nodes.
Note: You can cancel an ongoing installation in standalone Foundation but not in CVM Foundation.
Results
After all the operations are completed, the Installation finished page appears.
Note: If you missed any configuration, want to reconfigure, or perform the installation again, click Reset to return to
the Start page.
• Assign IP addresses to the hypervisor host, the Controller VMs, and the IPMI interfaces. Do not assign IP
addresses from a subnet that overlaps with the 192.168.5.0/24 address space on the default VLAN. Nutanix uses
Procedure
2. In a web browser inside the Foundation VM, browse to the http://localhost:8000/gui/index.html URL.
3. After you assign an IP address to the Foundation VM, browse to the http://<foundation-vm-ip-
address>:8000/gui/index.html URL from a web browser outside the VM .
» To configure automatically:
1. Download the Foundation configuration file from the https://install.nutanix.com portal.
Note: The configuration file only stores configuration details and not AOS or hypervisor images. You
can upload images only in the Foundation GUI.
2. Update this file with either partial or complete Foundation VM configuration details.
3. To select the updated file, click the import the configuration file link.
Option Description
Add Nodes Manually Add nodes manually if they are not already populated.
Add Compute-Only Nodes Add compute-only nodes. For more information about
compute only nodes, see the Compute-Only Node
Configuration (AHV Only) section in the Prism Web
Console Guide.
Select Only Failed Nodes Select all the failed nodes to debug the issues.
Note:
• The Cluster Virtual IP field is essential for Hyper-V clusters but optional for ESXi and AHV
clusters.
• To provide multiple DNS or NTP servers, enter a comma-separated list of IP addresses. For best
practices in configuring NTP servers, see the Recommendations for Time Synchronization section
in the Prism Web Console Guide.
7. On the AOS page, upload an AOS image or view the existing installed version of AOS image on each node.
Note: Skip updating CVMs with AOS if all discovered nodes' CVMs already have the same AOS version that
you want to use.
Note:
• You can select one or more nodes to be storage-only nodes, which host AHV only. You must
image rest of the nodes with another hypervisor and from a multi-hypervisor cluster.
• For discovered nodes, if you skip updating CVMs with AOS, you can still reimage the hypervisors.
However, imaging CVMs with AOS without imaging hypervisors are not supported.
• [Hyper-V only] If you choose Hyper-V, from the Choose Hyper-V SKU list, select the SKU
that you want to use.
Four Hyper-V SKUs are supported: Standard, Datacenter, Standard with GUI, and
Datacenter with GUI.
9. (For standalone Foundation only) On the IPMI page, provide the IPMI access credentials for each node.
Standalone Foundation needs IPMI remote access to all the nodes.
To provide credentials for all the nodes, select the Tools drop-down list to either use the Range Autofill
option or assign a vendor's default IPMI credentials to all nodes.
Note: You can cancel an ongoing installation in standalone Foundation but not in CVM Foundation.
Results
After all the operations are completed, the Installation finished page appears.
Note: If you missed any configuration, want to reconfigure, or perform the installation again, click Reset to return to
the Start page.
Procedure
1. Verify that the cluster has passed the latest Nutanix Cluster Check (NCC) tests.
a. Check the installed NCC version and update it if a recent version is available (see the "Software and Firmware
Upgrades" section).
b. Run NCC if you downloaded a newer version or did not run it as part of the install.
Run NCC from a command line. Open a command window, log on to any Controller VM in the cluster.
Establish SSH session, and then run the following command:
nutanix@cvm$ ncc health_checks run_all
If the check reports a status other than PASS, resolve the reported issues before proceeding. If you are unable
to resolve the issues, contact Nutanix Support for assistance.
c. Configure NCC so that the cluster checks run and emailed according to your desired frequency.
nutanix@cvm$ ncc --set_email_frequency=num_hrs
where num_hrs is a positive integer of at least 4 to specify how frequently NCC runs and results are emailed.
For example, to run NCC and email results every 12 hours, specify 12; or every 24 hours, specify 24, and so
on. For other commands related to automatically emailing NCC results, see "Automatically Emailing NCC
Results" in the Nutanix Cluster Check (NCC) Guide for your version of NCC.
Replace cluster_timezone with the timezone of the cluster (for example, America/Los_Angeles, Europe/
London, or Asia/Tokyo). Restart all Controller VMs in the cluster after changing the timezone. A cluster can
only tolerate outage of a single Controller VM at a time hence restart Controller VMs in a rolling fashion. Ensure
that a Controller VM is fully operational after a restart prior to proceeding with the restart of the next one. For
more information about using the nCLI, see the Command Reference.
3. Specify an outgoing SMTP server (see the "Configuring an SMTP Server" section).
CAUTION: Failing to enable remote support prevents Nutanix Support from directly addressing cluster issues.
Nutanix recommends that all customers allow email alerts at minimum because it allows proactive support of
customer issues.
5. If the site security policy allows Nutanix Support to collect cluster status information, enable the Pulse feature
(see the "Configuring Pulse" section).
This information is used by Nutanix Support to send automated hardware failure alerts, as well as diagnose
potential problems and assist proactively.
6. Add a list of alert email recipients, or if the security policy does not allow it, disable alert emails (see the
"Configuring Email Alerts" section).
You can also specify email recipients for specific alerts (see the "Configuring Alert Policies" section).
7. If the site security policy permits automatic downloads of upgrade software packages for cluster components,
enable the feature (see the "Software and Firmware Upgrades" section).
Note: To ensure that automatic download of updates can function, allow access to the following URLs through
your firewall:
• *.compute-*.amazonaws.com:80
• release-api.nutanix.com:80
9. For ESXi and Hyper-V clusters, add the hosts to the appropriate management interface.
Note: For the Lenovo Converged HX Series platform, use the custom ISOs that are available on the VMware website
(www.vmware.com) at Downloads > Product Downloads > vSphere > Custom ISOs.
Ensure to list the MD5 checksum of the hypervisor ISO image in the ISO whitelist file used by Foundation. See
Verify Hypervisor Support on page 39.
The following table describes the fields that appear in the iso_whitelist.json file for each ISO image.
Name Description
The following sample entries are from the whitelist for an ESX and an AHV image:
"iso_whitelist": {
"478e2c6f7a875dd3dacaaeb2b0b38228": {
"min_foundation": "2.1",
"hypervisor": "esx",
"min_nos": null,
"friendly_name": "ESX 6.0",
"version": "6.0",
"filesize": 329611264,
"unsupported_hardware": [],
"compatible_versions": {
"esx": ["^6\\.0.*"]
},
"a2a97a6af6a3e397b43e3a4c7a86ee37": {
"min_foundation": "3.0",
"hypervisor": "kvm",
"min_nos": null,
"friendly_name": "20160127",
"compatible_versions": {
"kvm": [
"^el6.nutanix.20160127$"
]
},
"version": "20160127",
"deprecated": "3.1",
"unsupported_hardware": []
},
Procedure
1. Obtain the MD5 checksum of the ISO that you want to use.
2. Open the downloaded whitelist file in a text editor and perform a search for the MD5 checksum.
What to do next
If the MD5 checksum is listed in the whitelist file, save the file to the workstation that hosts the Foundation
VM. If the whitelist file on the Foundation VM does not contain the MD5 checksum, you can replace that file
with the downloaded file before you begin installation.
Procedure
1. On the Foundation page, click Hypervisor and select a hypervisor from the drop-down list below Select a
hypervisor installer.
2. To upload a new whitelist.json file, click Manage Whitelist and click upload it.
Note: To verify if the iso_whitelist.json is updated successfully, open the Manage Whitelist menu and check
for the date of the newly updated file.
9
NETWORK REQUIREMENTS
When configuring a Nutanix block a set of IP addresses is required to be allocated to the cluster. Ensure that chosen
IP addresses do not overlap with any hosts or services within the environment. You will also need to make sure
to open the software ports that are used to manage cluster components and to enable communication between
components such as the Controller VM, Web console, Prism Central, hypervisor, and the Nutanix hardware. Nutanix
recommends that you specify information such as a DNS server and NTP server even if the cluster is not connected to
the Internet or runs in a non-production environment.
• Default gateway
• Network mask
• DNS server
• NTP server
Check whether a proxy server is in place in the network. To perform this check, you need the IP address and port
number of that server when enabling Nutanix Support on the cluster.
New IP Addresses
Each node in a Nutanix cluster requires three IP addresses, one for each of the following components:
• IPMI interface
• Hypervisor host
• Nutanix Controller VM
All Controller VMs and hypervisor hosts must be on the same subnet. No systems other than the Controller VMs and
hypervisor hosts can be on this network.
Note: If you have Volume Shadow Copy Service (VSS) based back up tool (for example Veeam), functional level
of Active Directory must be 2008 or higher.
• Install and run Active Directory Web Services (ADWS). By default, connections are made over TCP port 9389
and firewall policies enable an exception on this port for ADWS.
To test that ADWS is installed and run on a domain controller, log on as a domain administrator to a Windows
host in the same domain with the RSAT-AD-Powershell feature installed. Execute the following PowerShell
command:
> (Get-ADDomainController).Name
If the command prints the primary name of the domain controller, then ADWS is installed and the port is open.
Note: If any of the preceding requirements are not met, you must manually create an Active Directory computer
object for the Nutanix storage in the Active Directory, and add a DNS entry for the name.
• Ensure that the Active Directory domain is configured correctly for consistent time synchronization.
Accounts and Privileges:
• An Active Directory account with permission to create new Active Directory computer objects for either a storage
container or Organizational Unit (OU) where Nutanix nodes are placed. The credentials of this account are not
stored anywhere.
• An account that has sufficient privileges to join a Windows host to a domain. The credentials of this account are
not stored anywhere. These credentials are only used to join the hosts to the domain.
The following additional information are required:
Note: The primary domain controller IP address is set as the primary DNS server on all the Nutanix hosts. It is also
set as the NTP server in the Nutanix storage cluster to synchronize time between Controller VMs, hosts and Active
Directory.
• The fully qualified domain name to which the Nutanix hosts and the storage cluster is going to be joined.
Requirements:
• The SCVMM version must be 2012 R2 or later. The tool must be installed on a Windows Server 2012 or a later
version.
• The SCVMM server must allow PowerShell remote execution.
To test this scenario, log on by using the SCVMM administrator account in a Windows host and run the following
PowerShell command on a Windows host that is different to the SCVMM host (for example, run the command
from the domain controller). If the command returns the name of the SCVMM server, then PowerShell remote
execution on the SCVMM server is permitted.
> Invoke-Command -ComputerName scvmm_server -ScriptBlock {hostname} -Credential MYDOMAIN
\username
Replace scvmm_server with the SCVMM host name and MYDOMAIN with Active Directory domain name.
Note: If the SCVMM server does not allow PowerShell remote execution, you can perform the SCVMM setup
manually by using the SCVMM user interface.
• The ipconfig command must run in a PowerShell window on the SCVMM server. To verify run the following
command.
> Invoke-Command -ComputerName scvmm_server_name -ScriptBlock {ipconfig} -Credential
MYDOMAIN\username
Replace scvmm_server_name with the SCVMM host name and MYDOMAIN with Active Directory domain name.
• The SMB client configuration in the SCVMM server should have RequireSecuritySignature set to False. To
verify, run the following command.
> Invoke-Command -ComputerName scvmm_server_name -ScriptBlock {Get-SMBClientConfiguration |
FL RequireSecuritySignature}
If you are changing it from True to False, it is important to confirm that the policies on the SCVMM host have
the correct value. Changing the value of RequireSecuritySignature may not take effect if there is a policy with
the opposite value. On the SCVMM host, run rsop.msc to review the resultant set of policy details, and verify
the value by going to, Servername > Computer Configuration > Windows Settings > Security
Settings > Local Policies > Security Options: Policy Microsoft network client: Digitally sign
communications (always). The value displayed in RSOP must be, Disabled or Not Defined for the
change to persist. Also, if the RSOP shows it as Enabled, the group policies that are configured in the domain to
apply to the SCVMM server must be updated to Disabled. Otherwise, the RequireSecuritySignature will revert
back to the value of True. After setting the policy in Active Directory and propagating to the domain controllers,
refresh the SCVMM server policy by running the command gpupdate /force. Confirm in RSOP that the value
is Disabled.
Note: If security signing is mandatory, then you must enable Kerberos in the Nutanix cluster. In this case, it is
important to ensure that the time remains synchronized between the Active Directory server, the Nutanix hosts, and
the Nutanix Controller VMs. The Nutanix hosts and the Controller VMs use the Active Directory server as the NTP
server. So, ensure that Active Directory domain is configured correctly for consistent time synchronization.
• When adding a host or a cluster to the SCVMM, the run-as account that is used to manage the host or the cluster
must be different from the service account that was used to install SCVMM.
• Run-as account must be a domain account and must have local administrator privileges on the Nutanix hosts.
This can be a domain administrator account. When the Nutanix hosts are joined to the domain, the domain
administrator account automatically takes administrator privileges on the host. If the domain account used as the
run-as account in SCVMM is not a domain administrator account, you must manually add the run-as account to
the list of local administrators on each host by running sconfig.
• SCVMM domain account with administrator privileges on SCVMM and PowerShell remote execution
privileges.
• If you want to install SCVMM server, a service account with local administrator privileges on the SCVMM
server.
IP Addresses
Note: For N nodes, (3*N + 2) IP addresses are required. All IP addresses must be in the same subnet.
DNS Requirements
• Each Nutanix host must be assigned a name of 15 characters or less, which gets automatically added to the DNS
server during domain joining.
• The Nutanix storage cluster must be assigned a name of 15 characters or less. Then, add the name to the DNS
server when the storage cluster joins the domain.
• The Hyper-V failover cluster must be assigned a name of 15 characters or less, which gets automatically added to
the DNS server when the failover cluster is created.
• After the Hyper-V configuration, all names must resolve to an IP address in the Nutanix hosts, the SCVMM server
(if applicable), or any other host that needs access to the Nutanix storage, for example, a host running the Hyper-V
Manager.
• Virtual machine and virtual disk paths must always refer to the Nutanix storage cluster by FQDN and not the
external IP address. If you use the IP address, it directs all the I/O to a single node in the cluster and compromises
performance and scalability.
Note: For external non-Nutanix hosts that require access to Nutanix SMB shares, see Nutanix SMB Shares
Connection Requirements from Outside the Cluster..
• When applying Windows updates to the Nutanix hosts, the hosts should be restarted one at a time, ensuring that
Nutanix services of the Controller VM on the restarted host come up fully prior to proceeding with the update of
Note: Ensure that automatic Windows updates are not enabled for the Nutanix hosts in the domain policies.
• If you place a host that is managed by SCVMM in maintenance mode, the Controller VM running on the host is
placed in the saved state by default , which might create issues. To properly place a host in the maintenance mode
refer to SCVMM Operation in the Hyper-V Administration for Acropolis Guide.
11
SETTING IPMI STATIC IP ADDRESS
You can assign a static IP address for an IPMI port by resetting the BIOS configuration.
Note: Do not perform the following procedure for HPE DX series. The label on the HPE DX chassis contains the iLO
MAC address that is also used to perform MAC-based imaging.
To configure a static IP address for the IPMI port on a node, do the following:
Procedure
3. Press the Delete key during boot up when prompted to enter the BIOS setup mode.
The BIOS Setup Utility screen appears.
6. Select Update IPMI LAN Configuration, press Enter, and then select Yes in the dialog box.
8. Select Station IP Address, press Enter, and then enter the IP address for the IPMI port on that node in the
dialog box.
9. Select Subnet Mask, press Enter, and then enter the corresponding submask value in the dialog box.
10. Select Gateway IP Address, press Enter, and then enter the IP address for the node's network gateway in
the dialog box.
11. When all the field entries are updated with the correct values, press the F4 key to save the settings and exit the
BIOS setup mode.
12
TROUBLESHOOTING
This section provides guidance for fixing problems that might occur during a Foundation installation.
• For issues related to IPMI configuration of bare-metal nodes, see Fixing IPMI Configuration Issues on
page 49.
• For issues with the imaging, see Fixing Imaging Issues on page 50.
• For answers to other common questions, see Frequently Asked Questions (FAQ) on page 51.
• One or more IPMI MAC addresses are invalid or there are conflicting IP addresses. Go to the IPMI screen and
correct the IPMI MAC and IP addresses as needed.
• There is a user name/password mismatch. Go to the IPMI page and correct the IPMI username and password
fields as required.
• One or more nodes are connected to the switch through the wrong network interface. Verify that the first 1 GbE
network interface of each node is connected to the switch (see Setting Up the Network on page 21).
• The Foundation VM is not in the same broadcast domain as the Controller VMs for discovered nodes or the
IPMI interface for added (bare-metal or undiscovered) nodes. This problem typically occurs because (a) a non-
flat switch is used, (b) IP addresses of the node are not in the same subnet as the Foundation VM, and (c) multi-
homing is not configured.
• If all the nodes are in the Foundation VM subnet, go to the Node page and correct the IP addresses as needed.
• If the nodes are in multiple subnets, go to the Cluster page and configure multihoming.
• The IPMI interface is not set to failover. Go to BIOS settings to verify the interface configuration.
To identify and resolve IPMI port configuration problems, do the following:
Foundation | Troubleshooting | 49
Procedure
1. Go to the Block & Node Config screen and review the problematic IP address for the failed nodes (nodes with a
red X next to the IPMI address field).
Hovering the cursor over the address displays a pop-up message with troubleshooting information. This can help
to diagnose the problem. See the service.log file (in /home/nutanix/foundation/log) and the individual
node log files for more detailed information.
2. When the issue is resolved, click the Configure IPMI button at the top of the screen.
3. Repeat the preceding steps as necessary to fix all the IPMI configuration errors.
4. When all nodes have green check marks in the IPMI address column, click Image Nodes at the top of the
screen to begin the imaging step.
If you cannot fix the IPMI configuration problem for one or more of the nodes, click Proceed bypass those
nodes and continue to the imaging step for the other nodes. In this case, configure the IPMI port address manually
for each bypassed node (see Setting IPMI Static IP Address on page 47).
• A type failure was detected. Check connectivity to the IPMI (bare-metal workflow).
• There are network connectivity issues such as the following:
• The connection is dropping intermittently. If intermittent failures persist, look for conflicting IPs.
• [Hyper-V only] SAMBA service is not running. If Hyper-V displays a warning that it failed to mount the
install share, restart SAMBA with the command "sudo service smb restart".
• Foundation ran out of disk space during the hypervisor or Phoenix preparation phase. Free up some space by
deleting unwanted ISO images. In addition, a Foundation crash could leave a /tmp/tmp* directory that contains a
copy of an ISO image that you can unmount (if necessary) and delete. Foundation requires 9 GB of free space for
Hyper-V and 3 GB for ESXi or AHV.
• The host boots but returns the error indicating an issue with reaching the Foundation VM. The message varies per
hypervisor. For example, on ESXi you might see a ks.cfg:line 12: "/.pre" script returned with an
Foundation | Troubleshooting | 50
error error message. Ensure that you assign the host an IP address on the same subnet as the Foundation VM or
multihoming is configured. Also check for IP address conflicts.
To identify and resolve imaging problems, do the following:
Procedure
1. See the individual log file for any failed nodes for information about the problem.
2. When the issue is resolved, click the Image Nodes (bare metal workflow) button.
3. Repeat the preceding steps as necessary to resolve any other imaging errors.
If you are unable to resolve the issue for one or more of the nodes, it is possible to image these nodes one at a time
(Contact Support for help).
Installation Issues
Foundation | Troubleshooting | 51
• I want to troubleshoot the operating system installation during cluster creation.
Point a VNC console to the hypervisor host IP address of the target node at port 5901.
• I need to restart Foundation on the Controller VM.
To restart Foundation, log on to the Controller VM with SSH and then run the following command:
nutanix@cvm$ pkill foundation && genesis restart
• My installation hangs, and the service log complains about type detection.
Verify that all of your IPMI IPs are reachable through Foundation. (On a rare occasion the IPMI IP assignment
will take some time.) If you see an authentication error, double-check your password. If the problem persists, try
resetting the BMC.
• Installation fails with an error where Foundation cannot ping the configured IPMI IP addresses.
Verify that the LAN interface is set to failover mode in the IPMI settings for each node. You can find this setting
by logging into IPMI and going to Configuration > Network > LAN Interface. Verify that the setting is
Failover (not Dedicate).
• The diagnostic box was checked to run after installation, but that test (diagnostics.py) does not complete
(hangs, fails, times out).
Running this test can result in timeouts or low IOPS if you are using 1GbE port. Connection via 1GbE port may
not provide the performance necessary to run this test at a reasonable speed.
• Foundation seems to be preparing the ISOs properly, but the nodes boot into <previous hypervisor> and the install
hangs.
The boot order for one or more nodes might be set incorrectly with the precedence of the USB over SATA DOM
as the first boot device instead of the CD-ROM. To fix this, boot the nodes into BIOS mode and either select
"restore optimized defaults" (F3 as of BIOS version 3.0.2) or give the CD-ROM boot priority. Reboot the nodes
and retry the installation.
• I have misconfigured the IP addresses in the Foundation configuration page. How long is the timeout for the call
back function, and is there a way to avoid the wait?
The call back timeout is 60 minutes. To stop the Foundation process and restart it, open up the terminal in the
Foundation VM and enter the following commands:
$ sudo /etc/init.d/foundation_service stop
$ cd ~/foundation/
$ mv persisted_config.json persisted_config.json.bak
$ sudo /etc/init.d/foundation_service start
Refresh the Foundation web page. If the nodes are still stuck, reboot them.
• I need to reset a block to the default state.
Using the bare-metal imaging workflow, download the desired Phoenix ISO image for AHV from the support
portal (see https://portal.nutanix.com/#/page/phoenix/list). Boot each node in the block to that ISO and follow the
prompts until the re-imaging process is complete. You should then be able to use Foundation as usual.
• The cluster create step is not working.
If you are installing NOS 3.5 or later, check the service.log file for messages about the problem. Next, check
the relevant cluster log (cluster_X.log) for cluster-specific messages. The cluster create step in Foundation is
not supported for earlier releases and will fail if you are using Foundation to image a pre-3.5 NOS release. You
must create the cluster manually (after imaging) for earlier NOS releases.
• I want to re-image nodes that are part of an existing cluster.
Do a cluster destroy prior to discovery. (Nodes in an existing cluster are ignored during discovery.)
Foundation | Troubleshooting | 52
• My Foundation VM displays an error that it is out of disk space. What can I delete to make room?
Unmount any temporarily-mounted file systems using the following commands:
$ sudo fusermount -u /home/nutanix/foundation/tmp/fuse
$ sudo umount /tmp/tmp*
$ sudo rm -rf /tmp/tmp*
If more space is needed, delete some of the Phoenix ISO images from the Foundation VM.
• I keep seeing the message tar: Exiting with failure status due to previous errors'tar rf /
home/nutanix/foundation/log/archive/log-archive-20140604-131859.tar -C /home/nutanix/
foundation ./persisted_config.json' failed; error ignored.
This is a benign message. Foundation archives persisted configuration file (persisted_config.json) alongside the
logs. Occasionally, there is no configuration file to back up. This is expected, and you may ignore this message
with no ill consequences.
• Imaging fails after changing the language pack.
Do not change the language pack. Only the default English language pack is supported. Changing the language
pack can cause some scripts to fail during Foundation imaging. Even after imaging, character set changes can
cause problems for NOS.
• [ESXi] Foundation is booting into pre-install Phoenix, but not the ESXi installer.
Check the BIOS version and verify it is supported. If it is not a supported version, upgrade it.
Foundation | Troubleshooting | 53
• The switch is dropping my IPMI connections in the middle of imaging.
If your network connection seems to be dropping out in the middle of imaging, try using an unmanaged switch
with Spanning Tree Protocol (STP) disabled.
• Foundation is stalled on the ping home phase.
The ping test will wait up to two minutes per NIC to receive a response, so a long delay in the ping phase indicates
a network connection issue. Check that your 10GbE ports are unplugged and your 1GbE connection can reach
Foundation.
• Can I image nodes connected via 10/100 Mbps switch ports?
A 10/100Mbps switch is not recommended, but it can be used for a few nodes. You may see timeouts. It is highly
recommended that you use a 1GbE or 10GbE switch if possible.
Informational Topics
• How can I determine whether a node was imaged with Foundation or standalone Phoenix?
• A node imaged using standalone Phoenix will have the file /etc/nutanix/foundation_version in it, but
the contents will be “unknown” instead of a valid Foundation version.
• A node imaged using Foundation will have the file /etc/nutanix/foundation_version in it with a valid
Foundation version.
• Does the first boot work when run more than once?
First boot creates a marker failure marker file whenever it fails and a success marker file whenever it succeeds. If
the first boot script needs to be executed again, delete these marker files and just manually execute the script.
• Do the first boot marker files contain anything?
They are just empty files.
• Why might first boot fail?
Possible reasons include the following:
• First boot may take more time than expected, in which case Foundation might time out.
• NIC teaming fails.
• The Controller VM has a kernel panic when it boots.
• Hostd service does not start on time.
• What is the timeout for first boot?
The timeout is 90 minutes. A node may restart several times (requirements from certain driver installations)
during the execution of the first boot script, and this can increase the overall first boot time.
• How does the Foundation process differ on a Dell system?
Foundation uses a different tool called racadm to talk to the IPMI interface of a Dell system, and the files that
have the hardware layout details are different. However, the overall Foundation workflow (series of steps) remains
the same.
Foundation | Troubleshooting | 54
• How does the Foundation service start in the Controller VM-based and standalone versions?
• Standalone: Manually start the Foundation service using foundation_service start (in the ~/
foundation/bin directory).
• Controller VM-based: Genesis service takes care of starting the Foundation service. If the Foundation service
is not already running, use the genesis restart command to start Foundation. If the Foundation service
is already running, a genesis restart will not restart Foundation. You must manually kill the Foundation
service that is running currently before executing genesis restart. The genesis status command lists
the services running currently along with their PIDs.
• Why doesn’t the genesis restart command stop Foundation?
Genesis only restarts services that are required for a cluster to be up and running. Stopping Foundation could
cause failures to current imaging sessions. For example, when expanding a cluster Foundation may be in the
process of imaging a node, which must not be disrupted by restarting Genesis.
• How is the installer VM created?
The Qemu library is part of Phoenix. The qemu command starts the VM by taking a hypervisor ISO and disk
details as input. This command is simply executed on Phoenix to launch the installer VM.
• How do you validate that installation is complete and the node is ready with regards to firstboot?
This can be validated by checking the presence of a first boot success marker file. The marker file varies per
hypervisor:
• ESXi: /bootbank/first_boot.log
• AHV: /root/.firstboot_success
• Hyper-V: D:\markers\firstboot_success
• Does the Repair CVM process re-create partitions?
Repair CVM task only images AOS and recreates the partitions on the SSD. It does not alter any of the data on the
SATADOM, which contains the hypervisor.
• Can I use older Phoenix ISOs for manual imaging?
Use a Phoenix ISO that contains the AOS installation bundle and hypervisor ISO in it. Makefiles has a separate
target for building such a standalone Phoenix ISO.
• What are the pre-checks run when a node is added?
• The hypervisor type and version should match between the existing cluster and the new node.
• The AOS version should match between the existing cluster and the new node.
• Can I get a map of percent completion to step?
No. The percent completion does not have a one-to-one mapping to the step. Percent completion depends on the
different tasks that are executed during imaging.
• Do the log folders contain past imaging session logs?
Yes. All the previous imaging session logs are compressed (on a session basis) and archived in the folder ~/
foundation/log/archive.
• If I have two clusters in my lab, can I use one to do bare-metal imaging on the other?
No. The tools and packages that are required for bare-metal imaging are typically not present in the Controller
VM.
Foundation | Troubleshooting | 55
• How do you add an already imaged node to an existing cluster?
If the AOS version of the node matches or is below the same of the cluster you can use the "Expand Cluster"
option in the Prism web console. This option employs Foundation to image the new node (if required) and then
adds it to the existing cluster. You can also add the node through the nCLI: ncli cluster add-node node-
uuid=<uuid>. The UUID value can be found in the factory_config.json file on the node. Always check
compatibility matrix to verify that the hypervisor and AOS versions are compatible prior to expanding the cluster.
• Is it required to supply IPMI details when using the Controller VM-based Foundation?
It is optional to provide IPMI details in the Controller VM-based Foundation. If IPMI information is provided,
Foundation will attempt to configure the IPMI interface as well.
• Can I use a file share to hold AOS installation bundles and hypervisor ISO files?
AOS installation bundles and hypervisor ISO files can be stored externally. A link to the location must be added
to (as appropriate) ~/foundation/nos or ~/foundation/isos/hypervisor/[esx|kvm|hyperv]/ to the
share location. Foundation will discover files as described in these locations only. As long as the files' location is
described and is accessible from the Foundation using a link, Foundation will be able to leverage the files.
• Where is Foundation located in the Controller VM?
/home/nutanix/foundation
• How can I determine if a particular (standalone) Foundation VM can image a given cluster?
Execute the following command on the Foundation VM and see whether it returns successfully (exit status 0):
ipmitool –H ipmi_ip -U ipmi_username -P ipmi_password fru
If this command is successful, the Foundation VM can be used to image the node. This is the command used by
Foundation to get hardware details from the IPMI interface of the node. The exact tool used for communicating to
the SMC IPMI interface is the following:
java -jar SMCIPMITool.jar ipmi_ip username password shell
If this command is able to open a shell, imaging will not fail because of an IPMI issue. Any other errors like
violating minimum requirements will be displayed only after Foundation starts imaging the node.
• How do I determine whether a particular hypervisor ISO will work?
The md5 hash of all qualified hypervisor ISO images are listed in the iso_whitelist.json file, which is
located in ~/foundation/config/. The latest version of the iso_whitelist.json file is available from the
Nutanix support portal (see Hypervisor ISO Images on page 38).
• How does Foundation mount an ISO over IPMI?
The java command starts a shell with access to the remote IPMI interface. The vmwa command mounts the
ISO file virtually over IPMI. Foundation then opens another terminal and uses the following commands to set
the first boot device to CD-ROM and restarts the node.
ipmitool –H ipmi_ip -U ipmi_username -P ipmi_password chassis bootdev cdrom
ipmitool –H ipmi_ip -U ipmi_username -P ipmi_password chassis power reset
Foundation | Troubleshooting | 56
racadm -r ipmi_ip -u ipmi_username -p ipmi_password remoteimage -c -
l nfs_share_path_to_iso_file -u nutanix –p nutanix/4uracadm -r ipmi_ip -u ipmi_username -
p ipmi_password remoteimage –s
racadm -r ipmi_ip -u ipmi_username -p ipmi_password remoteimage –s
racadm -r ipmi_ip -u ipmi_username -p ipmi_password config -g cfgServerInfo -o
cfgServerBootOnce 1
racadm -r ipmi_ip -u ipmi_username -p ipmi_password config -g cfgServerInfo -o
cfgServerFirstBootDevice vCD-DVD
The node can be rebooted using the following commands:
racadm -r ipmi_ip -u ipmi_username -p ipmi_password serveraction powerdown
racadm -r ipmi_ip -u ipmi_username -p ipmi_password serveraction powerup
Foundation | Troubleshooting | 57
Appendix
A
A APPENDIX: SINGLE-NODE
CONFIGURATION (PHOENIX)
To configure a single node, to reinstall the hypervisor after you replace a hypervisor boot drive, or to install
or repair a Nutanix Controller VM, use Phoenix, which is an ISO installer.
For more information about using Phoenix, see KB 5591. To use Phoenix, contact Nutanix Support.
Warning:
• Nutanix does not support the use of Phoenix to reimage or reinstall AOS with the Action titled "Install
CVM" on a node that is already part of a cluster as it can lead to data loss.
• Use of Phoenix to repair the AOS software on a node with the Action titled "Repair CVM" must be
done only with the direct assistance of Nutanix Support.
• Use of Phoenix to recover a node after a hypervisor boot disk failure is not necessary usually. To
see how this recovery is automated through Prism for your platform model and AOS version, see
the Hardware Replacement Documentation.
COPYRIGHT
Copyright 2020 Nutanix, Inc.
Nutanix, Inc.
1740 Technology Drive, Suite 150
San Jose, CA 95110
All rights reserved. This product is protected by U.S. and international copyright and intellectual property
laws. Nutanix and the Nutanix logo are registered trademarks of Nutanix, Inc. in the United States and/or other
jurisdictions. All other brand and product names mentioned herein are for identification purposes only and may be
trademarks of their respective holders.
License
The provision of this software to you does not grant any licenses or other rights under any Microsoft patents with
respect to anything other than the file server implementation portion of the binaries for this software, including no
licenses or any other rights in any hardware or any devices or software that are used to communicate with or in
connection with this software.
Conventions
Convention Description
user@host$ command The commands are executed as a non-privileged user (such as nutanix)
in the system shell.
root@host# command The commands are executed as the root user in the vSphere or Acropolis
host shell.
> command The commands are executed in the Hyper-V host shell.
Foundation |
Interface Target Username Password
Version
Last modified: October 1, 2020 (2020-10-01T14:33:05+05:30)