Installation & Configuration
This forum is the best way to get up and running with the Nutanix platform
- 1,101 Topics
- 2,925 Replies
Hello All, For one of the customers, I did a fresh install and on that one, when I check the data resiliency status, it shows me that for Extend Group, it is 0. The explanation is as follows: "Based on placement of extend group replicas the cluster can tolerate a maximum of 0 node failure." Do you know what this can be about? There is no error or warning on the cluster. Cheers..
VDi (Virtual Desktop Infrastructure) is one of the earliest applications that was used in hyper-convergent systems. The closer the storage to the cpu and memory the better the performance. The plugin provides the means for getting more information on Citrix director from Citrix generated desktops running on a Nutanix cluster. Citrix also has its own plugin that can be used. This conversation is about nutanix provided plugin and how to install it. Nutanix AHV Plug-in for Citrix Director creates customized reports for the following VM-level performance statistics: VM IOPS VM I/O Bandwidth VM Average I/O Latency Nutanix AHV Plug-in for Citrix Director gets the data directly from the Nutanix AHV hosts to generate performance statistics of the virtual machines. The link: https://portal.nutanix.com/#/page/docs/details?targetId=AHV-Plugin-Citrix-Director-Installation-Guide-v1110:AHV-Plugin-Citrix-Director-Installation-Guide-v1110 have step by step instructions for installation of this plugin
As we have a strict policy in our company to be always at the latest patches for all services, we are required to upgrade to latest version of VMware "ESXi 6.7 U3" which was released last month, we are already running 6.7 U2 now. we have already upgraded our vCenter to 6.7 U3 which we do not think it has any issues with Nutanix. But to upgrade ESXi itself we do not know about its compatibility, i know about the case of patching ESXi to a higher patch level than the Nutanix certified ones "which we are regularly doing", but for ESXi 6.7 U3 we are afraid of breaking something or being incompatible, and we do not know either if ESXi U's are considered as patches or different versions according to Nutanix.. I searched over Internet and this forum about anything related to 6.7 U3 compatibility but found nothing, at least if we have a road-map from Nutanix about planned compatible version we can wait till its release. Thanks
hi all Im trying to install 4 node nutanix but I got this issue Do you have any idea what is the issue? this error is the same in this 4 nodes [code]20190902 16:16:19 INFO Setting cdrom as boot device for next boot 20190902 16:16:35 INFO Next boot device is set to optical 20190902 16:16:35 INFO Power status is off 20190902 16:16:35 INFO Powering up node 20190902 16:16:58 INFO Exiting SMCIPMITool 20190902 16:16:58 ERROR Exception in ) @bcb0> Traceback (most recent call last): File "foundation\decorators.py", line 77, in wrap_method File "foundation\imaging_step_init_ipmi.py", line 309, in run File "foundation\imaging_step_init_ipmi.py", line 156, in boot_phoenix File "site-packages\bmc_utils\boot_media.py", line 47, in wrapped File "site-packages\bmc_utils\remote_boot_rmh.py", line 193, in boot File "site-packages\bmc_utils\remote_boot_rmh.py", line 240, in boot_from_iso File "site-packages\pyghmi\ipmi\command.py", line 314, in set_power IpmiException: timeout 20190
Hi, everyone I’m trying to use LCM to update firmware of host machines on my cluster. It all went well until the post action phase. Error message said: ‘Operation failed. Reason: LCM failed performing action reboot_from_phoenix in phase PostActions on ip address xx’. I searched the KB and found KB9177，but it’s about ‘Mixed Hypervisor cluster‘, my cluster uses solely AHV so it doesn’t applied. Anyway I still tried to follow the KB9177’s suggestion and upgraded my cluster’s foundation to 4.5.3 and retried the LCM firmware update process on another host but still got the same ‘ LCM failed performing action reboot_from_phoenix’ error. I used the workaround provided in that KB to make the affected two hosts’ CVM out of maintenance mode. It works and the cluster is back to normal. Then I logged on to the affected hosts’ IMM and found out that actually the primary IMM2 firmware has already been updated by the LCM( the backup IMM2 firmware is not upgraded), and when I go to the LCM section
Hello, comunity I’m been trying to figure out why the configuration with RF2 & N+1 reserves more capacity than just RF2. I understand that with RF2 all the data is duplicated along the cluster and the failure of one node is tolerated. In my scenario I have a 3-node cluster with 34TiB of effective capacity, so with RF2, I would have only 17TiB, everything is clear until that point. But the extent store assuming RF2 & N+1 gives me 11.3TiB, which is 1/3 of my available capacity. So...the question is: If I reserve capacity, by tripling all the data, that is for tolerate 1 additional node failure besides the 1st one (tolerated by RF2)? A silly conclussion would be that with RF2 & N+1 the cluster is able to tolerate 2 nodes and continue operating just with one node, but I know that is not posible. So, why assuming RF2 & N+1 reserves more TiB’s than just RF2? Please, I would appreciate the help. Thanks in advanced!
Dears, i have below situation happened to my nutanix block my college changged hostname of all ESXi servers and then try to change CVM IPs without stopping cluster after that he rebooted CVM but cluster is not started with error message unable to reach node where genesis is started i have check genesis service in all node and it's running correclty and all nodes are reachable suing hosts in /etc/hosts file and all of them in same lan how can i get original hostname of esxi and does change of it's hostname cause this issue and also hos i resolve this issue thanks in advance
Suppose, it is discovered that a Nutanix cluster is configured with an incorrect time zone. While like all good things in life this requires a little planning, fear not – this can be helped!First things first, what to expect:Timestamps of Nutanix logs events will remain in the incorrect time zone until cluster services are restarted or the CMVs are rebooted. Cluster services restart on a cluster with active workload may impact availability of the workload. Changing the timezone for PE and PC is the same procedure, is not disruptive, but it requires a reboot of CVMs. Reboot can be performed on one CVM at a time. For a seamless experience it is recommended to evacuate VMs from the node prior to CVM reboot. Verify cluster Data Resiliency health prior to restart of cluster services or CVMs. Ensure Data Resiliency comes back healthy prior to proceeding with another CVM reboot. It is necessary to remove DR schedules prior to the change. Schedules can be re-created post change. Command for ch
At the end of guest OS installation the VM is rebooted but it doesn't automatically boot to the vHDD. How can I set it to boot automatically from the disk with installed guest OS? By default, AHV uses CD-ROM as the first boot device for VMs, followed by hard disk, then NIC. Please follow steps described below to override this behavior and choose a specific boot device. Things to consider:CD-ROM or hard disk can be selected as boot device in Prism UI. Network adapter can only be configured as boot device via acli. Selecting NIC as boot device may be required to boot VM from PXE server. Power cycle is required to apply boot device change. Resetting or rebooting VM is not enough. If boot device not explicitly selected, then all VMs will try to boot from all CD-ROM devices. If all CD-ROMs are empty or do not contain bootable image, then VM will only try to boot from hard disk with lowest index. If VM is configured to boot from network and there is no PXE server available, VM will not try t
While much is shared and posted about LACP configuration in AOS* the questions “How?” remains open. Let’s take a look at the workflow:Connect to CVM via SSH. Make sure you are connected to correct CVM by checking its name and IP. Follow steps from KB 2852 to verify cluster health. Do not proceed if cluster cannot tolerate failure of at least 1 node. Put the node and CVM in the maintenance mode. Connect to host via IPMI, as LACP configuration process might cause network disconnect. Perform required configuration steps (enable/disable LACP). Once configuration is completed, make sure both host and CVM are accessible via network. Please also make sure that all NICs in bond are operational, by shutting down links one by one and verifying connectivity. If all tests are successfully completed remove CVM and node from maintenance mode.For detailed instructions as well as commands examples please refer to KB-3263.For complete command reference Acropolis Command-Line Interface (aCLI).* For more
I attempted an upgrade of NCC however it is stuck. Is there a way to restart the upgrade? [img]https://d1qy7qyune0vt1.cloudfront.net/nutanix-us/attachment/d9c1c7da-1011-4da7-98c3-5a0d014baade.png[/img] [img]https://d1qy7qyune0vt1.cloudfront.net/nutanix-us/attachment/279e1192-82b4-4c34-b612-2c9269707901.png[/img]
Hi All First I run this: allssh manage_ovs --bridge_name br0 --interfaces eth0,eth1 --bond_name bond0 update_uplinks and the expectation I would have Bond0 entry somewhere, however I run the "ovs-appctl bond/show /bond0" command and I get this answer: FIPS mode initialized Nutanix AHV no such bond ovs-appctl: ovs-vswitchd: server returned an error What am I missing?
We recently had multiple Nutanix Blocks installed and I have started configuring them for Active Directory Authentication; however, logging on using AD Accounts is super slow and takes several minutes to logon. I have configured the Authentication to IP Addresses, FQDN's and DOMAIN but all are still unacceptably slow. ldap://192.168.1.1:389 ldap://server.domain.org:389 ldap://domain.org:389 For the Prism Role mapping, I have configured AD Groups and Single Users and the logon is still super slow. There was a post about change recursive authentication to be off; however, there was no command string associated with NCLI. Anyone experiencing this issue? Would like to know the best practice for configuration AD Authentication. Thanks for any assistance... David
Companies merge, hosts are inherited and repurposed, administrators come and go and not necessarily leave things in order after themselves. Or what if you forgot the password and the record of it is not recoverable? Where there’s a will, there’s a way. I touched the other day on how ipmitool can be accessed from a CVM CLI but would require authentication details to pass on to the IPMI. Well, the host or rather the hypervisor installed on the host does not have to do so. Hence even when ADMIN account password (or any other account with full privileges) is lost the IPMI is still accessible from the hypervisor CLI and you can still reset the password, set up new user entirely, or set a new password for a user. The process consists of several steps: Log in to the hypervisor on the host (AHV or ESXi). List user records using ipmitool. Take note of the user ID. Set the password for the user ID. The task is complete. Conveniently, commands syntax between AHV and ESXi differs only in the
This article contains IPMI commands for checking and setting interfaces to dedicated or shared mode. For example, after a BMC upgrade, the IPMI might not be accessible. So, you need to verify and change the interfaces to dedicated or shared mode. Note: To run ipmitool commands on an ESXI host, prefix all commands with a forward slash (/). Note: To run ipmitool commands from a remote system such as the CVM (Controller VM), add the "-I lanplus", "-H <IPMI IP>", "-U <username>" and "-P <password>" parameters to the ipmitool command. For example: nutanix@cvm$ ipmitool -I lanplus –H x.x.x.x –U ADMIN –P <password> <command> Quanta Platform Use these commands for an NX-3400 (Quanta) platform. All commands are executed dynamically and a restart is not required. Check the status. [root@host]# ipmitool raw 0x0c 0x02 0x01 0xff 0 0 An output similar to the following is displayed.1100 :00 - Shared port 1101 :01 - D
Let’s say you are running AOS version 5.10.7, and want to upgrade to a different Nutanix Files version or confused regarding which Prism Central version you should install or upgrade to?Nutanix gives you a feature, Software Product Interoperability, to easily identify the compatible versions between software like Prism Central and Nutanix files with the AOS.All we need to do is first select the Software Type(AOS, Nutanix Files or Prism Central) and then the target software type and then we can find the AOS version compatible.Points to be taken into consideration Prism Central does not follow the Long Term Support Release (LTS) and Short Term Support Release (STS) tracks. Nutanix Files (formerly known as AFS) does not follow the Long Term Support Release (LTS) and Short Term Support Release (STS) tracks. Refer to the following link to know more about LTS and STS tracksLTS/STS InformationInterested to know more?How upgrades happen in Nutanix environment Let’s learn about upgrades Prism
In this topic I will share how to configure an SMTP server on your Nutanix cluster. This can be done from the Prism UI. This is pretty simple and straight forward. Below is the link shared, that discusses how to configure a SMTP server- https://portal.nutanix.com/#/page/kbs/details?targetId=kA032000000TTWtCAO After the SMTP server is created, the email alerts should be tested if they are being sent from the cluster. SMTP always uses port 80 and 8443, so make sure that these firewall ports are open. You could send test emails to verify correct SMTP configuration. This can be done using the KB2773. I will attach a link to the KB below. https://portal.nutanix.com/#/page/kbs/details?targetId=kA03200000097qFCAQ Hope this piece of information is useful!
Hi there!I’ve installed NutanixCE (twice!) on my existing hardware (custom SuperMicro servers, not Nutanix hardware, though they meet the required specs). We found it useful, and decided to spring for Ultimate Licenses for the non-Community Edition.So I’ve got those licenses, but now I’m having a heck of a time figuring out what to download and install! For CE, there was a nice procedure to create an install image on a thumb drive, but I’m not seeing the same thing for non-CE.Can anyone point me to documentation that spells out what needs to be downloaded and how to install it?I’ve already found the AHV admin guide, but that doesn't have installation instructions.I’ve found the “Field Installation Guide - Foundation 4.5.x”, but the “https://portal.nutanix.com/#/page/docs/details?targetId=Field-Installation-Guide-v4_5:Prepare%20Bare-Metal%20Nodes%20for%20Imaging” section seems impossibly complicated; do I really need to install and configure a VM on virtual box on a laptop in order to
Nutanix Cluster is comprised of several critical software and hardware components. All communication happens on a IP network. CVM - Controller Virtual Machine (AOS) Hypervisor - (AHV | ESXi | Hyper-V | XEN) IPMI - (Out of band management for the hosts) Each CVM communicates with its respective Hypervisor Host on a private IP network. Hosts and CVMs must be on the same subnet. Before proceeding with any network related changes, on the CVM or Hypervisor hosts, it is recommended to perform a cluster health check. Following Nutanix KB outlines the steps to ensure your Nutanix Cluster is in a healthy and consistent state: Checklist on verifying cluster health status To change the IP Addresses, Net-Mask (subnet mask), Default gateway for your AHV hosts due to hardware re-location, network design changes, please follow the guidance from AHV Advanced Administration Guide: CHANGING THE IP ADDRESS OF AN ACROPOLIS HOST Changing IPMI Address of a Hypervisor Host CHANGING THE CONTROLLER VM IP A
Hello everyone, my environment: 6 * 3060-G4 NODE per node : 2*ssd 480G， 4* HDD（how to find out if it`s sas or sata drive ?）1TB RF=2 --------------------------------------------------------------------------------------------------------- how to calculate? I see prism report 22.18TiB MaxCapacity(Physical)。but I think physical should be 1T*4*6+480G*2*6=24T+5760G if RF2 need two complete copy(like RAID1) ?, the usable capacity is just half of 24T+5760G, and if after one node failure, for N+1 , I should also reserve another extra node capacity(4T+480G*2) ? also I have a basic question and guess, if I have three NODE in all(assume not 6 node ), and one node failed, lost data will be replicate to another any node of remaining two, because of RF. so I have two copy again. but if the second node also fails later(now I have only one node), will the cluster be down ? data lost ? I think data shouldn`t be lost because one of the two copy exists(3 node ,2copy-> 2 node 2copy-> 1node 1copy)
Here’s my problem. Node A & B are synchronizing to the wrong host. Node C is synchronizing to the right host. This is the behavior I see on the CVMs. When I do “hostssh ntpq -pn”, the hypervisors (AHV) are reporting the correct ntp server. How do I bring nodes A & B back inline to what they should be? I tried manually correcting their ntp.conf files with the correct IP and restarted the ntpd service. No change and eventually the ntp.conf files changed to the wrong settings. Not sure how to wrestle this one to the ground.
Login to the community
Login with your account
Enter your username or e-mail address. We'll send you an e-mail with instructions to reset your password.