How It works
Have questions about how the Nutanix Platform works? Looking to get started - start here!
- 1,185 Topics
- 1,843 Replies
The Nutanix REST API allows administrators to create and run scripts on their Nutanix clusters. On a Nutanix cluster: API v1 and v2 runs through the Prism service. API v3 runs through Aplos services. Both of the above are services running on a CVM/PC-VM in a Nutanix cluster. Prism: Prism is the management gateway for component and administrators to configure and monitor the Nutanix cluster. This includes Ncli, the HTML5 UI, and REST API. Prism runs on every node in the cluster and uses an elected leader like all components in the cluster. All v1 and v2 API calls runs through Prism. Aplos: It’s an intentful orchestration engine + intentful API proxy. All v3 REST API calls runs through Aplos. Looking at the logs of the above 2 services can tell us a lot of things about our REST API and HTTP issues on PC-VM/CVM. The logs for these services are located on a CVM/PC-VM at: API v3 logs: ~/data/logs/aplos.out API v1 and v2 logs: ~/data/logs/prism_gateway.log HTTP defines a set of reque
We are using Image configuration to upload and convert the VHDXs and so far it works except when we attempt to upload 4K sector aligned VHDXs from an old HyperV cluster, the upload completes successfully but the conversion fails. Is it possible to migrate and convert 4K sector aligned VHDXs to a Nutanix AHV cluster? Error:
Below are new knowledge base articles published on the week of June 14-20, 2020. KB 8507 - Cosmetic high latency spikes may be observed in Prism at time of low IOPS on a cluster KB 8932 - NCC Health Check: pc_vm_resource_resize_check KB 9423 - Beam Cost Governance for Nutanix On-Prem stops reporting cost analytics post Cluster upgrade to AOS 5.15 KB 9488 - AHV | CentOS/RHEL 6.8 may hang durin the boot when tboot package is used KB 9496 - Prism Central UI showing license expired red banner even after disabling prism pro features KB 9503 - Genesis crashing on a ESXI Node | No services are up | pyVmomi issue | ParserError: 'xml document KB 9516 - Using Nutanix objects as static web server for LCM dark sites KB 9519 - Nutanix Move | How to Configure HTTP(s) Proxy on Move KB 9534 - Hyper-V: Hyper-V 2019 fails during Foundation with "InstallerVM timeout occurred, current retry 0" KB 9538 - Prism logins fail if service account AD permissions are default KB 9545 - Copier reports inc
Hi, A customer moved his Nutanix Cluster (with Hyper-V) from a DC to another, after powering the Nodes up, the IPs of all Hyper-V host and CMs released, I logged it locally to Hyper-V hosts and configure the internal IP (192.168.5.1/28) and the external IP same like before shutting the cluster down. I repeated the previous step with CVMs, I went through cd/etc/sysconfigs/ and edit network-scripts file and added the the external IP in the eth0 and the internal one (192.168.5.2/28) in eth1. Now Hyper-V FC is working fine but cannot start VMs due to the Nutanix cluster issue, whenever I tried to start cluster from any CVM, I get this message “WARNING genesis_utils.py:1211 Failed to reach a node where Genesis is up. Retrying” Is there any way to fix this issue or to repair cluster configuration without disrupting existing data? Thanks in advance
Ques- There is 5 node cluster with 200 TB raw disk data, administrator want to enable eraser coding to obtain more space. What will we the usable space after enable eraser coding Options – 100, 125, 150, 175 TB ? Please share the calculation formula ? Ques2 - Administrator want to perform DR test after 6 months, which snapshot he should use? Option – latest snapshot, oldest one
Below are new knowledge base articles published on the week of June 7-13, 2020. KB 9456 - Alert - A400114 - PolicyEngineServiceDown KB 9468 - Different Docker hosts can see volumes created with Nutanix DVP KB 9478 - How to clear stuck LCM inventory tasks ? KB 9480 - Windows 10, version 2004 or Windows Server, version 2004 VMs may fail to boot on AHV KB 9487 - AHV | VM update operations initiated from Prism Central may fail with "Entity CAS version mismatch" error due to missing machine_type attribute KB 9492 - Move VM IP address may change post deployment KB 9494 - Disks from ISCSI volumes change drive letters on Windows VMs after upgrade AHV 2016* to 2017* Note: You may need to log in to the Support Portal to view some of these articles.
Hi all, A newbie question. It seems I still have an old crash dump directory on one of my AHV hosts. A ls -lahtr /var/crash shows the single directory from back in April. At the time the issue was resolved, and the faulty DIMM that was causing the issue was replaced. That said, clearly the dump file was not removed. In terms of cleaning this up, is it OK to delete the dump directory within the /var/crash directory and then rerun the ncc health check. Or is there a better method for clearing crash dumps from Nutanix clusters. Many thanks, Rob
Please answer the below questions : When one cvm goes down(in a 10 node cluster with RF3) for 20 mints, then Guest VM’s new write and read IO would be served by anther's CVM and all these will be traveling across 10 g network. A : Will those new IOs served from local copy via another CVM or IOs will be served from the replica copy B: Will cluster starts to build a new replica to accommodate the missing copy or RF3 Ques 2 - When new Write IO request comes, first it will write the data on Oplog then synchronously send to other CVM’s Oplog. All host in clusters having 2 SSD and 6 HDDs. Will Write IO process by both SSD in every host or only one SSD is hsving Oplog partition active at the same time ? I think there is only one oplog per CVM/host whether the host is aving all flash drive or 2 SSD and remaining HDD. But I am not sure about this statement, Please clarify..
Is it possible to recover data from an overwritten entity - does the overwritten VM or its storage still exist anywhere? I'm asking as I recovered a VM from a backup but realised in hindsight that the logs from the VM in its failed state would be useful for diagnosing the issue. Thanks in advance.
What is the DIMM error? A memory error is an event that leads to the logical state of one or multiple bits being read differently from how they were last written. For example, If 1 was written in a memory cell and while reading the same memory cell, it returns 0. Memory errors can be classified into two types: Soft errors, which randomly corrupt bits but do not leave physical damage. Soft errors are transient in nature and are not repeatable. Soft errors can be because of electrical or magnetic interference (e.g. due to cosmic rays, alpha particles, leakage, random noise). Hard errors, which corrupt bits in a repeatable manner because of a physical/hardware defect or an environmental problem. Hard error can also occur if DIMM is not seated properly. All memory systems in use in servers today are protected by error detection and correction codes. These server machines employ error correcting codes (ECC), which allows the detection and correction of one or m
First, let us understand what NTP (Network time protocol) is. An NTP server is a time server that is used to keep/sync the time in your cluster. An NTP server can be public or private depending on the strictness of your environment. To know how to configure NTP in your Nutanix cluster, take a look at- https://support-portal.nutanix.com/#/page/docs/details?targetId=Web-Console-Guide-Prism-v5_16:wc-system-ntp-servers-wc-t.html After the NTP server is configured, the genesis leader becomes the NTP leader, which means that the genesis leader is syncing time to the NTP server and other CVMs are syncing time with the genesis leader. How NTP works in AHV:- It’s as simple as it gets. The AHV hypervisor takes the same server configured on the cluster and syncs the time with it individually. There are no extra steps required to configure the NTP server on the AHV hosts. How NTP works in ESXi:- The ESXi cluster does not take the server configured on the Nutanix cluster and it needs to be
Hello, I’m working with Rest API, to “automate” some operations with Ansible, trought URI Module and Jinja2. I need to create a Project, and set permission to a specific user. How can i do this with API V3? I can add user to a Project, but i can’t set the Role.. When i check trought the Web Interface, i see the user without role. Thanks!
Below are new knowledge base articles published on the week of May 31-June 6, 2020. KB 9442 - LCM BIOS/BMC Upgrade fails when node does not respond to IPMI power reset KB 9460 - Move: ESXi-AHV migration connection limits KB 9463 - Pre-upgrade check: Hypervisor Upgrade (test_host_upgrade_versions_compatible) KB 9467 - How to enable Karbonctl in Karbon darksite environment Note: You may need to log in to the Support Portal to view some of these articles.
I have a W2k19 VM I created as a patched image. Idea is to patch it monthly. Shut it down and clone it. Then sysprep the clone and copy the disk to the Image Service. Then I can create VM’s off of it. I will repeat the colne after patching the ~master image monthly in order to keep a patched disk ready to spin up VM’s from. We are using AHV, version 5.10. Mt question is the “CLONE” a full, independant copy of the VM ? TIA -S
Hi community, I’m fairly new to interacting with Nutanix for automation and was hoping my VM build task would be a simple process but I’ve hit a wall with the guest tools install. I’m building a RDS environment where I need as little human interaction as I can get to make administration easier so I’m building my host VM using SCCM. I’ve multiple good reasons for approach but its mainly in support for application deployment via happening at build time. I’ve tried to install the guest tools manually but I believe they are failing due to the files being copied from a mounted ISO which has some dynamically created certificates involved so this isn’t going to work. I can see the ISO can be mounted using the API calls but every reference I’ve seen looks to need a username/password passing in the script and I’m not overly keen in hard coding creds into script. Most of what i’ve seen is a few years old so I wondered if there has been any changes to the approach to mounting guest tools? It
Below are the top knowledge base articles for the month of May 2020. KB 7503 - G6, G7 platforms - DIMM Error handling and replacement policy KB 4116 - Alert - A1187, A1188 - ECCErrorsLast1Day, ECCErrorsLast10Days KB 1540 - What to do when /home partition or /home/nutanix directory is full KB 7604 - Disk space usage for root on Controller VM has exceeded 80% KB 4141 - Alert - A1046 - PowerSupplyDown KB 4409 - LCM: (LifeCycle Manager) Troubleshooting Guide KB 4158 - Alert - A1104 - PhysicalDiskBad KB 1113 - HDD/SSD Troubleshooting KB 2090 - AHV | Host and Guest Networking KB 2473 - NCC Health Check: cvm_memory_usage_check KB 8792 - NCC checks: same_hypervisor_version_check, duplicate_cvm_ip_check, same_timezone_check, esx_sioc_status_check, power_supply_check, orphan_vm_snapshot_check giving ERR KB 4519 - NCC Health Check: check_ntp KB 2486 - NCC Health Check: cvm_mtu_check KB 4273 - NCC Health Check: aged_third_party_backup_snapshot_check KB 3357 - NCC Health Check: ipmi_
Below are new knowledge base articles published on the week of May 24-30, 2020. KB 9390 - Alert - A130338 - ServiceBadScore KB 9431 - LCM operation on ESXi hosts fails with "LAG is configured with X uplinks however there are Y NICs added in the teaming policy" Note: You may need to log in to the Support Portal to view some of these articles.
Giving thought to change your replication factor from 2 to 3? What are the impacts and things to consider? First, let’s take a look at what replication factor is. Redundancy factor is a configurable option that allows a Nutanix cluster to withstand the failure of nodes or drives in different blocks. By default, Nutanix clusters have redundancy factor 2, which means they can tolerate the failure of a single node or drive. So RF3 means cluster can tolerate the failure of 2 nodes or drive… Basic Maths isn’t it? Redundancy factor 3 has the following requirements: Redundancy factor 3 can be enabled at the time of cluster creation or after creation too. A cluster must have at least five nodes for redundancy factor 3 to be enabled. For guest VMs to tolerate the simultaneous failure of two nodes or drives in different blocks, the data must be stored on containers with replication factor 3. Controller VMs must be configured with a minimum of 28 GB(20 GB default+8 GB for the featur
Below are new knowledge base articles published on the week of May 17-23, 2020. KB 4781 - A Windows 7 or Windows 10 User VM Fails to Power On with an NVIDIA vGPU Profile in ESXi 6.5b KB 9218 - Prism Element alternative UPN login fails KB 9357 - Oops - Server error when updating Container settings in Prism KB 9373 - AHV | 10Gbps NIC shows as 1Gbps despite auto negotiation enabled on Intel X550 NIC cards KB 9394 - Objects - Deployment might fail with Asymmetric routing / Policy Based routing configured in environment | Deployed cluster greyed out | IAM service is not healthy KB 9400 - AHV | Troubleshooting Virtual Machine boot failures KB 9403 - LEAP - Entity Sync tasks generated continously after PC upgrade to 5.16.1 KB 9407 - [Objects] Objects 2.1 Darksite Deployment failing at deploying metadata store KB 9420 - /tmp filled up by /tmp/paramiko_logs file KB 9422 - Error updating the current checks parameter. Health check schedule interval is invalid. KB 9425 - Official Guidan
Below are new knowledge base articles published on the week of May 10-16, 2020. KB 9304 - Object Store Deployment Failure: type[CREATE]:code: 400, message: Primary MSP deployment in failed state status_code: 400 KB 9329 - ESXi | 1 click upgrade failing with 'Could not find a trusted signer' KB 9333 - Prism Central - Adding same email address under both "report settings" as well as "schedule settings" causes duplicate emails to be sent to recipient KB 9349 - [Nutanix Objects][Common Deployment Failure Scenario] Failed to create Envoy VM | Prism Central is not able to reach MSP and Envoy VMs or slow image download KB 9354 - Error: Node XXX cannot be removed: Cluster needs at least 5 usable nodes KB 9368 - Karbon UI stuck at loading in case of old browser version KB 9377 - [Karbon] Darksite Deployment fails in Proxy Environment in ETCD Deployment Stage KB 9383 - AHV | manage_ovs tool shows error "Failed to send RPC request. Retrying." when trying to change bridge configuration KB 9389 - D
Login to the community
Login with your account
Enter your username or e-mail address. We'll send you an e-mail with instructions to reset your password.