After change ahv network port got NIC down alert | Nutanix Community
Skip to main content
Solved

After change ahv network port got NIC down alert

  • October 17, 2024
  • 6 replies
  • 132 views

Forum|alt.badge.img+1

Hi All,

Our nutanix 1175S-G9 is a 4-port 10G BaseT, initially connected to the  two ports
eth0, eth2,  later, I change  to eth0 and eth1, and then NCC check will show the message eth2 down. Is there any solution ?

Best answer by DylanH

Jon Jordan wrote:

You will have to remove it from PE web interface as manage_ovs commands will not work with the virtual switch created.

 

The alert should go away after 24 hours. NCC will alert on a NIC port going down for 24 hours, after this it will assume that its expected to be down and it should stop alerting.

 

https://portal.nutanix.com/kb/2566

 

This generally happens when you bring down an interface, then the NIC link down alert is displayed and cluster health turns red.

After the NIC is in a downstate, the NIC status will stay in the file check_cvm_health_job_state.json for 24 hours.

From NCC 4.3.0 onwards, nic_link_down_check will detect the NIC status changed to a downstate and then raise the alert. After 24 hours, if the NIC is still in the downstate and never goes in an upstate within the day, the NIC will be removed from file check_cvm_health_job_state.json and no more alert would be generated nor failure detected by nic_link_down_check.

Not working ? The vswitch will just be inconsistent. With “acli net.disable_virtual_switch” before and “acli net.migrate_br_to_virtual_switch xx vs_name=vsX” after ? 🤔

Regards,

View original
Did this topic help you find an answer to your question?
This topic has been closed for comments

6 replies

DylanH
Forum|alt.badge.img+2
  • Trendsetter
  • 46 replies
  • October 17, 2024

Hello, if i understood well. You just must remove the eth2 interface from the bonding. either from the prism element web interface, or from the cli on a CVM with the manage_ovs command. If you want more details/help, don't hesitate.

Regards,


Forum|alt.badge.img
  • Nutanix Employee
  • 5 replies
  • October 17, 2024

You will have to remove it from PE web interface as manage_ovs commands will not work with the virtual switch created.

 

The alert should go away after 24 hours. NCC will alert on a NIC port going down for 24 hours, after this it will assume that its expected to be down and it should stop alerting.

 

https://portal.nutanix.com/kb/2566

 

This generally happens when you bring down an interface, then the NIC link down alert is displayed and cluster health turns red.

After the NIC is in a downstate, the NIC status will stay in the file check_cvm_health_job_state.json for 24 hours.

From NCC 4.3.0 onwards, nic_link_down_check will detect the NIC status changed to a downstate and then raise the alert. After 24 hours, if the NIC is still in the downstate and never goes in an upstate within the day, the NIC will be removed from file check_cvm_health_job_state.json and no more alert would be generated nor failure detected by nic_link_down_check.


DylanH
Forum|alt.badge.img+2
  • Trendsetter
  • 46 replies
  • Answer
  • October 17, 2024
Jon Jordan wrote:

You will have to remove it from PE web interface as manage_ovs commands will not work with the virtual switch created.

 

The alert should go away after 24 hours. NCC will alert on a NIC port going down for 24 hours, after this it will assume that its expected to be down and it should stop alerting.

 

https://portal.nutanix.com/kb/2566

 

This generally happens when you bring down an interface, then the NIC link down alert is displayed and cluster health turns red.

After the NIC is in a downstate, the NIC status will stay in the file check_cvm_health_job_state.json for 24 hours.

From NCC 4.3.0 onwards, nic_link_down_check will detect the NIC status changed to a downstate and then raise the alert. After 24 hours, if the NIC is still in the downstate and never goes in an upstate within the day, the NIC will be removed from file check_cvm_health_job_state.json and no more alert would be generated nor failure detected by nic_link_down_check.

Not working ? The vswitch will just be inconsistent. With “acli net.disable_virtual_switch” before and “acli net.migrate_br_to_virtual_switch xx vs_name=vsX” after ? 🤔

Regards,


Forum|alt.badge.img
  • Nutanix Employee
  • 5 replies
  • October 17, 2024

That is correct Dylan, its just a lot of extra steps.

Manage_ovs commands to update things will not work if vswitch is created, and is a legacy way of doing things.

  • You can disable vswitch
  • Put the host in MM
  • Update it with manage_ovs
  • Take host out of MM
  • Then recreate the vswitch.

That is a lot more steps than just updating that one server in the GUI and letting automated processes handle everything.

if a CLI method is preferred you can use ACLI commands such as acli net.update_virtual_switch vs# host_uplink_config=
 

Going into acli context mode and using tab will help to learn how to use the commands and options. 


Forum|alt.badge.img+1
  • Author
  • Trailblazer
  • 10 replies
  • October 17, 2024

Hi @DylanH & @Jon Jordan ,

 

Thank you both.


JeroenTielen
Forum|alt.badge.img+8
  • Vanguard
  • 1363 replies
  • October 18, 2024

I had the same issue. But the alert will be gone after 24 hours. Just wait.