«

»

Jul 03 2014

Nutanix One-Click Upgrade Demo Video (Simple & Easy)

As the name says, it’s a One-Click NOS upgrade for the entire Nutanix cluster. Nutanix one-click upgrade automatically indicates when a new NOS version is available and it will auto-download the binaries if the auto-download option is enabled. With a single-click to upgrade all nodes in a cluster Nutanix will use a highly parallel process and reboot one CVM at a time using a rolling upgrade mechanism. The entire cluster upgrade can be fully monitored by the administrator.

During the upgrade process a feature called Autopathing kicks in redirecting VM I/Os on that Hypervisor host to another CVM (Nutanix Controller) on a different hypervisor host. This has several benefits including allowing you to upgrade the NOS on a CVM without disruption.

The new and improved CVM AutoPathing 2.0 prevents performance loss during rolling upgrades minimizing I/O timeout by pre-emptively redirecting storage traffic to other CVMs. Failover traffic is automatically load-balanced with the rest of the cluster based on node load.

Simple and Easy – Consumer-grade technology applied to enterprises while hiding all the complexity. As my colleague Michael Webster says it best…
 

Screen Shot 2014-07-03 at 11.59.21 AM

 
[Watch in 1080p full-screen]


 
This article was first published by Andre Leibovici (@andreleibovici) at myvirtualcloud.net.

7 comments

Skip to comment form

  1. comotoza

    after upgrade nodes to 4.0.1 Prism show me:
    Critical i RESILIENCY STATUS
    Yes REBUILD CAPACITY AVAILABLE
    Yes AUTO REBUILD IN PROGRESS

    Domain Type : NODE
    Component Type : EXTENT_GROUPS
    Current Fault Tolerance : 0
    Fault Tolerance Details : Based on placement of extent group replicas the
    cluster can tolerate a maximum of 0 node failure(s)

    is it normal? what if one of my node/disk failed till next time partial scan?
    partial scan is one in minutes
    or I show manually run something like

    # Description: Starts a Curator full scan from the CLI

    for i in `svmips`;do wget -O – “http://$i:2010/master/api/client/StartCuratorTasks?task_type=1”; done

    in console?

  2. Andre Leibovici

    comotoza, how many nodes do you have in your cluster? did you run ncc?

  3. comotoza

    3350 – 3 nodes, ncc and health in prism tell that everything is OK
    After 15-30 min after upgrade to 4.0.1 resiliency status changed to OK

    I’ve got answer on NEXT
    The UI is just being conservative. When the CVM goes down there could be writes happening to that CVM so it marks it as unhealthy until the next partial scan happens so it can confirm. When a CVM goes down the Oplog does have it’s own recovery process and will fix itself right away.
    So I have to wait till partial scan? What if till scan one node failed?
    Or with more nodes rebuild is more quicker?

  4. Andre Leibovici

    The good NEXT… I should write about it to get more people using it. Glad it’s all sorted.

  5. Bonn

    I have the same problem in our Proof of concept. Every time i reboot a CVM, i get the resiliency status critical (extend groups). I takes an hour until the status is good. I have a block with 4 nodes. So if i want to update my ESX-Hosts, it takes over 4 hours. THat is unreasonable. The question if i can shutdown another CVM while the status in this case is critical is not answered.

    Soory my bad english. Greetings from germany.

  6. Andre Leibovici

    Bonn, it’s just Prism showing wrong data, curator will show the real resiliency. It’s improved in NOS 4.1. However, I would encourage you to log a support request as to get it investigated since hypervisor upgrade should not take one hour.

  7. Bonn

    Thank you for your reply. I will test it with version 4.1 again.

Leave a Reply