你好!
我对Nutanix的世界还很陌生。已经处理标准服务器+存储超过十年。这里我们有2个集群,每个站点有3个节点,每个站点都有地铁可用性。每个站点(活动)中有3个激活的保护域,它们被复制到另一个站点(被动),反之亦然。
Site1:
Node1 - Node3 - Node5
pd site1 (active): PROD_001, DEV_001, INFRA_001
pd site1 (passive): PROD_002, DEV_002, INFRA_002
Site2:
Node2 - Node4 - Node6
pd site2 (active): PROD_002, DEV_002, INFRA_002
pd site2(被动):PROD_001, DEV_001, INFRA_001
在vCenter集群配置中,我们对site1中的虚拟机/主机和site2中的虚拟机/主机有明显的亲和性规则,防止运行在“奇数”节点上的虚拟机被存储在“偶数”节点上。
有时我们必须将虚拟机从一个站点迁移到另一个站点。所以我们做了一个完整的vmotion(计算和存储)。在迁移之后,我们开始不断收到这条消息的警告:
租户infr_001快照状态:Failed。租户infr_001中存在被其他租户保护的虚拟机:VM = SXXXX96租户= (PROD_002)。请先解除对租户的保护,再对该租户进行快照。
当我们将虚拟机数据文件从一个数据存储存储到同一站点的另一个数据存储时,也会发生这种情况。我搜索了互联网和nutanix的文档,没有发现如何处理这些错误。它说“unprotect vm to vstore before snapshot this vstore”,但我怎么做?它是在ncli上完成的吗?棱镜?vCenter吗?我们是不是有什么事没做?最佳实践是什么?
任何帮助都将不胜感激。
谢谢
恩里克
最佳答案伊万诺夫
I have checked the history of your support cases and I have found a performance related case that was regarding the bug in VMware - when there are more than 5 NFS datastores connected via the same IP, the storage performance degrades over time. This issue is addressed in ESXi versions 6.5U3, 6.7U3 and newer. We have also applied a workaround from the AOS side and simply upgrading AOS to 5.10.4 and newer applies the fix, but the hosts need a reboot after that. That is what i can see happened in your situation - fix was already applied, but the reboot was pending. As i can see from the case, the issue was resolved after the hosts reboots were completed.<\/p>
Here is the information about that VMware bug:\u00a0https:\/\/kb.vmware.com\/s\/article\/67129<\/a><\/p> We also have a KB about this issue with more details:\u00a0https:\/\/portal.nutanix.com\/kb\/6961<\/a><\/p> \u00a0<\/p>","className":"post__content__best_answer"}">