解决了

磁盘故障Tollance.

  • 2016年3月25日
  • 3回复
  • 4406观点

徽章 +3
嗨,我对Nutanix技术印象非常深刻!

但是我是一个问题,如果我有一个带有Thre节点的块(基本要开始),请使用RF2配置,我理解我可以松开1个节点,我的所有系统仍然可以启动和运行。

如果我在同一节点内的磁盘松开,也让我继续运行没有问题,但如果我在不同节点上的同时松开两个磁盘,是什么堆堆?

我明白,在这种情况下,我松开了两个丢失的磁盘上的数据,所以在这种情况下,所有MI进一步都会下降?

可能我错过了一些东西......
图标

最好的答案j2016年3月31日,10:47

\n
\nNutanix does not use RAID to protect data, we store data in a \"Replication Factor\", which stores individual blocks of data in a redundant fashion across two or more nodes in a cluster (i.e RF2 two copies or RF3, three copies).
\n
\n
\nIf you have a drive fail, let's say it was a 1TB drive but only 200GB full.
\n
\nFor the sake of easy math, let's maintain a three node cluster.
\n
\nThat means (roughly) 200GB of information was on that disk, and approximately 200GB of information is spread across on all of the disks in the other two nodes, roughly 100GB per node
\n
\nIn a traditional storage system, you'd have to:
\nRebuild an entire 4TB Drive map, on to a hot spare (idle drive) within the system, regardless of data
\nRebuild that data parity from the \"RAID Pack\" the drive failed from, which trashes performance of that RAID pack and other workloads on it, and takes forever to do the operation.
\n
\nIn Nutanix, No Raid, so you only have to rebuild\/reprotect 200GB worth of information, instead of 1TB. Also, that 200GB is spread out across the entire cluster, so all disks and nodes participate in the rebuild, spreading out the rebuild task, and making it very low impact on the cluster and performance (if at all).
\n
\n
\nThe end result?
\nDrives fail, and rebuilds happen very quickly, as the rebuild eats into the free capacity of the cluster. This means no idle\/wasted hot spares.
\nThis means that the data is re-protected faster, so the likely hood of that second drive failure taking out data is minimized (not zero, but minimized).
\n
\n
\nIf you are concerned with dual disk failure, which some customers are for business critical operations, you'd want to go with an RF3 setup, which is basically N+2, so you can have any two components fail without worry.
\n
\n
\nAnyhow, you can read more about cluster resiliency in the nutanix bible:
\nhttp:\/\/nutanixbible.com\/<\/a>","className":"post__content__best_answer"}">
查看原版

此主题已关闭征询意见

3回复

UserLevel 6.
徽章 +29
真实的,但与传统的存储系统不同,这很少,如果有的话,问题,这里是高水平的“为什么”:

Nutanix不使用RAID来保护数据,我们将数据存储在“复制因子”中,该数据存储在群集中的两个或更多个节点中以冗余方式存储各个数据块(即RF2两个副本或RF3,三个副本)。


如果您有一个驱动器失败,让我们说这是一个1TB驱动器,但只有200GB完整。

为了简单的数学,让我们维护三个节点群集。

这意味着(粗略地)200GB的信息在该磁盘上,大约200GB的信息在其他两个节点中的所有磁盘上传播,每个节点大约为100GB

在传统的存储系统中,您必须:
无论数据如何,重建整个4TB驱动器地图,on系统内的热备用(空闲驱动器)
重建“RAID Pack”的数据奇偶校验,驱动器失败,从而删除了该RAID包和其它工作负载的性能,并将永远执行操作。

在Nutanix,没有raid,所以你只需要重建/重新调整200GB的信息,而不是1TB。此外,200GB在整个群集中展开,因此所有磁盘和节点都会参与重建,传播重建任务,并使其对群集和性能的影响非常低(如果有的话)。


最终结果?
驱动器失败,重建很快就会发生,因为重建进入群集的自由容量。这意味着没有空闲/浪费的热备件。
这意味着数据重新受到保护,因此第二个驱动器故障所取出数据的可能引擎盖被最小化(不为零,但最小化)。


如果您关注的是双磁盘故障,某些客户是企业关键操作,您希望使用基本上为N + 2的RF3设置,因此您可以在没有担心的情况下使用任何两个组件失败。


无论如何,您可以在Nutanix圣经中阅读有关集群弹性的更多信息:
http://nutanixbible.com/
徽章 +3
非常感谢您,为您彻底的回应。

这就是我所知道的。

谢谢!

当2个不同的主机上1个驱动器失败时,带有RF2的群集中的VM会发生什么?将是任何数据丢失还是没有?VM会重新启动吗?

Baidu