That's the slick part of this - there wouldn't be a performance impact if setup correctly. Stargate and Curator will just not use the other disks as a target for the data or it's replicas. This does mean that one application will fill up before others. When they start to run out of space I would say they should spin out a new cluster rather then expanding this cluster (makes it easier to manage and just nicer over all)
If setup incorrectly (for example by not keeping the VMs on the hosts that have their SP's disks) we would lose data locality, but that's about it.
And yes, Zookeeper and Cassandra will continue on as normal. Assuming 3 2-node blocks, you should keep ZKs on separate blocks but beyond that it doesn't matter too much. There's a couple of choices you can have for actual deployment:
1) Ability to shutdown block without affecting any layer
In this deployment any block can fail without taking down any of the layers, but if you need to shutdown one of the layers for maintenance you need to shutdown the whole cluster
2) Ability to shutdown application without affecting other applications:
In this case any application can go down without affecting the other applciations, but if they want to shutdown a block for some reason {EDIT:} an application layer needs to go down {not the whole cluster}
Just to add another wrech to the deployment, you could do an RF3 cluster with RF2 containers. In that case you can lose 2 nodes without affecting the cluster as a whole, but if both nodes failures are from 1 application that application will go down without affecting the other applications. This plus the setup from the first scenario up there give you the ability to withstand both failure scenarios. There is a small performance hit here. They also wouldn't be able to spin out the nodes into a new cluster easily here (as RF3 requires 5 nodes) which makes this a less then ideal solution.
Thanks for this Sharing !:)
I'll give you feed back on this.
To explain this choice of design, you need to understand some contraints that I have for building this blackbox :
- Implementation will be spread across multiple entities, including massive (1000+ VM + on demand IAAS to little (no more than 50 VMs) deployment.
- As a matter of simplification (CAPEX volume and OPEX installation team), choice was made on only one DELL model,
- Since we need 3D support and FC child card for some entities, RC730 16G will be the only option (for nutanix).
- Nutanix is balanced with VSAN on my client choice.
- model that will be tested : 2xE5_2680v3 CPU / 24 x 16 Go of RAM / 4 x 400Go SSD drive / 12x 1,2 To SAS 10k HDD / 2xK2 card when 3D is needed...
=> this is a good fat box so I'm trying to convice my client chosing the coolest product and this kind of short package for little client.
So my strategy will be :
__________
little BOX :
6/8 nodes with 3/4 storage pools / 1 vCenter & 3/4 clusters
middle BOX :
{9/12 nodes with 2 clusters, 1 storage pool for management / 1 for resources and 2/3 containers / 2 vCenter (management & resources w/ 2/3 Vclusters} OR
{9/12 nodes with 3/4 clusters, 1 storage pool each and 1 containers / 2 vCenter (management & resources w/ 2/3 clusters}
large BOX :
- 6+ nodes management / 1 cluster /{1 SP (for RF3 +ECx support).Maybe 2 SP if NSX edge segmentation is needed}.1 vCenter 1+Vcluster.
- 6+ nodes resources / 1 cluster / 2/3 SP. 1vCenter 2/3 clusters.
_____________
w/ your explanation (and a reserve since it will be taylored on paperboard not tested nor validated), and some pros and cons table I bet it will be fair enough for their choice to be made.
sheers,