Doug Stewart on 8 Aug 2018 19:37:06 -0700


[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]

Re: [PLUG] Virtualization clusters & shared storage


JP:
Merely conceptual overlap and not really practical for your purposes. I think I was trying to say that other people see the same issue as you do, just in a different venue. 

--
Doug Stewart

> On Aug 8, 2018, at 9:23 PM, JP Vossen <jp@jpsdomain.org> wrote:
> 
> Firth, thanks to Lee and Andy for the term I didn't know for what I mean: hyper-converged infrastructure.
>    https://en.wikipedia.org/wiki/Hyper-converged_infrastructure
> 
> It looks like Proxmox does that:
>    https://pve.proxmox.com/wiki/Hyper-converged_Infrastructure
> 
> Thanks Keith, I had seen LizardFS but was not aware of the implications.
> 
> Doug, Kubernetes is not in play, though it might be in the future.  Or I may be missing your point.
> 
> Lee, thanks for the insights and VMware updates, as always. :-)
> 
> I've used FreeNAS in the past and like it, but I'm not sure I'm explaining my thoughts as well as I'd like.  But let me try this:
> 
> (I  wrote this next part before I learned that "hyper-converged" is what I mean, but I'm leaving it here in case it's still useful.)
> 
> Assume for simplicity that I have 5 R710s with 24M RAM, 8 CPUs and 6x 2TB drives in RAID5 for 10TB local storage each.
> 
> 3 node cluster with redundant FreeNAS:
>    1. VM node1: CPU/RAM used for VMs, 10TB local space wasted
>    2. VM node2: CPU/RAM used, 10TB local space wasted
>    3. VM node3: CPU/RAM used, 10TB local space wasted
>    4. FreeNAS 1: CPU/RAM somewhat wasted, 10TB space available
>    5. FreeNAS 2: CPU/RAM somewhat wasted, 0TB space available since it's a mirror
> 
> 3 node cluster with -on-node-shared-storage- hyper-converged storage:
>    1. VM node1: CPU/RAM used, 20TB local/shared space available +
>    2. VM node2: CPU/RAM used, 20TB local/shared space available +
>    3. VM node3: CPU/RAM used, 20TB local/shared space available +
> + For 20TB I'm assuming (3-1) * 10TB, for some kind of parity space loss.  If it was REALLY smart, it would keep the store for the local VMs local while still replicating, but that would require a lot more hooks into the entire system, not just some kind of replicated system.
> 
> But the point here is that with my idea I have 2x the disk with 3/5ths the servers.  Or put another way, I can now do a 5 node cluster with even more CPU, RAM and space dedicated to actually running VMs, and not lose 2/5ths of the nodes to just storing the VMs.
> 
> That said, I'd thought about the rebuild overhead, but not in depth, and that--and general "parity" or redundancy however implemented--are significant.  So my 2/5ths comparisons are not 100% fair.  But still, the idea apparently does have merit.
> 
> 
>> On 08/08/2018 07:50 PM, Lee H. Marzke wrote:
>> JP, if you want cheap storage for your lab , I think you can't beat FreeNAS or
>> equiv rackmount solutions from https://www.ixsystems.com. I run my lab on
>> FreeNAS and a Dell 2950 server with 6x2TB disks and 2 x SSD. If you put storage
>> into the servers you will find all sorts of edge cases that you hadn't planned on.
>> Just taking down a server for quick RAM swap, will cause it to need to rebuild
>> using lots of network/CPU.   If you have TB of fast SSD storage on multiple servers and
>> don't have 10GB connectivity between hosts, or you have slow HD's you will have pain.
>> Generally you try to migrate data off nodes prior to maintenance - which may take several days.
>> The VMware solutions have changed a lot, and though it does not meet JP's
>> needs for *free*,  they may fit someone else's needs for a reliable / highly
>> available solution.    Of course ESXi and VSAN are free for 60 day trial after
>> install for lab use.
>> First there is a VMware Migrate *both* mode where you can migrate both the Hypervisor and
>> then storage in one go, where the two storage units are not connected across Hypervisors.
>> Needless to say this takes a long time to sequentially move memory , then disk to the remote
>> server, and it doesn't help improve HA.
>> Next VMware VSAN is caching on really fast and VMware is hiring like mad
>> to fill new VSAN technical sales roles Nationwide.   VSAN uses storage in each
>> host ( minimum of one SSD and one HD ) and uses high-performance object
>> storage on each compute node.  All VM objects are stored on two hosts
>> minimum, with vSAN taking care of all the distribution.  The hosts must be
>> linked on a 1GB ( pref 10GB ) private network for back-end communication.
>> Writes are sent and committed to two nodes before being acknowledged.
>> You get one big storage pool - and allocate storage to VM's as you like -
>> with no sub LUNs or anything else to manage.   If you have 4 or more hosts
>> instead of mirroring data over 2 hosts,  you can do erasure coding ( equiv
>> of RAID 5/6 but with disks spread out across hosts.   So now your not
>> losing 50% of your storage , but have more intensive CPU and network
>> operations.   The vSAN software is pre-installed into ESX these days - just
>> need to activate it and apply a license after the 60 free day trial.
>> Not sure why you say FreeNAS is wasting CPU in more nodes,  as those CPU cycles
>> would be used locally in the Hyperconverged solutions as well ( perhaps taking 10%
>> to 20% cycles away from a host for storage and replication ) so you may need more / larger
>> hosts in a hyperconverged solution to make up for that.   Remember mirroring
>> takes little CPU, but waste's 50% of your storage,  any erasure coding is
>> much more CPU intensive, and more network intensive.
>> The other solutions mentioned except a ZFS server are likely way too complex
>> for a lab storage solution.   Is a company really going to give
>> a lab team 6 months of effort to put together storage that may or may
>> not perform ?  Can you do a business justification to spend dozens of MM of
>> effort just to save the $20K on an entry level  TrueNAS ZFS ?
>> Lee
>> ----- Original Message -----
>>> From: "Vossen JP" <jp@jpsdomain.org>
>>> To: "Philadelphia Linux User's Group Discussion List" <plug@lists.phillylinux.org>
>>> Sent: Wednesday, 8 August, 2018 17:13:17
>>> Subject: [PLUG] Virtualization clusters & shared storage
>>> I have a question about virtualization cluster solutions.  One thing
>>> that has always bugged me is that VM vMotion/LiveMigration features
>>> require shared storage, which makes sense, but they always seem to
>>> assume that shared storage is external, as in a NAS or SAN.  What would
>>> be REALLY cool is a system that uses the cluster members "local" storage
>>> as JBOD that becomes the shared storage.  Maybe that's how some of
>>> solutions work (via Ceph, GlusterFS or ZFS?) and I've missed it, but
>>> that seems to me to be a great solution for the lab & SOHO market.
>>> 
>>> What I mean is, say I have at least 2 nodes in a cluster, though 3+
>>> would be better.  Each node would have at least 2 partitions, one for
>>> the OS/Hypervisor/whatever and the other for shared & replicated
>>> storage.  The "shared & replicated" partition would be, well, shared &
>>> replicated across the cluster, providing shared storage without needing
>>> an external NAS/SAN.
>>> 
>>> This is important to me because we have a lot of hardware sitting around
>>> that has a lot of local storage.  It's basically all R710/720/730 with
>>> PERC RAID and 6x or 8x drive bays full of 1TB to 4TB drives.  While I
>>> *can* allocate some nodes for FreeNAS or something, that increases my
>>> required node count and wastes the CPU & RAM in the NAS nodes while also
>>> wasting a ton of local storage on the host nodes.  It would be more
>>> resource efficient to just use the "local" storage that's already
>>> spinning.  The alternative we're using now (that sucks) is that the
>>> hypervisors are all just stand-alone with local storage.  I'd rather get
>>> all the cluster advantages without the NAS/SAN issues
>>> (connectivity/speed, resilience, yet more rack space & boxes).
>>> 
>>> Are there solutions that work that way and I've just missed it?
>>> 
>>> 
>>> Related, I'm aware of these virtualization environment tools, any more
>>> good ones?
>>> 1. OpenStack, but this is way too complicated and overkill
>>> 2. Proxmox sounds very cool
>>> 3. Cloudstack likewise, except it's Java! :-(
>>> 4. Ganeti was interesting but it looks like it may have stalled out
>>> around 2016
>>> 5. https://en.wikipedia.org/wiki/OVirt except it's Java and too limited
>>> 6. https://en.wikipedia.org/wiki/OpenNebula with some Java and might do
>>> on-node-shared-storage?
>>> 7. Like AWS: https://en.wikipedia.org/wiki/Eucalyptus_(software) except
>>> it's Java
>>> 
>>> I'm asking partly for myself to replace my free but not F/OSS ESXi
>>> server at home and partly for a work lab that my team needs to rebuild
>>> in the next few months.  We have a mishmash right now, much of it ESXi.
>>> We have a lot of hardware laying around, but we have *no budget* for
>>> licenses for anything.  I know Lee will talk about the VMware starter
>>> packs and deals like that but we not only have no budget, that kind of
>>> thing is a nightmare politically and procedurally and is a no-go; it's
>>> free or nothing.  And yes I know that free costs money in terms of
>>> people time, but that's already paid for and while we're already busy,
>>> this is something that has to happen.
>>> 
>>> Also we might like to branch out from ESXi anyway...  We are doing a
>>> some work in AWS, but that's not a solution here, though cross cloud
>>> tools like Terraform (and Ansible) are in use and the more we can use
>>> them here too the better.
> Thanks,
> JP
> --  -------------------------------------------------------------------
> JP Vossen, CISSP | http://www.jpsdomain.org/ | http://bashcookbook.com/
> ___________________________________________________________________________
> Philadelphia Linux Users Group         --        http://www.phillylinux.org
> Announcements - http://lists.phillylinux.org/mailman/listinfo/plug-announce
> General Discussion  --   http://lists.phillylinux.org/mailman/listinfo/plug
___________________________________________________________________________
Philadelphia Linux Users Group         --        http://www.phillylinux.org
Announcements - http://lists.phillylinux.org/mailman/listinfo/plug-announce
General Discussion  --   http://lists.phillylinux.org/mailman/listinfo/plug