Doug Stewart on 8 Aug 2018 16:55:15 -0700 |
[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]
Re: [PLUG] Virtualization clusters & shared storage |
OpenShift has container-native storage as a part of their native solution now, which isn’t 100% of what you’re saying, but conceptually is there. Basically it sets up a containerized Gluster cluster on the Kubernetes cluster. -- Doug Stewart > On Aug 8, 2018, at 7:50 PM, Lee H. Marzke <lee@marzke.net> wrote: > > > JP, if you want cheap storage for your lab , I think you can't beat FreeNAS or > equiv rackmount solutions from https://www.ixsystems.com. I run my lab on > FreeNAS and a Dell 2950 server with 6x2TB disks and 2 x SSD. If you put storage > into the servers you will find all sorts of edge cases that you hadn't planned on. > > Just taking down a server for quick RAM swap, will cause it to need to rebuild > using lots of network/CPU. If you have TB of fast SSD storage on multiple servers and > don't have 10GB connectivity between hosts, or you have slow HD's you will have pain. > Generally you try to migrate data off nodes prior to maintenance - which may take several days. > > The VMware solutions have changed a lot, and though it does not meet JP's > needs for *free*, they may fit someone else's needs for a reliable / highly > available solution. Of course ESXi and VSAN are free for 60 day trial after > install for lab use. > > First there is a VMware Migrate *both* mode where you can migrate both the Hypervisor and > then storage in one go, where the two storage units are not connected across Hypervisors. > Needless to say this takes a long time to sequentially move memory , then disk to the remote > server, and it doesn't help improve HA. > > Next VMware VSAN is caching on really fast and VMware is hiring like mad > to fill new VSAN technical sales roles Nationwide. VSAN uses storage in each > host ( minimum of one SSD and one HD ) and uses high-performance object > storage on each compute node. All VM objects are stored on two hosts > minimum, with vSAN taking care of all the distribution. The hosts must be > linked on a 1GB ( pref 10GB ) private network for back-end communication. > Writes are sent and committed to two nodes before being acknowledged. > You get one big storage pool - and allocate storage to VM's as you like - > with no sub LUNs or anything else to manage. If you have 4 or more hosts > instead of mirroring data over 2 hosts, you can do erasure coding ( equiv > of RAID 5/6 but with disks spread out across hosts. So now your not > losing 50% of your storage , but have more intensive CPU and network > operations. The vSAN software is pre-installed into ESX these days - just > need to activate it and apply a license after the 60 free day trial. > > Not sure why you say FreeNAS is wasting CPU in more nodes, as those CPU cycles > would be used locally in the Hyperconverged solutions as well ( perhaps taking 10% > to 20% cycles away from a host for storage and replication ) so you may need more / larger > hosts in a hyperconverged solution to make up for that. Remember mirroring > takes little CPU, but waste's 50% of your storage, any erasure coding is > much more CPU intensive, and more network intensive. > > The other solutions mentioned except a ZFS server are likely way too complex > for a lab storage solution. Is a company really going to give > a lab team 6 months of effort to put together storage that may or may > not perform ? Can you do a business justification to spend dozens of MM of > effort just to save the $20K on an entry level TrueNAS ZFS ? > > Lee > > > ----- Original Message ----- >> From: "Vossen JP" <jp@jpsdomain.org> >> To: "Philadelphia Linux User's Group Discussion List" <plug@lists.phillylinux.org> >> Sent: Wednesday, 8 August, 2018 17:13:17 >> Subject: [PLUG] Virtualization clusters & shared storage > >> I have a question about virtualization cluster solutions. One thing >> that has always bugged me is that VM vMotion/LiveMigration features >> require shared storage, which makes sense, but they always seem to >> assume that shared storage is external, as in a NAS or SAN. What would >> be REALLY cool is a system that uses the cluster members "local" storage >> as JBOD that becomes the shared storage. Maybe that's how some of >> solutions work (via Ceph, GlusterFS or ZFS?) and I've missed it, but >> that seems to me to be a great solution for the lab & SOHO market. >> >> What I mean is, say I have at least 2 nodes in a cluster, though 3+ >> would be better. Each node would have at least 2 partitions, one for >> the OS/Hypervisor/whatever and the other for shared & replicated >> storage. The "shared & replicated" partition would be, well, shared & >> replicated across the cluster, providing shared storage without needing >> an external NAS/SAN. >> >> This is important to me because we have a lot of hardware sitting around >> that has a lot of local storage. It's basically all R710/720/730 with >> PERC RAID and 6x or 8x drive bays full of 1TB to 4TB drives. While I >> *can* allocate some nodes for FreeNAS or something, that increases my >> required node count and wastes the CPU & RAM in the NAS nodes while also >> wasting a ton of local storage on the host nodes. It would be more >> resource efficient to just use the "local" storage that's already >> spinning. The alternative we're using now (that sucks) is that the >> hypervisors are all just stand-alone with local storage. I'd rather get >> all the cluster advantages without the NAS/SAN issues >> (connectivity/speed, resilience, yet more rack space & boxes). >> >> Are there solutions that work that way and I've just missed it? >> >> >> Related, I'm aware of these virtualization environment tools, any more >> good ones? >> 1. OpenStack, but this is way too complicated and overkill >> 2. Proxmox sounds very cool >> 3. Cloudstack likewise, except it's Java! :-( >> 4. Ganeti was interesting but it looks like it may have stalled out >> around 2016 >> 5. https://en.wikipedia.org/wiki/OVirt except it's Java and too limited >> 6. https://en.wikipedia.org/wiki/OpenNebula with some Java and might do >> on-node-shared-storage? >> 7. Like AWS: https://en.wikipedia.org/wiki/Eucalyptus_(software) except >> it's Java >> >> I'm asking partly for myself to replace my free but not F/OSS ESXi >> server at home and partly for a work lab that my team needs to rebuild >> in the next few months. We have a mishmash right now, much of it ESXi. >> We have a lot of hardware laying around, but we have *no budget* for >> licenses for anything. I know Lee will talk about the VMware starter >> packs and deals like that but we not only have no budget, that kind of >> thing is a nightmare politically and procedurally and is a no-go; it's >> free or nothing. And yes I know that free costs money in terms of >> people time, but that's already paid for and while we're already busy, >> this is something that has to happen. >> >> Also we might like to branch out from ESXi anyway... We are doing a >> some work in AWS, but that's not a solution here, though cross cloud >> tools like Terraform (and Ansible) are in use and the more we can use >> them here too the better. >> >> Thanks, >> JP >> -- ------------------------------------------------------------------- >> JP Vossen, CISSP | http://www.jpsdomain.org/ | http://bashcookbook.com/ >> ___________________________________________________________________________ >> Philadelphia Linux Users Group -- http://www.phillylinux.org >> Announcements - http://lists.phillylinux.org/mailman/listinfo/plug-announce >> General Discussion -- http://lists.phillylinux.org/mailman/listinfo/plug > > -- > "Between subtle shading and the absence of light lies the nuance of iqlusion..." - Kryptos > > Lee Marzke, lee@marzke.net http://marzke.net/lee/ > IT Consultant, VMware, VCenter, SAN storage, infrastructure, SW CM > +1 800-393-5217 voice/text > +1 484-348-2230 fax > ___________________________________________________________________________ > Philadelphia Linux Users Group -- http://www.phillylinux.org > Announcements - http://lists.phillylinux.org/mailman/listinfo/plug-announce > General Discussion -- http://lists.phillylinux.org/mailman/listinfo/plug ___________________________________________________________________________ Philadelphia Linux Users Group -- http://www.phillylinux.org Announcements - http://lists.phillylinux.org/mailman/listinfo/plug-announce General Discussion -- http://lists.phillylinux.org/mailman/listinfo/plug