[Gluster-users] Backups
Joe Julian
joe at julianfamily.org
Thu Mar 23 19:40:51 UTC 2017
I always use raw images. And yes, sharding would also be good.
On 03/23/17 12:36, Gandalf Corvotempesta wrote:
> Georep expose to another problem:
> When using gluster as storage for VM, the VM file is saved as qcow.
> Changes are inside the qcow, thus rsync has to sync the whole file
> every time
>
> A little workaround would be sharding, as rsync has to sync only the
> changed shards, but I don't think this is a good solution
>
> Il 23 mar 2017 8:33 PM, "Joe Julian" <joe at julianfamily.org
> <mailto:joe at julianfamily.org>> ha scritto:
>
> In many cases, a full backup set is just not feasible. Georep to
> the same or different DC may be an option if the bandwidth can
> keep up with the change set. If not, maybe breaking the data up
> into smaller more manageable volumes where you only keep a smaller
> set of critical data and just back that up. Perhaps an object
> store (swift?) might handle fault tolerance distribution better
> for some workloads.
>
> There's no one right answer.
>
>
> On 03/23/17 12:23, Gandalf Corvotempesta wrote:
>> Backing up from inside each VM doesn't solve the problem
>> If you have to backup 500VMs you just need more than 1 day and
>> what if you have to restore the whole gluster storage?
>>
>> How many days do you need to restore 1PB?
>>
>> Probably the only solution should be a georep in the same
>> datacenter/rack with a similiar cluster,
>> ready to became the master storage.
>> In this case you don't need to restore anything as data are
>> already there,
>> only a little bit back in time but this double the TCO
>>
>> Il 23 mar 2017 6:39 PM, "Serkan Çoban" <cobanserkan at gmail.com
>> <mailto:cobanserkan at gmail.com>> ha scritto:
>>
>> Assuming a backup window of 12 hours, you need to send data
>> at 25GB/s
>> to backup solution.
>> Using 10G Ethernet on hosts you need at least 25 host to
>> handle 25GB/s.
>> You can create an EC gluster cluster that can handle this
>> rates, or
>> you just backup valuable data from inside VMs using open
>> source backup
>> tools like borg,attic,restic , etc...
>>
>> On Thu, Mar 23, 2017 at 7:48 PM, Gandalf Corvotempesta
>> <gandalf.corvotempesta at gmail.com
>> <mailto:gandalf.corvotempesta at gmail.com>> wrote:
>> > Let's assume a 1PB storage full of VMs images with each
>> brick over ZFS,
>> > replica 3, sharding enabled
>> >
>> > How do you backup/restore that amount of data?
>> >
>> > Backing up daily is impossible, you'll never finish the
>> backup that the
>> > following one is starting (in other words, you need more
>> than 24 hours)
>> >
>> > Restoring is even worse. You need more than 24 hours with
>> the whole cluster
>> > down
>> >
>> > You can't rely on ZFS snapshot due to sharding (the
>> snapshot took from one
>> > node is useless without all other node related at the same
>> shard) and you
>> > still have the same restore speed
>> >
>> > How do you backup this?
>> >
>> > Even georep isn't enough, if you have to restore the whole
>> storage in case
>> > of disaster
>> >
>> > _______________________________________________
>> > Gluster-users mailing list
>> > Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>> > http://lists.gluster.org/mailman/listinfo/gluster-users
>> <http://lists.gluster.org/mailman/listinfo/gluster-users>
>>
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>> http://lists.gluster.org/mailman/listinfo/gluster-users
>> <http://lists.gluster.org/mailman/listinfo/gluster-users>
> _______________________________________________ Gluster-users
> mailing list Gluster-users at gluster.org
> <mailto:Gluster-users at gluster.org>
> http://lists.gluster.org/mailman/listinfo/gluster-users
> <http://lists.gluster.org/mailman/listinfo/gluster-users>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170323/e995f055/attachment.html>
More information about the Gluster-users
mailing list