[Gluster-users] Join multiple Gluster Cluster

Holger Rojahn iceholger at gmail.com
Tue Aug 24 06:59:15 UTC 2021


Ok, thanks.
I setup the Cluster from Scratch with the neccesary Downtime in the eve ..

Greets
Holger

Strahil Nikolov <hunter86_bg at yahoo.com> schrieb am Di., 24. Aug. 2021,
08:55:

> I guess you can manipulate the peer list and merge them into a single TSP,
> but I don't know if that would be 'supported'.
>
> Give a try on a setup of VMs and just stop all glusterd and start them
> after merging the files.
>
> Best Regards,
> Strahil Nikolov
>
> Sent from Yahoo Mail on Android
> <https://go.onelink.me/107872968?pid=InProduct&c=Global_Internal_YGrowth_AndroidEmailSig__AndroidUsers&af_wl=ym&af_sub1=Internal&af_sub2=Global_YGrowth&af_sub3=EmailSignature>
>
> On Mon, Aug 23, 2021 at 13:41, Holger Rojahn
> <iceholger at gmail.com> wrote:
> Hi,
> thats not nice because i have a big downtime then..
> I hope there is a unjoin/join cluster ...
>
> Tnx for help. Must c how i handle this situation...
>
> Strahil Nikolov <hunter86_bg at yahoo.com> schrieb am Mo., 23. Aug. 2021,
> 12:33:
>
> I guess you have to canibilize the clusters, pick the one that has most
> data and start shrinking the one with most empty space.
> Once you can't shrink more, you will need downtime to set the 'donor'
> volume read only and start copying the data to the 'recieving' volume.
> Once data is moved, move the remaining bricks to the 'recieving' volume
> and then rebalance.
>
> Repeat again for the other 'donor' volume.
>
> Best Regards,
> Strahil Nikolov
>
> Sent from Yahoo Mail on Android
> <https://go.onelink.me/107872968?pid=InProduct&c=Global_Internal_YGrowth_AndroidEmailSig__AndroidUsers&af_wl=ym&af_sub1=Internal&af_sub2=Global_YGrowth&af_sub3=EmailSignature>
>
> On Fri, Aug 20, 2021 at 15:01, Holger Rojahn
> <iceholger at gmail.com> wrote:
>
> Hi,
>
> I Configured 3 Clusters (For several Shared Homes between different
> maschines) .. So far no Problem.
>
> Now it need a Volume that going across this maschines, but the nodes are
> bound to there own clusters so peer probe fails …
>
>
>
> How i can build one big cluster with all nodes but without dataloss (And
> best without Downtime J)
>
>
>
> Hope there is some Pro that can help J
>
>
>
> Greets from Germany
>
>
>
> Holger
>
>
> ________
>
>
>
> Community Meeting Calendar:
>
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://meet.google.com/cpu-eiue-hvk
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20210824/0dc4ef0a/attachment.html>


More information about the Gluster-users mailing list