<div dir="ltr">Not so fast with my solution!<div>After shutting the other node in the head, get FAULTY stat again...</div><div>The only failure I saw in this thing regarding xattr value... </div><div><br></div><div>[2020-10-27 19:20:07.718897] E [syncdutils(worker /DATA/vms):110:gf_mount_ready] <top>: failed to get the xattr value <br></div><div><br></div><div>Don't know if I am looking at the right log: <span style="color:rgb(0,0,0);font-family:monospace">/var/log/glusterfs/geo-replication/VMS_gluster03_VMS-SLAVE/gsyncd.log </span></div><div><br></div><div><span style="font-family:monospace"><span style="color:rgb(0,0,0)">[2020-10-27 19:20:03.867749] I [gsyncdstatus(monitor):248:set_worker_status] GeorepStatus: Worker Status Change [{status=Initializing...}]
</span><br>[2020-10-27 19:20:03.868206] I [monitor(monitor):160:monitor] Monitor: starting gsyncd worker [{brick=/DATA/vms}, {slave_node=gluster03}]
<br>[2020-10-27 19:20:04.397444] I [resource(worker /DATA/vms):1387:connect_remote] SSH: Initializing SSH connection between master and slave...
<br>[2020-10-27 19:20:06.337282] I [resource(worker /DATA/vms):1436:connect_remote] SSH: SSH connection between master and slave established. [{duration=1.9385}]
<br>[2020-10-27 19:20:06.337854] I [resource(worker /DATA/vms):1116:connect] GLUSTER: Mounting gluster volume locally...
<br>[2020-10-27 19:20:07.718897] E [syncdutils(worker /DATA/vms):110:gf_mount_ready] <top>: failed to get the xattr value
<br>[2020-10-27 19:20:07.720089] I [resource(worker /DATA/vms):1139:connect] GLUSTER: Mounted gluster volume [{duration=1.3815}]
<br>[2020-10-27 19:20:07.720644] I [subcmds(worker /DATA/vms):84:subcmd_worker] <top>: Worker spawn successful. Acknowledging back to monitor
<br>[2020-10-27 19:20:09.757677] I [master(worker /DATA/vms):1645:register] _GMaster: Working dir [{path=/var/lib/misc/gluster/gsyncd/VMS_gluster03_VMS-SLAVE/DATA-vms}]
<br>[2020-10-27 19:20:09.758440] I [resource(worker /DATA/vms):1292:service_loop] GLUSTER: Register time [{time=1603826409}]
<br>[2020-10-27 19:20:09.925364] I [gsyncdstatus(worker /DATA/vms):281:set_active] GeorepStatus: Worker Status Change [{status=Active}]
<br>[2020-10-27 19:20:10.407319] I [gsyncdstatus(worker /DATA/vms):253:set_worker_crawl_status] GeorepStatus: Crawl Status Change [{status=History Crawl}]
<br>[2020-10-27 19:20:10.420385] I [master(worker /DATA/vms):1559:crawl] _GMaster: starting history crawl [{turns=1}, {stime=(1603821702, 0)}, {etime=1603826410}, {entry_<br>stime=(1603822857, 0)}]
<br>[2020-10-27 19:20:10.424286] E [resource(worker /DATA/vms):1312:service_loop] GLUSTER: Changelog History Crawl failed [{error=[Errno 0] Success}]
<br>[2020-10-27 19:20:10.731317] I [monitor(monitor):228:monitor] Monitor: worker died in startup phase [{brick=/DATA/vms}]
<br>[2020-10-27 19:20:10.740046] I [gsyncdstatus(monitor):248:set_worker_status] GeorepStatus: Worker Status Change [{status=Faulty}]<br>
<br></span><div><br clear="all"><div><div dir="ltr" class="gmail_signature" data-smartmail="gmail_signature"><div dir="ltr"><div dir="ltr"><div dir="ltr"><div dir="ltr"><div dir="ltr"><div>---</div><div><div><div>Gilberto Nunes Ferreira</div></div><div><span style="font-size:12.8px">(47) 99676-7530 - Whatsapp / Telegram</span><br></div><div><p style="font-size:12.8px;margin:0px"></p><p style="font-size:12.8px;margin:0px"><br></p><p style="font-size:12.8px;margin:0px"><br></p></div></div><div><br></div></div></div></div></div></div></div></div><br></div></div></div><br><div class="gmail_quote"><div dir="ltr" class="gmail_attr">Em ter., 27 de out. de 2020 às 16:06, Strahil Nikolov <<a href="mailto:hunter86_bg@yahoo.com">hunter86_bg@yahoo.com</a>> escreveu:<br></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex">It could be a "simple" bug - software has bugs and regressions.<br>
<br>
I would recommend you to ping the debian mailing list - at least it won't hurt.<br>
<br>
Best Regards,<br>
Strahil Nikolov<br>
<br>
<br>
<br>
<br>
<br>
<br>
В вторник, 27 октомври 2020 г., 20:10:39 Гринуич+2, Gilberto Nunes <<a href="mailto:gilberto.nunes32@gmail.com" target="_blank">gilberto.nunes32@gmail.com</a>> написа: <br>
<br>
<br>
<br>
<br>
<br>
[SOLVED]<br>
<br>
Well... It seems to me that pure Debian Linux 10 has some problem with XFS, which is the FS that I used.<br>
It's not accept attr2 mount options.<br>
<br>
Interestingly enough, I have now used Proxmox 6.x, which is Debian based, I am now able to use the attr2 mount point option.<br>
Then the Faulty status of geo-rep has gone.<br>
Perhaps Proxmox staff has compiled xfs from scratch... Don't know....<br>
But now I am happy ' cause the main reason to use geo-rep to me is to use it over Proxmox....<br>
<br>
cat /etc/fstab # <file system> <mount point> <type> <options> <dump> <pass> /dev/pve/root / xfs defaults 0 1 /dev/pve/swap none swap sw 0 0 /dev/sdb1 /DATA xfs attr2 0 0 gluster01:VMS /vms glusterfs defaults,_netdev,x-systemd.automount,backupvolfile-server=gluster02 0 0 proc /proc proc defaults 0 0<br>
<br>
<br>
---<br>
Gilberto Nunes Ferreira<br>
<br>
<br>
<br>
<br>
<br>
<br>
<br>
Em ter., 27 de out. de 2020 às 09:39, Gilberto Nunes <<a href="mailto:gilberto.nunes32@gmail.com" target="_blank">gilberto.nunes32@gmail.com</a>> escreveu:<br>
>>> IIUC you're begging for split-brain ...<br>
> Not at all!<br>
> I have used this configuration and there isn't any split brain at all!<br>
> But if I do not use it, then I get a split brain.<br>
> Regarding count 2 I will see it!<br>
> Thanks<br>
> <br>
> ---<br>
> Gilberto Nunes Ferreira<br>
> <br>
> <br>
> <br>
> <br>
> <br>
> Em ter., 27 de out. de 2020 às 09:37, Diego Zuccato <<a href="mailto:diego.zuccato@unibo.it" target="_blank">diego.zuccato@unibo.it</a>> escreveu:<br>
>> Il 27/10/20 13:15, Gilberto Nunes ha scritto:<br>
>>> I have applied this parameters to the 2-node gluster:<br>
>>> gluster vol set VMS cluster.heal-timeout 10<br>
>>> gluster volume heal VMS enable<br>
>>> gluster vol set VMS cluster.quorum-reads false<br>
>>> gluster vol set VMS cluster.quorum-count 1<br>
>> Urgh!<br>
>> IIUC you're begging for split-brain ...<br>
>> I think you should leave quorum-count=2 for safe writes. If a node is<br>
>> down, obviously the volume becomes readonly. But if you planned the<br>
>> downtime you can reduce quorum-count just before shutting it down.<br>
>> You'll have to bring it back to 2 before re-enabling the downed server,<br>
>> then wait for heal to complete before being able to down the second server.<br>
>> <br>
>>> Then I mount the gluster volume putting this line in the fstab file:<br>
>>> In gluster01<br>
>>> gluster01:VMS /vms glusterfs<br>
>>> defaults,_netdev,x-systemd.automount,backupvolfile-server=gluster02 0 0<br>
>>> In gluster02<br>
>>> gluster02:VMS /vms glusterfs<br>
>>> defaults,_netdev,x-systemd.automount,backupvolfile-server=gluster01 0 0<br>
>> Isn't it preferrable to use the 'hostlist' syntax?<br>
>> gluster01,gluster02:VMS /vms glusterfs defaults,_netdev 0 0<br>
>> A / at the beginning is optional, but can be useful if you're trying to<br>
>> use the diamond freespace collector (w/o the initial slash, it ignores<br>
>> glusterfs mountpoints).<br>
>> <br>
>> -- <br>
>> Diego Zuccato<br>
>> DIFA - Dip. di Fisica e Astronomia<br>
>> Servizi Informatici<br>
>> Alma Mater Studiorum - Università di Bologna<br>
>> V.le Berti-Pichat 6/2 - 40127 Bologna - Italy<br>
>> tel.: +39 051 20 95786<br>
>> <br>
> <br>
________<br>
<br>
<br>
<br>
Community Meeting Calendar:<br>
<br>
Schedule -<br>
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC<br>
Bridge: <a href="https://bluejeans.com/441850968" rel="noreferrer" target="_blank">https://bluejeans.com/441850968</a><br>
<br>
Gluster-users mailing list<br>
<a href="mailto:Gluster-users@gluster.org" target="_blank">Gluster-users@gluster.org</a><br>
<a href="https://lists.gluster.org/mailman/listinfo/gluster-users" rel="noreferrer" target="_blank">https://lists.gluster.org/mailman/listinfo/gluster-users</a><br>
</blockquote></div>