[Gluster-users] Geo-replication status Faulty

Gilberto Nunes gilberto.nunes32 at gmail.com
Mon Oct 26 20:23:35 UTC 2020


Hi there...

I'd created a 2 gluster vol and another 1 gluster server acting as a backup
server, using geo-replication.
So in gluster01 I'd issued the command:

gluster peer probe gluster02;gluster peer probe gluster03
gluster vol create DATA replica 2 gluster01:/DATA/master01-data
gluster02:/DATA/master01-data/

Then in gluster03 server:

gluster vol create DATA-SLAVE gluster03:/DATA/slave-data/

I'd setted the ssh powerless session between this 3 servers.

Then I'd used this script

https://github.com/gilbertoferreira/georepsetup

like this

georepsetup
/usr/local/lib/python2.7/dist-packages/paramiko-2.7.2-py2.7.egg/paramiko/transport.py:33:
CryptographyDeprecationWarning: Python 2 is no longer supp
orted by the Python core team. Support for it is now deprecated in
cryptography, and will be removed in a future release.
 from cryptography.hazmat.backends import default_backend
usage: georepsetup [-h] [--force] [--no-color] MASTERVOL SLAVE SLAVEVOL
georepsetup: error: too few arguments
gluster01:~# georepsetup DATA gluster03 DATA-SLAVE
/usr/local/lib/python2.7/dist-packages/paramiko-2.7.2-py2.7.egg/paramiko/transport.py:33:
CryptographyDeprecationWarning: Python 2 is no longer supp
orted by the Python core team. Support for it is now deprecated in
cryptography, and will be removed in a future release.
 from cryptography.hazmat.backends import default_backend
Geo-replication session will be established between DATA and
gluster03::DATA-SLAVE
Root password of gluster03 is required to complete the setup. NOTE:
Password will not be stored.

root at gluster03's password:
[    OK] gluster03 is Reachable(Port 22)
[    OK] SSH Connection established root at gluster03
[    OK] Master Volume and Slave Volume are compatible (Version: 8.2)
[    OK] Common secret pub file present at
/var/lib/glusterd/geo-replication/common_secret.pem.pub
[    OK] common_secret.pem.pub file copied to gluster03
[    OK] Master SSH Keys copied to all Up Slave nodes
[    OK] Updated Master SSH Keys to all Up Slave nodes authorized_keys file
[    OK] Geo-replication Session Established

Then I reboot the 3 servers...
After a while everything works ok, but after a few minutes, I get Faulty
status in gluster01....

There's the log


[2020-10-26 20:16:41.362584] I
[gsyncdstatus(monitor):248:set_worker_status] GeorepStatus: Worker Status
Change [{status=Initializing...}]
[2020-10-26 20:16:41.362937] I [monitor(monitor):160:monitor] Monitor:
starting gsyncd worker [{brick=/DATA/master01-data},
{slave_node=gluster03}]
[2020-10-26 20:16:41.508884] I [resource(worker
/DATA/master01-data):1387:connect_remote] SSH: Initializing SSH connection
between master and slave.
..
[2020-10-26 20:16:42.996678] I [resource(worker
/DATA/master01-data):1436:connect_remote] SSH: SSH connection between
master and slave established.
[{duration=1.4873}]
[2020-10-26 20:16:42.997121] I [resource(worker
/DATA/master01-data):1116:connect] GLUSTER: Mounting gluster volume
locally...
[2020-10-26 20:16:44.170661] E [syncdutils(worker
/DATA/master01-data):110:gf_mount_ready] <top>: failed to get the xattr
value
[2020-10-26 20:16:44.171281] I [resource(worker
/DATA/master01-data):1139:connect] GLUSTER: Mounted gluster volume
[{duration=1.1739}]
[2020-10-26 20:16:44.171772] I [subcmds(worker
/DATA/master01-data):84:subcmd_worker] <top>: Worker spawn successful.
Acknowledging back to monitor
[2020-10-26 20:16:46.200603] I [master(worker
/DATA/master01-data):1645:register] _GMaster: Working dir
[{path=/var/lib/misc/gluster/gsyncd/DATA_glu
ster03_DATA-SLAVE/DATA-master01-data}]
[2020-10-26 20:16:46.201798] I [resource(worker
/DATA/master01-data):1292:service_loop] GLUSTER: Register time
[{time=1603743406}]
[2020-10-26 20:16:46.226415] I [gsyncdstatus(worker
/DATA/master01-data):281:set_active] GeorepStatus: Worker Status Change
[{status=Active}]
[2020-10-26 20:16:46.395112] I [gsyncdstatus(worker
/DATA/master01-data):253:set_worker_crawl_status] GeorepStatus: Crawl
Status Change [{status=His
tory Crawl}]
[2020-10-26 20:16:46.396491] I [master(worker
/DATA/master01-data):1559:crawl] _GMaster: starting history crawl
[{turns=1}, {stime=(1603742506, 0)},
{etime=1603743406}, {entry_stime=(1603743226, 0)}]
[2020-10-26 20:16:46.399292] E [resource(worker
/DATA/master01-data):1312:service_loop] GLUSTER: Changelog History Crawl
failed [{error=[Errno 0] Su
cesso}]
[2020-10-26 20:16:47.177205] I [monitor(monitor):228:monitor] Monitor:
worker died in startup phase [{brick=/DATA/master01-data}]
[2020-10-26 20:16:47.184525] I
[gsyncdstatus(monitor):248:set_worker_status] GeorepStatus: Worker Status
Change [{status=Faulty}]


Any advice will be welcome.

Thanks

---
Gilberto Nunes Ferreira
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20201026/a7c1927c/attachment-0001.html>


More information about the Gluster-users mailing list