[Gluster-users] 答复: geo-replication status partial faulty

vyyy杨雨阳 yuyangyang at Ctrip.com
Thu May 19 02:23:35 UTC 2016


Hello,

I have tried to config a geo-replication volume , all the master nodes configuration are the same, When I start this volume, the status shows partial faulty as following:

gluster volume geo-replication filews glusterfs01.sh3.ctripcorp.com::filews_slave status

MASTER NODE      MASTER VOL    MASTER BRICK          SLAVE                                          STATUS     CHECKPOINT STATUS    CRAWL STATUS
-------------------------------------------------------------------------------------------------------------------------------------------------
SVR8048HW2285    filews        /export/sdb/filews    glusterfs01.sh3.ctripcorp.com::filews_slave    faulty     N/A                  N/A
SVR8050HW2285    filews        /export/sdb/filews    glusterfs03.sh3.ctripcorp.com::filews_slave    Passive    N/A                  N/A
SVR8047HW2285    filews        /export/sdb/filews    glusterfs01.sh3.ctripcorp.com::filews_slave    Active     N/A                  Hybrid Crawl
SVR8049HW2285    filews        /export/sdb/filews    glusterfs05.sh3.ctripcorp.com::filews_slave    Active     N/A                  Hybrid Crawl
SH02SVR5951      filews        /export/sdb/brick1    glusterfs06.sh3.ctripcorp.com::filews_slave    Passive    N/A                  N/A
SH02SVR5953      filews        /export/sdb/brick1    glusterfs01.sh3.ctripcorp.com::filews_slave    faulty     N/A                  N/A
SVR6995HW2285    filews        /export/sdb/filews    glusterfs01.sh3.ctripcorp.com::filews_slave    faulty     N/A                  N/A
SH02SVR5954      filews        /export/sdb/brick1    glusterfs01.sh3.ctripcorp.com::filews_slave    faulty     N/A                  N/A
SVR6994HW2285    filews        /export/sdb/filews    glusterfs02.sh3.ctripcorp.com::filews_slave    Passive    N/A                  N/A
SVR6993HW2285    filews        /export/sdb/filews    glusterfs01.sh3.ctripcorp.com::filews_slave    faulty     N/A                  N/A
SH02SVR5952      filews        /export/sdb/brick1    glusterfs01.sh3.ctripcorp.com::filews_slave    faulty     N/A                  N/A
SVR6996HW2285    filews        /export/sdb/filews    glusterfs04.sh3.ctripcorp.com::filews_slave    Passive    N/A                  N/A

On the faulty node, log file /var/log/glusterfs/geo-replication/filews shows worker(/export/sdb/filews) died before establishing connection

[2016-05-18 16:55:46.402622] I [monitor(monitor):215:monitor] Monitor: ------------------------------------------------------------
[2016-05-18 16:55:46.402930] I [monitor(monitor):216:monitor] Monitor: starting gsyncd worker
[2016-05-18 16:55:46.517460] I [changelogagent(agent):72:__init__] ChangelogAgent: Agent listining...
[2016-05-18 16:55:46.518066] I [repce(agent):92:service_loop] RepceServer: terminating on reaching EOF.
[2016-05-18 16:55:46.518279] I [syncdutils(agent):214:finalize] <top>: exiting.
[2016-05-18 16:55:46.518194] I [monitor(monitor):267:monitor] Monitor: worker(/export/sdb/filews) died before establishing connection
[2016-05-18 16:55:56.697036] I [monitor(monitor):215:monitor] Monitor: ------------------------------------------------------------

Any advice and suggestions will be greatly appreciated.





Best Regards
杨雨阳 Yuyang Yang


-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160519/0685b624/attachment.html>


More information about the Gluster-users mailing list