[Gluster-users] Geo Replication OSError: [Errno 107] Transport endpoint is not connected

Michael Roth michael.roth at tuwien.ac.at
Thu Oct 25 12:10:40 UTC 2018


I've a big Problem.

If I start geo-replication everything seems fine, but after replicating 
2.5TB I got errors, it's starting over an over again with the same errors.

I've two nodes with a replicated volume and a third arbiter node.
The destination node is a single node.
The firewall between all nodes ist open.

Master Log

[2018-10-25 07:08:59.619699] D 
[master(/gluster/owncloud/brick2):1665:Xcrawl] _GMaster: entering 
./data/fa/files/backup/research/projects/2011-Regularity/2012-03-Gain-of-Regularity-linearWFP
[2018-10-25 07:08:59.619874] E 
[syncdutils(/gluster/owncloud/brick2):325:log_raise_exception] <top>: 
glusterfs session went down        error=ENOTCONN
[2018-10-25 07:08:59.620109] E 
[syncdutils(/gluster/owncloud/brick2):331:log_raise_exception] <top>: 
FULL EXCEPTION TRACE:
Traceback (most recent call last):
   File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 210, 
in main
     main_i()
   File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 801, 
in main_i
     local.service_loop(*[r for r in [remote] if r])
   File "/usr/libexec/glusterfs/python/syncdaemon/resource.py", line 
1679, in service_loop
     g1.crawlwrap(oneshot=True, register_time=register_time)
   File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 597, 
in crawlwrap
     self.crawl()
   File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1555, 
in crawl
     self.process([item[1]], 0)
   File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1204, 
in process
     self.process_change(change, done, retry)
   File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1143, 
in process_change
     st = lstat(go[0])
   File "/usr/libexec/glusterfs/python/syncdaemon/syncdutils.py", line 
553, in lstat
     return errno_wrap(os.lstat, [e], [ENOENT], [ESTALE, EBUSY])
   File "/usr/libexec/glusterfs/python/syncdaemon/syncdutils.py", line 
535, in errno_wrap
     return call(*arg)
OSError: [Errno 107] Transport endpoint is not connected: 
'.gfid/5c143d64-165f-44b1-98ed-71e491376a76'
[2018-10-25 07:08:59.627846] D 
[master(/gluster/owncloud/brick2):1665:Xcrawl] _GMaster: entering 
./data/fa/files/backup/research/projects/2011-Regularity/resources
[2018-10-25 07:08:59.632826] D 
[master(/gluster/owncloud/brick2):1665:Xcrawl] _GMaster: entering 
./data/fa/files/backup/research/projects/2011-Regularity/add material
[2018-10-25 07:08:59.633582] D 
[master(/gluster/owncloud/brick2):1665:Xcrawl] _GMaster: entering 
./data/fa/files/backup/research/projects/2011-Regularity/add material/Maple
[2018-10-25 07:08:59.636306] D 
[master(/gluster/owncloud/brick2):1665:Xcrawl] _GMaster: entering 
./data/fa/files/backup/research/projects/2011-Regularity/add material/notes
[2018-10-25 07:08:59.637303] I 
[syncdutils(/gluster/owncloud/brick2):271:finalize] <top>: exiting.
[2018-10-25 07:08:59.640778] I 
[repce(/gluster/owncloud/brick2):92:service_loop] RepceServer: 
terminating on reaching EOF.
[2018-10-25 07:08:59.641222] I 
[syncdutils(/gluster/owncloud/brick2):271:finalize] <top>: exiting.
[2018-10-25 07:09:00.314140] I [monitor(monitor):363:monitor] Monitor: 
worker died in startup phase brick=/gluster/owncloud/brick2
[2018-10-25 07:09:00.315172] I 
[gsyncdstatus(monitor):243:set_worker_status] GeorepStatus: Worker 
Status Change status=Faulty

Slave Log

[2018-10-25 07:08:44.206372] I [resource(slave):1502:connect] GLUSTER: 
Mounting gluster volume locally...
[2018-10-25 07:08:45.229620] I [resource(slave):1515:connect] GLUSTER: 
Mounted gluster volume   duration=1.0229
[2018-10-25 07:08:45.230180] I [resource(slave):1012:service_loop] 
GLUSTER: slave listening
[2018-10-25 07:08:59.641242] I [repce(slave):92:service_loop] 
RepceServer: terminating on reaching EOF.
[2018-10-25 07:08:59.655611] I [syncdutils(slave):271:finalize] <top>: 
exiting.

Volume Info

Volume Name: datacloud
Type: Replicate
Volume ID: 6cc79599-7a5c-4b02-bd86-13020a9d91db
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: 172.17.45.11:/gluster/datacloud/brick2
Brick2: 172.17.45.12:/gluster/datacloud/brick2
Brick3: 172.17.45.13:/gluster/datacloud/brick2 (arbiter)
Options Reconfigured:
cluster.server-quorum-type: server
cluster.shd-max-threads: 32
cluster.self-heal-readdir-size: 64KB
cluster.quorum-type: fixed
transport.address-family: inet
diagnostics.brick-log-level: INFO
changelog.capture-del-path: on
storage.build-pgfid: on
changelog.changelog: on
geo-replication.ignore-pid-check: on
server.statedump-path: /tmp/gluster
cluster.self-heal-window-size: 32
geo-replication.indexing: on
nfs.trusted-sync: off
diagnostics.dump-fd-stats: off
nfs.disable: on
cluster.self-heal-daemon: enable
cluster.background-self-heal-count: 16
cluster.heal-timeout: 120
cluster.data-self-heal-algorithm: full
cluster.consistent-metadata: on
network.ping-timeout: 20
cluster.granular-entry-heal: enable
cluster.server-quorum-ratio: 51%
cluster.enable-shared-storage: enable

Best regards,
Michael


-- 
Michael Roth  | michael.roth at tuwien.ac.at
IT Solutions - Application Management
Technische Universität Wien - Operngasse 11, 1040 Wien
T +43-1-58801-42091



More information about the Gluster-users mailing list