[Bugs] [Bug 1287107] New: [georep][sharding] Unable to resume geo-rep session after previous errors

bugzilla at redhat.com bugzilla at redhat.com
Tue Dec 1 14:07:53 UTC 2015


https://bugzilla.redhat.com/show_bug.cgi?id=1287107

            Bug ID: 1287107
           Summary: [georep][sharding] Unable to resume geo-rep session
                    after previous errors
           Product: GlusterFS
           Version: mainline
         Component: geo-replication
          Severity: high
          Assignee: bugs at gluster.org
          Reporter: sabose at redhat.com
                CC: bugs at gluster.org, gluster-bugs at redhat.com



Created attachment 1100927
  --> https://bugzilla.redhat.com/attachment.cgi?id=1100927&action=edit
georep-master-log

Description of problem:

Geo-replication session that was running on a sharded volume resulted in
failures due to lack of space at slave volume.

Geo-rep session was stopped, slave volume disk space was extended (using
lvextend on underlying brick mount point), and geo-replication session was
resumed.

But looking at geo-rep status detail shows failures and it seems that files are
not being synced.

Status detail and volume info in Additional info

Version-Release number of selected component (if applicable):


How reproducible:


Steps to Reproduce:
1. Setup geo-replication session between master and slave (slave volume has
lesser capacity than master)
2. Start geo-rep 
3. Create data in master volume more than slave capacity
4. geo-rep status will report failures (seen in status detail as failure count)
5. stop geo-rep session
6. Increase capacity of slave volume (in my case, I extended brick lv by adding
additional vdisk to VM hosting slave)
7. start geo-rep session again

Actual results:


Expected results:


Additional info:

# gluster vol geo-replication data1 10.70.40.112::hc-slavevol  status detail

MASTER NODE                              MASTER VOL    MASTER BRICK      SLAVE
USER    SLAVE                        SLAVE NODE      STATUS     CRAWL STATUS   
 LAST_SYNCED            ENTRY    DATA    META    FAILURES    CHECKPOINT TIME   
    CHECKPOINT COMPLETED    CHECKPOINT COMPLETION TIME   
---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
rhsdev-docker1.lab.eng.blr.redhat.com    data1         /rhgs/data1/b1    root  
       10.70.40.112::hc-slavevol    10.70.40.112    Passive    N/A             
N/A                    N/A      N/A     N/A     N/A         N/A                
   N/A                     N/A                          
rhsdev9.lab.eng.blr.redhat.com           data1         /rhgs/data1/b1    root  
       10.70.40.112::hc-slavevol    10.70.40.112    Active     History Crawl   
2015-11-26 15:18:56    0        7226    0       107         2015-12-01 18:09:11
   No                      N/A                          
rhsdev-docker2.lab.eng.blr.redhat.com    data1         /rhgs/data1/b1    root  
       10.70.40.112::hc-slavevol    10.70.40.112    Passive    N/A             
N/A                    N/A      N/A     N/A     N/A         N/A                
   N/A

Master volume:
Volume Name: data1
Type: Replicate
Volume ID: 55bd10b0-f05a-446b-a481-6590cc400263
Status: Started
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: rhsdev9.lab.eng.blr.redhat.com:/rhgs/data1/b1
Brick2: rhsdev-docker2.lab.eng.blr.redhat.com:/rhgs/data1/b1
Brick3: rhsdev-docker1.lab.eng.blr.redhat.com:/rhgs/data1/b1
Options Reconfigured:
performance.readdir-ahead: on
performance.low-prio-threads: 32
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
cluster.eager-lock: enable
network.remote-dio: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
storage.owner-uid: 36
storage.owner-gid: 36
features.shard: on
features.shard-block-size: 512MB
geo-replication.indexing: on
geo-replication.ignore-pid-check: on
changelog.changelog: on

Slave volume:
Volume Name: hc-slavevol
Type: Distribute
Volume ID: 56a3d4d9-51bc-4daf-9257-bd13e10511ae
Status: Started
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: 10.70.40.112:/brick/hc1
Options Reconfigured:
storage.owner-gid: 36
storage.owner-uid: 36
cluster.server-quorum-type: server
cluster.quorum-type: auto
network.remote-dio: enable
cluster.eager-lock: enable
performance.stat-prefetch: off
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
performance.readdir-ahead: on
features.shard: on
features.shard-block-size: 512MB

-- 
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.


More information about the Bugs mailing list