[Bugs] [Bug 1532868] New: gluster upgrade causes vm disk errors

bugzilla at redhat.com bugzilla at redhat.com
Tue Jan 9 22:56:11 UTC 2018


https://bugzilla.redhat.com/show_bug.cgi?id=1532868

            Bug ID: 1532868
           Summary: gluster upgrade causes vm disk errors
           Product: GlusterFS
           Version: 3.12
         Component: fuse
          Severity: medium
          Assignee: bugs at gluster.org
          Reporter: bugs at josemaas.net
                CC: bugs at gluster.org



Description of problem:
We have a gluster replica x3 that is used for vm storage. After upgrading from
3.10 to 3.12 ovirt started logging "VM <vm name> has been paused due to unknown
storage error." and pausing the effected vms. In order to upgrade I used the
gluster upgrade guide. 

After looking into it some more I found the following lines in our mount log.
To me this seems that there is an issue with sharding. Note copying the vm
image and using the new one seems to fix the problem

[2018-01-09 15:45:47.098380] E [shard.c:426:shard_modify_size_and_block_count]
(-->/usr/lib64/glusterfs/3.12.3/xlator/cluster/distribute.so(+0x6abed)
[0x7f7241b3dbed]
-->/usr/lib64/glusterfs/3.12.3/xlator/features/shard.so(+0xbafe)
[0x7f72418bbafe]
-->/usr/lib64/glusterfs/3.12.3/xlator/features/shard.so(+0xb35b)
[0x7f72418bb35b] ) 0-tortoise-shard: Failed to get
trusted.glusterfs.shard.file-size for 6493ab88-f4a8-4696-a52e-6425a595fc80
[2018-01-09 15:45:47.098419] W [fuse-bridge.c:874:fuse_attr_cbk]
0-glusterfs-fuse: 4369015: STAT()
/aee19709-5859-4e48-b761-c4f8a140ea61/images/855cfe92-22ad-4f40-94f8-2547fa5e0f8e/e29713bf-557a-4943-a7c5-c29edc141c01
=> -1 (Invalid argument)

Version-Release number of selected component (if applicable):
centos-release-gluster310.noarch   1.0-1.el7.centos           @extras           
centos-release-gluster312.noarch   1.0-1.el7.centos           @extras           
glusterfs.x86_64                   3.12.3-1.el7              
@centos-gluster312
glusterfs-api.x86_64               3.12.3-1.el7              
@centos-gluster312
glusterfs-cli.x86_64               3.12.3-1.el7              
@centos-gluster312
glusterfs-client-xlators.x86_64    3.12.3-1.el7              
@centos-gluster312
glusterfs-fuse.x86_64              3.12.3-1.el7              
@centos-gluster312
glusterfs-gnfs.x86_64              3.12.3-1.el7              
@centos-gluster312
glusterfs-libs.x86_64              3.12.3-1.el7              
@centos-gluster312
glusterfs-server.x86_64            3.12.3-1.el7              
@centos-gluster312
libntirpc.x86_64                   1.5.3-1.el7               
@centos-gluster312
nfs-ganesha.x86_64                 2.5.3-1.el6               
@centos-gluster312
nfs-ganesha-gluster.x86_64         2.5.3-1.el6               
@centos-gluster312
userspace-rcu.x86_64               0.10.0-3.el7              
@centos-gluster312

previous version:
Dec 06 03:50:42 Updated: glusterfs-libs.x86_64 3.10.8-1.el7
Dec 06 03:50:42 Updated: glusterfs.x86_64 3.10.8-1.el7
Dec 06 03:50:42 Updated: glusterfs-client-xlators.x86_64 3.10.8-1.el7
Dec 06 03:50:42 Updated: glusterfs-api.x86_64 3.10.8-1.el7
Dec 06 03:50:42 Updated: glusterfs-fuse.x86_64 3.10.8-1.el7
Dec 06 03:50:42 Updated: glusterfs-cli.x86_64 3.10.8-1.el7
Dec 06 03:50:45 Updated: glusterfs-server.x86_64 3.10.8-1.el7
Dec 06 03:50:45 Updated: glusterfs-ganesha.x86_64 3.10.8-1.el

How reproducible:
Happened on multiple vm images. I do not have another cluster to try it on

Steps to Reproduce:
1.
2.
3.

Actual results:


Expected results:


Additional info:
Volume Name: tortoise
Type: Replicate
Volume ID: d4c00537-f1e8-4c43-b21d-90c9a6c5dee9
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: stor1.example.com:/tort/brick
Brick2: stor2.example.com:/tort/brick
Brick3: stor3.example.com:/tort/brick (arbiter)
Options Reconfigured:
user.cifs: off
features.shard: on
cluster.shd-wait-qlength: 10000
cluster.shd-max-threads: 8
cluster.locking-scheme: granular
cluster.data-self-heal-algorithm: full
cluster.server-quorum-type: server
cluster.eager-lock: enable
performance.low-prio-threads: 32
performance.stat-prefetch: off
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
auth.allow: *
network.ping-timeout: 10
cluster.quorum-type: auto
server.allow-insecure: on
storage.owner-gid: 36
storage.owner-uid: 36
performance.strict-o-direct: enable
network.remote-dio: enable
transport.address-family: inet
nfs.disable: on
features.shard-block-size: 256MB
cluster.enable-shared-storage: disable

-- 
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.


More information about the Bugs mailing list