[Bugs] [Bug 1232430] [SNAPSHOT] : Snapshot delete fails with error - Snap might not be in an usable state

bugzilla at redhat.com bugzilla at redhat.com
Fri Jul 17 07:48:02 UTC 2015


Richard Neuboeck <hawk at tbi.univie.ac.at> changed:

           What    |Removed                     |Added
                 CC|                            |hawk at tbi.univie.ac.at

--- Comment #3 from Richard Neuboeck <hawk at tbi.univie.ac.at> ---
I can confirm that the 'Snap might not be in an usable state' problem exists in
glusterfs 3.7.2

Version: glusterfs-3.7.2-3.el7.x86_64 (gluster repo)
OS: CentOS 7.1 64bit

Steps to recreate:

# gluster snapshot create snap1 plexus description 'test snapshot'
snapshot create: success: Snap snap1_GMT-2015.07.16-11.16.03 created

# gluster snapshot list

# gluster snapshot info
Snapshot                  : snap1_GMT-2015.07.16-11.16.03
Snap UUID                 : 6ddce064-2bd0-4770-9995-583147e1a35c
Description               : test snapshot
Created                   : 2015-07-16 11:16:03
Snap Volumes:

    Snap Volume Name          : e905ba76967f43efa0220c2283c87057
    Origin Volume name        : plexus
    Snaps taken for plexus      : 1
    Snaps available for plexus  : 255
    Status                    : Stopped

# gluster snapshot delete snap1_GMT-2015.07.16-11.16.03
Deleting snap will erase all the information about the snap. Do you still want
to continue? (y/n) y
snapshot delete: failed: Snapshot snap1_GMT-2015.07.16-11.16.03 might not be in
an usable state.
Snapshot command failed

# gluster snapshot delete all
System contains 1 snapshot(s).
Do you still want to continue and delete them?  (y/n) y
snapshot delete: failed: Snapshot snap1_GMT-2015.07.16-11.16.03 might not be in
an usable state.
Snapshot command failed

Setup on the machines I've tested this:

- CentOS 7.1 minimal installation
- Thinly provisioned as follows:
# lvs --all
  LV                                 VG                Attr       LSize  Pool  
  Origin   Data%  Meta%  Move Log Cpy%Sync Convert
  e905ba76967f43efa0220c2283c87057_0 storage_vg        Vwi-aotz-- 45.00t
thinpool thindata 0.04                                   
  [lvol0_pmspare]                    storage_vg        ewi------- 10.00g        
  thindata                           storage_vg        Vwi-aotz-- 45.00t
thinpool          0.05                                   
  thinpool                           storage_vg        twi-aotz-- 49.95t       
           0.05   0.43                            
  [thinpool_tdata]                   storage_vg        Twi-ao---- 49.95t        
  [thinpool_tmeta]                   storage_vg        ewi-ao---- 10.00g

- Gluster setup for this test consists of two machines, one brick each. The
brick is a (hardware) RAID 5 volume. Since I've got a lot of NFS related error
messages and I didn't use NFS in this case 'nfs.disable' is on.

# gluster volume info

Volume Name: plexus
Type: Replicate
Volume ID: 105559c1-c6d9-4557-8488-2197ad86d92d
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Brick1: sphere-one:/srv/gluster/brick
Brick2: sphere-two:/srv/gluster/brick
Options Reconfigured:
features.barrier: disable
nfs.disable: on
performance.readdir-ahead: on

Attached are the logs from both nodes from /var/log/glusterfs.

You are receiving this mail because:
You are on the CC list for the bug.
Unsubscribe from this bug https://bugzilla.redhat.com/token.cgi?t=PVOREHgsAb&a=cc_unsubscribe

More information about the Bugs mailing list