[Gluster-users] Volume info out of sync

Brian Candler B.Candler at pobox.com
Mon Jun 11 08:50:50 UTC 2012


(glusterfs 3.3.0, ubuntu 12.04)

I have two nodes in my test setup: dev-storage1 and dev-storage2.

While dev-storage2 was powered down, I added a new volume "single1" on
dev-storage1 (using only bricks on dev-storage1).

However, when I brought dev-storage2 back online, "gluster volume info" on
that node doesn't show the newly-created volume.

So my question is, how do I go about getting the node info in sync?

I found "gluster volume sync", but this by itself doesn't seem to be enough,
regardless of which node I run it on:

    root at dev-storage1:~# gluster volume sync all
    please delete all the volumes before full sync
    oot at dev-storage1:~# gluster volume sync single1
    please delete all the volumes before full sync

    root at dev-storage2:~# gluster volume sync single1
    please delete all the volumes before full sync
    root at dev-storage2:~# gluster volume sync all
    please delete all the volumes before full sync

Hmm, I don't want to delete all volumes in case the deletions replicate to
the other nodes too.  What am I supposed to do now?  Perhaps "peer detach"
first, then delete everything on dev-storage2, then sync?

The volume info is pasted below.

(Side question: why is there even an option to sync a single volume rather
than "all", if you have to delete all the voluems first anyway?)

Thanks,

Brian.

---xxx---

root at dev-storage1:~# gluster volume info
 
Volume Name: fast
Type: Distribute
Volume ID: 864fd12d-d879-4310-abaa-a2cb99b7f695
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: dev-storage1:/disk/storage1/fast
Brick2: dev-storage2:/disk/storage2/fast
 
Volume Name: single1
Type: Distribute
Volume ID: 74d62eb4-176e-4671-8471-779d909e19f0
Status: Started
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: dev-storage1:/disk/storage1/single1
 
Volume Name: safe
Type: Replicate
Volume ID: 47a8f326-0e48-4a71-9cfe-f9ef8d555db7
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: dev-storage1:/disk/storage1/safe
Brick2: dev-storage2:/disk/storage2/safe
root at dev-storage1:~# gluster peer status
Number of Peers: 1

Hostname: dev-storage2
Uuid: e95a9441-aec3-41a5-8d3d-615b61f3f2d3
State: Peer in Cluster (Connected)

---xxx---

root at dev-storage2:~# gluster volume info
 
Volume Name: safe
Type: Replicate
Volume ID: 47a8f326-0e48-4a71-9cfe-f9ef8d555db7
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: dev-storage1:/disk/storage1/safe
Brick2: dev-storage2:/disk/storage2/safe
 
Volume Name: fast
Type: Distribute
Volume ID: 864fd12d-d879-4310-abaa-a2cb99b7f695
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: dev-storage1:/disk/storage1/fast
Brick2: dev-storage2:/disk/storage2/fast
root at dev-storage2:~# gluster peer status
Number of Peers: 1

Hostname: 10.0.1.1
Uuid: a1a8a3cd-468e-44d5-8c08-23821db4c80f
State: Peer in Cluster (Connected)




More information about the Gluster-users mailing list