[Gluster-users] how to restore snapshot LV's
Mohammed Rafi K C
rkavunga at redhat.com
Fri May 19 06:34:21 UTC 2017
I do not know how you ended up in this state. This usually happens when
there is a commit failure. To recover from this state you can change the
value of "status" from
the path /var/lib/glusterd/snaps/<snapname>/<snapuid>/info . From this
file change the status to 0 in nodes where the values are one. Then
restart glusterd on those node where we changed manually.
Then try to activate it.
Regards
Rafi KC
On 05/18/2017 09:38 AM, Pranith Kumar Karampuri wrote:
> +Rafi, +Raghavendra Bhat
>
> On Tue, May 16, 2017 at 11:55 AM, WoongHee Han <polishedwh at gmail.com
> <mailto:polishedwh at gmail.com>> wrote:
>
> Hi, all!
>
> I erased the VG having snapshot LV related to gluster volumes
> and then, I tried to restore volume;
>
> 1. vgcreate vg_cluster /dev/sdb
> 2. lvcreate --size=10G --type=thin-pool -n tp_cluster vg_cluster
> 3. lvcreate -V 5G --thinpool vg_cluster/tp_cluster -n test_vol
> vg_cluster
> 4. gluster v stop test_vol
> 5. getfattr -n trusted.glusterfs.volume-id /volume/test_vol ( in
> other node)
> 6. setfattr -n trusted.glusterfs.volume-id -v
> 0sKtUJWIIpTeKWZx+S5PyXtQ== /volume/test_vol (already mounted)
> 7. gluster v start test_vol
> 8. restart glusterd
> 9. lvcreate -s vg_cluster/test_vol --setactivationskip=n
> --name 6564c50651484d09a36b912962c573df_0
> 10. lvcreate -s vg_cluster/test_vol --setactivationskip=n
> --name ee8c32a1941e4aba91feab21fbcb3c6c_0
> 11. lvcreate -s vg_cluster/test_vol --setactivationskip=n
> --name bf93dc34233646128f0c5f84c3ac1f83_0
> 12. reboot
>
> It works, but bricks for snapshot is not working.
>
> ------------------------------------------------------------------------------------------------------------------------------------------
> ~]# glsuter snpshot status
> Brick Path :
> 192.225.3.35:/var/run/gluster/snaps/bf93dc34233646128f0c5f84c3ac1f83/brick1
> Volume Group : vg_cluster
> Brick Running : No
> Brick PID : N/A
> Data Percentage : 0.22
> LV Size : 5.00g
>
>
> Brick Path :
> 192.225.3.36:/var/run/gluster/snaps/bf93dc34233646128f0c5f84c3ac1f83/brick2
> Volume Group : vg_cluster
> Brick Running : No
> Brick PID : N/A
> Data Percentage : 0.22
> LV Size : 5.00g
>
>
> Brick Path :
> 192.225.3.37:/var/run/gluster/snaps/bf93dc34233646128f0c5f84c3ac1f83/brick3
> Volume Group : vg_cluster
> Brick Running : No
> Brick PID : N/A
> Data Percentage : 0.22
> LV Size : 5.00g
>
>
> Brick Path :
> 192.225.3.38:/var/run/gluster/snaps/bf93dc34233646128f0c5f84c3ac1f83/brick4
> Volume Group : vg_cluster
> Brick Running : Tes
> Brick PID : N/A
> Data Percentage : 0.22
> LV Size : 5.00g
>
> ~]# gluster snapshot deactivate t3_GMT-2017.05.15-08.01.37
> Deactivating snap will make its data inaccessible. Do you want to
> continue? (y/n) y
> snapshot deactivate: failed: Pre Validation failed on
> 192.225.3.36. Snapshot t3_GMT-2017.05.15-08.01.37 is already
> deactivated.
> Snapshot command failed
>
> ~]# gluster snapshot activate t3_GMT-2017.05.15-08.01.37
> snapshot activate: failed: Snapshot t3_GMT-2017.05.15-08.01.37 is
> already activated
>
> ------------------------------------------------------------------------------------------------------------------------------------------
>
>
> how to restore snapshot LV's ?
>
> my nodes consist of four nodes and distributed, replicated (2x2)
>
>
> thank you.
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
> http://lists.gluster.org/mailman/listinfo/gluster-users
> <http://lists.gluster.org/mailman/listinfo/gluster-users>
>
>
>
>
> --
> Pranith
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170519/9f37eb6a/attachment.html>
More information about the Gluster-users
mailing list