[Gluster-users] how to restore snapshot LV's

Mohammed Rafi K C rkavunga at redhat.com
Mon May 29 06:54:17 UTC 2017


On 05/27/2017 09:22 AM, WoongHee Han wrote:
> Ih, i'm sorry for my late reply
>
> I've tried to solve it using your answer. It worked as well thanks. it
> means the snapshot was activated.
> and then i was restore the snapshot.
>
> but, after i restored the snapshot ,there was nothing  in the
> volume(like files) 
> can't it recover automatically?
I remember you were saying that you had reconfigured the vg's. Did you
had mount for the snapshot brick path active ?

Rafi KC


>
> Thank you agin for your answer.
>
>
> Best regards
>
>
>
> 2017-05-19 15:34 GMT+09:00 Mohammed Rafi K C <rkavunga at redhat.com
> <mailto:rkavunga at redhat.com>>:
>
>     I do not know how you ended up in this state. This usually happens
>     when there is a commit failure. To recover from this state you can
>     change the value of "status" from
>
>     the path /var/lib/glusterd/snaps/<snapname>/<snapuid>/info . From
>     this file change the status to 0 in nodes where the values are
>     one. Then restart glusterd on those node where we changed manually.
>
>     Then try to activate it.
>
>
>     Regards
>
>     Rafi KC
>
>
>     On 05/18/2017 09:38 AM, Pranith Kumar Karampuri wrote:
>>     +Rafi, +Raghavendra Bhat
>>
>>     On Tue, May 16, 2017 at 11:55 AM, WoongHee Han
>>     <polishedwh at gmail.com <mailto:polishedwh at gmail.com>> wrote:
>>
>>         Hi, all!
>>
>>         I erased the VG having snapshot LV related to gluster volumes
>>         and then, I tried to restore volume;
>>
>>         1. vgcreate vg_cluster /dev/sdb
>>         2. lvcreate --size=10G --type=thin-pool -n tp_cluster vg_cluster
>>         3. lvcreate -V 5G --thinpool vg_cluster/tp_cluster -n
>>         test_vol vg_cluster
>>         4. gluster v stop test_vol
>>         5. getfattr -n trusted.glusterfs.volume-id /volume/test_vol (
>>         in other node)
>>         6. setfattr -n trusted.glusterfs.volume-id -v
>>          0sKtUJWIIpTeKWZx+S5PyXtQ== /volume/test_vol (already mounted)
>>         7. gluster v start test_vol
>>         8. restart glusterd
>>         9. lvcreate -s vg_cluster/test_vol --setactivationskip=n
>>         --name 6564c50651484d09a36b912962c573df_0
>>         10. lvcreate -s vg_cluster/test_vol --setactivationskip=n
>>         --name ee8c32a1941e4aba91feab21fbcb3c6c_0
>>         11. lvcreate -s vg_cluster/test_vol --setactivationskip=n
>>         --name bf93dc34233646128f0c5f84c3ac1f83_0 
>>         12. reboot
>>
>>         It works, but bricks for snapshot is not working.
>>
>>         ------------------------------------------------------------------------------------------------------------------------------------------
>>         ~]# glsuter snpshot status
>>         Brick Path        :  
>>         192.225.3.35:/var/run/gluster/snaps/bf93dc34233646128f0c5f84c3ac1f83/brick1
>>         Volume Group      :   vg_cluster
>>         Brick Running     :   No
>>         Brick PID         :   N/A
>>         Data Percentage   :   0.22
>>         LV Size           :   5.00g
>>
>>
>>         Brick Path        :  
>>         192.225.3.36:/var/run/gluster/snaps/bf93dc34233646128f0c5f84c3ac1f83/brick2
>>         Volume Group      :   vg_cluster
>>         Brick Running     :   No
>>         Brick PID         :   N/A
>>         Data Percentage   :   0.22
>>         LV Size           :   5.00g
>>
>>
>>         Brick Path        :  
>>         192.225.3.37:/var/run/gluster/snaps/bf93dc34233646128f0c5f84c3ac1f83/brick3
>>         Volume Group      :   vg_cluster
>>         Brick Running     :   No
>>         Brick PID         :   N/A
>>         Data Percentage   :   0.22
>>         LV Size           :   5.00g
>>
>>
>>         Brick Path        :  
>>         192.225.3.38:/var/run/gluster/snaps/bf93dc34233646128f0c5f84c3ac1f83/brick4
>>         Volume Group      :   vg_cluster
>>         Brick Running     :   Tes
>>         Brick PID         :   N/A
>>         Data Percentage   :   0.22
>>         LV Size           :   5.00g
>>
>>         ~]# gluster snapshot deactivate t3_GMT-2017.05.15-08.01.37
>>         Deactivating snap will make its data inaccessible. Do you
>>         want to continue? (y/n) y
>>         snapshot deactivate: failed: Pre Validation failed on
>>         192.225.3.36. Snapshot t3_GMT-2017.05.15-08.01.37 is already
>>         deactivated.
>>         Snapshot command failed
>>
>>         ~]# gluster snapshot activate t3_GMT-2017.05.15-08.01.37
>>         snapshot activate: failed: Snapshot
>>         t3_GMT-2017.05.15-08.01.37 is already activated
>>
>>         ------------------------------------------------------------------------------------------------------------------------------------------
>>
>>
>>         how to  restore snapshot LV's ?
>>
>>         my nodes consist of four nodes and  distributed, replicated (2x2)
>>
>>
>>         thank you.
>>
>>
>>
>>
>>         _______________________________________________
>>         Gluster-users mailing list
>>         Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>>         http://lists.gluster.org/mailman/listinfo/gluster-users
>>         <http://lists.gluster.org/mailman/listinfo/gluster-users>
>>
>>
>>
>>
>>     -- 
>>     Pranith
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170529/b0224700/attachment.html>


More information about the Gluster-users mailing list