[Gluster-devel] [glusterfs-3.6.0beta3-0.11.gitd01b00a] gluster volume status is running even though the Disk is detached
Kiran Patil
kiran at fractalio.com
Tue Oct 28 06:14:57 UTC 2014
I changed git fetch git://review.gluster.org/glusterfs to git fetch
http://review.gluster.org/glusterfs and now it works.
Thanks,
Kiran.
On Tue, Oct 28, 2014 at 11:13 AM, Kiran Patil <kiran at fractalio.com> wrote:
> Hi Niels,
>
> I am getting "fatal: Couldn't find remote ref refs/changes/13/8213/9"
> error.
>
> Steps to reproduce the issue.
>
> 1) # git clone git://review.gluster.org/glusterfs
> Initialized empty Git repository in /root/gluster-3.6/glusterfs/.git/
> remote: Counting objects: 84921, done.
> remote: Compressing objects: 100% (48307/48307), done.
> remote: Total 84921 (delta 57264), reused 63233 (delta 36254)
> Receiving objects: 100% (84921/84921), 23.23 MiB | 192 KiB/s, done.
> Resolving deltas: 100% (57264/57264), done.
>
> 2) # cd glusterfs
> # git branch
> * master
>
> 3) # git fetch git://review.gluster.org/glusterfs refs/changes/13/8213/9
> && git checkout FETCH_HEAD
> fatal: Couldn't find remote ref refs/changes/13/8213/9
>
> Note: I also tried the above steps on git repo
> https://github.com/gluster/glusterfs and the result is same as above.
>
> Please let me know if I miss any steps.
>
> Thanks,
> Kiran.
>
> On Mon, Oct 27, 2014 at 5:53 PM, Niels de Vos <ndevos at redhat.com> wrote:
>
>> On Mon, Oct 27, 2014 at 05:19:13PM +0530, Kiran Patil wrote:
>> > Hi,
>> >
>> > I created replicated vol with two bricks on the same node and copied
>> some
>> > data to it.
>> >
>> > Now removed the disk which has hosted one of the brick of the volume.
>> >
>> > Storage.health-check-interval is set to 30 seconds.
>> >
>> > I could see the disk is unavailable using zpool command of zfs on linux
>> but
>> > the gluster volume status still displays the brick process running which
>> > should have been shutdown by this time.
>> >
>> > Is this a bug in 3.6 since it is mentioned as feature "
>> >
>> https://github.com/gluster/glusterfs/blob/release-3.6/doc/features/brick-failure-detection.md
>> "
>> > or am I doing any mistakes here?
>>
>> The initial detection of brick failures did not work for all
>> filesystems. It may not work for ZFS too. A fix has been posted, but it
>> has not been merged into the master branch yet. When the change has been
>> merged, it can get backported to 3.6 and 3.5.
>>
>> You may want to test with the patch applied, and add your "+1 Verified"
>> to the change in case it makes it functional for you:
>> - http://review.gluster.org/8213
>>
>> Cheers,
>> Niels
>>
>> >
>> > [root at fractal-c92e gluster-3.6]# gluster volume status
>> > Status of volume: repvol
>> > Gluster process Port Online Pid
>> >
>> ------------------------------------------------------------------------------
>> > Brick 192.168.1.246:/zp1/brick1 49154 Y 17671
>> > Brick 192.168.1.246:/zp2/brick2 49155 Y 17682
>> > NFS Server on localhost 2049 Y 17696
>> > Self-heal Daemon on localhost N/A Y 17701
>> >
>> > Task Status of Volume repvol
>> >
>> ------------------------------------------------------------------------------
>> > There are no active volume tasks
>> >
>> >
>> > [root at fractal-c92e gluster-3.6]# gluster volume info
>> >
>> > Volume Name: repvol
>> > Type: Replicate
>> > Volume ID: d4f992b1-1393-43b8-9fda-2e2b6e3b5039
>> > Status: Started
>> > Number of Bricks: 1 x 2 = 2
>> > Transport-type: tcp
>> > Bricks:
>> > Brick1: 192.168.1.246:/zp1/brick1
>> > Brick2: 192.168.1.246:/zp2/brick2
>> > Options Reconfigured:
>> > storage.health-check-interval: 30
>> >
>> > [root at fractal-c92e gluster-3.6]# zpool status zp2
>> > pool: zp2
>> > state: UNAVAIL
>> > status: One or more devices are faulted in response to IO failures.
>> > action: Make sure the affected devices are connected, then run 'zpool
>> > clear'.
>> > see: http://zfsonlinux.org/msg/ZFS-8000-HC
>> > scan: none requested
>> > config:
>> >
>> > NAME STATE READ WRITE CKSUM
>> > zp2 UNAVAIL 0 0 0 insufficient replicas
>> > sdb UNAVAIL 0 0 0
>> >
>> > errors: 2 data errors, use '-v' for a list
>> >
>> >
>> > Thanks,
>> > Kiran.
>>
>> > _______________________________________________
>> > Gluster-devel mailing list
>> > Gluster-devel at gluster.org
>> > http://supercolony.gluster.org/mailman/listinfo/gluster-devel
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-devel/attachments/20141028/6f2757ac/attachment.html>
More information about the Gluster-devel
mailing list