[Gluster-users] [ovirt-users] open error -13 = sanlock
Ravishankar N
ravishankar at redhat.com
Thu Mar 3 10:50:30 UTC 2016
On 03/03/2016 02:53 PM, paf1 at email.cz wrote:
> This is replica 2, only , with following settings
>
> Options Reconfigured:
> performance.quick-read: off
> performance.read-ahead: off
> performance.io-cache: off
> performance.stat-prefetch: off
> cluster.eager-lock: enable
> network.remote-dio: enable
> cluster.quorum-type: fixed
Not sure why you have set this option.
Ideally replica 3 or arbiter volumes are recommended for gluster+ovirt
use. (client) quorum does not make sense for a 2 node setup. I have a
detailed write up here which explains things
http://gluster.readthedocs.org/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/
which explains things.
> cluster.server-quorum-type: none
> storage.owner-uid: 36
> storage.owner-gid: 36
> cluster.quorum-count: 1
> cluster.self-heal-daemon: enable
>
> If I'll create "ids" file manually ( eg. " sanlock direct init -s
> 3c34ad63-6c66-4e23-ab46-084f3d70b147:0:/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids:0
> " ) on both bricks,
> vdsm is writing only to half of them ( that with 2 links = correct )
> "ids" file has correct permittions, owner, size on both bricks.
> brick 1: -rw-rw---- 1 vdsm kvm 1048576 2. bře 18.56
> /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids -
> not updated
Okay, so this one has link count =1 which means the .glusterfs hardlink
is missing. Can you try deleting this file from the brick and perform a
stat on the file from the mount? That should heal (i.e recreate it ) on
this brick from the other brick with the appropriate .glusterfs hard link.
> brick 2: -rw-rw---- 2 vdsm kvm 1048576 3. bře 10.16
> /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids -
> is continually updated
>
> What happens when I'll restart vdsm ? Will oVirt storages go to
> "disable " state ??? = disconnect VMs storages ?
No idea on this one...
-Ravi
>
> regs.Pa.
>
> On 3.3.2016 02:02, Ravishankar N wrote:
>> On 03/03/2016 12:43 AM, Nir Soffer wrote:
>>>
>>> PS: # find /STORAGES -samefile
>>> /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids
>>> -print
>>> /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids
>>> = missing "shadowfile" in " .gluster " dir.
>>> How can I fix it ?? - online !
>>>
>>>
>>> Ravi?
>> Is this the case in all 3 bricks of the replica?
>> BTW, you can just stat the file on the brick and see the link count
>> (it must be 2) instead of running the more expensive find command.
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160303/3daea443/attachment.html>
More information about the Gluster-users
mailing list