[Gluster-users] [Gluster-devel] 3.7.13 & proxmox/qemu

David Gossage dgossage at carouselchecks.com
Wed Jul 27 13:37:34 UTC 2016


On Tue, Jul 26, 2016 at 9:38 PM, Krutika Dhananjay <kdhananj at redhat.com>
wrote:

> Yes please, could you file a bug against glusterfs for this issue?
>

https://bugzilla.redhat.com/show_bug.cgi?id=1360785


>
>
> -Krutika
>
> On Wed, Jul 27, 2016 at 1:39 AM, David Gossage <
> dgossage at carouselchecks.com> wrote:
>
>> Has a bug report been filed for this issue or should l I create one with
>> the logs and results provided so far?
>>
>> *David Gossage*
>> *Carousel Checks Inc. | System Administrator*
>> *Office* 708.613.2284
>>
>> On Fri, Jul 22, 2016 at 12:53 PM, David Gossage <
>> dgossage at carouselchecks.com> wrote:
>>
>>>
>>>
>>>
>>> On Fri, Jul 22, 2016 at 9:37 AM, Vijay Bellur <vbellur at redhat.com>
>>> wrote:
>>>
>>>> On Fri, Jul 22, 2016 at 10:03 AM, Samuli Heinonen <
>>>> samppah at neutraali.net> wrote:
>>>> > Here is a quick way how to test this:
>>>> > GlusterFS 3.7.13 volume with default settings with brick on ZFS
>>>> dataset. gluster-test1 is server and gluster-test2 is client mounting with
>>>> FUSE.
>>>> >
>>>> > Writing file with oflag=direct is not ok:
>>>> > [root at gluster-test2 gluster]# dd if=/dev/zero of=file oflag=direct
>>>> count=1 bs=1024000
>>>> > dd: failed to open ‘file’: Invalid argument
>>>> >
>>>> > Enable network.remote-dio on Gluster Volume:
>>>> > [root at gluster-test1 gluster]# gluster volume set gluster
>>>> network.remote-dio enable
>>>> > volume set: success
>>>> >
>>>> > Writing small file with oflag=direct is ok:
>>>> > [root at gluster-test2 gluster]# dd if=/dev/zero of=file oflag=direct
>>>> count=1 bs=1024000
>>>> > 1+0 records in
>>>> > 1+0 records out
>>>> > 1024000 bytes (1.0 MB) copied, 0.0103793 s, 98.7 MB/s
>>>> >
>>>> > Writing bigger file with oflag=direct is ok:
>>>> > [root at gluster-test2 gluster]# dd if=/dev/zero of=file3 oflag=direct
>>>> count=100 bs=1M
>>>> > 100+0 records in
>>>> > 100+0 records out
>>>> > 104857600 bytes (105 MB) copied, 1.10583 s, 94.8 MB/s
>>>> >
>>>> > Enable Sharding on Gluster Volume:
>>>> > [root at gluster-test1 gluster]# gluster volume set gluster
>>>> features.shard enable
>>>> > volume set: success
>>>> >
>>>> > Writing small file  with oflag=direct is ok:
>>>> > [root at gluster-test2 gluster]# dd if=/dev/zero of=file3 oflag=direct
>>>> count=1 bs=1M
>>>> > 1+0 records in
>>>> > 1+0 records out
>>>> > 1048576 bytes (1.0 MB) copied, 0.0115247 s, 91.0 MB/s
>>>> >
>>>> > Writing bigger file with oflag=direct is not ok:
>>>> > [root at gluster-test2 gluster]# dd if=/dev/zero of=file3 oflag=direct
>>>> count=100 bs=1M
>>>> > dd: error writing ‘file3’: Operation not permitted
>>>> > dd: closing output file ‘file3’: Operation not permitted
>>>> >
>>>>
>>>>
>>>> Thank you for these tests! would it be possible to share the brick and
>>>> client logs?
>>>>
>>>
>>> Not sure if his tests are same as my setup but here is what I end up with
>>>
>>> Volume Name: glustershard
>>> Type: Replicate
>>> Volume ID: 0cc4efb6-3836-4caa-b24a-b3afb6e407c3
>>> Status: Started
>>> Number of Bricks: 1 x 3 = 3
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: 192.168.71.10:/gluster1/shard1/1
>>> Brick2: 192.168.71.11:/gluster1/shard2/1
>>> Brick3: 192.168.71.12:/gluster1/shard3/1
>>> Options Reconfigured:
>>> features.shard-block-size: 64MB
>>> features.shard: on
>>> server.allow-insecure: on
>>> storage.owner-uid: 36
>>> storage.owner-gid: 36
>>> cluster.server-quorum-type: server
>>> cluster.quorum-type: auto
>>> network.remote-dio: enable
>>> cluster.eager-lock: enable
>>> performance.stat-prefetch: off
>>> performance.io-cache: off
>>> performance.quick-read: off
>>> cluster.self-heal-window-size: 1024
>>> cluster.background-self-heal-count: 16
>>> nfs.enable-ino32: off
>>> nfs.addr-namelookup: off
>>> nfs.disable: on
>>> performance.read-ahead: off
>>> performance.readdir-ahead: on
>>>
>>>
>>>
>>>  dd if=/dev/zero of=/rhev/data-center/mnt/glusterSD/192.168.71.11\:_glustershard/
>>> oflag=direct count=100 bs=1M
>>> 81e19cd3-ae45-449c-b716-ec3e4ad4c2f0/ __DIRECT_IO_TEST__
>>>    .trashcan/
>>> [root at ccengine2 ~]# dd if=/dev/zero of=/rhev/data-center/mnt/glusterSD/
>>> 192.168.71.11\:_glustershard/81e19cd3-ae45-449c-b716-ec3e4ad4c2f0/images/test
>>> oflag=direct count=100 bs=1M
>>> dd: error writing ‘/rhev/data-center/mnt/glusterSD/192.168.71.11:_glustershard/81e19cd3-ae45-449c-b716-ec3e4ad4c2f0/images/test’:
>>> Operation not permitted
>>>
>>> creates the 64M file in expected location then the shard is 0
>>>
>>> # file:
>>> gluster1/shard1/1/81e19cd3-ae45-449c-b716-ec3e4ad4c2f0/images/test
>>>
>>> security.selinux=0x73797374656d5f753a6f626a6563745f723a756e6c6162656c65645f743a733000
>>> trusted.afr.dirty=0x000000000000000000000000
>>> trusted.bit-rot.version=0x0200000000000000579231f3000e16e7
>>> trusted.gfid=0xec6de302b35f427985639ca3e25d9df0
>>> trusted.glusterfs.shard.block-size=0x0000000004000000
>>>
>>> trusted.glusterfs.shard.file-size=0x0000000004000000000000000000000000000000000000010000000000000000
>>>
>>>
>>> # file: gluster1/shard1/1/.shard/ec6de302-b35f-4279-8563-9ca3e25d9df0.1
>>>
>>> security.selinux=0x73797374656d5f753a6f626a6563745f723a756e6c6162656c65645f743a733000
>>> trusted.afr.dirty=0x000000000000000000000000
>>> trusted.gfid=0x2bfd3cc8a727489b9a0474241548fe80
>>>
>>>
>>>> Regards,
>>>> Vijay
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org
>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>
>>>
>>>
>>
>> _______________________________________________
>> Gluster-devel mailing list
>> Gluster-devel at gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-devel
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160727/bd58e53d/attachment.html>


More information about the Gluster-users mailing list