[Gluster-users] Quota trouble
Avra Sengupta
asengupt at redhat.com
Tue Apr 21 09:04:23 UTC 2015
Hi Kondo,
Can u also mention the version of gluster you are using.
+Adding gluster-users
Regards,
Avra
On 04/21/2015 02:27 PM, Avra Sengupta wrote:
> Hi Kondo,
>
> I went through the gluster13 logs you had sent. Seems like something
> on that machine is holding the lock and is not releasing it. There are
> ways in which the system might end up in this scenario. I will try and
> explain the same with an example.
>
> Let's say I have gluster 11, gluster12, and gluster 13 in my cluster.
> I initiate a command from gluster11. Now the first thing that command
> does is, it holds a lock on all the nodes in the cluster on behalf of
> gluster11. Once the command does what's intended, it's last act before
> ending is to unlock all the nodes in the cluster. Now, only the node
> that has issued the lock, can issue the unlock.
>
> In your case what has happened is some command after successfully
> acquired the lock on gluster13. Now the node which initiated the
> command, went down or glusterd on that node went down before it could
> complete the command and it never got to send the unlock to gluster13.
>
> There's a workaround to it. You can restart glusterd on gluster13 and
> it should work fine.
>
> Regards,
> Avra
>
> On 04/20/2015 06:55 PM, kenji kondo wrote:
>> Hello Vijay,
>> Maybe this is very rare case. But is there any idea?
>>
>> Thanks,
>> Kondo
>>
>> 2015-04-15 9:47 GMT+09:00 Vijaikumar M <vmallika at redhat.com
>> <mailto:vmallika at redhat.com>>:
>>
>> Adding Avra...
>>
>> Thanks,
>> Vijay
>>
>>
>> -------- Forwarded Message --------
>> Subject: Re: [Gluster-users] Quota trouble
>> Date: Wed, 15 Apr 2015 00:27:26 +0900
>> From: kenji kondo <kkay.jp at gmail.com> <mailto:kkay.jp at gmail.com>
>> To: Vijaikumar M <vmallika at redhat.com> <mailto:vmallika at redhat.com>
>>
>>
>>
>> Hi Vijay,
>>
>> Thanks for your comments.
>>
>>
>> The lock error occurs at one server it's called "gluster13".
>>
>> In the gluster13, I tried to create new volume and start quota.
>> But it failed as below,
>>
>>
>> In both host gluster10 and gluster13, ran below
>>
>> $ sudo mkdir /export11/testbrick1
>>
>> $ sudo mkdir /export11/testbrick2
>>
>> In gluster13, ran below
>>
>> $ sudo /usr/sbin/gluster volume create testvol2
>> gluster13:/export11/testbrick1 gluster13:/export11/testbrick2
>>
>> volume create: testvol2: failed: Locking failed on gluster13.
>> Please check log file for details.
>>
>> $ sudo /usr/sbin/gluster volume create testvol2
>> gluster10:/export11/testbrick1 gluster10:/export11/testbrick2
>>
>> volume create: testvol2: failed: Locking failed on gluster13.
>> Please check log file for details.
>>
>> But I recived error messages above.
>>
>> On the other hand, in gluster10, it was success.
>>
>> Again, in gluster13, I tried to run quota, but it failed as below.
>>
>> $ sudo /usr/sbin/gluster volume quota testvol2 enable
>>
>> quota command failed : Locking failed on gluster13. Please check
>> log file for details.
>>
>>
>> Could you find attached?
>>
>> We can find error messages in the log of gluster13.
>>
>>
>> Best regards,
>>
>> Kondo
>>
>>
>>
>> 2015-04-13 19:38 GMT+09:00 Vijaikumar M <vmallika at redhat.com
>> <mailto:vmallika at redhat.com>>:
>>
>> Hi Kondo,
>>
>> The lock error you mentioned is because, another operation is
>> still running on the volume and hence not able to acquire the
>> lock.
>> This is bug of not displaying proper error message, we are
>> working on fixing this issue.
>>
>> I was not able to find any clue on why quotad is not running.
>>
>> I wanted to check, if we can manually start quotad something
>> like below:
>>
>> # /usr/local/sbin/glusterfs -s localhost --volfile-id
>> gluster/quotad -p /var/lib/glusterd/quotad/run/quotad.pid -l
>> /var/log/glusterfs/quotad.log -S
>> /var/run/gluster/myquotad.socket --xlator-option
>> *replicate*.data-self-heal=off --xlator-option
>> *replicate*.metadata-self-heal=off --xlator-option
>> *replicate*.entry-self-heal=off
>>
>> or
>>
>> create a new temporary volume, and enable quota on this
>> volume. (quotad will be same for all the volume which has
>> quota enabled)
>>
>>
>> Thanks,
>> Vijay
>>
>>
>> On Sunday 12 April 2015 07:05 PM, kenji kondo wrote:
>>> Hi Vijay,
>>>
>>> Thank you for your suggestion. But I'm sorry, it's difficult
>>> to access from outside because my glusterfs system is closed.
>>> I will give up if there is no clue information in attached log.
>>>
>>> Best regards,
>>> Kondo
>>>
>>>
>>> 2015-04-09 15:40 GMT+09:00 Vijaikumar M <vmallika at redhat.com
>>> <mailto:vmallika at redhat.com>>:
>>>
>>>
>>>
>>> On Thursday 09 April 2015 11:58 AM, Vijaikumar M wrote:
>>>>
>>>>
>>>> On Wednesday 08 April 2015 09:57 PM, kenji kondo wrote:
>>>>> Hi Vijay,
>>>>>
>>>>> I checked the all of the setting.
>>>>> The all are 'features.quota=on' when I set quota
>>>>> enable and the all are 'features.quota=off' when I set
>>>>> quota disable.
>>>>>
>>>>> But I could find new issue.
>>>>> When I checked a volume status for all server, in one
>>>>> of the servers I received the error message as below.
>>>>>
>>>>> $ sudo /usr/sbin/gluster volume status testvol
>>>>> Locking failed on gluster13. Please check log file for
>>>>> details.
>>>>>
>>>>> In etc-glusterfs-glusterd.vol.log of problem server, I
>>>>> found error messages as below.
>>>>> [2015-04-08 08:40:04.782644] I
>>>>> [mem-pool.c:545:mem_pool_destroy] 0-management:
>>>>> size=588 max=0 total=0
>>>>> [2015-04-08 08:40:04.782685] I
>>>>> [mem-pool.c:545:mem_pool_destroy] 0-management:
>>>>> size=124 max=0 total=0
>>>>> [2015-04-08 08:40:04.782848] W
>>>>> [socket.c:611:__socket_rwv] 0-management: readv on
>>>>> /var/run/14b05cd492843e6e288e290c2d63093c.socket
>>>>> failed (Invalid arguments)
>>>>> [2015-04-08 08:40:04.805407] I [MSGID: 106006]
>>>>> [glusterd-handler.c:4257:__glusterd_nodesvc_rpc_notify] 0-management:
>>>>> nfs has disconnected from glusterd.
>>>>> [2015-04-08 08:43:02.439001] I
>>>>> [glusterd-handler.c:3803:__glusterd_handle_status_volume]
>>>>> 0-management: Received status volume req for volume
>>>>> testvol
>>>>> [2015-04-08 08:43:02.460581] E
>>>>> [glusterd-utils.c:148:glusterd_lock] 0-management:
>>>>> Unable to get lock for uuid:
>>>>> 03a32bce-ec63-4dc3-a287-4901a55dd8c9, lock held by:
>>>>> 03a32bce-ec63-4dc3-a287-4901a55dd8c9
>>>>> [2015-04-08 08:43:02.460632] E
>>>>> [glusterd-op-sm.c:6584:glusterd_op_sm] 0-management:
>>>>> handler returned: -1
>>>>> [2015-04-08 08:43:02.460654] E
>>>>> [glusterd-syncop.c:105:gd_collate_errors] 0-: Locking
>>>>> failed on gluster13. Please check log file for details.
>>>>> [2015-04-08 08:43:02.461409] E
>>>>> [glusterd-syncop.c:1602:gd_sync_task_begin]
>>>>> 0-management: Locking Peers Failed.
>>>>> [2015-04-08 08:43:43.698168] I
>>>>> [glusterd-handler.c:3803:__glusterd_handle_status_volume]
>>>>> 0-management: Received status volume req for volume
>>>>> testvol
>>>>> [2015-04-08 08:43:43.698813] E
>>>>> [glusterd-utils.c:148:glusterd_lock] 0-management:
>>>>> Unable to get lock for uuid:
>>>>> 03a32bce-ec63-4dc3-a287-4901a55dd8c9, lock held by:
>>>>> 03a32bce-ec63-4dc3-a287-4901a55dd8c9
>>>>> [2015-04-08 08:43:43.698898] E
>>>>> [glusterd-op-sm.c:6584:glusterd_op_sm] 0-management:
>>>>> handler returned: -1
>>>>> [2015-04-08 08:43:43.698994] E
>>>>> [glusterd-syncop.c:105:gd_collate_errors] 0-: Locking
>>>>> failed on gluster13. Please check log file for details.
>>>>> [2015-04-08 08:43:43.702126] E
>>>>> [glusterd-syncop.c:1602:gd_sync_task_begin]
>>>>> 0-management: Locking Peers Failed.
>>>>> [2015-04-08 08:44:01.277139] I
>>>>> [glusterd-handler.c:3803:__glusterd_handle_status_volume]
>>>>> 0-management: Received status volume req for volume
>>>>> testvol
>>>>> [2015-04-08 08:44:01.277560] E
>>>>> [glusterd-utils.c:148:glusterd_lock] 0-management:
>>>>> Unable to get lock for uuid:
>>>>> 03a32bce-ec63-4dc3-a287-4901a55dd8c9, lock held by:
>>>>> 03a32bce-ec63-4dc3-a287-4901a55dd8c9
>>>>> [2015-04-08 08:44:01.277639] E
>>>>> [glusterd-op-sm.c:6584:glusterd_op_sm] 0-management:
>>>>> handler returned: -1
>>>>> [2015-04-08 08:44:01.277676] E
>>>>> [glusterd-syncop.c:105:gd_collate_errors] 0-: Locking
>>>>> failed on gluster13. Please check log file for details.
>>>>> [2015-04-08 08:44:01.281514] E
>>>>> [glusterd-syncop.c:1602:gd_sync_task_begin]
>>>>> 0-management: Locking Peers Failed.
>>>>> [2015-04-08 08:45:42.599796] I
>>>>> [glusterd-handler.c:3803:__glusterd_handle_status_volume]
>>>>> 0-management: Received status volume req for volume
>>>>> testvol
>>>>> [2015-04-08 08:45:42.600343] E
>>>>> [glusterd-utils.c:148:glusterd_lock] 0-management:
>>>>> Unable to get lock for uuid:
>>>>> 03a32bce-ec63-4dc3-a287-4901a55dd8c9, lock held by:
>>>>> 03a32bce-ec63-4dc3-a287-4901a55dd8c9
>>>>> [2015-04-08 08:45:42.600417] E
>>>>> [glusterd-op-sm.c:6584:glusterd_op_sm] 0-management:
>>>>> handler returned: -1
>>>>> [2015-04-08 08:45:42.600482] E
>>>>> [glusterd-syncop.c:105:gd_collate_errors] 0-: Locking
>>>>> failed on gluster13. Please check log file for details.
>>>>> [2015-04-08 08:45:42.601039] E
>>>>> [glusterd-syncop.c:1602:gd_sync_task_begin]
>>>>> 0-management: Locking Peers Failed.
>>>>>
>>>>> Does this situation relate to my quota problems?
>>>>>
>>>>
>>>> This is a glusterd different issue. Can we get the
>>>> glusterd logs from gluster13?
>>>> Can get access to these machines, so that we can debug
>>>> live?
>>>>
>>>> Thanks,
>>>> Vijay
>>>>
>>> Regarding quota issue, quota feature is enabled
>>> successfully. I am wondering why quotad is not started.
>>> If we get the access to the machine, it will be easier
>>> to debug the issue.
>>>
>>> Thanks,
>>> Vijay
>>>
>>>
>>>>>
>>>>> Best regards,
>>>>> Kondo
>>>>>
>>>>>
>>>>> 2015-04-08 15:14 GMT+09:00 Vijaikumar M
>>>>> <vmallika at redhat.com <mailto:vmallika at redhat.com>>:
>>>>>
>>>>> Hi Kondo,
>>>>>
>>>>> I suspect, in one of the node quota feature is not
>>>>> set for some reason and hence quotad is not starting.
>>>>>
>>>>> On all the nodes can you check if below option is
>>>>> set to 'on'
>>>>>
>>>>> # grep quota /var/lib/glusterd/vols/<volname>/info
>>>>> features.quota=on
>>>>>
>>>>>
>>>>> Also can I get brick logs from all the nodes?
>>>>>
>>>>> Also can you create a temporary volume and enable
>>>>> the quota here and see if see quota works fine
>>>>> with this volume?
>>>>>
>>>>>
>>>>> Thanks,
>>>>> Vijay
>>>>>
>>>>> On Tuesday 07 April 2015 08:34 PM, kenji kondo wrote:
>>>>>> Hi Vijay,
>>>>>>
>>>>>> Could you find attached?
>>>>>> I got logs of server and client.
>>>>>> As same as before, I could not create a file
>>>>>> after quota usage-limit setting.
>>>>>>
>>>>>> Best regards,
>>>>>> Kondo
>>>>>>
>>>>>>
>>>>>> 2015-04-07 18:34 GMT+09:00 Vijaikumar M
>>>>>> <vmallika at redhat.com <mailto:vmallika at redhat.com>>:
>>>>>>
>>>>>> Hi Konda,
>>>>>>
>>>>>> Can we get all the log files?
>>>>>>
>>>>>> # gluster volume quota <volname> disable
>>>>>> # gluster volume quota <volname> enable
>>>>>>
>>>>>>
>>>>>> Now copy all the logs files.
>>>>>>
>>>>>> Thanks,
>>>>>> Vijay
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Tuesday 07 April 2015 12:39 PM, K.Kondo wrote:
>>>>>>> Thank you very much ! Vijay
>>>>>>> I want to use a quota because each volume
>>>>>>> became too big.
>>>>>>>
>>>>>>> Best regard
>>>>>>> Kondo
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> 2015/04/07 15:18、Vijaikumar M
>>>>>>> <vmallika at redhat.com
>>>>>>> <mailto:vmallika at redhat.com>> のメッセージ:
>>>>>>>
>>>>>>>> Hi Kondo,
>>>>>>>>
>>>>>>>> I couldn’t find clue from the logs. I will
>>>>>>>> discuss about this issue with my colleagues
>>>>>>>> today.
>>>>>>>>
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Vijay
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Monday 06 April 2015 10:56 PM, kenji
>>>>>>>> kondo wrote:
>>>>>>>>> Hello Vijay,
>>>>>>>>> Is there something idea for this?
>>>>>>>>> Best regards,
>>>>>>>>> Kondo
>>>>>>>>>
>>>>>>>>> 2015-03-31 22:46 GMT+09:00 kenji kondo
>>>>>>>>> <kkay.jp at gmail.com
>>>>>>>>> <mailto:kkay.jp at gmail.com>>:
>>>>>>>>>
>>>>>>>>> Hi Vijay,
>>>>>>>>>
>>>>>>>>> I'm sorry for late reply.
>>>>>>>>> I could get the debug mode log as
>>>>>>>>> attached.
>>>>>>>>> In this test, unfortunately the quota
>>>>>>>>> did not work as same as before.
>>>>>>>>>
>>>>>>>>> Could you find the cause of my problem?
>>>>>>>>>
>>>>>>>>> Best regards,
>>>>>>>>> Kondo
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> 2015-03-25 17:20 GMT+09:00 Vijaikumar
>>>>>>>>> M <vmallika at redhat.com
>>>>>>>>> <mailto:vmallika at redhat.com>>:
>>>>>>>>>
>>>>>>>>> Hi Kondo,
>>>>>>>>>
>>>>>>>>> For some reason quota enable was
>>>>>>>>> not successful. We may have re-try
>>>>>>>>> enabling quota.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>> Vijay
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Tuesday 24 March 2015 07:08 PM,
>>>>>>>>> kenji kondo wrote:
>>>>>>>>>> Hi Vijay,
>>>>>>>>>> Thanks for your checking.
>>>>>>>>>> Unfortunately, currently I can't
>>>>>>>>>> stop the service because many
>>>>>>>>>> users are using.
>>>>>>>>>> But, I want to know this cause of
>>>>>>>>>> this trouble, so I will plan to
>>>>>>>>>> stop. Please wait to get the log.
>>>>>>>>>>
>>>>>>>>>> Best regards,
>>>>>>>>>> Kondo
>>>>>>>>>>
>>>>>>>>>> 2015-03-24 17:01 GMT+09:00
>>>>>>>>>> Vijaikumar M <vmallika at redhat.com
>>>>>>>>>> <mailto:vmallika at redhat.com>>:
>>>>>>>>>>
>>>>>>>>>> Hi Kondo,
>>>>>>>>>>
>>>>>>>>>> I couldn't find much clue in
>>>>>>>>>> the glusterd logs, other than
>>>>>>>>>> the error message you
>>>>>>>>>> mentioned below.
>>>>>>>>>> Can you try disabling and
>>>>>>>>>> enabling the quota again and
>>>>>>>>>> see if this start quotad?
>>>>>>>>>>
>>>>>>>>>> Try below command:
>>>>>>>>>> # gluster volume quota
>>>>>>>>>> <volname> disable
>>>>>>>>>>
>>>>>>>>>> wait for all quota process to
>>>>>>>>>> terminate
>>>>>>>>>> #ps -ef | quota
>>>>>>>>>>
>>>>>>>>>> # service glusterd stop
>>>>>>>>>> # glusterd -LDEBUG
>>>>>>>>>> # gluster volume quota
>>>>>>>>>> <volname> enable
>>>>>>>>>>
>>>>>>>>>> Now verify if quotad is running
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Thanks,
>>>>>>>>>> Vijay
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Monday 23 March 2015 06:24
>>>>>>>>>> PM, kenji kondo wrote:
>>>>>>>>>>> Hi Vijay,
>>>>>>>>>>> As you pointed out, the
>>>>>>>>>>> quotad is not running in the
>>>>>>>>>>> all of server.
>>>>>>>>>>> I checked the volume status
>>>>>>>>>>> and got following log.
>>>>>>>>>>>
>>>>>>>>>>> Quota Daemon on gluster25N/ANN/A
>>>>>>>>>>>
>>>>>>>>>>> So, I attached requested log
>>>>>>>>>>> 'etc-glusterfs-glusterd.vol.log'.
>>>>>>>>>>> The error messages can be
>>>>>>>>>>> found in the log.
>>>>>>>>>>>
>>>>>>>>>>> [2015-03-19 11:51:07.457697]
>>>>>>>>>>> E
>>>>>>>>>>> [glusterd-quota.c:1467:glusterd_op_stage_quota]
>>>>>>>>>>> 0-management: Quota is
>>>>>>>>>>> disabled, please enable quota
>>>>>>>>>>>
>>>>>>>>>>> If you want more some
>>>>>>>>>>> information to solve this
>>>>>>>>>>> problems, please ask me.
>>>>>>>>>>>
>>>>>>>>>>> Best regards,
>>>>>>>>>>> Kondo
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> 2015-03-23 16:04 GMT+09:00
>>>>>>>>>>> Vijaikumar M
>>>>>>>>>>> <vmallika at redhat.com
>>>>>>>>>>> <mailto:vmallika at redhat.com>>:
>>>>>>>>>>>
>>>>>>>>>>> Hi Kondo,
>>>>>>>>>>>
>>>>>>>>>>> Can you please verify if
>>>>>>>>>>> quotad is running?
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> root at rh1:~ *# gluster
>>>>>>>>>>> volume status*
>>>>>>>>>>> Status of volume: vol1
>>>>>>>>>>> Gluster process TCP
>>>>>>>>>>> Port RDMA Port Online Pid
>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>> Brick
>>>>>>>>>>> rh1:/var/opt/gluster/bricks/b1/dir
>>>>>>>>>>> 49152 0 Y 1858
>>>>>>>>>>> NFS Server on localhost
>>>>>>>>>>> 2049 0 Y 1879
>>>>>>>>>>> *Quota Daemon on
>>>>>>>>>>> localhost N/A N/A
>>>>>>>>>>> Y 1914 **
>>>>>>>>>>> *
>>>>>>>>>>> Task Status of Volume vol1
>>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>> There are no active
>>>>>>>>>>> volume tasks
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> root at rh1:~ # *ps -ef |
>>>>>>>>>>> grep quotad*
>>>>>>>>>>> root 1914 1 0 12:29
>>>>>>>>>>> ? 00:00:00
>>>>>>>>>>> /usr/local/sbin/glusterfs -s
>>>>>>>>>>> localhost --volfile-id
>>>>>>>>>>> gluster/quotad -p
>>>>>>>>>>> /var/lib/glusterd/quotad/run/quotad.pid
>>>>>>>>>>> -l
>>>>>>>>>>> */var/log/glusterfs/quotad.log*-S
>>>>>>>>>>> /var/run/gluster/bb6ab82f70f555fd5c0e188fa4e09584.socket
>>>>>>>>>>> --xlator-option
>>>>>>>>>>> *replicate*.data-self-heal=off
>>>>>>>>>>> --xlator-option
>>>>>>>>>>> *replicate*.metadata-self-heal=off
>>>>>>>>>>> --xlator-option
>>>>>>>>>>> *replicate*.entry-self-heal=off
>>>>>>>>>>> root 1970 1511 0 12:31
>>>>>>>>>>> pts/1 00:00:00 grep quotad
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> root at rh1:~ # *gluster
>>>>>>>>>>> volume info*
>>>>>>>>>>> Volume Name: vol1
>>>>>>>>>>> Type: Distribute
>>>>>>>>>>> Volume ID:
>>>>>>>>>>> a55519ec-65d1-4741-9ad3-f94020fc9b21
>>>>>>>>>>> Status: Started
>>>>>>>>>>> Number of Bricks: 1
>>>>>>>>>>> Transport-type: tcp
>>>>>>>>>>> Bricks:
>>>>>>>>>>> Brick1:
>>>>>>>>>>> rh1:/var/opt/gluster/bricks/b1/dir
>>>>>>>>>>> Options Reconfigured:
>>>>>>>>>>> *features.quota: on**
>>>>>>>>>>> *
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> If quotad is not
>>>>>>>>>>> running, can you please
>>>>>>>>>>> provide glusterd logs
>>>>>>>>>>> 'usr-local-etc-glusterfs-glusterd.vol.log'.
>>>>>>>>>>> I will check is there
>>>>>>>>>>> are any issues starting
>>>>>>>>>>> quotad.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Thanks,
>>>>>>>>>>> Vihay
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Monday 23 March 2015
>>>>>>>>>>> 11:54 AM, K.Kondo wrote:
>>>>>>>>>>>> Hi Vijay,
>>>>>>>>>>>> I could not find
>>>>>>>>>>>> the"quotad.log" in
>>>>>>>>>>>> directory
>>>>>>>>>>>> /var/log/glusterfs in
>>>>>>>>>>>> both servers and
>>>>>>>>>>>> client. But other test
>>>>>>>>>>>> server has the log.
>>>>>>>>>>>> Do you know why there
>>>>>>>>>>>> is no the file?
>>>>>>>>>>>> Thanks,
>>>>>>>>>>>> Kondo
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> 2015/03/23 13:41、
>>>>>>>>>>>> Vijaikumar M
>>>>>>>>>>>> <vmallika at redhat.com
>>>>>>>>>>>> <mailto:vmallika at redhat.com>>
>>>>>>>>>>>> のメッセージ:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi Kondo,
>>>>>>>>>>>>>
>>>>>>>>>>>>> log file 'quotad.log'
>>>>>>>>>>>>> is missing in the
>>>>>>>>>>>>> attachment.Can you
>>>>>>>>>>>>> provide this log file
>>>>>>>>>>>>> as well?
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>> Vijay
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Monday 23 March
>>>>>>>>>>>>> 2015 09:50 AM, kenji
>>>>>>>>>>>>> kondo wrote:
>>>>>>>>>>>>>> Hi Vijay,
>>>>>>>>>>>>>> Could you find the
>>>>>>>>>>>>>> attached?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Best regards,
>>>>>>>>>>>>>> Kondo
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 2015-03-23 12:53
>>>>>>>>>>>>>> GMT+09:00 Vijaikumar
>>>>>>>>>>>>>> M
>>>>>>>>>>>>>> <vmallika at redhat.com
>>>>>>>>>>>>>> <mailto:vmallika at redhat.com>>:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hi Kondo,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Can you please
>>>>>>>>>>>>>> provide below
>>>>>>>>>>>>>> mentioned
>>>>>>>>>>>>>> gluterfs logs?
>>>>>>>>>>>>>> client logs
>>>>>>>>>>>>>> (name of this log
>>>>>>>>>>>>>> will be prefixed
>>>>>>>>>>>>>> with mount-point
>>>>>>>>>>>>>> dirname)
>>>>>>>>>>>>>> brick logs
>>>>>>>>>>>>>> quotad logs
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>> Vijay
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Friday 20
>>>>>>>>>>>>>> March 2015 06:31
>>>>>>>>>>>>>> PM, kenji kondo
>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>> Hi, Vijay and Peter
>>>>>>>>>>>>>>> Thanks for your
>>>>>>>>>>>>>>> reply.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I create new
>>>>>>>>>>>>>>> volume "testvol"
>>>>>>>>>>>>>>> with two bricks
>>>>>>>>>>>>>>> and set quota to
>>>>>>>>>>>>>>> simplify this
>>>>>>>>>>>>>>> problem.
>>>>>>>>>>>>>>> I got the
>>>>>>>>>>>>>>> glusterfs log as
>>>>>>>>>>>>>>> following after
>>>>>>>>>>>>>>> try to create a
>>>>>>>>>>>>>>> directory and file.
>>>>>>>>>>>>>>> BTW, my glusterd
>>>>>>>>>>>>>>> was upgraded
>>>>>>>>>>>>>>> from older
>>>>>>>>>>>>>>> version,
>>>>>>>>>>>>>>> although I don't
>>>>>>>>>>>>>>> know related to it.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Best regards,
>>>>>>>>>>>>>>> Kondo
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.931016]
>>>>>>>>>>>>>>> I [MSGID:
>>>>>>>>>>>>>>> 100030]
>>>>>>>>>>>>>>> [glusterfsd.c:1998:main]
>>>>>>>>>>>>>>> 0-/usr/sbin/glusterfs:
>>>>>>>>>>>>>>> Started running
>>>>>>>>>>>>>>> /usr/sbin/glusterfs
>>>>>>>>>>>>>>> version 3.6.0.29
>>>>>>>>>>>>>>> (args:
>>>>>>>>>>>>>>> /usr/sbin/glusterfs
>>>>>>>>>>>>>>> --volfile-server=gluster10
>>>>>>>>>>>>>>> --volfile-id=testvol
>>>>>>>>>>>>>>> testvol)
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.944850]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [dht-shared.c:337:dht_init_regex]
>>>>>>>>>>>>>>> 0-testvol-dht:
>>>>>>>>>>>>>>> using regex
>>>>>>>>>>>>>>> rsync-hash-regex
>>>>>>>>>>>>>>> = ^\.(.+)\.[^.]+$
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.946256]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client.c:2280:notify]
>>>>>>>>>>>>>>> 0-testvol-client-0:
>>>>>>>>>>>>>>> parent
>>>>>>>>>>>>>>> translators are
>>>>>>>>>>>>>>> ready,
>>>>>>>>>>>>>>> attempting
>>>>>>>>>>>>>>> connect on transport
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.950674]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client.c:2280:notify]
>>>>>>>>>>>>>>> 0-testvol-client-1:
>>>>>>>>>>>>>>> parent
>>>>>>>>>>>>>>> translators are
>>>>>>>>>>>>>>> ready,
>>>>>>>>>>>>>>> attempting
>>>>>>>>>>>>>>> connect on transport
>>>>>>>>>>>>>>> Final graph:
>>>>>>>>>>>>>>> +------------------------------------------------------------------------------+
>>>>>>>>>>>>>>> 1: volume
>>>>>>>>>>>>>>> testvol-client-0
>>>>>>>>>>>>>>> 2: type
>>>>>>>>>>>>>>> protocol/client
>>>>>>>>>>>>>>> 3: option
>>>>>>>>>>>>>>> ping-timeout 42
>>>>>>>>>>>>>>> 4: option
>>>>>>>>>>>>>>> remote-host
>>>>>>>>>>>>>>> gluster24
>>>>>>>>>>>>>>> 5: option
>>>>>>>>>>>>>>> remote-subvolume
>>>>>>>>>>>>>>> /export25/brick
>>>>>>>>>>>>>>> 6: option
>>>>>>>>>>>>>>> transport-type
>>>>>>>>>>>>>>> socket
>>>>>>>>>>>>>>> 7: option
>>>>>>>>>>>>>>> send-gids true
>>>>>>>>>>>>>>> 8: end-volume
>>>>>>>>>>>>>>> 9:
>>>>>>>>>>>>>>> 10: volume
>>>>>>>>>>>>>>> testvol-client-1
>>>>>>>>>>>>>>> 11: type
>>>>>>>>>>>>>>> protocol/client
>>>>>>>>>>>>>>> 12: option
>>>>>>>>>>>>>>> ping-timeout 42
>>>>>>>>>>>>>>> 13: option
>>>>>>>>>>>>>>> remote-host
>>>>>>>>>>>>>>> gluster25
>>>>>>>>>>>>>>> 14: option
>>>>>>>>>>>>>>> remote-subvolume
>>>>>>>>>>>>>>> /export25/brick
>>>>>>>>>>>>>>> 15: option
>>>>>>>>>>>>>>> transport-type
>>>>>>>>>>>>>>> socket
>>>>>>>>>>>>>>> 16: option
>>>>>>>>>>>>>>> send-gids true
>>>>>>>>>>>>>>> 17: end-volume
>>>>>>>>>>>>>>> 18:
>>>>>>>>>>>>>>> 19: volume
>>>>>>>>>>>>>>> testvol-dht
>>>>>>>>>>>>>>> 20: type
>>>>>>>>>>>>>>> cluster/distribute
>>>>>>>>>>>>>>> 21: subvolumes
>>>>>>>>>>>>>>> testvol-client-0
>>>>>>>>>>>>>>> testvol-client-1
>>>>>>>>>>>>>>> 22: end-volume
>>>>>>>>>>>>>>> 23:
>>>>>>>>>>>>>>> 24: volume
>>>>>>>>>>>>>>> testvol-write-behind
>>>>>>>>>>>>>>> 25: type
>>>>>>>>>>>>>>> performance/write-behind
>>>>>>>>>>>>>>> 26: subvolumes
>>>>>>>>>>>>>>> testvol-dht
>>>>>>>>>>>>>>> 27: end-volume
>>>>>>>>>>>>>>> 28:
>>>>>>>>>>>>>>> 29: volume
>>>>>>>>>>>>>>> testvol-read-ahead
>>>>>>>>>>>>>>> 30: type
>>>>>>>>>>>>>>> performance/read-ahead
>>>>>>>>>>>>>>> 31: subvolumes
>>>>>>>>>>>>>>> testvol-write-behind
>>>>>>>>>>>>>>> 32: end-volume
>>>>>>>>>>>>>>> 33:
>>>>>>>>>>>>>>> 34: volume
>>>>>>>>>>>>>>> testvol-io-cache
>>>>>>>>>>>>>>> 35: type
>>>>>>>>>>>>>>> performance/io-cache
>>>>>>>>>>>>>>> 36: subvolumes
>>>>>>>>>>>>>>> testvol-read-ahead
>>>>>>>>>>>>>>> 37: end-volume
>>>>>>>>>>>>>>> 38:
>>>>>>>>>>>>>>> 39: volume
>>>>>>>>>>>>>>> testvol-quick-read
>>>>>>>>>>>>>>> 40: type
>>>>>>>>>>>>>>> performance/quick-read
>>>>>>>>>>>>>>> 41: subvolumes
>>>>>>>>>>>>>>> testvol-io-cache
>>>>>>>>>>>>>>> 42: end-volume
>>>>>>>>>>>>>>> 43:
>>>>>>>>>>>>>>> 44: volume
>>>>>>>>>>>>>>> testvol-md-cache
>>>>>>>>>>>>>>> 45: type
>>>>>>>>>>>>>>> performance/md-cache
>>>>>>>>>>>>>>> 46: subvolumes
>>>>>>>>>>>>>>> testvol-quick-read
>>>>>>>>>>>>>>> 47: end-volume
>>>>>>>>>>>>>>> 48:
>>>>>>>>>>>>>>> 49: volume testvol
>>>>>>>>>>>>>>> 50: type
>>>>>>>>>>>>>>> debug/io-stats
>>>>>>>>>>>>>>> 51: option
>>>>>>>>>>>>>>> latency-measurement
>>>>>>>>>>>>>>> off
>>>>>>>>>>>>>>> 52: option
>>>>>>>>>>>>>>> count-fop-hits off
>>>>>>>>>>>>>>> 53: subvolumes
>>>>>>>>>>>>>>> testvol-md-cache
>>>>>>>>>>>>>>> 54: end-volume
>>>>>>>>>>>>>>> 55:
>>>>>>>>>>>>>>> 56: volume
>>>>>>>>>>>>>>> meta-autoload
>>>>>>>>>>>>>>> 57: type meta
>>>>>>>>>>>>>>> 58: subvolumes
>>>>>>>>>>>>>>> testvol
>>>>>>>>>>>>>>> 59: end-volume
>>>>>>>>>>>>>>> 60:
>>>>>>>>>>>>>>> +------------------------------------------------------------------------------+
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.955337]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [rpc-clnt.c:1759:rpc_clnt_reconfig]
>>>>>>>>>>>>>>> 0-testvol-client-0:
>>>>>>>>>>>>>>> changing port to
>>>>>>>>>>>>>>> 49155 (from 0)
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.957549]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [rpc-clnt.c:1759:rpc_clnt_reconfig]
>>>>>>>>>>>>>>> 0-testvol-client-1:
>>>>>>>>>>>>>>> changing port to
>>>>>>>>>>>>>>> 49155 (from 0)
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.959889]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:1415:select_server_supported_programs]
>>>>>>>>>>>>>>> 0-testvol-client-0:
>>>>>>>>>>>>>>> Using Program
>>>>>>>>>>>>>>> GlusterFS 3.3,
>>>>>>>>>>>>>>> Num (1298437),
>>>>>>>>>>>>>>> Version (330)
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.960090]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:1415:select_server_supported_programs]
>>>>>>>>>>>>>>> 0-testvol-client-1:
>>>>>>>>>>>>>>> Using Program
>>>>>>>>>>>>>>> GlusterFS 3.3,
>>>>>>>>>>>>>>> Num (1298437),
>>>>>>>>>>>>>>> Version (330)
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.960376]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:1200:client_setvolume_cbk]
>>>>>>>>>>>>>>> 0-testvol-client-0:
>>>>>>>>>>>>>>> Connected to
>>>>>>>>>>>>>>> testvol-client-0, attached
>>>>>>>>>>>>>>> to remote volume
>>>>>>>>>>>>>>> '/export25/brick'.
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.960405]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:1212:client_setvolume_cbk]
>>>>>>>>>>>>>>> 0-testvol-client-0:
>>>>>>>>>>>>>>> Server and
>>>>>>>>>>>>>>> Client
>>>>>>>>>>>>>>> lk-version
>>>>>>>>>>>>>>> numbers are not
>>>>>>>>>>>>>>> same, reopening
>>>>>>>>>>>>>>> the fds
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.960471]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:1200:client_setvolume_cbk]
>>>>>>>>>>>>>>> 0-testvol-client-1:
>>>>>>>>>>>>>>> Connected to
>>>>>>>>>>>>>>> testvol-client-1, attached
>>>>>>>>>>>>>>> to remote volume
>>>>>>>>>>>>>>> '/export25/brick'.
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.960478]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:1212:client_setvolume_cbk]
>>>>>>>>>>>>>>> 0-testvol-client-1:
>>>>>>>>>>>>>>> Server and
>>>>>>>>>>>>>>> Client
>>>>>>>>>>>>>>> lk-version
>>>>>>>>>>>>>>> numbers are not
>>>>>>>>>>>>>>> same, reopening
>>>>>>>>>>>>>>> the fds
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.962288]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [fuse-bridge.c:5042:fuse_graph_setup]
>>>>>>>>>>>>>>> 0-fuse: switched
>>>>>>>>>>>>>>> to graph 0
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.962351]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:188:client_set_lk_version_cbk]
>>>>>>>>>>>>>>> 0-testvol-client-1:
>>>>>>>>>>>>>>> Server lk
>>>>>>>>>>>>>>> version = 1
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.962362]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:188:client_set_lk_version_cbk]
>>>>>>>>>>>>>>> 0-testvol-client-0:
>>>>>>>>>>>>>>> Server lk
>>>>>>>>>>>>>>> version = 1
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:42:52.962424]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [fuse-bridge.c:3971:fuse_init]
>>>>>>>>>>>>>>> 0-glusterfs-fuse: FUSE
>>>>>>>>>>>>>>> inited with
>>>>>>>>>>>>>>> protocol
>>>>>>>>>>>>>>> versions:
>>>>>>>>>>>>>>> glusterfs 7.22
>>>>>>>>>>>>>>> kernel 7.14
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:13.352234]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [glusterfsd-mgmt.c:56:mgmt_cbk_spec]
>>>>>>>>>>>>>>> 0-mgmt: Volume
>>>>>>>>>>>>>>> file changed
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.518667]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [dht-shared.c:337:dht_init_regex]
>>>>>>>>>>>>>>> 2-testvol-dht:
>>>>>>>>>>>>>>> using regex
>>>>>>>>>>>>>>> rsync-hash-regex
>>>>>>>>>>>>>>> = ^\.(.+)\.[^.]+$
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.520034]
>>>>>>>>>>>>>>> W
>>>>>>>>>>>>>>> [graph.c:344:_log_if_unknown_option]
>>>>>>>>>>>>>>> 2-testvol-quota:
>>>>>>>>>>>>>>> option 'timeout'
>>>>>>>>>>>>>>> is not recognized
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.520091]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client.c:2280:notify]
>>>>>>>>>>>>>>> 2-testvol-client-0:
>>>>>>>>>>>>>>> parent
>>>>>>>>>>>>>>> translators are
>>>>>>>>>>>>>>> ready,
>>>>>>>>>>>>>>> attempting
>>>>>>>>>>>>>>> connect on transport
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.524546]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client.c:2280:notify]
>>>>>>>>>>>>>>> 2-testvol-client-1:
>>>>>>>>>>>>>>> parent
>>>>>>>>>>>>>>> translators are
>>>>>>>>>>>>>>> ready,
>>>>>>>>>>>>>>> attempting
>>>>>>>>>>>>>>> connect on transport
>>>>>>>>>>>>>>> Final graph:
>>>>>>>>>>>>>>> +------------------------------------------------------------------------------+
>>>>>>>>>>>>>>> 1: volume
>>>>>>>>>>>>>>> testvol-client-0
>>>>>>>>>>>>>>> 2: type
>>>>>>>>>>>>>>> protocol/client
>>>>>>>>>>>>>>> 3: option
>>>>>>>>>>>>>>> ping-timeout 42
>>>>>>>>>>>>>>> 4: option
>>>>>>>>>>>>>>> remote-host
>>>>>>>>>>>>>>> gluster24
>>>>>>>>>>>>>>> 5: option
>>>>>>>>>>>>>>> remote-subvolume
>>>>>>>>>>>>>>> /export25/brick
>>>>>>>>>>>>>>> 6: option
>>>>>>>>>>>>>>> transport-type
>>>>>>>>>>>>>>> socket
>>>>>>>>>>>>>>> 7: option
>>>>>>>>>>>>>>> send-gids true
>>>>>>>>>>>>>>> 8: end-volume
>>>>>>>>>>>>>>> 9:
>>>>>>>>>>>>>>> 10: volume
>>>>>>>>>>>>>>> testvol-client-1
>>>>>>>>>>>>>>> 11: type
>>>>>>>>>>>>>>> protocol/client
>>>>>>>>>>>>>>> 12: option
>>>>>>>>>>>>>>> ping-timeout 42
>>>>>>>>>>>>>>> 13: option
>>>>>>>>>>>>>>> remote-host
>>>>>>>>>>>>>>> gluster25
>>>>>>>>>>>>>>> 14: option
>>>>>>>>>>>>>>> remote-subvolume
>>>>>>>>>>>>>>> /export25/brick
>>>>>>>>>>>>>>> 15: option
>>>>>>>>>>>>>>> transport-type
>>>>>>>>>>>>>>> socket
>>>>>>>>>>>>>>> 16: option
>>>>>>>>>>>>>>> send-gids true
>>>>>>>>>>>>>>> 17: end-volume
>>>>>>>>>>>>>>> 18:
>>>>>>>>>>>>>>> 19: volume
>>>>>>>>>>>>>>> testvol-dht
>>>>>>>>>>>>>>> 20: type
>>>>>>>>>>>>>>> cluster/distribute
>>>>>>>>>>>>>>> 21: subvolumes
>>>>>>>>>>>>>>> testvol-client-0
>>>>>>>>>>>>>>> testvol-client-1
>>>>>>>>>>>>>>> 22: end-volume
>>>>>>>>>>>>>>> 23:
>>>>>>>>>>>>>>> 24: volume
>>>>>>>>>>>>>>> testvol-quota
>>>>>>>>>>>>>>> 25: type
>>>>>>>>>>>>>>> features/quota
>>>>>>>>>>>>>>> 26: option
>>>>>>>>>>>>>>> timeout 0
>>>>>>>>>>>>>>> 27: option
>>>>>>>>>>>>>>> deem-statfs off
>>>>>>>>>>>>>>> 28: subvolumes
>>>>>>>>>>>>>>> testvol-dht
>>>>>>>>>>>>>>> 29: end-volume
>>>>>>>>>>>>>>> 30:
>>>>>>>>>>>>>>> 31: volume
>>>>>>>>>>>>>>> testvol-write-behind
>>>>>>>>>>>>>>> 32: type
>>>>>>>>>>>>>>> performance/write-behind
>>>>>>>>>>>>>>> 33: subvolumes
>>>>>>>>>>>>>>> testvol-quota
>>>>>>>>>>>>>>> 34: end-volume
>>>>>>>>>>>>>>> 35:
>>>>>>>>>>>>>>> 36: volume
>>>>>>>>>>>>>>> testvol-read-ahead
>>>>>>>>>>>>>>> 37: type
>>>>>>>>>>>>>>> performance/read-ahead
>>>>>>>>>>>>>>> 38: subvolumes
>>>>>>>>>>>>>>> testvol-write-behind
>>>>>>>>>>>>>>> 39: end-volume
>>>>>>>>>>>>>>> 40:
>>>>>>>>>>>>>>> 41: volume
>>>>>>>>>>>>>>> testvol-io-cache
>>>>>>>>>>>>>>> 42: type
>>>>>>>>>>>>>>> performance/io-cache
>>>>>>>>>>>>>>> 43: subvolumes
>>>>>>>>>>>>>>> testvol-read-ahead
>>>>>>>>>>>>>>> 44: end-volume
>>>>>>>>>>>>>>> 45:
>>>>>>>>>>>>>>> 46: volume
>>>>>>>>>>>>>>> testvol-quick-read
>>>>>>>>>>>>>>> 47: type
>>>>>>>>>>>>>>> performance/quick-read
>>>>>>>>>>>>>>> 48: subvolumes
>>>>>>>>>>>>>>> testvol-io-cache
>>>>>>>>>>>>>>> 49: end-volume
>>>>>>>>>>>>>>> 50:
>>>>>>>>>>>>>>> 51: volume
>>>>>>>>>>>>>>> testvol-md-cache
>>>>>>>>>>>>>>> 52: type
>>>>>>>>>>>>>>> performance/md-cache
>>>>>>>>>>>>>>> 53: subvolumes
>>>>>>>>>>>>>>> testvol-quick-read
>>>>>>>>>>>>>>> 54: end-volume
>>>>>>>>>>>>>>> 55:
>>>>>>>>>>>>>>> 56: volume testvol
>>>>>>>>>>>>>>> 57: type
>>>>>>>>>>>>>>> debug/io-stats
>>>>>>>>>>>>>>> 58: option
>>>>>>>>>>>>>>> latency-measurement
>>>>>>>>>>>>>>> off
>>>>>>>>>>>>>>> 59: option
>>>>>>>>>>>>>>> count-fop-hits off
>>>>>>>>>>>>>>> 60: subvolumes
>>>>>>>>>>>>>>> testvol-md-cache
>>>>>>>>>>>>>>> 61: end-volume
>>>>>>>>>>>>>>> 62:
>>>>>>>>>>>>>>> 63: volume
>>>>>>>>>>>>>>> meta-autoload
>>>>>>>>>>>>>>> 64: type meta
>>>>>>>>>>>>>>> 65: subvolumes
>>>>>>>>>>>>>>> testvol
>>>>>>>>>>>>>>> 66: end-volume
>>>>>>>>>>>>>>> 67:
>>>>>>>>>>>>>>> +------------------------------------------------------------------------------+
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.530005]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [rpc-clnt.c:1759:rpc_clnt_reconfig]
>>>>>>>>>>>>>>> 2-testvol-client-1:
>>>>>>>>>>>>>>> changing port to
>>>>>>>>>>>>>>> 49155 (from 0)
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.530047]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [rpc-clnt.c:1759:rpc_clnt_reconfig]
>>>>>>>>>>>>>>> 2-testvol-client-0:
>>>>>>>>>>>>>>> changing port to
>>>>>>>>>>>>>>> 49155 (from 0)
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.539062]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:1415:select_server_supported_programs]
>>>>>>>>>>>>>>> 2-testvol-client-1:
>>>>>>>>>>>>>>> Using Program
>>>>>>>>>>>>>>> GlusterFS 3.3,
>>>>>>>>>>>>>>> Num (1298437),
>>>>>>>>>>>>>>> Version (330)
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.539299]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:1415:select_server_supported_programs]
>>>>>>>>>>>>>>> 2-testvol-client-0:
>>>>>>>>>>>>>>> Using Program
>>>>>>>>>>>>>>> GlusterFS 3.3,
>>>>>>>>>>>>>>> Num (1298437),
>>>>>>>>>>>>>>> Version (330)
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.539462]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:1200:client_setvolume_cbk]
>>>>>>>>>>>>>>> 2-testvol-client-1:
>>>>>>>>>>>>>>> Connected to
>>>>>>>>>>>>>>> testvol-client-1, attached
>>>>>>>>>>>>>>> to remote volume
>>>>>>>>>>>>>>> '/export25/brick'.
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.539485]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:1212:client_setvolume_cbk]
>>>>>>>>>>>>>>> 2-testvol-client-1:
>>>>>>>>>>>>>>> Server and
>>>>>>>>>>>>>>> Client
>>>>>>>>>>>>>>> lk-version
>>>>>>>>>>>>>>> numbers are not
>>>>>>>>>>>>>>> same, reopening
>>>>>>>>>>>>>>> the fds
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.539729]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:1200:client_setvolume_cbk]
>>>>>>>>>>>>>>> 2-testvol-client-0:
>>>>>>>>>>>>>>> Connected to
>>>>>>>>>>>>>>> testvol-client-0, attached
>>>>>>>>>>>>>>> to remote volume
>>>>>>>>>>>>>>> '/export25/brick'.
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.539751]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:1212:client_setvolume_cbk]
>>>>>>>>>>>>>>> 2-testvol-client-0:
>>>>>>>>>>>>>>> Server and
>>>>>>>>>>>>>>> Client
>>>>>>>>>>>>>>> lk-version
>>>>>>>>>>>>>>> numbers are not
>>>>>>>>>>>>>>> same, reopening
>>>>>>>>>>>>>>> the fds
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.542878]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [fuse-bridge.c:5042:fuse_graph_setup]
>>>>>>>>>>>>>>> 0-fuse: switched
>>>>>>>>>>>>>>> to graph 2
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.542959]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:188:client_set_lk_version_cbk]
>>>>>>>>>>>>>>> 2-testvol-client-1:
>>>>>>>>>>>>>>> Server lk
>>>>>>>>>>>>>>> version = 1
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:47:15.542987]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client-handshake.c:188:client_set_lk_version_cbk]
>>>>>>>>>>>>>>> 2-testvol-client-0:
>>>>>>>>>>>>>>> Server lk
>>>>>>>>>>>>>>> version = 1
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:48:04.586291]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client.c:2289:notify]
>>>>>>>>>>>>>>> 0-testvol-client-0:
>>>>>>>>>>>>>>> current graph is
>>>>>>>>>>>>>>> no longer
>>>>>>>>>>>>>>> active,
>>>>>>>>>>>>>>> destroying
>>>>>>>>>>>>>>> rpc_client
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:48:04.586360]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client.c:2289:notify]
>>>>>>>>>>>>>>> 0-testvol-client-1:
>>>>>>>>>>>>>>> current graph is
>>>>>>>>>>>>>>> no longer
>>>>>>>>>>>>>>> active,
>>>>>>>>>>>>>>> destroying
>>>>>>>>>>>>>>> rpc_client
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:48:04.586378]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client.c:2215:client_rpc_notify]
>>>>>>>>>>>>>>> 0-testvol-client-0:
>>>>>>>>>>>>>>> disconnected
>>>>>>>>>>>>>>> from
>>>>>>>>>>>>>>> testvol-client-0. Client
>>>>>>>>>>>>>>> process will
>>>>>>>>>>>>>>> keep trying to
>>>>>>>>>>>>>>> connect to
>>>>>>>>>>>>>>> glusterd until
>>>>>>>>>>>>>>> brick's port is
>>>>>>>>>>>>>>> available
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:48:04.586430]
>>>>>>>>>>>>>>> I
>>>>>>>>>>>>>>> [client.c:2215:client_rpc_notify]
>>>>>>>>>>>>>>> 0-testvol-client-1:
>>>>>>>>>>>>>>> disconnected
>>>>>>>>>>>>>>> from
>>>>>>>>>>>>>>> testvol-client-1. Client
>>>>>>>>>>>>>>> process will
>>>>>>>>>>>>>>> keep trying to
>>>>>>>>>>>>>>> connect to
>>>>>>>>>>>>>>> glusterd until
>>>>>>>>>>>>>>> brick's port is
>>>>>>>>>>>>>>> available
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:48:04.589552]
>>>>>>>>>>>>>>> W
>>>>>>>>>>>>>>> [client-rpc-fops.c:306:client3_3_mkdir_cbk]
>>>>>>>>>>>>>>> 2-testvol-client-0:
>>>>>>>>>>>>>>> remote operation
>>>>>>>>>>>>>>> failed:
>>>>>>>>>>>>>>> Transport
>>>>>>>>>>>>>>> endpoint is not
>>>>>>>>>>>>>>> connected. Path:
>>>>>>>>>>>>>>> /test/a
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:48:04.589608]
>>>>>>>>>>>>>>> W
>>>>>>>>>>>>>>> [fuse-bridge.c:481:fuse_entry_cbk]
>>>>>>>>>>>>>>> 0-glusterfs-fuse: 78:
>>>>>>>>>>>>>>> MKDIR() /test/a
>>>>>>>>>>>>>>> => -1 (Transport
>>>>>>>>>>>>>>> endpoint is not
>>>>>>>>>>>>>>> connected)
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:48:11.073349]
>>>>>>>>>>>>>>> W
>>>>>>>>>>>>>>> [client-rpc-fops.c:2212:client3_3_create_cbk]
>>>>>>>>>>>>>>> 2-testvol-client-1:
>>>>>>>>>>>>>>> remote operation
>>>>>>>>>>>>>>> failed:
>>>>>>>>>>>>>>> Transport
>>>>>>>>>>>>>>> endpoint is not
>>>>>>>>>>>>>>> connected. Path:
>>>>>>>>>>>>>>> /test/f
>>>>>>>>>>>>>>> [2015-03-20
>>>>>>>>>>>>>>> 03:48:11.073419]
>>>>>>>>>>>>>>> W
>>>>>>>>>>>>>>> [fuse-bridge.c:1937:fuse_create_cbk]
>>>>>>>>>>>>>>> 0-glusterfs-fuse: 82:
>>>>>>>>>>>>>>> /test/f => -1
>>>>>>>>>>>>>>> (Transport
>>>>>>>>>>>>>>> endpoint is not
>>>>>>>>>>>>>>> connected)
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> 2015-03-20 11:27
>>>>>>>>>>>>>>> GMT+09:00
>>>>>>>>>>>>>>> Vijaikumar M
>>>>>>>>>>>>>>> <vmallika at redhat.com
>>>>>>>>>>>>>>> <mailto:vmallika at redhat.com>>:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Hi Kondo,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Can you
>>>>>>>>>>>>>>> please
>>>>>>>>>>>>>>> provide all
>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>> glusterfs
>>>>>>>>>>>>>>> log files?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>> Vijay
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Friday 20
>>>>>>>>>>>>>>> March 2015
>>>>>>>>>>>>>>> 07:33 AM,
>>>>>>>>>>>>>>> K.Kondo wrote:
>>>>>>>>>>>>>>>> Hello, experts
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> I had a
>>>>>>>>>>>>>>>> trouble
>>>>>>>>>>>>>>>> about quota.
>>>>>>>>>>>>>>>> I set quota
>>>>>>>>>>>>>>>> to one
>>>>>>>>>>>>>>>> distributed
>>>>>>>>>>>>>>>> volume
>>>>>>>>>>>>>>>> "vol12" as
>>>>>>>>>>>>>>>> bellow.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> gluster>
>>>>>>>>>>>>>>>> volume
>>>>>>>>>>>>>>>> quota vol12
>>>>>>>>>>>>>>>> enable
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> volume
>>>>>>>>>>>>>>>> quota : success
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> gluster>
>>>>>>>>>>>>>>>> volume
>>>>>>>>>>>>>>>> quota vol12
>>>>>>>>>>>>>>>> limit-usage
>>>>>>>>>>>>>>>> /test 10GB
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> volume
>>>>>>>>>>>>>>>> quota : success
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> But I
>>>>>>>>>>>>>>>> couldn't
>>>>>>>>>>>>>>>> create a
>>>>>>>>>>>>>>>> file and
>>>>>>>>>>>>>>>> directory
>>>>>>>>>>>>>>>> with below
>>>>>>>>>>>>>>>> error message.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> In a client
>>>>>>>>>>>>>>>> host,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> $cd test
>>>>>>>>>>>>>>>> (mounted
>>>>>>>>>>>>>>>> using fuse)
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> $mkdir a
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> mkdir:
>>>>>>>>>>>>>>>> cannot
>>>>>>>>>>>>>>>> create
>>>>>>>>>>>>>>>> directory
>>>>>>>>>>>>>>>> `a':
>>>>>>>>>>>>>>>> Transport
>>>>>>>>>>>>>>>> endpoint is
>>>>>>>>>>>>>>>> not connected
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Additionally,
>>>>>>>>>>>>>>>> I couldn't
>>>>>>>>>>>>>>>> check quota
>>>>>>>>>>>>>>>> status
>>>>>>>>>>>>>>>> using
>>>>>>>>>>>>>>>> gluster
>>>>>>>>>>>>>>>> command.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> gluster>
>>>>>>>>>>>>>>>> volume
>>>>>>>>>>>>>>>> quota vol12
>>>>>>>>>>>>>>>> list
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Path
>>>>>>>>>>>>>>>> Hard-limit
>>>>>>>>>>>>>>>> Soft-limit
>>>>>>>>>>>>>>>> Used
>>>>>>>>>>>>>>>> Available
>>>>>>>>>>>>>>>> Soft-limit
>>>>>>>>>>>>>>>> exceeded?
>>>>>>>>>>>>>>>> Hard-limit
>>>>>>>>>>>>>>>> exceeded?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> ---------------------------------------------------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Here, this
>>>>>>>>>>>>>>>> command
>>>>>>>>>>>>>>>> stops, so I
>>>>>>>>>>>>>>>> have to do
>>>>>>>>>>>>>>>> Ctrl-C.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Gluster
>>>>>>>>>>>>>>>> version is
>>>>>>>>>>>>>>>> 3.6.1 and
>>>>>>>>>>>>>>>> 3.6.0.29
>>>>>>>>>>>>>>>> for server
>>>>>>>>>>>>>>>> and client
>>>>>>>>>>>>>>>> respectively.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Any idea
>>>>>>>>>>>>>>>> for this?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Best regards,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> K. Kondo
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>> Gluster-users mailing list
>>>>>>>>>>>>>>>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>>>>>>>>>>>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>
>>>
>>>
>>
>>
>>
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150421/365d8914/attachment.html>
More information about the Gluster-users
mailing list