[Gluster-users] CentOS Freeze with GlusterFS Error

Deepak Shetty dpkshetty at gmail.com
Tue Jan 20 07:36:14 UTC 2015


Your response is still confusing as you say its SAN and use the word
'mounted' !
Anyways, given the above stack trace, i feel you should follow this up on
centos-devel since it mostly looks like a kernel panic to me.

I would leave other experts on this list to comment / suggest more.

thanx,
deepak

On Tue, Jan 20, 2015 at 12:36 PM, chamara samarakoon <chthsa123 at gmail.com>
wrote:

> HI,
>
>
> We have created a separate volume on SAN and it is mounted on Openstack
> controller node, there we have created multiple bricks on mounted location
> as  folders.
>
> Regards,
> Chamara
>
> On Tue, Jan 20, 2015 at 12:20 PM, Deepak Shetty <dpkshetty at gmail.com>
> wrote:
>
>> Just trying to understand the setup better....
>>
>> So you create the bricks on controller node and have the gluster volume
>> present there ?
>> Do you use the mpathXX devices as bricks ? Whats the brick type (LV or
>> dir) ?
>>
>> On Tue, Jan 20, 2015 at 12:11 PM, chamara samarakoon <chthsa123 at gmail.com
>> > wrote:
>>
>>> HI Deepak,
>>>
>>>
>>> Yes, it is  Openstack controller and SAN is configured for redundancy
>>> through multi-path.
>>>
>>> Regards,
>>> Chamara
>>>
>>>
>>> On Tue, Jan 20, 2015 at 11:55 AM, Deepak Shetty <dpkshetty at gmail.com>
>>> wrote:
>>>
>>>> What does "Controller" mean, the openstack controller node or somethign
>>>> else (like HBA ) ?
>>>> You picture says its SAN but the text says multi-path mount.. SAN would
>>>> mean block devices, so I am assuming you have redundant block devices on
>>>> the compute host, mkfs'ing it and then creating bricks for gluster ?
>>>>
>>>>
>>>> The stack trace looks like you hit a kernel bug and glusterfsd happens
>>>> to be running on the CPU at the time... my 2 cents
>>>>
>>>> thanx,
>>>> deepak
>>>>
>>>> On Tue, Jan 20, 2015 at 11:29 AM, chamara samarakoon <
>>>> chthsa123 at gmail.com> wrote:
>>>>
>>>>> Hi All,
>>>>>
>>>>>
>>>>> We have setup Openstack cloud as below. And the
>>>>> "/va/lib/nova/instances" is a Gluster volume.
>>>>>
>>>>> CentOS - 6.5
>>>>> Kernel -  2.6.32-431.29.2.el6.x86_64
>>>>> GlusterFS - glusterfs 3.5.2 built on Jul 31 2014 18:47:54
>>>>> OpenStack - RDO using Packstack
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>>>>>>
>>>>>
>>>>> Recently Controller node freezes with following error (Which required
>>>>> hard reboot), as a result Gluster volumes on compute node can not reach the
>>>>> controller and due to that all the instances on compute nodes  become to
>>>>> read-only status  which causes to restart all instances.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> *BUG: scheduling while atomic : glusterfsd/42725/0xffffffff*
>>>>> *BUG: unable to handle kernel paging request at 0000000038a60d0a8*
>>>>> *IP: [<fffffffff81058e5d>] task_rq_lock+0x4d/0xa0*
>>>>> *PGD 1065525067 PUD 0*
>>>>> *Oops: 0000 [#1] SMP*
>>>>> *last sysfs file :
>>>>> /sys/device/pci0000:80/0000:80:02.0/0000:86:00.0/host2/port-2:0/end_device-2:0/target2:0:0/2:0:0:1/state*
>>>>> *CPU 0*
>>>>> *Modules linked in : xtconntrack iptable_filter ip_tables ipt_REDIRECT
>>>>> fuse ipv openvswitch vxlan iptable_mangle *
>>>>>
>>>>> Please advice on above incident , also feedback on the Openstack +
>>>>> GlusterFS setup is appreciated.
>>>>>
>>>>> Thank You,
>>>>> Chamara
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Gluster-users mailing list
>>>>> Gluster-users at gluster.org
>>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> chthsa
>>>
>>
>>
>
>
> --
> chthsa
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150120/0370b6d9/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: openstak-gluster.png
Type: image/png
Size: 240027 bytes
Desc: not available
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150120/0370b6d9/attachment-0001.png>


More information about the Gluster-users mailing list