[Gluster-users] Command "/etc/init.d/glusterd start" failed
Chalcogen
chalcogen_eg_oxygen at yahoo.com
Sat Apr 19 16:49:58 UTC 2014
I have been plagued by errors of this kind every so often, mainly
because we are in a development phase and we reboot our servers so
frequently. If you start glusterd in debug mode:
sh$ glusterd --debug
you can easily pinpoint exactly which volume/peer data is causing the
initialization failure for mgmt/glusterd.
In addition, from my own experiences, two of the leading reasons for
failure include:
a) Bad peer data if glusterd is somehow killed during an active peer
probe operation, and
b) I have noticed that if glusterd needs to update info for volume/brick
(say "info" for volume testvol) in /var/lib/glusterd, it first renames
/var/lib/glusterd/vols/testvol/info to info.tmp, and then creates a new
file info, which is probably written into _freshly_. If glusterd were to
crash at this point, it would cause failures in glusterd startup till
this is manually resolved. Usually, moving info.tmp into info works for me.
Thanks,
Anirban
On Saturday 12 April 2014 08:45 AM, 吴保川 wrote:
> It is tcp.
>
> [root at server1 wbc]# gluster volume info
>
> Volume Name: gv_replica
> Type: Replicate
> Volume ID: 81014863-ee59-409b-8897-6485d411d14d
> Status: Started
> Number of Bricks: 1 x 2 = 2
> Transport-type: tcp
> Bricks:
> Brick1: 192.168.1.3:/home/wbc/vdir/gv_replica
> Brick2: 192.168.1.4:/home/wbc/vdir/gv_replica
>
> Volume Name: gv1
> Type: Distribute
> Volume ID: cfe2b8a0-284b-489d-a153-21182933f266
> Status: Started
> Number of Bricks: 2
> Transport-type: tcp
> Bricks:
> Brick1: 192.168.1.4:/home/wbc/vdir/gv1
> Brick2: 192.168.1.3:/home/wbc/vdir/gv1
>
> Thanks,
> Baochuan Wu
>
>
>
> 2014-04-12 10:11 GMT+08:00 Nagaprasad Sathyanarayana
> <nsathyan at redhat.com <mailto:nsathyan at redhat.com>>:
>
> If you run
>
> # gluster volume info
>
> What is the value set for transport-type?
>
> Thanks
> Naga
>
>
> On 12-Apr-2014, at 7:33 am, 吴保川 <wildpointercs at gmail.com
> <mailto:wildpointercs at gmail.com>> wrote:
>
>> Thanks, Joe. I found one of my machine has been assigned wrong IP
>> address. This leads to the error.
>> Originally, I thought the following error is critical:
>> [2014-04-11 18:12:03.433371] E
>> [rpc-transport.c:269:rpc_transport_load] 0-rpc-transport:
>> /usr/local/lib/glusterfs/3.4.3/rpc-transport/rdma.so: cannot open
>> shared object file: No such file or directory
>>
>>
>> 2014-04-12 5:34 GMT+08:00 Joe Julian <joe at julianfamily.org
>> <mailto:joe at julianfamily.org>>:
>>
>> On 04/11/2014 11:18 AM, 吴保川 wrote:
>>
>> [2014-04-11 18:12:05.165989] E
>> [glusterd-store.c:2663:glusterd_resolve_all_bricks]
>> 0-glusterd: resolve brick failed in restore
>>
>> I'm pretty sure that means that one of the bricks isn't
>> resolved in your list of peers.
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140419/6f8e3768/attachment.html>
More information about the Gluster-users
mailing list