[Gluster-users] [ovirt-users] Gluster command [<UNKNOWN>] failed on server...

Punit Dambiwal hypunit at gmail.com
Tue Dec 9 09:54:30 UTC 2014


Dear Kaushal,

I tried various method...but still the same error...it seems it's gluster
bug..is there any body can suggest work-around here ??

Thanks,
Punit

On Sun, Dec 7, 2014 at 8:40 PM, Punit Dambiwal <hypunit at gmail.com> wrote:

> Hi Kaushal,
>
> Still the same error...even try your suggested workaround :-
>
> -------------------
> Can you replace 'Before=network-online.target' with
> 'Wants=network-online.target' and try the boot again? This should
> force the network to be online before starting GlusterD.
> -------------------
>
> Thanks,
> Punit
>
> On Sat, Dec 6, 2014 at 11:44 AM, Punit Dambiwal <hypunit at gmail.com> wrote:
>
>> Hi Kaushal,
>>
>> I already have all the hosts entry in the /etc/hosts for the easy
>> resolution....i will try your method in the glusterd.services and check and
>> let you know....weather problem solve or not.....
>>
>> On Fri, Dec 5, 2014 at 9:50 PM, Kaushal M <kshlmster at gmail.com> wrote:
>>
>>> Can you replace 'Before=network-online.target' with
>>> 'Wants=network-online.target' and try the boot again? This should
>>> force the network to be online before starting GlusterD.
>>>
>>> If even that fails, you could try adding an entry into /etc/hosts with
>>> the hostname of the system. This should prevent any more failures.
>>>
>>> I still don't believe it's a problem with Gluster. Gluster uses apis
>>> provided by the system to perform name resolution. These definitely
>>> work correctly because you can start GlusterD later. Since the
>>> resolution failure only happens during boot, it points to system or
>>> network setup issues during boot. To me it seems like the network
>>> isn't completely setup at that point of time.
>>>
>>> ~kaushal
>>>
>>> On Fri, Dec 5, 2014 at 12:47 PM, Punit Dambiwal <hypunit at gmail.com>
>>> wrote:
>>> > Hi Kaushal,
>>> >
>>> > It seems it's bug in glusterfs 3.6....even i manage my systemd to
>>> start the
>>> > network service before glusterd...but it's still fail...
>>> >
>>> > ---------------
>>> > [Unit]
>>> > Description=GlusterFS, a clustered file-system server
>>> > After=network.target rpcbind.service
>>> > Before=network-online.target
>>> >
>>> > [Service]
>>> > Type=forking
>>> > PIDFile=/var/run/glusterd.pid
>>> > LimitNOFILE=65536
>>> > ExecStartPre=/etc/rc.d/init.d/network start
>>> > ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
>>> > KillMode=process
>>> >
>>> > [Install]
>>> > WantedBy=multi-user.target
>>> > ----------------
>>> >
>>> > Thanks,
>>> > Punit
>>> >
>>> > On Wed, Dec 3, 2014 at 8:56 PM, Kaushal M <kshlmster at gmail.com> wrote:
>>> >>
>>> >> I just remembered this.
>>> >>
>>> >> There was another user having a similar issue of GlusterD failing to
>>> >> start on the mailing list a while back. The cause of his problem was
>>> >> the way his network was brought up.
>>> >> IIRC, he was using a static network configuration. The problem
>>> >> vanished when he began using dhcp. Or it might have been he was using
>>> >> dhcp.service and it got solved after switching to NetworkManager.
>>> >>
>>> >> This could be one more thing you could look at.
>>> >>
>>> >> I'll try to find the mail thread to see if it was the same problem as
>>> you.
>>> >>
>>> >> ~kaushal
>>> >>
>>> >> On Wed, Dec 3, 2014 at 6:22 PM, Kaushal M <kshlmster at gmail.com>
>>> wrote:
>>> >> > I don't know much about how the network target is brought up in
>>> >> > CentOS7, but I'll try as much as I can.
>>> >> >
>>> >> > It seems to me that, after the network has been brought up and by
>>> the
>>> >> > time GlusterD is started,
>>> >> > a. The machine hasn't yet recieved it's hostname, or
>>> >> > b. It hasn't yet registered with the name server.
>>> >> >
>>> >> > This is causing name resolution failures.
>>> >> >
>>> >> > I don't know if the network target could come up without the machine
>>> >> > getting its hostname, so I'm pretty sure it's not a.
>>> >> >
>>> >> > So it seems to be b. But these kind of signing in happens only in
>>> DDNS
>>> >> > systems, which doesn't seem to be the case for you.
>>> >> >
>>> >> > Both of these reasons might be wrong (most likely wrong). You'd do
>>> >> > good if you could ask for help from someone with more experience in
>>> >> > systemd + networking.
>>> >> >
>>> >> > ~kaushal
>>> >> >
>>> >> > On Wed, Dec 3, 2014 at 10:54 AM, Punit Dambiwal <hypunit at gmail.com>
>>> >> > wrote:
>>> >> >> Hi Kaushal,
>>> >> >>
>>> >> >> This is the host...which i rebooted...would you mind to let me
>>> know how
>>> >> >> i
>>> >> >> can make the glusterd sevice come up after network...i am using
>>> >> >> centos7...if
>>> >> >> network is the issue...
>>> >> >>
>>> >> >> On Wed, Dec 3, 2014 at 11:54 AM, Kaushal M <kshlmster at gmail.com>
>>> wrote:
>>> >> >>>
>>> >> >>> This peer cannot be identified.
>>> >> >>>
>>> >> >>> " [2014-12-03 02:29:25.998153] D
>>> >> >>> [glusterd-peer-utils.c:121:glusterd_peerinfo_find_by_hostname]
>>> >> >>> 0-management:
>>> >> >>> Unable to find friend: cpu05.zne01.hkg1.ovt.36stack.com"
>>> >> >>>
>>> >> >>> I don't know why this address is not being resolved during boot
>>> time.
>>> >> >>> If
>>> >> >>> this is a valid peer, the the only reason I can think of this
>>> that the
>>> >> >>> network is not up.
>>> >> >>>
>>> >> >>> If you had previously detached the peer forcefully, the that could
>>> >> >>> have
>>> >> >>> left stale entries in some volumes. In this case as well, GlusterD
>>> >> >>> will fail
>>> >> >>> to identify the peer.
>>> >> >>>
>>> >> >>> Do either of these reasons seem a possibility to you?
>>> >> >>>
>>> >> >>> On Dec 3, 2014 8:07 AM, "Punit Dambiwal" <hypunit at gmail.com>
>>> wrote:
>>> >> >>>>
>>> >> >>>> Hi Kaushal,
>>> >> >>>>
>>> >> >>>> Please find the logs here :- http://ur1.ca/iyoe5 and
>>> >> >>>> http://ur1.ca/iyoed
>>> >> >>>>
>>> >> >>>> On Tue, Dec 2, 2014 at 10:43 PM, Kaushal M <kshlmster at gmail.com>
>>> >> >>>> wrote:
>>> >> >>>>>
>>> >> >>>>> Hey Punit,
>>> >> >>>>> In the logs you've provided, GlusterD appears to be running
>>> >> >>>>> correctly.
>>> >> >>>>> Could you provide the logs for the time period when GlusterD
>>> >> >>>>> attempts to
>>> >> >>>>> start but fails.
>>> >> >>>>>
>>> >> >>>>> ~kaushal
>>> >> >>>>>
>>> >> >>>>> On Dec 2, 2014 8:03 PM, "Punit Dambiwal" <hypunit at gmail.com>
>>> wrote:
>>> >> >>>>>>
>>> >> >>>>>> Hi Kaushal,
>>> >> >>>>>>
>>> >> >>>>>> Please find the logs here :- http://ur1.ca/iyhs5 and
>>> >> >>>>>> http://ur1.ca/iyhue
>>> >> >>>>>>
>>> >> >>>>>> Thanks,
>>> >> >>>>>> punit
>>> >> >>>>>>
>>> >> >>>>>>
>>> >> >>>>>> On Tue, Dec 2, 2014 at 12:00 PM, Kaushal M <
>>> kshlmster at gmail.com>
>>> >> >>>>>> wrote:
>>> >> >>>>>>>
>>> >> >>>>>>> Hey Punit,
>>> >> >>>>>>> Could you start Glusterd in debug mode and provide the logs
>>> here?
>>> >> >>>>>>> To start it in debug mode, append '-LDEBUG' to the ExecStart
>>> line
>>> >> >>>>>>> in
>>> >> >>>>>>> the service file.
>>> >> >>>>>>>
>>> >> >>>>>>> ~kaushal
>>> >> >>>>>>>
>>> >> >>>>>>> On Mon, Dec 1, 2014 at 9:05 AM, Punit Dambiwal <
>>> hypunit at gmail.com>
>>> >> >>>>>>> wrote:
>>> >> >>>>>>> > Hi,
>>> >> >>>>>>> >
>>> >> >>>>>>> > Can Any body help me on this ??
>>> >> >>>>>>> >
>>> >> >>>>>>> > On Thu, Nov 27, 2014 at 9:29 AM, Punit Dambiwal
>>> >> >>>>>>> > <hypunit at gmail.com>
>>> >> >>>>>>> > wrote:
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Hi Kaushal,
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Thanks for the detailed reply....let me explain my setup
>>> first
>>> >> >>>>>>> >> :-
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> 1. Ovirt Engine
>>> >> >>>>>>> >> 2. 4* host as well as storage machine (Host and gluster
>>> >> >>>>>>> >> combined)
>>> >> >>>>>>> >> 3. Every host has 24 bricks...
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Now whenever the host machine reboot...it can come up but
>>> can
>>> >> >>>>>>> >> not
>>> >> >>>>>>> >> join the
>>> >> >>>>>>> >> cluster again and through the following error "Gluster
>>> command
>>> >> >>>>>>> >> [<UNKNOWN>]
>>> >> >>>>>>> >> failed on server.."
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Please check my comment in line :-
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> 1. Use the same string for doing the peer probe and for the
>>> >> >>>>>>> >> brick
>>> >> >>>>>>> >> address
>>> >> >>>>>>> >> during volume create/add-brick. Ideally, we suggest you use
>>> >> >>>>>>> >> properly
>>> >> >>>>>>> >> resolvable FQDNs everywhere. If that is not possible, then
>>> use
>>> >> >>>>>>> >> only
>>> >> >>>>>>> >> IP
>>> >> >>>>>>> >> addresses. Try to avoid short names.
>>> >> >>>>>>> >> ---------------
>>> >> >>>>>>> >> [root at cpu05 ~]# gluster peer status
>>> >> >>>>>>> >> Number of Peers: 3
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Hostname: cpu03.stack.com
>>> >> >>>>>>> >> Uuid: 5729b8c4-e80d-4353-b456-6f467bddbdfb
>>> >> >>>>>>> >> State: Peer in Cluster (Connected)
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Hostname: cpu04.stack.com
>>> >> >>>>>>> >> Uuid: d272b790-c4b2-4bed-ba68-793656e6d7b0
>>> >> >>>>>>> >> State: Peer in Cluster (Connected)
>>> >> >>>>>>> >> Other names:
>>> >> >>>>>>> >> 10.10.0.8
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Hostname: cpu02.stack.com
>>> >> >>>>>>> >> Uuid: 8d8a7041-950e-40d0-85f9-58d14340ca25
>>> >> >>>>>>> >> State: Peer in Cluster (Connected)
>>> >> >>>>>>> >> [root at cpu05 ~]#
>>> >> >>>>>>> >> ----------------
>>> >> >>>>>>> >> 2. During boot up, make sure to launch glusterd only after
>>> the
>>> >> >>>>>>> >> network is
>>> >> >>>>>>> >> up. This will allow the new peer identification mechanism
>>> to do
>>> >> >>>>>>> >> its
>>> >> >>>>>>> >> job correctly.
>>> >> >>>>>>> >> >> I think the service itself doing the same job....
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> [root at cpu05 ~]# cat
>>> /usr/lib/systemd/system/glusterd.service
>>> >> >>>>>>> >> [Unit]
>>> >> >>>>>>> >> Description=GlusterFS, a clustered file-system server
>>> >> >>>>>>> >> After=network.target rpcbind.service
>>> >> >>>>>>> >> Before=network-online.target
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> [Service]
>>> >> >>>>>>> >> Type=forking
>>> >> >>>>>>> >> PIDFile=/var/run/glusterd.pid
>>> >> >>>>>>> >> LimitNOFILE=65536
>>> >> >>>>>>> >> ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
>>> >> >>>>>>> >> KillMode=process
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> [Install]
>>> >> >>>>>>> >> WantedBy=multi-user.target
>>> >> >>>>>>> >> [root at cpu05 ~]#
>>> >> >>>>>>> >> --------------------
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> gluster logs :-
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> [2014-11-24 09:22:22.147471] I [MSGID: 100030]
>>> >> >>>>>>> >> [glusterfsd.c:2018:main]
>>> >> >>>>>>> >> 0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd
>>> >> >>>>>>> >> version
>>> >> >>>>>>> >> 3.6.1
>>> >> >>>>>>> >> (args: /usr/sbin/glusterd -p /var/run/glusterd.pid)
>>> >> >>>>>>> >> [2014-11-24 09:22:22.151565] I [glusterd.c:1214:init]
>>> >> >>>>>>> >> 0-management:
>>> >> >>>>>>> >> Maximum allowed open file descriptors set to 65536
>>> >> >>>>>>> >> [2014-11-24 09:22:22.151599] I [glusterd.c:1259:init]
>>> >> >>>>>>> >> 0-management:
>>> >> >>>>>>> >> Using
>>> >> >>>>>>> >> /var/lib/glusterd as working directory
>>> >> >>>>>>> >> [2014-11-24 09:22:22.155216] W
>>> >> >>>>>>> >> [rdma.c:4195:__gf_rdma_ctx_create]
>>> >> >>>>>>> >> 0-rpc-transport/rdma: rdma_cm event channel creation
>>> failed (No
>>> >> >>>>>>> >> such device)
>>> >> >>>>>>> >> [2014-11-24 09:22:22.155264] E [rdma.c:4483:init]
>>> >> >>>>>>> >> 0-rdma.management:
>>> >> >>>>>>> >> Failed to initialize IB Device
>>> >> >>>>>>> >> [2014-11-24 09:22:22.155285] E
>>> >> >>>>>>> >> [rpc-transport.c:333:rpc_transport_load]
>>> >> >>>>>>> >> 0-rpc-transport: 'rdma' initialization failed
>>> >> >>>>>>> >> [2014-11-24 09:22:22.155354] W
>>> >> >>>>>>> >> [rpcsvc.c:1524:rpcsvc_transport_create]
>>> >> >>>>>>> >> 0-rpc-service: cannot create listener, initing the
>>> transport
>>> >> >>>>>>> >> failed
>>> >> >>>>>>> >> [2014-11-24 09:22:22.156290] I
>>> >> >>>>>>> >> [glusterd.c:413:glusterd_check_gsync_present] 0-glusterd:
>>> >> >>>>>>> >> geo-replication
>>> >> >>>>>>> >> module not installed in the system
>>> >> >>>>>>> >> [2014-11-24 09:22:22.161318] I
>>> >> >>>>>>> >> [glusterd-store.c:2043:glusterd_restore_op_version]
>>> 0-glusterd:
>>> >> >>>>>>> >> retrieved
>>> >> >>>>>>> >> op-version: 30600
>>> >> >>>>>>> >> [2014-11-24 09:22:22.821800] I
>>> >> >>>>>>> >> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>>> >> >>>>>>> >> 0-management:
>>> >> >>>>>>> >> connect returned 0
>>> >> >>>>>>> >> [2014-11-24 09:22:22.825810] I
>>> >> >>>>>>> >> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>>> >> >>>>>>> >> 0-management:
>>> >> >>>>>>> >> connect returned 0
>>> >> >>>>>>> >> [2014-11-24 09:22:22.828705] I
>>> >> >>>>>>> >> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>>> >> >>>>>>> >> 0-management:
>>> >> >>>>>>> >> connect returned 0
>>> >> >>>>>>> >> [2014-11-24 09:22:22.828771] I
>>> >> >>>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init]
>>> >> >>>>>>> >> 0-management: setting frame-timeout to 600
>>> >> >>>>>>> >> [2014-11-24 09:22:22.832670] I
>>> >> >>>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init]
>>> >> >>>>>>> >> 0-management: setting frame-timeout to 600
>>> >> >>>>>>> >> [2014-11-24 09:22:22.835919] I
>>> >> >>>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init]
>>> >> >>>>>>> >> 0-management: setting frame-timeout to 600
>>> >> >>>>>>> >> [2014-11-24 09:22:22.840209] E
>>> >> >>>>>>> >> [glusterd-store.c:4248:glusterd_resolve_all_bricks]
>>> 0-glusterd:
>>> >> >>>>>>> >> resolve
>>> >> >>>>>>> >> brick failed in restore
>>> >> >>>>>>> >> [2014-11-24 09:22:22.840233] E [xlator.c:425:xlator_init]
>>> >> >>>>>>> >> 0-management:
>>> >> >>>>>>> >> Initialization of volume 'management' failed, review your
>>> >> >>>>>>> >> volfile
>>> >> >>>>>>> >> again
>>> >> >>>>>>> >> [2014-11-24 09:22:22.840245] E
>>> >> >>>>>>> >> [graph.c:322:glusterfs_graph_init]
>>> >> >>>>>>> >> 0-management: initializing translator failed
>>> >> >>>>>>> >> [2014-11-24 09:22:22.840264] E
>>> >> >>>>>>> >> [graph.c:525:glusterfs_graph_activate]
>>> >> >>>>>>> >> 0-graph: init failed
>>> >> >>>>>>> >> [2014-11-24 09:22:22.840754] W
>>> >> >>>>>>> >> [glusterfsd.c:1194:cleanup_and_exit]
>>> >> >>>>>>> >> (-->
>>> >> >>>>>>> >> 0-: received signum (0), shutting down
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Thanks,
>>> >> >>>>>>> >> Punit
>>> >> >>>>>>> >>
>>> >> >>>>>>> >>
>>> >> >>>>>>> >>
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> On Wed, Nov 26, 2014 at 7:14 PM, Kaushal M
>>> >> >>>>>>> >> <kshlmster at gmail.com>
>>> >> >>>>>>> >> wrote:
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> Based on the logs I can guess that glusterd is being
>>> started
>>> >> >>>>>>> >>> before
>>> >> >>>>>>> >>> the network has come up and that the addresses given to
>>> bricks
>>> >> >>>>>>> >>> do
>>> >> >>>>>>> >>> not
>>> >> >>>>>>> >>> directly match the addresses used in during peer probe.
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> The gluster_after_reboot log has the line "[2014-11-25
>>> >> >>>>>>> >>> 06:46:09.972113] E
>>> >> >>>>>>> >>> [glusterd-store.c:2632:glusterd_resolve_all_bricks]
>>> >> >>>>>>> >>> 0-glusterd: resolve brick failed in restore".
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> Brick resolution fails when glusterd cannot match the
>>> address
>>> >> >>>>>>> >>> for
>>> >> >>>>>>> >>> the
>>> >> >>>>>>> >>> brick, with one of the peers. Brick resolution happens in
>>> two
>>> >> >>>>>>> >>> phases,
>>> >> >>>>>>> >>> 1. We first try to identify the peer by performing string
>>> >> >>>>>>> >>> comparisions
>>> >> >>>>>>> >>> with the brick address and the peer addresses (The peer
>>> names
>>> >> >>>>>>> >>> will
>>> >> >>>>>>> >>> be
>>> >> >>>>>>> >>> the names/addresses that were given when the peer was
>>> probed).
>>> >> >>>>>>> >>> 2. If we don't find a match from step 1, we will then
>>> resolve
>>> >> >>>>>>> >>> all
>>> >> >>>>>>> >>> the
>>> >> >>>>>>> >>> brick address and the peer addresses into addrinfo
>>> structs,
>>> >> >>>>>>> >>> and
>>> >> >>>>>>> >>> then
>>> >> >>>>>>> >>> compare these structs to find a match. This process should
>>> >> >>>>>>> >>> generally
>>> >> >>>>>>> >>> find a match if available. This will fail only if the
>>> network
>>> >> >>>>>>> >>> is
>>> >> >>>>>>> >>> not
>>> >> >>>>>>> >>> up yet as we cannot resolve addresses.
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> The above steps are applicable only to glusterfs versions
>>> >> >>>>>>> >>> >=3.6.
>>> >> >>>>>>> >>> They
>>> >> >>>>>>> >>> were introduced to reduce problems with peer
>>> identification,
>>> >> >>>>>>> >>> like
>>> >> >>>>>>> >>> the
>>> >> >>>>>>> >>> one you encountered
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> Since both of the steps failed to find a match in one
>>> run, but
>>> >> >>>>>>> >>> succeeded later, we can come to the conclusion that,
>>> >> >>>>>>> >>> a) the bricks don't have the exact same string used in
>>> peer
>>> >> >>>>>>> >>> probe
>>> >> >>>>>>> >>> for
>>> >> >>>>>>> >>> their addresses as step 1 failed, and
>>> >> >>>>>>> >>> b) the network was not up in the initial run, as step 2
>>> failed
>>> >> >>>>>>> >>> during
>>> >> >>>>>>> >>> the initial run, but passed in the second run.
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> Please let me know if my conclusion is correct.
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> If it is, you can solve your problem in two ways.
>>> >> >>>>>>> >>> 1. Use the same string for doing the peer probe and for
>>> the
>>> >> >>>>>>> >>> brick
>>> >> >>>>>>> >>> address during volume create/add-brick. Ideally, we
>>> suggest
>>> >> >>>>>>> >>> you
>>> >> >>>>>>> >>> use
>>> >> >>>>>>> >>> properly resolvable FQDNs everywhere. If that is not
>>> possible,
>>> >> >>>>>>> >>> then
>>> >> >>>>>>> >>> use only IP addresses. Try to avoid short names.
>>> >> >>>>>>> >>> 2. During boot up, make sure to launch glusterd only
>>> after the
>>> >> >>>>>>> >>> network
>>> >> >>>>>>> >>> is up. This will allow the new peer identification
>>> mechanism
>>> >> >>>>>>> >>> to do
>>> >> >>>>>>> >>> its
>>> >> >>>>>>> >>> job correctly.
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> If you have already followed these steps and yet still
>>> hit the
>>> >> >>>>>>> >>> problem, then please provide more information (setup,
>>> logs,
>>> >> >>>>>>> >>> etc.).
>>> >> >>>>>>> >>> It
>>> >> >>>>>>> >>> could be much different problem that you are facing.
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> ~kaushal
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> On Wed, Nov 26, 2014 at 4:01 PM, Punit Dambiwal
>>> >> >>>>>>> >>> <hypunit at gmail.com>
>>> >> >>>>>>> >>> wrote:
>>> >> >>>>>>> >>> > Is there any one can help on this ??
>>> >> >>>>>>> >>> >
>>> >> >>>>>>> >>> > Thanks,
>>> >> >>>>>>> >>> > punit
>>> >> >>>>>>> >>> >
>>> >> >>>>>>> >>> > On Wed, Nov 26, 2014 at 9:42 AM, Punit Dambiwal
>>> >> >>>>>>> >>> > <hypunit at gmail.com>
>>> >> >>>>>>> >>> > wrote:
>>> >> >>>>>>> >>> >>
>>> >> >>>>>>> >>> >> Hi,
>>> >> >>>>>>> >>> >>
>>> >> >>>>>>> >>> >> My Glusterfs version is :- glusterfs-3.6.1-1.el7
>>> >> >>>>>>> >>> >>
>>> >> >>>>>>> >>> >> On Wed, Nov 26, 2014 at 1:59 AM, Kanagaraj Mayilsamy
>>> >> >>>>>>> >>> >> <kmayilsa at redhat.com>
>>> >> >>>>>>> >>> >> wrote:
>>> >> >>>>>>> >>> >>>
>>> >> >>>>>>> >>> >>> [+Gluster-users at gluster.org]
>>> >> >>>>>>> >>> >>>
>>> >> >>>>>>> >>> >>> "Initialization of volume 'management' failed, review
>>> your
>>> >> >>>>>>> >>> >>> volfile
>>> >> >>>>>>> >>> >>> again", glusterd throws this error when the service is
>>> >> >>>>>>> >>> >>> started
>>> >> >>>>>>> >>> >>> automatically
>>> >> >>>>>>> >>> >>> after the reboot. But the service is successfully
>>> started
>>> >> >>>>>>> >>> >>> later
>>> >> >>>>>>> >>> >>> manually by
>>> >> >>>>>>> >>> >>> the user.
>>> >> >>>>>>> >>> >>>
>>> >> >>>>>>> >>> >>> can somebody from gluster-users please help on this?
>>> >> >>>>>>> >>> >>>
>>> >> >>>>>>> >>> >>> glusterfs version: 3.5.1
>>> >> >>>>>>> >>> >>>
>>> >> >>>>>>> >>> >>> Thanks,
>>> >> >>>>>>> >>> >>> Kanagaraj
>>> >> >>>>>>> >>> >>>
>>> >> >>>>>>> >>> >>> ----- Original Message -----
>>> >> >>>>>>> >>> >>> > From: "Punit Dambiwal" <hypunit at gmail.com>
>>> >> >>>>>>> >>> >>> > To: "Kanagaraj" <kmayilsa at redhat.com>
>>> >> >>>>>>> >>> >>> > Cc: users at ovirt.org
>>> >> >>>>>>> >>> >>> > Sent: Tuesday, November 25, 2014 7:24:45 PM
>>> >> >>>>>>> >>> >>> > Subject: Re: [ovirt-users] Gluster command
>>> [<UNKNOWN>]
>>> >> >>>>>>> >>> >>> > failed on
>>> >> >>>>>>> >>> >>> > server...
>>> >> >>>>>>> >>> >>> >
>>> >> >>>>>>> >>> >>> > Hi Kanagraj,
>>> >> >>>>>>> >>> >>> >
>>> >> >>>>>>> >>> >>> > Please check the attached log files....i didn't
>>> find any
>>> >> >>>>>>> >>> >>> > thing
>>> >> >>>>>>> >>> >>> > special....
>>> >> >>>>>>> >>> >>> >
>>> >> >>>>>>> >>> >>> > On Tue, Nov 25, 2014 at 12:12 PM, Kanagaraj
>>> >> >>>>>>> >>> >>> > <kmayilsa at redhat.com>
>>> >> >>>>>>> >>> >>> > wrote:
>>> >> >>>>>>> >>> >>> >
>>> >> >>>>>>> >>> >>> > >  Do you see any errors in
>>> >> >>>>>>> >>> >>> > > /var/log/glusterfs/etc-glusterfs-glusterd.vol.log
>>> or
>>> >> >>>>>>> >>> >>> > > vdsm.log
>>> >> >>>>>>> >>> >>> > > when
>>> >> >>>>>>> >>> >>> > > the
>>> >> >>>>>>> >>> >>> > > service is trying to start automatically after the
>>> >> >>>>>>> >>> >>> > > reboot?
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > > Thanks,
>>> >> >>>>>>> >>> >>> > > Kanagaraj
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > > On 11/24/2014 08:13 PM, Punit Dambiwal wrote:
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > > Hi Kanagaraj,
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >  Yes...once i will start the gluster service and
>>> then
>>> >> >>>>>>> >>> >>> > > vdsmd
>>> >> >>>>>>> >>> >>> > > ...the
>>> >> >>>>>>> >>> >>> > > host
>>> >> >>>>>>> >>> >>> > > can connect to cluster...but the question is why
>>> it's
>>> >> >>>>>>> >>> >>> > > not
>>> >> >>>>>>> >>> >>> > > started
>>> >> >>>>>>> >>> >>> > > even it
>>> >> >>>>>>> >>> >>> > > has chkconfig enabled...
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >  I have tested it in two host cluster
>>> >> >>>>>>> >>> >>> > > environment...(Centos 6.6
>>> >> >>>>>>> >>> >>> > > and
>>> >> >>>>>>> >>> >>> > > centos 7.0) on both hypervisior cluster..it's
>>> failed
>>> >> >>>>>>> >>> >>> > > to
>>> >> >>>>>>> >>> >>> > > reconnect
>>> >> >>>>>>> >>> >>> > > in
>>> >> >>>>>>> >>> >>> > > to
>>> >> >>>>>>> >>> >>> > > cluster after reboot....
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >  In both the environment glusterd enabled for next
>>> >> >>>>>>> >>> >>> > > boot....but
>>> >> >>>>>>> >>> >>> > > it's
>>> >> >>>>>>> >>> >>> > > failed with the same error....seems it's bug in
>>> either
>>> >> >>>>>>> >>> >>> > > gluster or
>>> >> >>>>>>> >>> >>> > > Ovirt ??
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >  Please help me to find the workaround here if
>>> can not
>>> >> >>>>>>> >>> >>> > > resolve
>>> >> >>>>>>> >>> >>> > > it...as
>>> >> >>>>>>> >>> >>> > > without this the Host machine can not connect
>>> after
>>> >> >>>>>>> >>> >>> > > reboot....that
>>> >> >>>>>>> >>> >>> > > means
>>> >> >>>>>>> >>> >>> > > engine will consider it as down and every time
>>> need to
>>> >> >>>>>>> >>> >>> > > manually
>>> >> >>>>>>> >>> >>> > > start
>>> >> >>>>>>> >>> >>> > > the
>>> >> >>>>>>> >>> >>> > > gluster service and vdsmd... ??
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >  Thanks,
>>> >> >>>>>>> >>> >>> > > Punit
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > > On Mon, Nov 24, 2014 at 10:20 PM, Kanagaraj
>>> >> >>>>>>> >>> >>> > > <kmayilsa at redhat.com>
>>> >> >>>>>>> >>> >>> > > wrote:
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >>  From vdsm.log "error: Connection failed. Please
>>> >> >>>>>>> >>> >>> > >> check if
>>> >> >>>>>>> >>> >>> > >> gluster
>>> >> >>>>>>> >>> >>> > >> daemon
>>> >> >>>>>>> >>> >>> > >> is operational."
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >> Starting glusterd service should fix this issue.
>>> >> >>>>>>> >>> >>> > >> 'service
>>> >> >>>>>>> >>> >>> > >> glusterd
>>> >> >>>>>>> >>> >>> > >> start'
>>> >> >>>>>>> >>> >>> > >> But i am wondering why the glusterd was not
>>> started
>>> >> >>>>>>> >>> >>> > >> automatically
>>> >> >>>>>>> >>> >>> > >> after
>>> >> >>>>>>> >>> >>> > >> the reboot.
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >> Thanks,
>>> >> >>>>>>> >>> >>> > >> Kanagaraj
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >> On 11/24/2014 07:18 PM, Punit Dambiwal wrote:
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >> Hi Kanagaraj,
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>  Please find the attached VDSM logs :-
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>  ----------------
>>> >> >>>>>>> >>> >>> > >>  Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:17,182::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>> >> >>>>>>> >>> >>> > >> Owner.cancelAll requests {}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:17,182::task::993::Storage.TaskManager.Task::(_decref)
>>> >> >>>>>>> >>> >>> > >> Task=`1691d409-9b27-4585-8281-5ec26154367a`::ref
>>> 0
>>> >> >>>>>>> >>> >>> > >> aborting
>>> >> >>>>>>> >>> >>> > >> False
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:32,393::task::595::Storage.TaskManager.Task::(_updateState)
>>> >> >>>>>>> >>> >>> > >>
>>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving
>>> >> >>>>>>> >>> >>> > >> from
>>> >> >>>>>>> >>> >>> > >> state
>>> >> >>>>>>> >>> >>> > >> init
>>> >> >>>>>>> >>> >>> > >> ->
>>> >> >>>>>>> >>> >>> > >> state preparing
>>> >> >>>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:32,393::logUtils::44::dispatcher::(wrapper) Run
>>> >> >>>>>>> >>> >>> > >> and
>>> >> >>>>>>> >>> >>> > >> protect:
>>> >> >>>>>>> >>> >>> > >> repoStats(options=None)
>>> >> >>>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:32,393::logUtils::47::dispatcher::(wrapper) Run
>>> >> >>>>>>> >>> >>> > >> and
>>> >> >>>>>>> >>> >>> > >> protect:
>>> >> >>>>>>> >>> >>> > >> repoStats, Return response: {}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:32,393::task::1191::Storage.TaskManager.Task::(prepare)
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::finished: {}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:32,394::task::595::Storage.TaskManager.Task::(_updateState)
>>> >> >>>>>>> >>> >>> > >>
>>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving
>>> >> >>>>>>> >>> >>> > >> from
>>> >> >>>>>>> >>> >>> > >> state
>>> >> >>>>>>> >>> >>> > >> preparing
>>> >> >>>>>>> >>> >>> > >> ->
>>> >> >>>>>>> >>> >>> > >> state finished
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:32,394::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>>> >> >>>>>>> >>> >>> > >> Owner.releaseAll requests {} resources {}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:32,394::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>> >> >>>>>>> >>> >>> > >> Owner.cancelAll requests {}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:32,394::task::993::Storage.TaskManager.Task::(_decref)
>>> >> >>>>>>> >>> >>> > >> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::ref
>>> 0
>>> >> >>>>>>> >>> >>> > >> aborting
>>> >> >>>>>>> >>> >>> > >> False
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,550::BindingXMLRPC::1132::vds::(wrapper)
>>> >> >>>>>>> >>> >>> > >> client
>>> >> >>>>>>> >>> >>> > >> [10.10.10.2]::call
>>> >> >>>>>>> >>> >>> > >> getCapabilities with () {}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,553::utils::738::root::(execCmd)
>>> >> >>>>>>> >>> >>> > >> /sbin/ip route show to 0.0.0.0/0 table all (cwd
>>> None)
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,560::utils::758::root::(execCmd)
>>> >> >>>>>>> >>> >>> > >> SUCCESS: <err> = ''; <rc> = 0
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,588::caps::728::root::(_getKeyPackages)
>>> rpm
>>> >> >>>>>>> >>> >>> > >> package
>>> >> >>>>>>> >>> >>> > >> ('gluster-swift',) not found
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,592::caps::728::root::(_getKeyPackages)
>>> rpm
>>> >> >>>>>>> >>> >>> > >> package
>>> >> >>>>>>> >>> >>> > >> ('gluster-swift-object',) not found
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,593::caps::728::root::(_getKeyPackages)
>>> rpm
>>> >> >>>>>>> >>> >>> > >> package
>>> >> >>>>>>> >>> >>> > >> ('gluster-swift-plugin',) not found
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,598::caps::728::root::(_getKeyPackages)
>>> rpm
>>> >> >>>>>>> >>> >>> > >> package
>>> >> >>>>>>> >>> >>> > >> ('gluster-swift-account',) not found
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,598::caps::728::root::(_getKeyPackages)
>>> rpm
>>> >> >>>>>>> >>> >>> > >> package
>>> >> >>>>>>> >>> >>> > >> ('gluster-swift-proxy',) not found
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,598::caps::728::root::(_getKeyPackages)
>>> rpm
>>> >> >>>>>>> >>> >>> > >> package
>>> >> >>>>>>> >>> >>> > >> ('gluster-swift-doc',) not found
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,599::caps::728::root::(_getKeyPackages)
>>> rpm
>>> >> >>>>>>> >>> >>> > >> package
>>> >> >>>>>>> >>> >>> > >> ('gluster-swift-container',) not found
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,599::caps::728::root::(_getKeyPackages)
>>> rpm
>>> >> >>>>>>> >>> >>> > >> package
>>> >> >>>>>>> >>> >>> > >> ('glusterfs-geo-replication',) not found
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,600::caps::646::root::(get)
>>> >> >>>>>>> >>> >>> > >> VirtioRNG DISABLED: libvirt version
>>> 0.10.2-29.el6_5.9
>>> >> >>>>>>> >>> >>> > >> required
>>> >> >>>>>>> >>> >>> > >> >=
>>> >> >>>>>>> >>> >>> > >> 0.10.2-31
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,603::BindingXMLRPC::1139::vds::(wrapper)
>>> >> >>>>>>> >>> >>> > >> return
>>> >> >>>>>>> >>> >>> > >> getCapabilities
>>> >> >>>>>>> >>> >>> > >> with {'status': {'message': 'Done', 'code': 0},
>>> >> >>>>>>> >>> >>> > >> 'info':
>>> >> >>>>>>> >>> >>> > >> {'HBAInventory':
>>> >> >>>>>>> >>> >>> > >> {'iSCSI': [{'InitiatorName':
>>> >> >>>>>>> >>> >>> > >> 'iqn.1994-05.com.redhat:32151ce183c8'}],
>>> >> >>>>>>> >>> >>> > >> 'FC':
>>> >> >>>>>>> >>> >>> > >> []}, 'packages2': {'kernel': {'release':
>>> >> >>>>>>> >>> >>> > >> '431.el6.x86_64',
>>> >> >>>>>>> >>> >>> > >> 'buildtime':
>>> >> >>>>>>> >>> >>> > >> 1385061309.0, 'version': '2.6.32'},
>>> 'glusterfs-rdma':
>>> >> >>>>>>> >>> >>> > >> {'release':
>>> >> >>>>>>> >>> >>> > >> '1.el6',
>>> >> >>>>>>> >>> >>> > >> 'buildtime': 1403622628L, 'version': '3.5.1'},
>>> >> >>>>>>> >>> >>> > >> 'glusterfs-fuse':
>>> >> >>>>>>> >>> >>> > >> {'release': '1.el6', 'buildtime': 1403622628L,
>>> >> >>>>>>> >>> >>> > >> 'version':
>>> >> >>>>>>> >>> >>> > >> '3.5.1'},
>>> >> >>>>>>> >>> >>> > >> 'spice-server': {'release': '6.el6_5.2',
>>> 'buildtime':
>>> >> >>>>>>> >>> >>> > >> 1402324637L,
>>> >> >>>>>>> >>> >>> > >> 'version': '0.12.4'}, 'vdsm': {'release':
>>> >> >>>>>>> >>> >>> > >> '1.gitdb83943.el6',
>>> >> >>>>>>> >>> >>> > >> 'buildtime':
>>> >> >>>>>>> >>> >>> > >> 1412784567L, 'version': '4.16.7'}, 'qemu-kvm':
>>> >> >>>>>>> >>> >>> > >> {'release':
>>> >> >>>>>>> >>> >>> > >> '2.415.el6_5.10', 'buildtime': 1402435700L,
>>> >> >>>>>>> >>> >>> > >> 'version':
>>> >> >>>>>>> >>> >>> > >> '0.12.1.2'},
>>> >> >>>>>>> >>> >>> > >> 'qemu-img': {'release': '2.415.el6_5.10',
>>> >> >>>>>>> >>> >>> > >> 'buildtime':
>>> >> >>>>>>> >>> >>> > >> 1402435700L,
>>> >> >>>>>>> >>> >>> > >> 'version': '0.12.1.2'}, 'libvirt': {'release':
>>> >> >>>>>>> >>> >>> > >> '29.el6_5.9',
>>> >> >>>>>>> >>> >>> > >> 'buildtime':
>>> >> >>>>>>> >>> >>> > >> 1402404612L, 'version': '0.10.2'}, 'glusterfs':
>>> >> >>>>>>> >>> >>> > >> {'release':
>>> >> >>>>>>> >>> >>> > >> '1.el6',
>>> >> >>>>>>> >>> >>> > >> 'buildtime': 1403622628L, 'version': '3.5.1'},
>>> 'mom':
>>> >> >>>>>>> >>> >>> > >> {'release':
>>> >> >>>>>>> >>> >>> > >> '2.el6',
>>> >> >>>>>>> >>> >>> > >> 'buildtime': 1403794344L, 'version': '0.4.1'},
>>> >> >>>>>>> >>> >>> > >> 'glusterfs-server':
>>> >> >>>>>>> >>> >>> > >> {'release': '1.el6', 'buildtime': 1403622628L,
>>> >> >>>>>>> >>> >>> > >> 'version':
>>> >> >>>>>>> >>> >>> > >> '3.5.1'}},
>>> >> >>>>>>> >>> >>> > >> 'numaNodeDistance': {'1': [20, 10], '0': [10,
>>> 20]},
>>> >> >>>>>>> >>> >>> > >> 'cpuModel':
>>> >> >>>>>>> >>> >>> > >> 'Intel(R)
>>> >> >>>>>>> >>> >>> > >> Xeon(R) CPU           X5650  @ 2.67GHz',
>>> 'liveMerge':
>>> >> >>>>>>> >>> >>> > >> 'false',
>>> >> >>>>>>> >>> >>> > >> 'hooks':
>>> >> >>>>>>> >>> >>> > >> {},
>>> >> >>>>>>> >>> >>> > >> 'cpuSockets': '2', 'vmTypes': ['kvm'], 'selinux':
>>> >> >>>>>>> >>> >>> > >> {'mode': '1'},
>>> >> >>>>>>> >>> >>> > >> 'kdumpStatus': 0, 'supportedProtocols': ['2.2',
>>> >> >>>>>>> >>> >>> > >> '2.3'],
>>> >> >>>>>>> >>> >>> > >> 'networks':
>>> >> >>>>>>> >>> >>> > >> {'ovirtmgmt': {'iface': u'bond0.10', 'addr':
>>> >> >>>>>>> >>> >>> > >> '43.252.176.16',
>>> >> >>>>>>> >>> >>> > >> 'bridged':
>>> >> >>>>>>> >>> >>> > >> False, 'ipv6addrs':
>>> ['fe80::62eb:69ff:fe20:b46c/64'],
>>> >> >>>>>>> >>> >>> > >> 'mtu':
>>> >> >>>>>>> >>> >>> > >> '1500',
>>> >> >>>>>>> >>> >>> > >> 'bootproto4': 'none', 'netmask': '255.255.255.0',
>>> >> >>>>>>> >>> >>> > >> 'ipv4addrs':
>>> >> >>>>>>> >>> >>> > >> ['
>>> >> >>>>>>> >>> >>> > >> 43.252.176.16/24' <http://43.252.176.16/24%27>],
>>> >> >>>>>>> >>> >>> > >> 'interface':
>>> >> >>>>>>> >>> >>> > >> u'bond0.10', 'ipv6gateway': '::', 'gateway':
>>> >> >>>>>>> >>> >>> > >> '43.25.17.1'},
>>> >> >>>>>>> >>> >>> > >> 'Internal':
>>> >> >>>>>>> >>> >>> > >> {'iface': 'Internal', 'addr': '', 'cfg':
>>> {'DEFROUTE':
>>> >> >>>>>>> >>> >>> > >> 'no',
>>> >> >>>>>>> >>> >>> > >> 'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >> 'no', 'MTU': '9000', 'DELAY': '0',
>>> 'NM_CONTROLLED':
>>> >> >>>>>>> >>> >>> > >> 'no',
>>> >> >>>>>>> >>> >>> > >> 'BOOTPROTO':
>>> >> >>>>>>> >>> >>> > >> 'none', 'STP': 'off', 'DEVICE': 'Internal',
>>> 'TYPE':
>>> >> >>>>>>> >>> >>> > >> 'Bridge',
>>> >> >>>>>>> >>> >>> > >> 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >> 'no'}, 'bridged': True, 'ipv6addrs':
>>> >> >>>>>>> >>> >>> > >> ['fe80::210:18ff:fecd:daac/64'],
>>> >> >>>>>>> >>> >>> > >> 'gateway': '', 'bootproto4': 'none', 'netmask':
>>> '',
>>> >> >>>>>>> >>> >>> > >> 'stp':
>>> >> >>>>>>> >>> >>> > >> 'off',
>>> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [], 'mtu': '9000', 'ipv6gateway':
>>> '::',
>>> >> >>>>>>> >>> >>> > >> 'ports':
>>> >> >>>>>>> >>> >>> > >> ['bond1.100']}, 'storage': {'iface': u'bond1',
>>> >> >>>>>>> >>> >>> > >> 'addr':
>>> >> >>>>>>> >>> >>> > >> '10.10.10.6',
>>> >> >>>>>>> >>> >>> > >> 'bridged': False, 'ipv6addrs':
>>> >> >>>>>>> >>> >>> > >> ['fe80::210:18ff:fecd:daac/64'],
>>> >> >>>>>>> >>> >>> > >> 'mtu':
>>> >> >>>>>>> >>> >>> > >> '9000', 'bootproto4': 'none', 'netmask':
>>> >> >>>>>>> >>> >>> > >> '255.255.255.0',
>>> >> >>>>>>> >>> >>> > >> 'ipv4addrs': ['
>>> >> >>>>>>> >>> >>> > >> 10.10.10.6/24' <http://10.10.10.6/24%27>],
>>> >> >>>>>>> >>> >>> > >> 'interface':
>>> >> >>>>>>> >>> >>> > >> u'bond1',
>>> >> >>>>>>> >>> >>> > >> 'ipv6gateway': '::', 'gateway': ''}, 'VMNetwork':
>>> >> >>>>>>> >>> >>> > >> {'iface':
>>> >> >>>>>>> >>> >>> > >> 'VMNetwork',
>>> >> >>>>>>> >>> >>> > >> 'addr': '', 'cfg': {'DEFROUTE': 'no', 'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >> 'no',
>>> >> >>>>>>> >>> >>> > >> 'MTU':
>>> >> >>>>>>> >>> >>> > >> '1500',
>>> >> >>>>>>> >>> >>> > >> 'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO':
>>> >> >>>>>>> >>> >>> > >> 'none',
>>> >> >>>>>>> >>> >>> > >> 'STP':
>>> >> >>>>>>> >>> >>> > >> 'off',
>>> >> >>>>>>> >>> >>> > >> 'DEVICE': 'VMNetwork', 'TYPE': 'Bridge',
>>> 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >> 'no'},
>>> >> >>>>>>> >>> >>> > >> 'bridged':
>>> >> >>>>>>> >>> >>> > >> True,
>>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
>>> >> >>>>>>> >>> >>> > >> 'gateway':
>>> >> >>>>>>> >>> >>> > >> '',
>>> >> >>>>>>> >>> >>> > >> 'bootproto4':
>>> >> >>>>>>> >>> >>> > >> 'none', 'netmask': '', 'stp': 'off',
>>> 'ipv4addrs': [],
>>> >> >>>>>>> >>> >>> > >> 'mtu':
>>> >> >>>>>>> >>> >>> > >> '1500',
>>> >> >>>>>>> >>> >>> > >> 'ipv6gateway': '::', 'ports': ['bond0.36']}},
>>> >> >>>>>>> >>> >>> > >> 'bridges':
>>> >> >>>>>>> >>> >>> > >> {'Internal':
>>> >> >>>>>>> >>> >>> > >> {'addr': '', 'cfg': {'DEFROUTE': 'no', 'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >> 'no',
>>> >> >>>>>>> >>> >>> > >> 'MTU':
>>> >> >>>>>>> >>> >>> > >> '9000',
>>> >> >>>>>>> >>> >>> > >> 'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO':
>>> >> >>>>>>> >>> >>> > >> 'none',
>>> >> >>>>>>> >>> >>> > >> 'STP':
>>> >> >>>>>>> >>> >>> > >> 'off',
>>> >> >>>>>>> >>> >>> > >> 'DEVICE': 'Internal', 'TYPE': 'Bridge', 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >> 'no'},
>>> >> >>>>>>> >>> >>> > >> 'ipv6addrs':
>>> >> >>>>>>> >>> >>> > >> ['fe80::210:18ff:fecd:daac/64'], 'mtu': '9000',
>>> >> >>>>>>> >>> >>> > >> 'netmask': '',
>>> >> >>>>>>> >>> >>> > >> 'stp':
>>> >> >>>>>>> >>> >>> > >> 'off', 'ipv4addrs': [], 'ipv6gateway': '::',
>>> >> >>>>>>> >>> >>> > >> 'gateway':
>>> >> >>>>>>> >>> >>> > >> '',
>>> >> >>>>>>> >>> >>> > >> 'opts':
>>> >> >>>>>>> >>> >>> > >> {'topology_change_detected': '0',
>>> >> >>>>>>> >>> >>> > >> 'multicast_last_member_count':
>>> >> >>>>>>> >>> >>> > >> '2',
>>> >> >>>>>>> >>> >>> > >> 'hash_elasticity': '4',
>>> >> >>>>>>> >>> >>> > >> 'multicast_query_response_interval':
>>> >> >>>>>>> >>> >>> > >> '999',
>>> >> >>>>>>> >>> >>> > >> 'multicast_snooping': '1',
>>> >> >>>>>>> >>> >>> > >> 'multicast_startup_query_interval':
>>> >> >>>>>>> >>> >>> > >> '3124',
>>> >> >>>>>>> >>> >>> > >> 'hello_timer': '31',
>>> 'multicast_querier_interval':
>>> >> >>>>>>> >>> >>> > >> '25496',
>>> >> >>>>>>> >>> >>> > >> 'max_age':
>>> >> >>>>>>> >>> >>> > >> '1999', 'hash_max': '512', 'stp_state': '0',
>>> >> >>>>>>> >>> >>> > >> 'root_id':
>>> >> >>>>>>> >>> >>> > >> '8000.001018cddaac', 'priority': '32768',
>>> >> >>>>>>> >>> >>> > >> 'multicast_membership_interval':
>>> >> >>>>>>> >>> >>> > >> '25996', 'root_path_cost': '0', 'root_port': '0',
>>> >> >>>>>>> >>> >>> > >> 'multicast_querier':
>>> >> >>>>>>> >>> >>> > >> '0',
>>> >> >>>>>>> >>> >>> > >> 'multicast_startup_query_count': '2',
>>> 'hello_time':
>>> >> >>>>>>> >>> >>> > >> '199',
>>> >> >>>>>>> >>> >>> > >> 'topology_change': '0', 'bridge_id':
>>> >> >>>>>>> >>> >>> > >> '8000.001018cddaac',
>>> >> >>>>>>> >>> >>> > >> 'topology_change_timer': '0', 'ageing_time':
>>> '29995',
>>> >> >>>>>>> >>> >>> > >> 'gc_timer':
>>> >> >>>>>>> >>> >>> > >> '31',
>>> >> >>>>>>> >>> >>> > >> 'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0',
>>> >> >>>>>>> >>> >>> > >> 'multicast_query_interval': '12498',
>>> >> >>>>>>> >>> >>> > >> 'multicast_last_member_interval':
>>> >> >>>>>>> >>> >>> > >> '99', 'multicast_router': '1', 'forward_delay':
>>> '0'},
>>> >> >>>>>>> >>> >>> > >> 'ports':
>>> >> >>>>>>> >>> >>> > >> ['bond1.100']}, 'VMNetwork': {'addr': '', 'cfg':
>>> >> >>>>>>> >>> >>> > >> {'DEFROUTE':
>>> >> >>>>>>> >>> >>> > >> 'no',
>>> >> >>>>>>> >>> >>> > >> 'HOTPLUG': 'no', 'MTU': '1500', 'DELAY': '0',
>>> >> >>>>>>> >>> >>> > >> 'NM_CONTROLLED':
>>> >> >>>>>>> >>> >>> > >> 'no',
>>> >> >>>>>>> >>> >>> > >> 'BOOTPROTO': 'none', 'STP': 'off', 'DEVICE':
>>> >> >>>>>>> >>> >>> > >> 'VMNetwork',
>>> >> >>>>>>> >>> >>> > >> 'TYPE':
>>> >> >>>>>>> >>> >>> > >> 'Bridge',
>>> >> >>>>>>> >>> >>> > >> 'ONBOOT': 'no'}, 'ipv6addrs':
>>> >> >>>>>>> >>> >>> > >> ['fe80::62eb:69ff:fe20:b46c/64'],
>>> >> >>>>>>> >>> >>> > >> 'mtu':
>>> >> >>>>>>> >>> >>> > >> '1500', 'netmask': '', 'stp': 'off',
>>> 'ipv4addrs': [],
>>> >> >>>>>>> >>> >>> > >> 'ipv6gateway':
>>> >> >>>>>>> >>> >>> > >> '::',
>>> >> >>>>>>> >>> >>> > >> 'gateway': '', 'opts':
>>> {'topology_change_detected':
>>> >> >>>>>>> >>> >>> > >> '0',
>>> >> >>>>>>> >>> >>> > >> 'multicast_last_member_count': '2',
>>> >> >>>>>>> >>> >>> > >> 'hash_elasticity':
>>> >> >>>>>>> >>> >>> > >> '4',
>>> >> >>>>>>> >>> >>> > >> 'multicast_query_response_interval': '999',
>>> >> >>>>>>> >>> >>> > >> 'multicast_snooping':
>>> >> >>>>>>> >>> >>> > >> '1',
>>> >> >>>>>>> >>> >>> > >> 'multicast_startup_query_interval': '3124',
>>> >> >>>>>>> >>> >>> > >> 'hello_timer':
>>> >> >>>>>>> >>> >>> > >> '131',
>>> >> >>>>>>> >>> >>> > >> 'multicast_querier_interval': '25496', 'max_age':
>>> >> >>>>>>> >>> >>> > >> '1999',
>>> >> >>>>>>> >>> >>> > >> 'hash_max':
>>> >> >>>>>>> >>> >>> > >> '512', 'stp_state': '0', 'root_id':
>>> >> >>>>>>> >>> >>> > >> '8000.60eb6920b46c',
>>> >> >>>>>>> >>> >>> > >> 'priority':
>>> >> >>>>>>> >>> >>> > >> '32768', 'multicast_membership_interval':
>>> '25996',
>>> >> >>>>>>> >>> >>> > >> 'root_path_cost':
>>> >> >>>>>>> >>> >>> > >> '0',
>>> >> >>>>>>> >>> >>> > >> 'root_port': '0', 'multicast_querier': '0',
>>> >> >>>>>>> >>> >>> > >> 'multicast_startup_query_count': '2',
>>> 'hello_time':
>>> >> >>>>>>> >>> >>> > >> '199',
>>> >> >>>>>>> >>> >>> > >> 'topology_change': '0', 'bridge_id':
>>> >> >>>>>>> >>> >>> > >> '8000.60eb6920b46c',
>>> >> >>>>>>> >>> >>> > >> 'topology_change_timer': '0', 'ageing_time':
>>> '29995',
>>> >> >>>>>>> >>> >>> > >> 'gc_timer':
>>> >> >>>>>>> >>> >>> > >> '31',
>>> >> >>>>>>> >>> >>> > >> 'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0',
>>> >> >>>>>>> >>> >>> > >> 'multicast_query_interval': '12498',
>>> >> >>>>>>> >>> >>> > >> 'multicast_last_member_interval':
>>> >> >>>>>>> >>> >>> > >> '99', 'multicast_router': '1', 'forward_delay':
>>> '0'},
>>> >> >>>>>>> >>> >>> > >> 'ports':
>>> >> >>>>>>> >>> >>> > >> ['bond0.36']}}, 'uuid':
>>> >> >>>>>>> >>> >>> > >> '44454C4C-4C00-1057-8053-B7C04F504E31',
>>> >> >>>>>>> >>> >>> > >> 'lastClientIface': 'bond1', 'nics': {'eth3':
>>> >> >>>>>>> >>> >>> > >> {'permhwaddr':
>>> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ae', 'addr': '', 'cfg': {'SLAVE':
>>> >> >>>>>>> >>> >>> > >> 'yes',
>>> >> >>>>>>> >>> >>> > >> 'NM_CONTROLLED':
>>> >> >>>>>>> >>> >>> > >> 'no', 'MTU': '9000', 'HWADDR':
>>> '00:10:18:cd:da:ae',
>>> >> >>>>>>> >>> >>> > >> 'MASTER':
>>> >> >>>>>>> >>> >>> > >> 'bond1',
>>> >> >>>>>>> >>> >>> > >> 'DEVICE': 'eth3', 'ONBOOT': 'no'}, 'ipv6addrs':
>>> [],
>>> >> >>>>>>> >>> >>> > >> 'mtu':
>>> >> >>>>>>> >>> >>> > >> '9000',
>>> >> >>>>>>> >>> >>> > >> 'netmask': '', 'ipv4addrs': [], 'hwaddr':
>>> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ac',
>>> >> >>>>>>> >>> >>> > >> 'speed':
>>> >> >>>>>>> >>> >>> > >> 1000}, 'eth2': {'permhwaddr':
>>> '00:10:18:cd:da:ac',
>>> >> >>>>>>> >>> >>> > >> 'addr': '',
>>> >> >>>>>>> >>> >>> > >> 'cfg':
>>> >> >>>>>>> >>> >>> > >> {'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU':
>>> >> >>>>>>> >>> >>> > >> '9000',
>>> >> >>>>>>> >>> >>> > >> 'HWADDR':
>>> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ac', 'MASTER': 'bond1', 'DEVICE':
>>> >> >>>>>>> >>> >>> > >> 'eth2',
>>> >> >>>>>>> >>> >>> > >> 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >> 'no'},
>>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': [], 'mtu': '9000', 'netmask': '',
>>> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [],
>>> >> >>>>>>> >>> >>> > >> 'hwaddr':
>>> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ac', 'speed': 1000}, 'eth1':
>>> >> >>>>>>> >>> >>> > >> {'permhwaddr':
>>> >> >>>>>>> >>> >>> > >> '60:eb:69:20:b4:6d', 'addr': '', 'cfg': {'SLAVE':
>>> >> >>>>>>> >>> >>> > >> 'yes',
>>> >> >>>>>>> >>> >>> > >> 'NM_CONTROLLED':
>>> >> >>>>>>> >>> >>> > >> 'no', 'MTU': '1500', 'HWADDR':
>>> '60:eb:69:20:b4:6d',
>>> >> >>>>>>> >>> >>> > >> 'MASTER':
>>> >> >>>>>>> >>> >>> > >> 'bond0',
>>> >> >>>>>>> >>> >>> > >> 'DEVICE': 'eth1', 'ONBOOT': 'yes'}, 'ipv6addrs':
>>> [],
>>> >> >>>>>>> >>> >>> > >> 'mtu':
>>> >> >>>>>>> >>> >>> > >> '1500',
>>> >> >>>>>>> >>> >>> > >> 'netmask': '', 'ipv4addrs': [], 'hwaddr':
>>> >> >>>>>>> >>> >>> > >> '60:eb:69:20:b4:6c',
>>> >> >>>>>>> >>> >>> > >> 'speed':
>>> >> >>>>>>> >>> >>> > >> 1000}, 'eth0': {'permhwaddr':
>>> '60:eb:69:20:b4:6c',
>>> >> >>>>>>> >>> >>> > >> 'addr': '',
>>> >> >>>>>>> >>> >>> > >> 'cfg':
>>> >> >>>>>>> >>> >>> > >> {'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU':
>>> >> >>>>>>> >>> >>> > >> '1500',
>>> >> >>>>>>> >>> >>> > >> 'HWADDR':
>>> >> >>>>>>> >>> >>> > >> '60:eb:69:20:b4:6c', 'MASTER': 'bond0', 'DEVICE':
>>> >> >>>>>>> >>> >>> > >> 'eth0',
>>> >> >>>>>>> >>> >>> > >> 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >> 'yes'},
>>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': [], 'mtu': '1500', 'netmask': '',
>>> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [],
>>> >> >>>>>>> >>> >>> > >> 'hwaddr':
>>> >> >>>>>>> >>> >>> > >> '60:eb:69:20:b4:6c', 'speed': 1000}},
>>> >> >>>>>>> >>> >>> > >> 'software_revision': '1',
>>> >> >>>>>>> >>> >>> > >> 'clusterLevels': ['3.0', '3.1', '3.2', '3.3',
>>> '3.4',
>>> >> >>>>>>> >>> >>> > >> '3.5'],
>>> >> >>>>>>> >>> >>> > >> 'cpuFlags':
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> u'fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,pdpe1gb,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,xtopology,nonstop_tsc,pni,pclmulqdq,dtes64,monitor,ds_cpl,vmx,smx,est,tm2,ssse3,cx16,xtpr,pdcm,pcid,dca,sse4_1,sse4_2,popcnt,aes,lahf_lm,tpr_shadow,vnmi,flexpriority,ept,vpid,model_Nehalem,model_Conroe,model_coreduo,model_core2duo,model_Penryn,model_Westmere,model_n270',
>>> >> >>>>>>> >>> >>> > >> 'ISCSIInitiatorName':
>>> >> >>>>>>> >>> >>> > >> 'iqn.1994-05.com.redhat:32151ce183c8',
>>> >> >>>>>>> >>> >>> > >> 'netConfigDirty': 'False', 'supportedENGINEs':
>>> >> >>>>>>> >>> >>> > >> ['3.0',
>>> >> >>>>>>> >>> >>> > >> '3.1',
>>> >> >>>>>>> >>> >>> > >> '3.2',
>>> >> >>>>>>> >>> >>> > >> '3.3',
>>> >> >>>>>>> >>> >>> > >> '3.4', '3.5'], 'autoNumaBalancing': 2,
>>> 'reservedMem':
>>> >> >>>>>>> >>> >>> > >> '321',
>>> >> >>>>>>> >>> >>> > >> 'bondings':
>>> >> >>>>>>> >>> >>> > >> {'bond4': {'addr': '', 'cfg': {}, 'mtu': '1500',
>>> >> >>>>>>> >>> >>> > >> 'netmask': '',
>>> >> >>>>>>> >>> >>> > >> 'slaves':
>>> >> >>>>>>> >>> >>> > >> [], 'hwaddr': '00:00:00:00:00:00'}, 'bond0':
>>> {'addr':
>>> >> >>>>>>> >>> >>> > >> '',
>>> >> >>>>>>> >>> >>> > >> 'cfg':
>>> >> >>>>>>> >>> >>> > >> {'HOTPLUG': 'no', 'MTU': '1500', 'NM_CONTROLLED':
>>> >> >>>>>>> >>> >>> > >> 'no',
>>> >> >>>>>>> >>> >>> > >> 'BONDING_OPTS':
>>> >> >>>>>>> >>> >>> > >> 'mode=4 miimon=100', 'DEVICE': 'bond0', 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >> 'yes'},
>>> >> >>>>>>> >>> >>> > >> 'ipv6addrs':
>>> >> >>>>>>> >>> >>> > >> ['fe80::62eb:69ff:fe20:b46c/64'], 'mtu': '1500',
>>> >> >>>>>>> >>> >>> > >> 'netmask': '',
>>> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [], 'hwaddr': '60:eb:69:20:b4:6c',
>>> >> >>>>>>> >>> >>> > >> 'slaves':
>>> >> >>>>>>> >>> >>> > >> ['eth0',
>>> >> >>>>>>> >>> >>> > >> 'eth1'],
>>> >> >>>>>>> >>> >>> > >> 'opts': {'miimon': '100', 'mode': '4'}}, 'bond1':
>>> >> >>>>>>> >>> >>> > >> {'addr':
>>> >> >>>>>>> >>> >>> > >> '10.10.10.6',
>>> >> >>>>>>> >>> >>> > >> 'cfg': {'DEFROUTE': 'no', 'IPADDR': '10.10.10.6',
>>> >> >>>>>>> >>> >>> > >> 'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >> 'no',
>>> >> >>>>>>> >>> >>> > >> 'MTU':
>>> >> >>>>>>> >>> >>> > >> '9000', 'NM_CONTROLLED': 'no', 'NETMASK':
>>> >> >>>>>>> >>> >>> > >> '255.255.255.0',
>>> >> >>>>>>> >>> >>> > >> 'BOOTPROTO':
>>> >> >>>>>>> >>> >>> > >> 'none', 'BONDING_OPTS': 'mode=4 miimon=100',
>>> >> >>>>>>> >>> >>> > >> 'DEVICE':
>>> >> >>>>>>> >>> >>> > >> 'bond1',
>>> >> >>>>>>> >>> >>> > >> 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >> 'no'}, 'ipv6addrs':
>>> ['fe80::210:18ff:fecd:daac/64'],
>>> >> >>>>>>> >>> >>> > >> 'mtu':
>>> >> >>>>>>> >>> >>> > >> '9000',
>>> >> >>>>>>> >>> >>> > >> 'netmask': '255.255.255.0', 'ipv4addrs':
>>> >> >>>>>>> >>> >>> > >> ['10.10.10.6/24'
>>> >> >>>>>>> >>> >>> > >> <http://10.10.10.6/24%27>], 'hwaddr':
>>> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ac',
>>> >> >>>>>>> >>> >>> > >> 'slaves':
>>> >> >>>>>>> >>> >>> > >> ['eth2', 'eth3'], 'opts': {'miimon': '100',
>>> 'mode':
>>> >> >>>>>>> >>> >>> > >> '4'}},
>>> >> >>>>>>> >>> >>> > >> 'bond2':
>>> >> >>>>>>> >>> >>> > >> {'addr': '', 'cfg': {}, 'mtu': '1500',
>>> 'netmask': '',
>>> >> >>>>>>> >>> >>> > >> 'slaves':
>>> >> >>>>>>> >>> >>> > >> [],
>>> >> >>>>>>> >>> >>> > >> 'hwaddr': '00:00:00:00:00:00'}, 'bond3':
>>> {'addr': '',
>>> >> >>>>>>> >>> >>> > >> 'cfg': {},
>>> >> >>>>>>> >>> >>> > >> 'mtu':
>>> >> >>>>>>> >>> >>> > >> '1500', 'netmask': '', 'slaves': [], 'hwaddr':
>>> >> >>>>>>> >>> >>> > >> '00:00:00:00:00:00'}},
>>> >> >>>>>>> >>> >>> > >> 'software_version': '4.16', 'memSize': '24019',
>>> >> >>>>>>> >>> >>> > >> 'cpuSpeed':
>>> >> >>>>>>> >>> >>> > >> '2667.000',
>>> >> >>>>>>> >>> >>> > >> 'numaNodes': {u'1': {'totalMemory': '12288',
>>> 'cpus':
>>> >> >>>>>>> >>> >>> > >> [6,
>>> >> >>>>>>> >>> >>> > >> 7, 8,
>>> >> >>>>>>> >>> >>> > >> 9,
>>> >> >>>>>>> >>> >>> > >> 10, 11,
>>> >> >>>>>>> >>> >>> > >> 18, 19, 20, 21, 22, 23]}, u'0': {'totalMemory':
>>> >> >>>>>>> >>> >>> > >> '12278',
>>> >> >>>>>>> >>> >>> > >> 'cpus':
>>> >> >>>>>>> >>> >>> > >> [0,
>>> >> >>>>>>> >>> >>> > >> 1, 2,
>>> >> >>>>>>> >>> >>> > >> 3, 4, 5, 12, 13, 14, 15, 16, 17]}},
>>> 'version_name':
>>> >> >>>>>>> >>> >>> > >> 'Snow
>>> >> >>>>>>> >>> >>> > >> Man',
>>> >> >>>>>>> >>> >>> > >> 'vlans':
>>> >> >>>>>>> >>> >>> > >> {'bond0.10': {'iface': 'bond0', 'addr':
>>> >> >>>>>>> >>> >>> > >> '43.25.17.16',
>>> >> >>>>>>> >>> >>> > >> 'cfg':
>>> >> >>>>>>> >>> >>> > >> {'DEFROUTE':
>>> >> >>>>>>> >>> >>> > >> 'yes', 'VLAN': 'yes', 'IPADDR': '43.25.17.16',
>>> >> >>>>>>> >>> >>> > >> 'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >> 'no',
>>> >> >>>>>>> >>> >>> > >> 'GATEWAY':
>>> >> >>>>>>> >>> >>> > >> '43.25.17.1', 'NM_CONTROLLED': 'no', 'NETMASK':
>>> >> >>>>>>> >>> >>> > >> '255.255.255.0',
>>> >> >>>>>>> >>> >>> > >> 'BOOTPROTO': 'none', 'DEVICE': 'bond0.10', 'MTU':
>>> >> >>>>>>> >>> >>> > >> '1500',
>>> >> >>>>>>> >>> >>> > >> 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >> 'yes'},
>>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
>>> >> >>>>>>> >>> >>> > >> 'vlanid':
>>> >> >>>>>>> >>> >>> > >> 10,
>>> >> >>>>>>> >>> >>> > >> 'mtu':
>>> >> >>>>>>> >>> >>> > >> '1500',
>>> >> >>>>>>> >>> >>> > >> 'netmask': '255.255.255.0', 'ipv4addrs':
>>> >> >>>>>>> >>> >>> > >> ['43.25.17.16/24']
>>> >> >>>>>>> >>> >>> > >> <http://43.25.17.16/24%27%5D>}, 'bond0.36':
>>> {'iface':
>>> >> >>>>>>> >>> >>> > >> 'bond0',
>>> >> >>>>>>> >>> >>> > >> 'addr':
>>> >> >>>>>>> >>> >>> > >> '', 'cfg': {'BRIDGE': 'VMNetwork', 'VLAN': 'yes',
>>> >> >>>>>>> >>> >>> > >> 'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >> 'no',
>>> >> >>>>>>> >>> >>> > >> 'MTU':
>>> >> >>>>>>> >>> >>> > >> '1500', 'NM_CONTROLLED': 'no', 'DEVICE':
>>> 'bond0.36',
>>> >> >>>>>>> >>> >>> > >> 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >> 'no'},
>>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
>>> >> >>>>>>> >>> >>> > >> 'vlanid':
>>> >> >>>>>>> >>> >>> > >> 36,
>>> >> >>>>>>> >>> >>> > >> 'mtu':
>>> >> >>>>>>> >>> >>> > >> '1500',
>>> >> >>>>>>> >>> >>> > >> 'netmask': '', 'ipv4addrs': []}, 'bond1.100':
>>> >> >>>>>>> >>> >>> > >> {'iface':
>>> >> >>>>>>> >>> >>> > >> 'bond1',
>>> >> >>>>>>> >>> >>> > >> 'addr':
>>> >> >>>>>>> >>> >>> > >> '', 'cfg': {'BRIDGE': 'Internal', 'VLAN': 'yes',
>>> >> >>>>>>> >>> >>> > >> 'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >> 'no',
>>> >> >>>>>>> >>> >>> > >> 'MTU':
>>> >> >>>>>>> >>> >>> > >> '9000', 'NM_CONTROLLED': 'no', 'DEVICE':
>>> 'bond1.100',
>>> >> >>>>>>> >>> >>> > >> 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >> 'no'},
>>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': ['fe80::210:18ff:fecd:daac/64'],
>>> >> >>>>>>> >>> >>> > >> 'vlanid':
>>> >> >>>>>>> >>> >>> > >> 100,
>>> >> >>>>>>> >>> >>> > >> 'mtu':
>>> >> >>>>>>> >>> >>> > >> '9000',
>>> >> >>>>>>> >>> >>> > >> 'netmask': '', 'ipv4addrs': []}}, 'cpuCores':
>>> '12',
>>> >> >>>>>>> >>> >>> > >> 'kvmEnabled':
>>> >> >>>>>>> >>> >>> > >> 'true',
>>> >> >>>>>>> >>> >>> > >> 'guestOverhead': '65', 'cpuThreads': '24',
>>> >> >>>>>>> >>> >>> > >> 'emulatedMachines':
>>> >> >>>>>>> >>> >>> > >> [u'rhel6.5.0', u'pc', u'rhel6.4.0', u'rhel6.3.0',
>>> >> >>>>>>> >>> >>> > >> u'rhel6.2.0',
>>> >> >>>>>>> >>> >>> > >> u'rhel6.1.0', u'rhel6.0.0', u'rhel5.5.0',
>>> >> >>>>>>> >>> >>> > >> u'rhel5.4.4',
>>> >> >>>>>>> >>> >>> > >> u'rhel5.4.0'],
>>> >> >>>>>>> >>> >>> > >> 'operatingSystem': {'release':
>>> '5.el6.centos.11.1',
>>> >> >>>>>>> >>> >>> > >> 'version':
>>> >> >>>>>>> >>> >>> > >> '6',
>>> >> >>>>>>> >>> >>> > >> 'name':
>>> >> >>>>>>> >>> >>> > >> 'RHEL'}, 'lastClient': '10.10.10.2'}}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,620::BindingXMLRPC::1132::vds::(wrapper)
>>> >> >>>>>>> >>> >>> > >> client
>>> >> >>>>>>> >>> >>> > >> [10.10.10.2]::call
>>> >> >>>>>>> >>> >>> > >> getHardwareInfo with () {}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,621::BindingXMLRPC::1139::vds::(wrapper)
>>> >> >>>>>>> >>> >>> > >> return
>>> >> >>>>>>> >>> >>> > >> getHardwareInfo
>>> >> >>>>>>> >>> >>> > >> with {'status': {'message': 'Done', 'code': 0},
>>> >> >>>>>>> >>> >>> > >> 'info':
>>> >> >>>>>>> >>> >>> > >> {'systemProductName': 'CS24-TY',
>>> >> >>>>>>> >>> >>> > >> 'systemSerialNumber':
>>> >> >>>>>>> >>> >>> > >> '7LWSPN1',
>>> >> >>>>>>> >>> >>> > >> 'systemFamily': 'Server', 'systemVersion': 'A00',
>>> >> >>>>>>> >>> >>> > >> 'systemUUID':
>>> >> >>>>>>> >>> >>> > >> '44454c4c-4c00-1057-8053-b7c04f504e31',
>>> >> >>>>>>> >>> >>> > >> 'systemManufacturer':
>>> >> >>>>>>> >>> >>> > >> 'Dell'}}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:41,733::BindingXMLRPC::1132::vds::(wrapper)
>>> >> >>>>>>> >>> >>> > >> client
>>> >> >>>>>>> >>> >>> > >> [10.10.10.2]::call
>>> >> >>>>>>> >>> >>> > >> hostsList with () {} flowID [222e8036]
>>> >> >>>>>>> >>> >>> > >> Thread-13::ERROR::2014-11-24
>>> >> >>>>>>> >>> >>> > >> 21:41:44,753::BindingXMLRPC::1148::vds::(wrapper)
>>> >> >>>>>>> >>> >>> > >> vdsm
>>> >> >>>>>>> >>> >>> > >> exception
>>> >> >>>>>>> >>> >>> > >> occured
>>> >> >>>>>>> >>> >>> > >> Traceback (most recent call last):
>>> >> >>>>>>> >>> >>> > >>   File "/usr/share/vdsm/rpc/BindingXMLRPC.py",
>>> line
>>> >> >>>>>>> >>> >>> > >> 1135,
>>> >> >>>>>>> >>> >>> > >> in
>>> >> >>>>>>> >>> >>> > >> wrapper
>>> >> >>>>>>> >>> >>> > >>     res = f(*args, **kwargs)
>>> >> >>>>>>> >>> >>> > >>   File "/usr/share/vdsm/gluster/api.py", line
>>> 54, in
>>> >> >>>>>>> >>> >>> > >> wrapper
>>> >> >>>>>>> >>> >>> > >>     rv = func(*args, **kwargs)
>>> >> >>>>>>> >>> >>> > >>   File "/usr/share/vdsm/gluster/api.py", line
>>> 251, in
>>> >> >>>>>>> >>> >>> > >> hostsList
>>> >> >>>>>>> >>> >>> > >>     return {'hosts':
>>> >> >>>>>>> >>> >>> > >> self.svdsmProxy.glusterPeerStatus()}
>>> >> >>>>>>> >>> >>> > >>   File "/usr/share/vdsm/supervdsm.py", line 50,
>>> in
>>> >> >>>>>>> >>> >>> > >> __call__
>>> >> >>>>>>> >>> >>> > >>     return callMethod()
>>> >> >>>>>>> >>> >>> > >>   File "/usr/share/vdsm/supervdsm.py", line 48,
>>> in
>>> >> >>>>>>> >>> >>> > >> <lambda>
>>> >> >>>>>>> >>> >>> > >>     **kwargs)
>>> >> >>>>>>> >>> >>> > >>   File "<string>", line 2, in glusterPeerStatus
>>> >> >>>>>>> >>> >>> > >>   File
>>> >> >>>>>>> >>> >>> > >>
>>> "/usr/lib64/python2.6/multiprocessing/managers.py",
>>> >> >>>>>>> >>> >>> > >> line
>>> >> >>>>>>> >>> >>> > >> 740,
>>> >> >>>>>>> >>> >>> > >> in
>>> >> >>>>>>> >>> >>> > >> _callmethod
>>> >> >>>>>>> >>> >>> > >>     raise convert_to_error(kind, result)
>>> >> >>>>>>> >>> >>> > >> GlusterCmdExecFailedException: Command execution
>>> >> >>>>>>> >>> >>> > >> failed
>>> >> >>>>>>> >>> >>> > >> error: Connection failed. Please check if gluster
>>> >> >>>>>>> >>> >>> > >> daemon
>>> >> >>>>>>> >>> >>> > >> is
>>> >> >>>>>>> >>> >>> > >> operational.
>>> >> >>>>>>> >>> >>> > >> return code: 1
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:50,949::task::595::Storage.TaskManager.Task::(_updateState)
>>> >> >>>>>>> >>> >>> > >>
>>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving
>>> >> >>>>>>> >>> >>> > >> from
>>> >> >>>>>>> >>> >>> > >> state
>>> >> >>>>>>> >>> >>> > >> init
>>> >> >>>>>>> >>> >>> > >> ->
>>> >> >>>>>>> >>> >>> > >> state preparing
>>> >> >>>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:50,950::logUtils::44::dispatcher::(wrapper) Run
>>> >> >>>>>>> >>> >>> > >> and
>>> >> >>>>>>> >>> >>> > >> protect:
>>> >> >>>>>>> >>> >>> > >> repoStats(options=None)
>>> >> >>>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:50,950::logUtils::47::dispatcher::(wrapper) Run
>>> >> >>>>>>> >>> >>> > >> and
>>> >> >>>>>>> >>> >>> > >> protect:
>>> >> >>>>>>> >>> >>> > >> repoStats, Return response: {}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:50,950::task::1191::Storage.TaskManager.Task::(prepare)
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::finished: {}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:50,950::task::595::Storage.TaskManager.Task::(_updateState)
>>> >> >>>>>>> >>> >>> > >>
>>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving
>>> >> >>>>>>> >>> >>> > >> from
>>> >> >>>>>>> >>> >>> > >> state
>>> >> >>>>>>> >>> >>> > >> preparing
>>> >> >>>>>>> >>> >>> > >> ->
>>> >> >>>>>>> >>> >>> > >> state finished
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:50,951::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>>> >> >>>>>>> >>> >>> > >> Owner.releaseAll requests {} resources {}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:50,951::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>> >> >>>>>>> >>> >>> > >> Owner.cancelAll requests {}
>>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> 21:41:50,951::task::993::Storage.TaskManager.Task::(_decref)
>>> >> >>>>>>> >>> >>> > >> Task=`c9042986-c978-4b08-adb2-616f5299e115`::ref
>>> 0
>>> >> >>>>>>> >>> >>> > >> aborting
>>> >> >>>>>>> >>> >>> > >> False
>>> >> >>>>>>> >>> >>> > >>  -------------------------------
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>  [root at compute4 ~]# service glusterd status
>>> >> >>>>>>> >>> >>> > >> glusterd is stopped
>>> >> >>>>>>> >>> >>> > >> [root at compute4 ~]# chkconfig --list | grep
>>> glusterd
>>> >> >>>>>>> >>> >>> > >> glusterd        0:off   1:off   2:on    3:on
>>> 4:on
>>> >> >>>>>>> >>> >>> > >> 5:on
>>> >> >>>>>>> >>> >>> > >> 6:off
>>> >> >>>>>>> >>> >>> > >> [root at compute4 ~]#
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>  Thanks,
>>> >> >>>>>>> >>> >>> > >> Punit
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >> On Mon, Nov 24, 2014 at 6:36 PM, Kanagaraj
>>> >> >>>>>>> >>> >>> > >> <kmayilsa at redhat.com>
>>> >> >>>>>>> >>> >>> > >> wrote:
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>>  Can you send the corresponding error in
>>> vdsm.log
>>> >> >>>>>>> >>> >>> > >>> from
>>> >> >>>>>>> >>> >>> > >>> the
>>> >> >>>>>>> >>> >>> > >>> host?
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>> Also check if glusterd service is running.
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>> Thanks,
>>> >> >>>>>>> >>> >>> > >>> Kanagaraj
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>> On 11/24/2014 03:39 PM, Punit Dambiwal wrote:
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>   Hi,
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>  After reboot my Hypervisior host can not
>>> activate
>>> >> >>>>>>> >>> >>> > >>> again
>>> >> >>>>>>> >>> >>> > >>> in the
>>> >> >>>>>>> >>> >>> > >>> cluster
>>> >> >>>>>>> >>> >>> > >>> and failed with the following error :-
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>  Gluster command [<UNKNOWN>] failed on server...
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>  Engine logs :-
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>  2014-11-24 18:05:28,397 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-64) START,
>>> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand(HostName =
>>> Compute4,
>>> >> >>>>>>> >>> >>> > >>> HostId
>>> >> >>>>>>> >>> >>> > >>> =
>>> >> >>>>>>> >>> >>> > >>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id:
>>> >> >>>>>>> >>> >>> > >>> 5f251c90
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:30,609 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-64) FINISH,
>>> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand,
>>> >> >>>>>>> >>> >>> > >>> return:
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> {26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity at d95203e0
>>> },
>>> >> >>>>>>> >>> >>> > >>> log id: 5f251c90
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,768 INFO
>>> >> >>>>>>> >>> >>> > >>>  [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>> >> >>>>>>> >>> >>> > >>> (ajp--127.0.0.1-8702-8)
>>> >> >>>>>>> >>> >>> > >>> [287d570d] Lock Acquired to object EngineLock
>>> >> >>>>>>> >>> >>> > >>> [exclusiveLocks=
>>> >> >>>>>>> >>> >>> > >>> key:
>>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a value: VDS
>>> >> >>>>>>> >>> >>> > >>> , sharedLocks= ]
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,795 INFO
>>> >> >>>>>>> >>> >>> > >>>  [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d]
>>> >> >>>>>>> >>> >>> > >>> Running
>>> >> >>>>>>> >>> >>> > >>> command:
>>> >> >>>>>>> >>> >>> > >>> ActivateVdsCommand internal: false. Entities
>>> >> >>>>>>> >>> >>> > >>> affected :
>>> >> >>>>>>> >>> >>> > >>> ID:
>>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type:
>>> VDSAction
>>> >> >>>>>>> >>> >>> > >>> group
>>> >> >>>>>>> >>> >>> > >>> MANIPULATE_HOST
>>> >> >>>>>>> >>> >>> > >>> with role type ADMIN
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,796 INFO
>>> >> >>>>>>> >>> >>> > >>>  [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d]
>>> >> >>>>>>> >>> >>> > >>> Before
>>> >> >>>>>>> >>> >>> > >>> acquiring
>>> >> >>>>>>> >>> >>> > >>> lock in
>>> >> >>>>>>> >>> >>> > >>> order to prevent monitoring for host Compute5
>>> from
>>> >> >>>>>>> >>> >>> > >>> data-center
>>> >> >>>>>>> >>> >>> > >>> SV_WTC
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,797 INFO
>>> >> >>>>>>> >>> >>> > >>>  [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d]
>>> Lock
>>> >> >>>>>>> >>> >>> > >>> acquired,
>>> >> >>>>>>> >>> >>> > >>> from
>>> >> >>>>>>> >>> >>> > >>> now a
>>> >> >>>>>>> >>> >>> > >>> monitoring of host will be skipped for host
>>> Compute5
>>> >> >>>>>>> >>> >>> > >>> from
>>> >> >>>>>>> >>> >>> > >>> data-center
>>> >> >>>>>>> >>> >>> > >>> SV_WTC
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,817 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d]
>>> >> >>>>>>> >>> >>> > >>> START,
>>> >> >>>>>>> >>> >>> > >>> SetVdsStatusVDSCommand(HostName = Compute5,
>>> HostId =
>>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>>> >> >>>>>>> >>> >>> > >>> status=Unassigned,
>>> >> >>>>>>> >>> >>> > >>> nonOperationalReason=NONE,
>>> >> >>>>>>> >>> >>> > >>> stopSpmFailureLogged=false),
>>> >> >>>>>>> >>> >>> > >>> log id:
>>> >> >>>>>>> >>> >>> > >>> 1cbc7311
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,820 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d]
>>> >> >>>>>>> >>> >>> > >>> FINISH,
>>> >> >>>>>>> >>> >>> > >>> SetVdsStatusVDSCommand, log id: 1cbc7311
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:34,086 INFO
>>> >> >>>>>>> >>> >>> > >>>  [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) Activate
>>> >> >>>>>>> >>> >>> > >>> finished.
>>> >> >>>>>>> >>> >>> > >>> Lock
>>> >> >>>>>>> >>> >>> > >>> released.
>>> >> >>>>>>> >>> >>> > >>> Monitoring can run now for host Compute5 from
>>> >> >>>>>>> >>> >>> > >>> data-center
>>> >> >>>>>>> >>> >>> > >>> SV_WTC
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:34,088 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) Correlation
>>> ID:
>>> >> >>>>>>> >>> >>> > >>> 287d570d,
>>> >> >>>>>>> >>> >>> > >>> Job
>>> >> >>>>>>> >>> >>> > >>> ID:
>>> >> >>>>>>> >>> >>> > >>> 5ef8e4d6-b2bc-469e-8e81-7ef74b2a001a, Call
>>> Stack:
>>> >> >>>>>>> >>> >>> > >>> null,
>>> >> >>>>>>> >>> >>> > >>> Custom
>>> >> >>>>>>> >>> >>> > >>> Event ID:
>>> >> >>>>>>> >>> >>> > >>> -1, Message: Host Compute5 was activated by
>>> admin.
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:34,090 INFO
>>> >> >>>>>>> >>> >>> > >>>  [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) Lock freed
>>> to
>>> >> >>>>>>> >>> >>> > >>> object
>>> >> >>>>>>> >>> >>> > >>> EngineLock
>>> >> >>>>>>> >>> >>> > >>> [exclusiveLocks= key:
>>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a
>>> >> >>>>>>> >>> >>> > >>> value:
>>> >> >>>>>>> >>> >>> > >>> VDS
>>> >> >>>>>>> >>> >>> > >>> , sharedLocks= ]
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:35,792 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-55) [3706e836]
>>> START,
>>> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand(HostName =
>>> Compute4,
>>> >> >>>>>>> >>> >>> > >>> HostId
>>> >> >>>>>>> >>> >>> > >>> =
>>> >> >>>>>>> >>> >>> > >>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id:
>>> >> >>>>>>> >>> >>> > >>> 48a0c832
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,064 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) START,
>>> >> >>>>>>> >>> >>> > >>> GetHardwareInfoVDSCommand(HostName = Compute5,
>>> >> >>>>>>> >>> >>> > >>> HostId =
>>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> vds=Host[Compute5,0bf6b00f-7947-4411-b55a-cc5eea2b381a]), log
>>> >> >>>>>>> >>> >>> > >>> id:
>>> >> >>>>>>> >>> >>> > >>> 6d560cc2
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,074 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) FINISH,
>>> >> >>>>>>> >>> >>> > >>> GetHardwareInfoVDSCommand, log
>>> >> >>>>>>> >>> >>> > >>> id: 6d560cc2
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,093 WARN
>>> >> >>>>>>> >>> >>> > >>>  [org.ovirt.engine.core.vdsbroker.VdsManager]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) Host
>>> Compute5 is
>>> >> >>>>>>> >>> >>> > >>> running
>>> >> >>>>>>> >>> >>> > >>> with
>>> >> >>>>>>> >>> >>> > >>> disabled
>>> >> >>>>>>> >>> >>> > >>> SELinux.
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,127 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf]
>>> >> >>>>>>> >>> >>> > >>> Running
>>> >> >>>>>>> >>> >>> > >>> command:
>>> >> >>>>>>> >>> >>> > >>> HandleVdsCpuFlagsOrClusterChangedCommand
>>> internal:
>>> >> >>>>>>> >>> >>> > >>> true.
>>> >> >>>>>>> >>> >>> > >>> Entities
>>> >> >>>>>>> >>> >>> > >>> affected
>>> >> >>>>>>> >>> >>> > >>> :  ID: 0bf6b00f-7947-4411-b55a-cc5eea2b381a
>>> Type:
>>> >> >>>>>>> >>> >>> > >>> VDS
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,147 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf]
>>> START,
>>> >> >>>>>>> >>> >>> > >>> GlusterServersListVDSCommand(HostName =
>>> Compute5,
>>> >> >>>>>>> >>> >>> > >>> HostId
>>> >> >>>>>>> >>> >>> > >>> =
>>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a), log id:
>>> >> >>>>>>> >>> >>> > >>> 4faed87
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,164 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf]
>>> >> >>>>>>> >>> >>> > >>> FINISH,
>>> >> >>>>>>> >>> >>> > >>> GlusterServersListVDSCommand, log id: 4faed87
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,189 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.bll.SetNonOperationalVdsCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>>> >> >>>>>>> >>> >>> > >>> Running
>>> >> >>>>>>> >>> >>> > >>> command:
>>> >> >>>>>>> >>> >>> > >>> SetNonOperationalVdsCommand internal: true.
>>> Entities
>>> >> >>>>>>> >>> >>> > >>> affected :
>>> >> >>>>>>> >>> >>> > >>> ID:
>>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,206 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>>> START,
>>> >> >>>>>>> >>> >>> > >>> SetVdsStatusVDSCommand(HostName = Compute5,
>>> HostId =
>>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>>> >> >>>>>>> >>> >>> > >>> status=NonOperational,
>>> >> >>>>>>> >>> >>> > >>> nonOperationalReason=GLUSTER_COMMAND_FAILED,
>>> >> >>>>>>> >>> >>> > >>> stopSpmFailureLogged=false),
>>> >> >>>>>>> >>> >>> > >>> log id: fed5617
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,209 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>>> >> >>>>>>> >>> >>> > >>> FINISH,
>>> >> >>>>>>> >>> >>> > >>> SetVdsStatusVDSCommand, log id: fed5617
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,223 ERROR
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>>> >> >>>>>>> >>> >>> > >>> Correlation ID:
>>> >> >>>>>>> >>> >>> > >>> 4a84c4e5,
>>> >> >>>>>>> >>> >>> > >>> Job
>>> >> >>>>>>> >>> >>> > >>> ID: 4bfd4a6d-c3ef-468f-a40e-a3a6ca13011b, Call
>>> >> >>>>>>> >>> >>> > >>> Stack:
>>> >> >>>>>>> >>> >>> > >>> null,
>>> >> >>>>>>> >>> >>> > >>> Custom
>>> >> >>>>>>> >>> >>> > >>> Event
>>> >> >>>>>>> >>> >>> > >>> ID: -1, Message: Gluster command [<UNKNOWN>]
>>> failed
>>> >> >>>>>>> >>> >>> > >>> on
>>> >> >>>>>>> >>> >>> > >>> server
>>> >> >>>>>>> >>> >>> > >>> Compute5.
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,243 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>>> >> >>>>>>> >>> >>> > >>> Correlation ID:
>>> >> >>>>>>> >>> >>> > >>> null,
>>> >> >>>>>>> >>> >>> > >>> Call
>>> >> >>>>>>> >>> >>> > >>> Stack: null, Custom Event ID: -1, Message:
>>> Status of
>>> >> >>>>>>> >>> >>> > >>> host
>>> >> >>>>>>> >>> >>> > >>> Compute5
>>> >> >>>>>>> >>> >>> > >>> was
>>> >> >>>>>>> >>> >>> > >>> set
>>> >> >>>>>>> >>> >>> > >>> to NonOperational.
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,272 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.bll.HandleVdsVersionCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [a0c8a7f]
>>> Running
>>> >> >>>>>>> >>> >>> > >>> command:
>>> >> >>>>>>> >>> >>> > >>> HandleVdsVersionCommand internal: true. Entities
>>> >> >>>>>>> >>> >>> > >>> affected :
>>> >> >>>>>>> >>> >>> > >>> ID:
>>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,274 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [a0c8a7f]
>>> Host
>>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a : Compute5
>>> is
>>> >> >>>>>>> >>> >>> > >>> already in
>>> >> >>>>>>> >>> >>> > >>> NonOperational status for reason
>>> >> >>>>>>> >>> >>> > >>> GLUSTER_COMMAND_FAILED.
>>> >> >>>>>>> >>> >>> > >>> SetNonOperationalVds command is skipped.
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:38,065 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-55) [3706e836]
>>> >> >>>>>>> >>> >>> > >>> FINISH,
>>> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand, return:
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> {26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity at 4e72a1b1
>>> },
>>> >> >>>>>>> >>> >>> > >>> log id: 48a0c832
>>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:43,243 INFO
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-35) START,
>>> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand(HostName =
>>> Compute4,
>>> >> >>>>>>> >>> >>> > >>> HostId
>>> >> >>>>>>> >>> >>> > >>> =
>>> >> >>>>>>> >>> >>> > >>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id:
>>> >> >>>>>>> >>> >>> > >>> 3ce13ebc
>>> >> >>>>>>> >>> >>> > >>> ^C
>>> >> >>>>>>> >>> >>> > >>> [root at ccr01 ~]#
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>  Thanks,
>>> >> >>>>>>> >>> >>> > >>> Punit
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>  _______________________________________________
>>> >> >>>>>>> >>> >>> > >>> Users mailing
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>> listUsers at ovirt.orghttp://
>>> lists.ovirt.org/mailman/listinfo/users
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >>
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> >
>>> >> >>>>>>> >>> >>
>>> >> >>>>>>> >>> >>
>>> >> >>>>>>> >>> >
>>> >> >>>>>>> >>> >
>>> >> >>>>>>> >>> > _______________________________________________
>>> >> >>>>>>> >>> > Gluster-users mailing list
>>> >> >>>>>>> >>> > Gluster-users at gluster.org
>>> >> >>>>>>> >>> >
>>> >> >>>>>>> >>> >
>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>> >> >>>>>>> >>
>>> >> >>>>>>> >>
>>> >> >>>>>>> >
>>> >> >>>>>>
>>> >> >>>>>>
>>> >> >>>>
>>> >> >>
>>> >
>>> >
>>>
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20141209/671c550f/attachment.html>


More information about the Gluster-users mailing list