Punit Dambiwal
2014-Dec-07 12:40 UTC
[Gluster-users] [ovirt-users] Gluster command [<UNKNOWN>] failed on server...
Hi Kaushal, Still the same error...even try your suggested workaround :- ------------------- Can you replace 'Before=network-online.target' with 'Wants=network-online.target' and try the boot again? This should force the network to be online before starting GlusterD. ------------------- Thanks, Punit On Sat, Dec 6, 2014 at 11:44 AM, Punit Dambiwal <hypunit at gmail.com> wrote:> Hi Kaushal, > > I already have all the hosts entry in the /etc/hosts for the easy > resolution....i will try your method in the glusterd.services and check and > let you know....weather problem solve or not..... > > On Fri, Dec 5, 2014 at 9:50 PM, Kaushal M <kshlmster at gmail.com> wrote: > >> Can you replace 'Before=network-online.target' with >> 'Wants=network-online.target' and try the boot again? This should >> force the network to be online before starting GlusterD. >> >> If even that fails, you could try adding an entry into /etc/hosts with >> the hostname of the system. This should prevent any more failures. >> >> I still don't believe it's a problem with Gluster. Gluster uses apis >> provided by the system to perform name resolution. These definitely >> work correctly because you can start GlusterD later. Since the >> resolution failure only happens during boot, it points to system or >> network setup issues during boot. To me it seems like the network >> isn't completely setup at that point of time. >> >> ~kaushal >> >> On Fri, Dec 5, 2014 at 12:47 PM, Punit Dambiwal <hypunit at gmail.com> >> wrote: >> > Hi Kaushal, >> > >> > It seems it's bug in glusterfs 3.6....even i manage my systemd to start >> the >> > network service before glusterd...but it's still fail... >> > >> > --------------- >> > [Unit] >> > Description=GlusterFS, a clustered file-system server >> > After=network.target rpcbind.service >> > Before=network-online.target >> > >> > [Service] >> > Type=forking >> > PIDFile=/var/run/glusterd.pid >> > LimitNOFILE=65536 >> > ExecStartPre=/etc/rc.d/init.d/network start >> > ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid >> > KillMode=process >> > >> > [Install] >> > WantedBy=multi-user.target >> > ---------------- >> > >> > Thanks, >> > Punit >> > >> > On Wed, Dec 3, 2014 at 8:56 PM, Kaushal M <kshlmster at gmail.com> wrote: >> >> >> >> I just remembered this. >> >> >> >> There was another user having a similar issue of GlusterD failing to >> >> start on the mailing list a while back. The cause of his problem was >> >> the way his network was brought up. >> >> IIRC, he was using a static network configuration. The problem >> >> vanished when he began using dhcp. Or it might have been he was using >> >> dhcp.service and it got solved after switching to NetworkManager. >> >> >> >> This could be one more thing you could look at. >> >> >> >> I'll try to find the mail thread to see if it was the same problem as >> you. >> >> >> >> ~kaushal >> >> >> >> On Wed, Dec 3, 2014 at 6:22 PM, Kaushal M <kshlmster at gmail.com> wrote: >> >> > I don't know much about how the network target is brought up in >> >> > CentOS7, but I'll try as much as I can. >> >> > >> >> > It seems to me that, after the network has been brought up and by the >> >> > time GlusterD is started, >> >> > a. The machine hasn't yet recieved it's hostname, or >> >> > b. It hasn't yet registered with the name server. >> >> > >> >> > This is causing name resolution failures. >> >> > >> >> > I don't know if the network target could come up without the machine >> >> > getting its hostname, so I'm pretty sure it's not a. >> >> > >> >> > So it seems to be b. But these kind of signing in happens only in >> DDNS >> >> > systems, which doesn't seem to be the case for you. >> >> > >> >> > Both of these reasons might be wrong (most likely wrong). You'd do >> >> > good if you could ask for help from someone with more experience in >> >> > systemd + networking. >> >> > >> >> > ~kaushal >> >> > >> >> > On Wed, Dec 3, 2014 at 10:54 AM, Punit Dambiwal <hypunit at gmail.com> >> >> > wrote: >> >> >> Hi Kaushal, >> >> >> >> >> >> This is the host...which i rebooted...would you mind to let me know >> how >> >> >> i >> >> >> can make the glusterd sevice come up after network...i am using >> >> >> centos7...if >> >> >> network is the issue... >> >> >> >> >> >> On Wed, Dec 3, 2014 at 11:54 AM, Kaushal M <kshlmster at gmail.com> >> wrote: >> >> >>> >> >> >>> This peer cannot be identified. >> >> >>> >> >> >>> " [2014-12-03 02:29:25.998153] D >> >> >>> [glusterd-peer-utils.c:121:glusterd_peerinfo_find_by_hostname] >> >> >>> 0-management: >> >> >>> Unable to find friend: cpu05.zne01.hkg1.ovt.36stack.com" >> >> >>> >> >> >>> I don't know why this address is not being resolved during boot >> time. >> >> >>> If >> >> >>> this is a valid peer, the the only reason I can think of this that >> the >> >> >>> network is not up. >> >> >>> >> >> >>> If you had previously detached the peer forcefully, the that could >> >> >>> have >> >> >>> left stale entries in some volumes. In this case as well, GlusterD >> >> >>> will fail >> >> >>> to identify the peer. >> >> >>> >> >> >>> Do either of these reasons seem a possibility to you? >> >> >>> >> >> >>> On Dec 3, 2014 8:07 AM, "Punit Dambiwal" <hypunit at gmail.com> >> wrote: >> >> >>>> >> >> >>>> Hi Kaushal, >> >> >>>> >> >> >>>> Please find the logs here :- http://ur1.ca/iyoe5 and >> >> >>>> http://ur1.ca/iyoed >> >> >>>> >> >> >>>> On Tue, Dec 2, 2014 at 10:43 PM, Kaushal M <kshlmster at gmail.com> >> >> >>>> wrote: >> >> >>>>> >> >> >>>>> Hey Punit, >> >> >>>>> In the logs you've provided, GlusterD appears to be running >> >> >>>>> correctly. >> >> >>>>> Could you provide the logs for the time period when GlusterD >> >> >>>>> attempts to >> >> >>>>> start but fails. >> >> >>>>> >> >> >>>>> ~kaushal >> >> >>>>> >> >> >>>>> On Dec 2, 2014 8:03 PM, "Punit Dambiwal" <hypunit at gmail.com> >> wrote: >> >> >>>>>> >> >> >>>>>> Hi Kaushal, >> >> >>>>>> >> >> >>>>>> Please find the logs here :- http://ur1.ca/iyhs5 and >> >> >>>>>> http://ur1.ca/iyhue >> >> >>>>>> >> >> >>>>>> Thanks, >> >> >>>>>> punit >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> On Tue, Dec 2, 2014 at 12:00 PM, Kaushal M <kshlmster at gmail.com >> > >> >> >>>>>> wrote: >> >> >>>>>>> >> >> >>>>>>> Hey Punit, >> >> >>>>>>> Could you start Glusterd in debug mode and provide the logs >> here? >> >> >>>>>>> To start it in debug mode, append '-LDEBUG' to the ExecStart >> line >> >> >>>>>>> in >> >> >>>>>>> the service file. >> >> >>>>>>> >> >> >>>>>>> ~kaushal >> >> >>>>>>> >> >> >>>>>>> On Mon, Dec 1, 2014 at 9:05 AM, Punit Dambiwal < >> hypunit at gmail.com> >> >> >>>>>>> wrote: >> >> >>>>>>> > Hi, >> >> >>>>>>> > >> >> >>>>>>> > Can Any body help me on this ?? >> >> >>>>>>> > >> >> >>>>>>> > On Thu, Nov 27, 2014 at 9:29 AM, Punit Dambiwal >> >> >>>>>>> > <hypunit at gmail.com> >> >> >>>>>>> > wrote: >> >> >>>>>>> >> >> >> >>>>>>> >> Hi Kaushal, >> >> >>>>>>> >> >> >> >>>>>>> >> Thanks for the detailed reply....let me explain my setup >> first >> >> >>>>>>> >> :- >> >> >>>>>>> >> >> >> >>>>>>> >> 1. Ovirt Engine >> >> >>>>>>> >> 2. 4* host as well as storage machine (Host and gluster >> >> >>>>>>> >> combined) >> >> >>>>>>> >> 3. Every host has 24 bricks... >> >> >>>>>>> >> >> >> >>>>>>> >> Now whenever the host machine reboot...it can come up but >> can >> >> >>>>>>> >> not >> >> >>>>>>> >> join the >> >> >>>>>>> >> cluster again and through the following error "Gluster >> command >> >> >>>>>>> >> [<UNKNOWN>] >> >> >>>>>>> >> failed on server.." >> >> >>>>>>> >> >> >> >>>>>>> >> Please check my comment in line :- >> >> >>>>>>> >> >> >> >>>>>>> >> 1. Use the same string for doing the peer probe and for the >> >> >>>>>>> >> brick >> >> >>>>>>> >> address >> >> >>>>>>> >> during volume create/add-brick. Ideally, we suggest you use >> >> >>>>>>> >> properly >> >> >>>>>>> >> resolvable FQDNs everywhere. If that is not possible, then >> use >> >> >>>>>>> >> only >> >> >>>>>>> >> IP >> >> >>>>>>> >> addresses. Try to avoid short names. >> >> >>>>>>> >> --------------- >> >> >>>>>>> >> [root at cpu05 ~]# gluster peer status >> >> >>>>>>> >> Number of Peers: 3 >> >> >>>>>>> >> >> >> >>>>>>> >> Hostname: cpu03.stack.com >> >> >>>>>>> >> Uuid: 5729b8c4-e80d-4353-b456-6f467bddbdfb >> >> >>>>>>> >> State: Peer in Cluster (Connected) >> >> >>>>>>> >> >> >> >>>>>>> >> Hostname: cpu04.stack.com >> >> >>>>>>> >> Uuid: d272b790-c4b2-4bed-ba68-793656e6d7b0 >> >> >>>>>>> >> State: Peer in Cluster (Connected) >> >> >>>>>>> >> Other names: >> >> >>>>>>> >> 10.10.0.8 >> >> >>>>>>> >> >> >> >>>>>>> >> Hostname: cpu02.stack.com >> >> >>>>>>> >> Uuid: 8d8a7041-950e-40d0-85f9-58d14340ca25 >> >> >>>>>>> >> State: Peer in Cluster (Connected) >> >> >>>>>>> >> [root at cpu05 ~]# >> >> >>>>>>> >> ---------------- >> >> >>>>>>> >> 2. During boot up, make sure to launch glusterd only after >> the >> >> >>>>>>> >> network is >> >> >>>>>>> >> up. This will allow the new peer identification mechanism >> to do >> >> >>>>>>> >> its >> >> >>>>>>> >> job correctly. >> >> >>>>>>> >> >> I think the service itself doing the same job.... >> >> >>>>>>> >> >> >> >>>>>>> >> [root at cpu05 ~]# cat >> /usr/lib/systemd/system/glusterd.service >> >> >>>>>>> >> [Unit] >> >> >>>>>>> >> Description=GlusterFS, a clustered file-system server >> >> >>>>>>> >> After=network.target rpcbind.service >> >> >>>>>>> >> Before=network-online.target >> >> >>>>>>> >> >> >> >>>>>>> >> [Service] >> >> >>>>>>> >> Type=forking >> >> >>>>>>> >> PIDFile=/var/run/glusterd.pid >> >> >>>>>>> >> LimitNOFILE=65536 >> >> >>>>>>> >> ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid >> >> >>>>>>> >> KillMode=process >> >> >>>>>>> >> >> >> >>>>>>> >> [Install] >> >> >>>>>>> >> WantedBy=multi-user.target >> >> >>>>>>> >> [root at cpu05 ~]# >> >> >>>>>>> >> -------------------- >> >> >>>>>>> >> >> >> >>>>>>> >> gluster logs :- >> >> >>>>>>> >> >> >> >>>>>>> >> [2014-11-24 09:22:22.147471] I [MSGID: 100030] >> >> >>>>>>> >> [glusterfsd.c:2018:main] >> >> >>>>>>> >> 0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd >> >> >>>>>>> >> version >> >> >>>>>>> >> 3.6.1 >> >> >>>>>>> >> (args: /usr/sbin/glusterd -p /var/run/glusterd.pid) >> >> >>>>>>> >> [2014-11-24 09:22:22.151565] I [glusterd.c:1214:init] >> >> >>>>>>> >> 0-management: >> >> >>>>>>> >> Maximum allowed open file descriptors set to 65536 >> >> >>>>>>> >> [2014-11-24 09:22:22.151599] I [glusterd.c:1259:init] >> >> >>>>>>> >> 0-management: >> >> >>>>>>> >> Using >> >> >>>>>>> >> /var/lib/glusterd as working directory >> >> >>>>>>> >> [2014-11-24 09:22:22.155216] W >> >> >>>>>>> >> [rdma.c:4195:__gf_rdma_ctx_create] >> >> >>>>>>> >> 0-rpc-transport/rdma: rdma_cm event channel creation failed >> (No >> >> >>>>>>> >> such device) >> >> >>>>>>> >> [2014-11-24 09:22:22.155264] E [rdma.c:4483:init] >> >> >>>>>>> >> 0-rdma.management: >> >> >>>>>>> >> Failed to initialize IB Device >> >> >>>>>>> >> [2014-11-24 09:22:22.155285] E >> >> >>>>>>> >> [rpc-transport.c:333:rpc_transport_load] >> >> >>>>>>> >> 0-rpc-transport: 'rdma' initialization failed >> >> >>>>>>> >> [2014-11-24 09:22:22.155354] W >> >> >>>>>>> >> [rpcsvc.c:1524:rpcsvc_transport_create] >> >> >>>>>>> >> 0-rpc-service: cannot create listener, initing the transport >> >> >>>>>>> >> failed >> >> >>>>>>> >> [2014-11-24 09:22:22.156290] I >> >> >>>>>>> >> [glusterd.c:413:glusterd_check_gsync_present] 0-glusterd: >> >> >>>>>>> >> geo-replication >> >> >>>>>>> >> module not installed in the system >> >> >>>>>>> >> [2014-11-24 09:22:22.161318] I >> >> >>>>>>> >> [glusterd-store.c:2043:glusterd_restore_op_version] >> 0-glusterd: >> >> >>>>>>> >> retrieved >> >> >>>>>>> >> op-version: 30600 >> >> >>>>>>> >> [2014-11-24 09:22:22.821800] I >> >> >>>>>>> >> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo] >> >> >>>>>>> >> 0-management: >> >> >>>>>>> >> connect returned 0 >> >> >>>>>>> >> [2014-11-24 09:22:22.825810] I >> >> >>>>>>> >> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo] >> >> >>>>>>> >> 0-management: >> >> >>>>>>> >> connect returned 0 >> >> >>>>>>> >> [2014-11-24 09:22:22.828705] I >> >> >>>>>>> >> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo] >> >> >>>>>>> >> 0-management: >> >> >>>>>>> >> connect returned 0 >> >> >>>>>>> >> [2014-11-24 09:22:22.828771] I >> >> >>>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init] >> >> >>>>>>> >> 0-management: setting frame-timeout to 600 >> >> >>>>>>> >> [2014-11-24 09:22:22.832670] I >> >> >>>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init] >> >> >>>>>>> >> 0-management: setting frame-timeout to 600 >> >> >>>>>>> >> [2014-11-24 09:22:22.835919] I >> >> >>>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init] >> >> >>>>>>> >> 0-management: setting frame-timeout to 600 >> >> >>>>>>> >> [2014-11-24 09:22:22.840209] E >> >> >>>>>>> >> [glusterd-store.c:4248:glusterd_resolve_all_bricks] >> 0-glusterd: >> >> >>>>>>> >> resolve >> >> >>>>>>> >> brick failed in restore >> >> >>>>>>> >> [2014-11-24 09:22:22.840233] E [xlator.c:425:xlator_init] >> >> >>>>>>> >> 0-management: >> >> >>>>>>> >> Initialization of volume 'management' failed, review your >> >> >>>>>>> >> volfile >> >> >>>>>>> >> again >> >> >>>>>>> >> [2014-11-24 09:22:22.840245] E >> >> >>>>>>> >> [graph.c:322:glusterfs_graph_init] >> >> >>>>>>> >> 0-management: initializing translator failed >> >> >>>>>>> >> [2014-11-24 09:22:22.840264] E >> >> >>>>>>> >> [graph.c:525:glusterfs_graph_activate] >> >> >>>>>>> >> 0-graph: init failed >> >> >>>>>>> >> [2014-11-24 09:22:22.840754] W >> >> >>>>>>> >> [glusterfsd.c:1194:cleanup_and_exit] >> >> >>>>>>> >> (--> >> >> >>>>>>> >> 0-: received signum (0), shutting down >> >> >>>>>>> >> >> >> >>>>>>> >> Thanks, >> >> >>>>>>> >> Punit >> >> >>>>>>> >> >> >> >>>>>>> >> >> >> >>>>>>> >> >> >> >>>>>>> >> >> >> >>>>>>> >> On Wed, Nov 26, 2014 at 7:14 PM, Kaushal M >> >> >>>>>>> >> <kshlmster at gmail.com> >> >> >>>>>>> >> wrote: >> >> >>>>>>> >>> >> >> >>>>>>> >>> Based on the logs I can guess that glusterd is being >> started >> >> >>>>>>> >>> before >> >> >>>>>>> >>> the network has come up and that the addresses given to >> bricks >> >> >>>>>>> >>> do >> >> >>>>>>> >>> not >> >> >>>>>>> >>> directly match the addresses used in during peer probe. >> >> >>>>>>> >>> >> >> >>>>>>> >>> The gluster_after_reboot log has the line "[2014-11-25 >> >> >>>>>>> >>> 06:46:09.972113] E >> >> >>>>>>> >>> [glusterd-store.c:2632:glusterd_resolve_all_bricks] >> >> >>>>>>> >>> 0-glusterd: resolve brick failed in restore". >> >> >>>>>>> >>> >> >> >>>>>>> >>> Brick resolution fails when glusterd cannot match the >> address >> >> >>>>>>> >>> for >> >> >>>>>>> >>> the >> >> >>>>>>> >>> brick, with one of the peers. Brick resolution happens in >> two >> >> >>>>>>> >>> phases, >> >> >>>>>>> >>> 1. We first try to identify the peer by performing string >> >> >>>>>>> >>> comparisions >> >> >>>>>>> >>> with the brick address and the peer addresses (The peer >> names >> >> >>>>>>> >>> will >> >> >>>>>>> >>> be >> >> >>>>>>> >>> the names/addresses that were given when the peer was >> probed). >> >> >>>>>>> >>> 2. If we don't find a match from step 1, we will then >> resolve >> >> >>>>>>> >>> all >> >> >>>>>>> >>> the >> >> >>>>>>> >>> brick address and the peer addresses into addrinfo structs, >> >> >>>>>>> >>> and >> >> >>>>>>> >>> then >> >> >>>>>>> >>> compare these structs to find a match. This process should >> >> >>>>>>> >>> generally >> >> >>>>>>> >>> find a match if available. This will fail only if the >> network >> >> >>>>>>> >>> is >> >> >>>>>>> >>> not >> >> >>>>>>> >>> up yet as we cannot resolve addresses. >> >> >>>>>>> >>> >> >> >>>>>>> >>> The above steps are applicable only to glusterfs versions >> >> >>>>>>> >>> >=3.6. >> >> >>>>>>> >>> They >> >> >>>>>>> >>> were introduced to reduce problems with peer >> identification, >> >> >>>>>>> >>> like >> >> >>>>>>> >>> the >> >> >>>>>>> >>> one you encountered >> >> >>>>>>> >>> >> >> >>>>>>> >>> Since both of the steps failed to find a match in one run, >> but >> >> >>>>>>> >>> succeeded later, we can come to the conclusion that, >> >> >>>>>>> >>> a) the bricks don't have the exact same string used in peer >> >> >>>>>>> >>> probe >> >> >>>>>>> >>> for >> >> >>>>>>> >>> their addresses as step 1 failed, and >> >> >>>>>>> >>> b) the network was not up in the initial run, as step 2 >> failed >> >> >>>>>>> >>> during >> >> >>>>>>> >>> the initial run, but passed in the second run. >> >> >>>>>>> >>> >> >> >>>>>>> >>> Please let me know if my conclusion is correct. >> >> >>>>>>> >>> >> >> >>>>>>> >>> If it is, you can solve your problem in two ways. >> >> >>>>>>> >>> 1. Use the same string for doing the peer probe and for the >> >> >>>>>>> >>> brick >> >> >>>>>>> >>> address during volume create/add-brick. Ideally, we suggest >> >> >>>>>>> >>> you >> >> >>>>>>> >>> use >> >> >>>>>>> >>> properly resolvable FQDNs everywhere. If that is not >> possible, >> >> >>>>>>> >>> then >> >> >>>>>>> >>> use only IP addresses. Try to avoid short names. >> >> >>>>>>> >>> 2. During boot up, make sure to launch glusterd only after >> the >> >> >>>>>>> >>> network >> >> >>>>>>> >>> is up. This will allow the new peer identification >> mechanism >> >> >>>>>>> >>> to do >> >> >>>>>>> >>> its >> >> >>>>>>> >>> job correctly. >> >> >>>>>>> >>> >> >> >>>>>>> >>> >> >> >>>>>>> >>> If you have already followed these steps and yet still hit >> the >> >> >>>>>>> >>> problem, then please provide more information (setup, logs, >> >> >>>>>>> >>> etc.). >> >> >>>>>>> >>> It >> >> >>>>>>> >>> could be much different problem that you are facing. >> >> >>>>>>> >>> >> >> >>>>>>> >>> ~kaushal >> >> >>>>>>> >>> >> >> >>>>>>> >>> On Wed, Nov 26, 2014 at 4:01 PM, Punit Dambiwal >> >> >>>>>>> >>> <hypunit at gmail.com> >> >> >>>>>>> >>> wrote: >> >> >>>>>>> >>> > Is there any one can help on this ?? >> >> >>>>>>> >>> > >> >> >>>>>>> >>> > Thanks, >> >> >>>>>>> >>> > punit >> >> >>>>>>> >>> > >> >> >>>>>>> >>> > On Wed, Nov 26, 2014 at 9:42 AM, Punit Dambiwal >> >> >>>>>>> >>> > <hypunit at gmail.com> >> >> >>>>>>> >>> > wrote: >> >> >>>>>>> >>> >> >> >> >>>>>>> >>> >> Hi, >> >> >>>>>>> >>> >> >> >> >>>>>>> >>> >> My Glusterfs version is :- glusterfs-3.6.1-1.el7 >> >> >>>>>>> >>> >> >> >> >>>>>>> >>> >> On Wed, Nov 26, 2014 at 1:59 AM, Kanagaraj Mayilsamy >> >> >>>>>>> >>> >> <kmayilsa at redhat.com> >> >> >>>>>>> >>> >> wrote: >> >> >>>>>>> >>> >>> >> >> >>>>>>> >>> >>> [+Gluster-users at gluster.org] >> >> >>>>>>> >>> >>> >> >> >>>>>>> >>> >>> "Initialization of volume 'management' failed, review >> your >> >> >>>>>>> >>> >>> volfile >> >> >>>>>>> >>> >>> again", glusterd throws this error when the service is >> >> >>>>>>> >>> >>> started >> >> >>>>>>> >>> >>> automatically >> >> >>>>>>> >>> >>> after the reboot. But the service is successfully >> started >> >> >>>>>>> >>> >>> later >> >> >>>>>>> >>> >>> manually by >> >> >>>>>>> >>> >>> the user. >> >> >>>>>>> >>> >>> >> >> >>>>>>> >>> >>> can somebody from gluster-users please help on this? >> >> >>>>>>> >>> >>> >> >> >>>>>>> >>> >>> glusterfs version: 3.5.1 >> >> >>>>>>> >>> >>> >> >> >>>>>>> >>> >>> Thanks, >> >> >>>>>>> >>> >>> Kanagaraj >> >> >>>>>>> >>> >>> >> >> >>>>>>> >>> >>> ----- Original Message ----- >> >> >>>>>>> >>> >>> > From: "Punit Dambiwal" <hypunit at gmail.com> >> >> >>>>>>> >>> >>> > To: "Kanagaraj" <kmayilsa at redhat.com> >> >> >>>>>>> >>> >>> > Cc: users at ovirt.org >> >> >>>>>>> >>> >>> > Sent: Tuesday, November 25, 2014 7:24:45 PM >> >> >>>>>>> >>> >>> > Subject: Re: [ovirt-users] Gluster command >> [<UNKNOWN>] >> >> >>>>>>> >>> >>> > failed on >> >> >>>>>>> >>> >>> > server... >> >> >>>>>>> >>> >>> > >> >> >>>>>>> >>> >>> > Hi Kanagraj, >> >> >>>>>>> >>> >>> > >> >> >>>>>>> >>> >>> > Please check the attached log files....i didn't find >> any >> >> >>>>>>> >>> >>> > thing >> >> >>>>>>> >>> >>> > special.... >> >> >>>>>>> >>> >>> > >> >> >>>>>>> >>> >>> > On Tue, Nov 25, 2014 at 12:12 PM, Kanagaraj >> >> >>>>>>> >>> >>> > <kmayilsa at redhat.com> >> >> >>>>>>> >>> >>> > wrote: >> >> >>>>>>> >>> >>> > >> >> >>>>>>> >>> >>> > > Do you see any errors in >> >> >>>>>>> >>> >>> > > /var/log/glusterfs/etc-glusterfs-glusterd.vol.log >> or >> >> >>>>>>> >>> >>> > > vdsm.log >> >> >>>>>>> >>> >>> > > when >> >> >>>>>>> >>> >>> > > the >> >> >>>>>>> >>> >>> > > service is trying to start automatically after the >> >> >>>>>>> >>> >>> > > reboot? >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > > Thanks, >> >> >>>>>>> >>> >>> > > Kanagaraj >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > > On 11/24/2014 08:13 PM, Punit Dambiwal wrote: >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > > Hi Kanagaraj, >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > > Yes...once i will start the gluster service and >> then >> >> >>>>>>> >>> >>> > > vdsmd >> >> >>>>>>> >>> >>> > > ...the >> >> >>>>>>> >>> >>> > > host >> >> >>>>>>> >>> >>> > > can connect to cluster...but the question is why >> it's >> >> >>>>>>> >>> >>> > > not >> >> >>>>>>> >>> >>> > > started >> >> >>>>>>> >>> >>> > > even it >> >> >>>>>>> >>> >>> > > has chkconfig enabled... >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > > I have tested it in two host cluster >> >> >>>>>>> >>> >>> > > environment...(Centos 6.6 >> >> >>>>>>> >>> >>> > > and >> >> >>>>>>> >>> >>> > > centos 7.0) on both hypervisior cluster..it's >> failed >> >> >>>>>>> >>> >>> > > to >> >> >>>>>>> >>> >>> > > reconnect >> >> >>>>>>> >>> >>> > > in >> >> >>>>>>> >>> >>> > > to >> >> >>>>>>> >>> >>> > > cluster after reboot.... >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > > In both the environment glusterd enabled for next >> >> >>>>>>> >>> >>> > > boot....but >> >> >>>>>>> >>> >>> > > it's >> >> >>>>>>> >>> >>> > > failed with the same error....seems it's bug in >> either >> >> >>>>>>> >>> >>> > > gluster or >> >> >>>>>>> >>> >>> > > Ovirt ?? >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > > Please help me to find the workaround here if can >> not >> >> >>>>>>> >>> >>> > > resolve >> >> >>>>>>> >>> >>> > > it...as >> >> >>>>>>> >>> >>> > > without this the Host machine can not connect after >> >> >>>>>>> >>> >>> > > reboot....that >> >> >>>>>>> >>> >>> > > means >> >> >>>>>>> >>> >>> > > engine will consider it as down and every time >> need to >> >> >>>>>>> >>> >>> > > manually >> >> >>>>>>> >>> >>> > > start >> >> >>>>>>> >>> >>> > > the >> >> >>>>>>> >>> >>> > > gluster service and vdsmd... ?? >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > > Thanks, >> >> >>>>>>> >>> >>> > > Punit >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > > On Mon, Nov 24, 2014 at 10:20 PM, Kanagaraj >> >> >>>>>>> >>> >>> > > <kmayilsa at redhat.com> >> >> >>>>>>> >>> >>> > > wrote: >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > >> From vdsm.log "error: Connection failed. Please >> >> >>>>>>> >>> >>> > >> check if >> >> >>>>>>> >>> >>> > >> gluster >> >> >>>>>>> >>> >>> > >> daemon >> >> >>>>>>> >>> >>> > >> is operational." >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> Starting glusterd service should fix this issue. >> >> >>>>>>> >>> >>> > >> 'service >> >> >>>>>>> >>> >>> > >> glusterd >> >> >>>>>>> >>> >>> > >> start' >> >> >>>>>>> >>> >>> > >> But i am wondering why the glusterd was not >> started >> >> >>>>>>> >>> >>> > >> automatically >> >> >>>>>>> >>> >>> > >> after >> >> >>>>>>> >>> >>> > >> the reboot. >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> Thanks, >> >> >>>>>>> >>> >>> > >> Kanagaraj >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> On 11/24/2014 07:18 PM, Punit Dambiwal wrote: >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> Hi Kanagaraj, >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> Please find the attached VDSM logs :- >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> ---------------- >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:17,182::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) >> >> >>>>>>> >>> >>> > >> Owner.cancelAll requests {} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:17,182::task::993::Storage.TaskManager.Task::(_decref) >> >> >>>>>>> >>> >>> > >> Task=`1691d409-9b27-4585-8281-5ec26154367a`::ref 0 >> >> >>>>>>> >>> >>> > >> aborting >> >> >>>>>>> >>> >>> > >> False >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:32,393::task::595::Storage.TaskManager.Task::(_updateState) >> >> >>>>>>> >>> >>> > >> >> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving >> >> >>>>>>> >>> >>> > >> from >> >> >>>>>>> >>> >>> > >> state >> >> >>>>>>> >>> >>> > >> init >> >> >>>>>>> >>> >>> > >> -> >> >> >>>>>>> >>> >>> > >> state preparing >> >> >>>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:32,393::logUtils::44::dispatcher::(wrapper) >> Run >> >> >>>>>>> >>> >>> > >> and >> >> >>>>>>> >>> >>> > >> protect: >> >> >>>>>>> >>> >>> > >> repoStats(options=None) >> >> >>>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:32,393::logUtils::47::dispatcher::(wrapper) >> Run >> >> >>>>>>> >>> >>> > >> and >> >> >>>>>>> >>> >>> > >> protect: >> >> >>>>>>> >>> >>> > >> repoStats, Return response: {} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:32,393::task::1191::Storage.TaskManager.Task::(prepare) >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::finished: {} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:32,394::task::595::Storage.TaskManager.Task::(_updateState) >> >> >>>>>>> >>> >>> > >> >> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving >> >> >>>>>>> >>> >>> > >> from >> >> >>>>>>> >>> >>> > >> state >> >> >>>>>>> >>> >>> > >> preparing >> >> >>>>>>> >>> >>> > >> -> >> >> >>>>>>> >>> >>> > >> state finished >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:32,394::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) >> >> >>>>>>> >>> >>> > >> Owner.releaseAll requests {} resources {} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:32,394::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) >> >> >>>>>>> >>> >>> > >> Owner.cancelAll requests {} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:32,394::task::993::Storage.TaskManager.Task::(_decref) >> >> >>>>>>> >>> >>> > >> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::ref 0 >> >> >>>>>>> >>> >>> > >> aborting >> >> >>>>>>> >>> >>> > >> False >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,550::BindingXMLRPC::1132::vds::(wrapper) >> >> >>>>>>> >>> >>> > >> client >> >> >>>>>>> >>> >>> > >> [10.10.10.2]::call >> >> >>>>>>> >>> >>> > >> getCapabilities with () {} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,553::utils::738::root::(execCmd) >> >> >>>>>>> >>> >>> > >> /sbin/ip route show to 0.0.0.0/0 table all (cwd >> None) >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,560::utils::758::root::(execCmd) >> >> >>>>>>> >>> >>> > >> SUCCESS: <err> = ''; <rc> = 0 >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,588::caps::728::root::(_getKeyPackages) >> rpm >> >> >>>>>>> >>> >>> > >> package >> >> >>>>>>> >>> >>> > >> ('gluster-swift',) not found >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,592::caps::728::root::(_getKeyPackages) >> rpm >> >> >>>>>>> >>> >>> > >> package >> >> >>>>>>> >>> >>> > >> ('gluster-swift-object',) not found >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,593::caps::728::root::(_getKeyPackages) >> rpm >> >> >>>>>>> >>> >>> > >> package >> >> >>>>>>> >>> >>> > >> ('gluster-swift-plugin',) not found >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,598::caps::728::root::(_getKeyPackages) >> rpm >> >> >>>>>>> >>> >>> > >> package >> >> >>>>>>> >>> >>> > >> ('gluster-swift-account',) not found >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,598::caps::728::root::(_getKeyPackages) >> rpm >> >> >>>>>>> >>> >>> > >> package >> >> >>>>>>> >>> >>> > >> ('gluster-swift-proxy',) not found >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,598::caps::728::root::(_getKeyPackages) >> rpm >> >> >>>>>>> >>> >>> > >> package >> >> >>>>>>> >>> >>> > >> ('gluster-swift-doc',) not found >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,599::caps::728::root::(_getKeyPackages) >> rpm >> >> >>>>>>> >>> >>> > >> package >> >> >>>>>>> >>> >>> > >> ('gluster-swift-container',) not found >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,599::caps::728::root::(_getKeyPackages) >> rpm >> >> >>>>>>> >>> >>> > >> package >> >> >>>>>>> >>> >>> > >> ('glusterfs-geo-replication',) not found >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,600::caps::646::root::(get) >> >> >>>>>>> >>> >>> > >> VirtioRNG DISABLED: libvirt version >> 0.10.2-29.el6_5.9 >> >> >>>>>>> >>> >>> > >> required >> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> 0.10.2-31 >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,603::BindingXMLRPC::1139::vds::(wrapper) >> >> >>>>>>> >>> >>> > >> return >> >> >>>>>>> >>> >>> > >> getCapabilities >> >> >>>>>>> >>> >>> > >> with {'status': {'message': 'Done', 'code': 0}, >> >> >>>>>>> >>> >>> > >> 'info': >> >> >>>>>>> >>> >>> > >> {'HBAInventory': >> >> >>>>>>> >>> >>> > >> {'iSCSI': [{'InitiatorName': >> >> >>>>>>> >>> >>> > >> 'iqn.1994-05.com.redhat:32151ce183c8'}], >> >> >>>>>>> >>> >>> > >> 'FC': >> >> >>>>>>> >>> >>> > >> []}, 'packages2': {'kernel': {'release': >> >> >>>>>>> >>> >>> > >> '431.el6.x86_64', >> >> >>>>>>> >>> >>> > >> 'buildtime': >> >> >>>>>>> >>> >>> > >> 1385061309.0, 'version': '2.6.32'}, >> 'glusterfs-rdma': >> >> >>>>>>> >>> >>> > >> {'release': >> >> >>>>>>> >>> >>> > >> '1.el6', >> >> >>>>>>> >>> >>> > >> 'buildtime': 1403622628L, 'version': '3.5.1'}, >> >> >>>>>>> >>> >>> > >> 'glusterfs-fuse': >> >> >>>>>>> >>> >>> > >> {'release': '1.el6', 'buildtime': 1403622628L, >> >> >>>>>>> >>> >>> > >> 'version': >> >> >>>>>>> >>> >>> > >> '3.5.1'}, >> >> >>>>>>> >>> >>> > >> 'spice-server': {'release': '6.el6_5.2', >> 'buildtime': >> >> >>>>>>> >>> >>> > >> 1402324637L, >> >> >>>>>>> >>> >>> > >> 'version': '0.12.4'}, 'vdsm': {'release': >> >> >>>>>>> >>> >>> > >> '1.gitdb83943.el6', >> >> >>>>>>> >>> >>> > >> 'buildtime': >> >> >>>>>>> >>> >>> > >> 1412784567L, 'version': '4.16.7'}, 'qemu-kvm': >> >> >>>>>>> >>> >>> > >> {'release': >> >> >>>>>>> >>> >>> > >> '2.415.el6_5.10', 'buildtime': 1402435700L, >> >> >>>>>>> >>> >>> > >> 'version': >> >> >>>>>>> >>> >>> > >> '0.12.1.2'}, >> >> >>>>>>> >>> >>> > >> 'qemu-img': {'release': '2.415.el6_5.10', >> >> >>>>>>> >>> >>> > >> 'buildtime': >> >> >>>>>>> >>> >>> > >> 1402435700L, >> >> >>>>>>> >>> >>> > >> 'version': '0.12.1.2'}, 'libvirt': {'release': >> >> >>>>>>> >>> >>> > >> '29.el6_5.9', >> >> >>>>>>> >>> >>> > >> 'buildtime': >> >> >>>>>>> >>> >>> > >> 1402404612L, 'version': '0.10.2'}, 'glusterfs': >> >> >>>>>>> >>> >>> > >> {'release': >> >> >>>>>>> >>> >>> > >> '1.el6', >> >> >>>>>>> >>> >>> > >> 'buildtime': 1403622628L, 'version': '3.5.1'}, >> 'mom': >> >> >>>>>>> >>> >>> > >> {'release': >> >> >>>>>>> >>> >>> > >> '2.el6', >> >> >>>>>>> >>> >>> > >> 'buildtime': 1403794344L, 'version': '0.4.1'}, >> >> >>>>>>> >>> >>> > >> 'glusterfs-server': >> >> >>>>>>> >>> >>> > >> {'release': '1.el6', 'buildtime': 1403622628L, >> >> >>>>>>> >>> >>> > >> 'version': >> >> >>>>>>> >>> >>> > >> '3.5.1'}}, >> >> >>>>>>> >>> >>> > >> 'numaNodeDistance': {'1': [20, 10], '0': [10, >> 20]}, >> >> >>>>>>> >>> >>> > >> 'cpuModel': >> >> >>>>>>> >>> >>> > >> 'Intel(R) >> >> >>>>>>> >>> >>> > >> Xeon(R) CPU X5650 @ 2.67GHz', >> 'liveMerge': >> >> >>>>>>> >>> >>> > >> 'false', >> >> >>>>>>> >>> >>> > >> 'hooks': >> >> >>>>>>> >>> >>> > >> {}, >> >> >>>>>>> >>> >>> > >> 'cpuSockets': '2', 'vmTypes': ['kvm'], 'selinux': >> >> >>>>>>> >>> >>> > >> {'mode': '1'}, >> >> >>>>>>> >>> >>> > >> 'kdumpStatus': 0, 'supportedProtocols': ['2.2', >> >> >>>>>>> >>> >>> > >> '2.3'], >> >> >>>>>>> >>> >>> > >> 'networks': >> >> >>>>>>> >>> >>> > >> {'ovirtmgmt': {'iface': u'bond0.10', 'addr': >> >> >>>>>>> >>> >>> > >> '43.252.176.16', >> >> >>>>>>> >>> >>> > >> 'bridged': >> >> >>>>>>> >>> >>> > >> False, 'ipv6addrs': >> ['fe80::62eb:69ff:fe20:b46c/64'], >> >> >>>>>>> >>> >>> > >> 'mtu': >> >> >>>>>>> >>> >>> > >> '1500', >> >> >>>>>>> >>> >>> > >> 'bootproto4': 'none', 'netmask': '255.255.255.0', >> >> >>>>>>> >>> >>> > >> 'ipv4addrs': >> >> >>>>>>> >>> >>> > >> [' >> >> >>>>>>> >>> >>> > >> 43.252.176.16/24' <http://43.252.176.16/24%27>], >> >> >>>>>>> >>> >>> > >> 'interface': >> >> >>>>>>> >>> >>> > >> u'bond0.10', 'ipv6gateway': '::', 'gateway': >> >> >>>>>>> >>> >>> > >> '43.25.17.1'}, >> >> >>>>>>> >>> >>> > >> 'Internal': >> >> >>>>>>> >>> >>> > >> {'iface': 'Internal', 'addr': '', 'cfg': >> {'DEFROUTE': >> >> >>>>>>> >>> >>> > >> 'no', >> >> >>>>>>> >>> >>> > >> 'HOTPLUG': >> >> >>>>>>> >>> >>> > >> 'no', 'MTU': '9000', 'DELAY': '0', >> 'NM_CONTROLLED': >> >> >>>>>>> >>> >>> > >> 'no', >> >> >>>>>>> >>> >>> > >> 'BOOTPROTO': >> >> >>>>>>> >>> >>> > >> 'none', 'STP': 'off', 'DEVICE': 'Internal', >> 'TYPE': >> >> >>>>>>> >>> >>> > >> 'Bridge', >> >> >>>>>>> >>> >>> > >> 'ONBOOT': >> >> >>>>>>> >>> >>> > >> 'no'}, 'bridged': True, 'ipv6addrs': >> >> >>>>>>> >>> >>> > >> ['fe80::210:18ff:fecd:daac/64'], >> >> >>>>>>> >>> >>> > >> 'gateway': '', 'bootproto4': 'none', 'netmask': >> '', >> >> >>>>>>> >>> >>> > >> 'stp': >> >> >>>>>>> >>> >>> > >> 'off', >> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [], 'mtu': '9000', 'ipv6gateway': >> '::', >> >> >>>>>>> >>> >>> > >> 'ports': >> >> >>>>>>> >>> >>> > >> ['bond1.100']}, 'storage': {'iface': u'bond1', >> >> >>>>>>> >>> >>> > >> 'addr': >> >> >>>>>>> >>> >>> > >> '10.10.10.6', >> >> >>>>>>> >>> >>> > >> 'bridged': False, 'ipv6addrs': >> >> >>>>>>> >>> >>> > >> ['fe80::210:18ff:fecd:daac/64'], >> >> >>>>>>> >>> >>> > >> 'mtu': >> >> >>>>>>> >>> >>> > >> '9000', 'bootproto4': 'none', 'netmask': >> >> >>>>>>> >>> >>> > >> '255.255.255.0', >> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [' >> >> >>>>>>> >>> >>> > >> 10.10.10.6/24' <http://10.10.10.6/24%27>], >> >> >>>>>>> >>> >>> > >> 'interface': >> >> >>>>>>> >>> >>> > >> u'bond1', >> >> >>>>>>> >>> >>> > >> 'ipv6gateway': '::', 'gateway': ''}, 'VMNetwork': >> >> >>>>>>> >>> >>> > >> {'iface': >> >> >>>>>>> >>> >>> > >> 'VMNetwork', >> >> >>>>>>> >>> >>> > >> 'addr': '', 'cfg': {'DEFROUTE': 'no', 'HOTPLUG': >> >> >>>>>>> >>> >>> > >> 'no', >> >> >>>>>>> >>> >>> > >> 'MTU': >> >> >>>>>>> >>> >>> > >> '1500', >> >> >>>>>>> >>> >>> > >> 'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO': >> >> >>>>>>> >>> >>> > >> 'none', >> >> >>>>>>> >>> >>> > >> 'STP': >> >> >>>>>>> >>> >>> > >> 'off', >> >> >>>>>>> >>> >>> > >> 'DEVICE': 'VMNetwork', 'TYPE': 'Bridge', 'ONBOOT': >> >> >>>>>>> >>> >>> > >> 'no'}, >> >> >>>>>>> >>> >>> > >> 'bridged': >> >> >>>>>>> >>> >>> > >> True, >> >> >>>>>>> >>> >>> > >> 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'], >> >> >>>>>>> >>> >>> > >> 'gateway': >> >> >>>>>>> >>> >>> > >> '', >> >> >>>>>>> >>> >>> > >> 'bootproto4': >> >> >>>>>>> >>> >>> > >> 'none', 'netmask': '', 'stp': 'off', 'ipv4addrs': >> [], >> >> >>>>>>> >>> >>> > >> 'mtu': >> >> >>>>>>> >>> >>> > >> '1500', >> >> >>>>>>> >>> >>> > >> 'ipv6gateway': '::', 'ports': ['bond0.36']}}, >> >> >>>>>>> >>> >>> > >> 'bridges': >> >> >>>>>>> >>> >>> > >> {'Internal': >> >> >>>>>>> >>> >>> > >> {'addr': '', 'cfg': {'DEFROUTE': 'no', 'HOTPLUG': >> >> >>>>>>> >>> >>> > >> 'no', >> >> >>>>>>> >>> >>> > >> 'MTU': >> >> >>>>>>> >>> >>> > >> '9000', >> >> >>>>>>> >>> >>> > >> 'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO': >> >> >>>>>>> >>> >>> > >> 'none', >> >> >>>>>>> >>> >>> > >> 'STP': >> >> >>>>>>> >>> >>> > >> 'off', >> >> >>>>>>> >>> >>> > >> 'DEVICE': 'Internal', 'TYPE': 'Bridge', 'ONBOOT': >> >> >>>>>>> >>> >>> > >> 'no'}, >> >> >>>>>>> >>> >>> > >> 'ipv6addrs': >> >> >>>>>>> >>> >>> > >> ['fe80::210:18ff:fecd:daac/64'], 'mtu': '9000', >> >> >>>>>>> >>> >>> > >> 'netmask': '', >> >> >>>>>>> >>> >>> > >> 'stp': >> >> >>>>>>> >>> >>> > >> 'off', 'ipv4addrs': [], 'ipv6gateway': '::', >> >> >>>>>>> >>> >>> > >> 'gateway': >> >> >>>>>>> >>> >>> > >> '', >> >> >>>>>>> >>> >>> > >> 'opts': >> >> >>>>>>> >>> >>> > >> {'topology_change_detected': '0', >> >> >>>>>>> >>> >>> > >> 'multicast_last_member_count': >> >> >>>>>>> >>> >>> > >> '2', >> >> >>>>>>> >>> >>> > >> 'hash_elasticity': '4', >> >> >>>>>>> >>> >>> > >> 'multicast_query_response_interval': >> >> >>>>>>> >>> >>> > >> '999', >> >> >>>>>>> >>> >>> > >> 'multicast_snooping': '1', >> >> >>>>>>> >>> >>> > >> 'multicast_startup_query_interval': >> >> >>>>>>> >>> >>> > >> '3124', >> >> >>>>>>> >>> >>> > >> 'hello_timer': '31', 'multicast_querier_interval': >> >> >>>>>>> >>> >>> > >> '25496', >> >> >>>>>>> >>> >>> > >> 'max_age': >> >> >>>>>>> >>> >>> > >> '1999', 'hash_max': '512', 'stp_state': '0', >> >> >>>>>>> >>> >>> > >> 'root_id': >> >> >>>>>>> >>> >>> > >> '8000.001018cddaac', 'priority': '32768', >> >> >>>>>>> >>> >>> > >> 'multicast_membership_interval': >> >> >>>>>>> >>> >>> > >> '25996', 'root_path_cost': '0', 'root_port': '0', >> >> >>>>>>> >>> >>> > >> 'multicast_querier': >> >> >>>>>>> >>> >>> > >> '0', >> >> >>>>>>> >>> >>> > >> 'multicast_startup_query_count': '2', >> 'hello_time': >> >> >>>>>>> >>> >>> > >> '199', >> >> >>>>>>> >>> >>> > >> 'topology_change': '0', 'bridge_id': >> >> >>>>>>> >>> >>> > >> '8000.001018cddaac', >> >> >>>>>>> >>> >>> > >> 'topology_change_timer': '0', 'ageing_time': >> '29995', >> >> >>>>>>> >>> >>> > >> 'gc_timer': >> >> >>>>>>> >>> >>> > >> '31', >> >> >>>>>>> >>> >>> > >> 'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0', >> >> >>>>>>> >>> >>> > >> 'multicast_query_interval': '12498', >> >> >>>>>>> >>> >>> > >> 'multicast_last_member_interval': >> >> >>>>>>> >>> >>> > >> '99', 'multicast_router': '1', 'forward_delay': >> '0'}, >> >> >>>>>>> >>> >>> > >> 'ports': >> >> >>>>>>> >>> >>> > >> ['bond1.100']}, 'VMNetwork': {'addr': '', 'cfg': >> >> >>>>>>> >>> >>> > >> {'DEFROUTE': >> >> >>>>>>> >>> >>> > >> 'no', >> >> >>>>>>> >>> >>> > >> 'HOTPLUG': 'no', 'MTU': '1500', 'DELAY': '0', >> >> >>>>>>> >>> >>> > >> 'NM_CONTROLLED': >> >> >>>>>>> >>> >>> > >> 'no', >> >> >>>>>>> >>> >>> > >> 'BOOTPROTO': 'none', 'STP': 'off', 'DEVICE': >> >> >>>>>>> >>> >>> > >> 'VMNetwork', >> >> >>>>>>> >>> >>> > >> 'TYPE': >> >> >>>>>>> >>> >>> > >> 'Bridge', >> >> >>>>>>> >>> >>> > >> 'ONBOOT': 'no'}, 'ipv6addrs': >> >> >>>>>>> >>> >>> > >> ['fe80::62eb:69ff:fe20:b46c/64'], >> >> >>>>>>> >>> >>> > >> 'mtu': >> >> >>>>>>> >>> >>> > >> '1500', 'netmask': '', 'stp': 'off', 'ipv4addrs': >> [], >> >> >>>>>>> >>> >>> > >> 'ipv6gateway': >> >> >>>>>>> >>> >>> > >> '::', >> >> >>>>>>> >>> >>> > >> 'gateway': '', 'opts': >> {'topology_change_detected': >> >> >>>>>>> >>> >>> > >> '0', >> >> >>>>>>> >>> >>> > >> 'multicast_last_member_count': '2', >> >> >>>>>>> >>> >>> > >> 'hash_elasticity': >> >> >>>>>>> >>> >>> > >> '4', >> >> >>>>>>> >>> >>> > >> 'multicast_query_response_interval': '999', >> >> >>>>>>> >>> >>> > >> 'multicast_snooping': >> >> >>>>>>> >>> >>> > >> '1', >> >> >>>>>>> >>> >>> > >> 'multicast_startup_query_interval': '3124', >> >> >>>>>>> >>> >>> > >> 'hello_timer': >> >> >>>>>>> >>> >>> > >> '131', >> >> >>>>>>> >>> >>> > >> 'multicast_querier_interval': '25496', 'max_age': >> >> >>>>>>> >>> >>> > >> '1999', >> >> >>>>>>> >>> >>> > >> 'hash_max': >> >> >>>>>>> >>> >>> > >> '512', 'stp_state': '0', 'root_id': >> >> >>>>>>> >>> >>> > >> '8000.60eb6920b46c', >> >> >>>>>>> >>> >>> > >> 'priority': >> >> >>>>>>> >>> >>> > >> '32768', 'multicast_membership_interval': '25996', >> >> >>>>>>> >>> >>> > >> 'root_path_cost': >> >> >>>>>>> >>> >>> > >> '0', >> >> >>>>>>> >>> >>> > >> 'root_port': '0', 'multicast_querier': '0', >> >> >>>>>>> >>> >>> > >> 'multicast_startup_query_count': '2', >> 'hello_time': >> >> >>>>>>> >>> >>> > >> '199', >> >> >>>>>>> >>> >>> > >> 'topology_change': '0', 'bridge_id': >> >> >>>>>>> >>> >>> > >> '8000.60eb6920b46c', >> >> >>>>>>> >>> >>> > >> 'topology_change_timer': '0', 'ageing_time': >> '29995', >> >> >>>>>>> >>> >>> > >> 'gc_timer': >> >> >>>>>>> >>> >>> > >> '31', >> >> >>>>>>> >>> >>> > >> 'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0', >> >> >>>>>>> >>> >>> > >> 'multicast_query_interval': '12498', >> >> >>>>>>> >>> >>> > >> 'multicast_last_member_interval': >> >> >>>>>>> >>> >>> > >> '99', 'multicast_router': '1', 'forward_delay': >> '0'}, >> >> >>>>>>> >>> >>> > >> 'ports': >> >> >>>>>>> >>> >>> > >> ['bond0.36']}}, 'uuid': >> >> >>>>>>> >>> >>> > >> '44454C4C-4C00-1057-8053-B7C04F504E31', >> >> >>>>>>> >>> >>> > >> 'lastClientIface': 'bond1', 'nics': {'eth3': >> >> >>>>>>> >>> >>> > >> {'permhwaddr': >> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ae', 'addr': '', 'cfg': {'SLAVE': >> >> >>>>>>> >>> >>> > >> 'yes', >> >> >>>>>>> >>> >>> > >> 'NM_CONTROLLED': >> >> >>>>>>> >>> >>> > >> 'no', 'MTU': '9000', 'HWADDR': >> '00:10:18:cd:da:ae', >> >> >>>>>>> >>> >>> > >> 'MASTER': >> >> >>>>>>> >>> >>> > >> 'bond1', >> >> >>>>>>> >>> >>> > >> 'DEVICE': 'eth3', 'ONBOOT': 'no'}, 'ipv6addrs': >> [], >> >> >>>>>>> >>> >>> > >> 'mtu': >> >> >>>>>>> >>> >>> > >> '9000', >> >> >>>>>>> >>> >>> > >> 'netmask': '', 'ipv4addrs': [], 'hwaddr': >> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ac', >> >> >>>>>>> >>> >>> > >> 'speed': >> >> >>>>>>> >>> >>> > >> 1000}, 'eth2': {'permhwaddr': '00:10:18:cd:da:ac', >> >> >>>>>>> >>> >>> > >> 'addr': '', >> >> >>>>>>> >>> >>> > >> 'cfg': >> >> >>>>>>> >>> >>> > >> {'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU': >> >> >>>>>>> >>> >>> > >> '9000', >> >> >>>>>>> >>> >>> > >> 'HWADDR': >> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ac', 'MASTER': 'bond1', 'DEVICE': >> >> >>>>>>> >>> >>> > >> 'eth2', >> >> >>>>>>> >>> >>> > >> 'ONBOOT': >> >> >>>>>>> >>> >>> > >> 'no'}, >> >> >>>>>>> >>> >>> > >> 'ipv6addrs': [], 'mtu': '9000', 'netmask': '', >> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [], >> >> >>>>>>> >>> >>> > >> 'hwaddr': >> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ac', 'speed': 1000}, 'eth1': >> >> >>>>>>> >>> >>> > >> {'permhwaddr': >> >> >>>>>>> >>> >>> > >> '60:eb:69:20:b4:6d', 'addr': '', 'cfg': {'SLAVE': >> >> >>>>>>> >>> >>> > >> 'yes', >> >> >>>>>>> >>> >>> > >> 'NM_CONTROLLED': >> >> >>>>>>> >>> >>> > >> 'no', 'MTU': '1500', 'HWADDR': >> '60:eb:69:20:b4:6d', >> >> >>>>>>> >>> >>> > >> 'MASTER': >> >> >>>>>>> >>> >>> > >> 'bond0', >> >> >>>>>>> >>> >>> > >> 'DEVICE': 'eth1', 'ONBOOT': 'yes'}, 'ipv6addrs': >> [], >> >> >>>>>>> >>> >>> > >> 'mtu': >> >> >>>>>>> >>> >>> > >> '1500', >> >> >>>>>>> >>> >>> > >> 'netmask': '', 'ipv4addrs': [], 'hwaddr': >> >> >>>>>>> >>> >>> > >> '60:eb:69:20:b4:6c', >> >> >>>>>>> >>> >>> > >> 'speed': >> >> >>>>>>> >>> >>> > >> 1000}, 'eth0': {'permhwaddr': '60:eb:69:20:b4:6c', >> >> >>>>>>> >>> >>> > >> 'addr': '', >> >> >>>>>>> >>> >>> > >> 'cfg': >> >> >>>>>>> >>> >>> > >> {'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU': >> >> >>>>>>> >>> >>> > >> '1500', >> >> >>>>>>> >>> >>> > >> 'HWADDR': >> >> >>>>>>> >>> >>> > >> '60:eb:69:20:b4:6c', 'MASTER': 'bond0', 'DEVICE': >> >> >>>>>>> >>> >>> > >> 'eth0', >> >> >>>>>>> >>> >>> > >> 'ONBOOT': >> >> >>>>>>> >>> >>> > >> 'yes'}, >> >> >>>>>>> >>> >>> > >> 'ipv6addrs': [], 'mtu': '1500', 'netmask': '', >> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [], >> >> >>>>>>> >>> >>> > >> 'hwaddr': >> >> >>>>>>> >>> >>> > >> '60:eb:69:20:b4:6c', 'speed': 1000}}, >> >> >>>>>>> >>> >>> > >> 'software_revision': '1', >> >> >>>>>>> >>> >>> > >> 'clusterLevels': ['3.0', '3.1', '3.2', '3.3', >> '3.4', >> >> >>>>>>> >>> >>> > >> '3.5'], >> >> >>>>>>> >>> >>> > >> 'cpuFlags': >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> u'fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,pdpe1gb,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,xtopology,nonstop_tsc,pni,pclmulqdq,dtes64,monitor,ds_cpl,vmx,smx,est,tm2,ssse3,cx16,xtpr,pdcm,pcid,dca,sse4_1,sse4_2,popcnt,aes,lahf_lm,tpr_shadow,vnmi,flexpriority,ept,vpid,model_Nehalem,model_Conroe,model_coreduo,model_core2duo,model_Penryn,model_Westmere,model_n270', >> >> >>>>>>> >>> >>> > >> 'ISCSIInitiatorName': >> >> >>>>>>> >>> >>> > >> 'iqn.1994-05.com.redhat:32151ce183c8', >> >> >>>>>>> >>> >>> > >> 'netConfigDirty': 'False', 'supportedENGINEs': >> >> >>>>>>> >>> >>> > >> ['3.0', >> >> >>>>>>> >>> >>> > >> '3.1', >> >> >>>>>>> >>> >>> > >> '3.2', >> >> >>>>>>> >>> >>> > >> '3.3', >> >> >>>>>>> >>> >>> > >> '3.4', '3.5'], 'autoNumaBalancing': 2, >> 'reservedMem': >> >> >>>>>>> >>> >>> > >> '321', >> >> >>>>>>> >>> >>> > >> 'bondings': >> >> >>>>>>> >>> >>> > >> {'bond4': {'addr': '', 'cfg': {}, 'mtu': '1500', >> >> >>>>>>> >>> >>> > >> 'netmask': '', >> >> >>>>>>> >>> >>> > >> 'slaves': >> >> >>>>>>> >>> >>> > >> [], 'hwaddr': '00:00:00:00:00:00'}, 'bond0': >> {'addr': >> >> >>>>>>> >>> >>> > >> '', >> >> >>>>>>> >>> >>> > >> 'cfg': >> >> >>>>>>> >>> >>> > >> {'HOTPLUG': 'no', 'MTU': '1500', 'NM_CONTROLLED': >> >> >>>>>>> >>> >>> > >> 'no', >> >> >>>>>>> >>> >>> > >> 'BONDING_OPTS': >> >> >>>>>>> >>> >>> > >> 'mode=4 miimon=100', 'DEVICE': 'bond0', 'ONBOOT': >> >> >>>>>>> >>> >>> > >> 'yes'}, >> >> >>>>>>> >>> >>> > >> 'ipv6addrs': >> >> >>>>>>> >>> >>> > >> ['fe80::62eb:69ff:fe20:b46c/64'], 'mtu': '1500', >> >> >>>>>>> >>> >>> > >> 'netmask': '', >> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [], 'hwaddr': '60:eb:69:20:b4:6c', >> >> >>>>>>> >>> >>> > >> 'slaves': >> >> >>>>>>> >>> >>> > >> ['eth0', >> >> >>>>>>> >>> >>> > >> 'eth1'], >> >> >>>>>>> >>> >>> > >> 'opts': {'miimon': '100', 'mode': '4'}}, 'bond1': >> >> >>>>>>> >>> >>> > >> {'addr': >> >> >>>>>>> >>> >>> > >> '10.10.10.6', >> >> >>>>>>> >>> >>> > >> 'cfg': {'DEFROUTE': 'no', 'IPADDR': '10.10.10.6', >> >> >>>>>>> >>> >>> > >> 'HOTPLUG': >> >> >>>>>>> >>> >>> > >> 'no', >> >> >>>>>>> >>> >>> > >> 'MTU': >> >> >>>>>>> >>> >>> > >> '9000', 'NM_CONTROLLED': 'no', 'NETMASK': >> >> >>>>>>> >>> >>> > >> '255.255.255.0', >> >> >>>>>>> >>> >>> > >> 'BOOTPROTO': >> >> >>>>>>> >>> >>> > >> 'none', 'BONDING_OPTS': 'mode=4 miimon=100', >> >> >>>>>>> >>> >>> > >> 'DEVICE': >> >> >>>>>>> >>> >>> > >> 'bond1', >> >> >>>>>>> >>> >>> > >> 'ONBOOT': >> >> >>>>>>> >>> >>> > >> 'no'}, 'ipv6addrs': >> ['fe80::210:18ff:fecd:daac/64'], >> >> >>>>>>> >>> >>> > >> 'mtu': >> >> >>>>>>> >>> >>> > >> '9000', >> >> >>>>>>> >>> >>> > >> 'netmask': '255.255.255.0', 'ipv4addrs': >> >> >>>>>>> >>> >>> > >> ['10.10.10.6/24' >> >> >>>>>>> >>> >>> > >> <http://10.10.10.6/24%27>], 'hwaddr': >> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ac', >> >> >>>>>>> >>> >>> > >> 'slaves': >> >> >>>>>>> >>> >>> > >> ['eth2', 'eth3'], 'opts': {'miimon': '100', >> 'mode': >> >> >>>>>>> >>> >>> > >> '4'}}, >> >> >>>>>>> >>> >>> > >> 'bond2': >> >> >>>>>>> >>> >>> > >> {'addr': '', 'cfg': {}, 'mtu': '1500', 'netmask': >> '', >> >> >>>>>>> >>> >>> > >> 'slaves': >> >> >>>>>>> >>> >>> > >> [], >> >> >>>>>>> >>> >>> > >> 'hwaddr': '00:00:00:00:00:00'}, 'bond3': {'addr': >> '', >> >> >>>>>>> >>> >>> > >> 'cfg': {}, >> >> >>>>>>> >>> >>> > >> 'mtu': >> >> >>>>>>> >>> >>> > >> '1500', 'netmask': '', 'slaves': [], 'hwaddr': >> >> >>>>>>> >>> >>> > >> '00:00:00:00:00:00'}}, >> >> >>>>>>> >>> >>> > >> 'software_version': '4.16', 'memSize': '24019', >> >> >>>>>>> >>> >>> > >> 'cpuSpeed': >> >> >>>>>>> >>> >>> > >> '2667.000', >> >> >>>>>>> >>> >>> > >> 'numaNodes': {u'1': {'totalMemory': '12288', >> 'cpus': >> >> >>>>>>> >>> >>> > >> [6, >> >> >>>>>>> >>> >>> > >> 7, 8, >> >> >>>>>>> >>> >>> > >> 9, >> >> >>>>>>> >>> >>> > >> 10, 11, >> >> >>>>>>> >>> >>> > >> 18, 19, 20, 21, 22, 23]}, u'0': {'totalMemory': >> >> >>>>>>> >>> >>> > >> '12278', >> >> >>>>>>> >>> >>> > >> 'cpus': >> >> >>>>>>> >>> >>> > >> [0, >> >> >>>>>>> >>> >>> > >> 1, 2, >> >> >>>>>>> >>> >>> > >> 3, 4, 5, 12, 13, 14, 15, 16, 17]}}, >> 'version_name': >> >> >>>>>>> >>> >>> > >> 'Snow >> >> >>>>>>> >>> >>> > >> Man', >> >> >>>>>>> >>> >>> > >> 'vlans': >> >> >>>>>>> >>> >>> > >> {'bond0.10': {'iface': 'bond0', 'addr': >> >> >>>>>>> >>> >>> > >> '43.25.17.16', >> >> >>>>>>> >>> >>> > >> 'cfg': >> >> >>>>>>> >>> >>> > >> {'DEFROUTE': >> >> >>>>>>> >>> >>> > >> 'yes', 'VLAN': 'yes', 'IPADDR': '43.25.17.16', >> >> >>>>>>> >>> >>> > >> 'HOTPLUG': >> >> >>>>>>> >>> >>> > >> 'no', >> >> >>>>>>> >>> >>> > >> 'GATEWAY': >> >> >>>>>>> >>> >>> > >> '43.25.17.1', 'NM_CONTROLLED': 'no', 'NETMASK': >> >> >>>>>>> >>> >>> > >> '255.255.255.0', >> >> >>>>>>> >>> >>> > >> 'BOOTPROTO': 'none', 'DEVICE': 'bond0.10', 'MTU': >> >> >>>>>>> >>> >>> > >> '1500', >> >> >>>>>>> >>> >>> > >> 'ONBOOT': >> >> >>>>>>> >>> >>> > >> 'yes'}, >> >> >>>>>>> >>> >>> > >> 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'], >> >> >>>>>>> >>> >>> > >> 'vlanid': >> >> >>>>>>> >>> >>> > >> 10, >> >> >>>>>>> >>> >>> > >> 'mtu': >> >> >>>>>>> >>> >>> > >> '1500', >> >> >>>>>>> >>> >>> > >> 'netmask': '255.255.255.0', 'ipv4addrs': >> >> >>>>>>> >>> >>> > >> ['43.25.17.16/24'] >> >> >>>>>>> >>> >>> > >> <http://43.25.17.16/24%27%5D>}, 'bond0.36': >> {'iface': >> >> >>>>>>> >>> >>> > >> 'bond0', >> >> >>>>>>> >>> >>> > >> 'addr': >> >> >>>>>>> >>> >>> > >> '', 'cfg': {'BRIDGE': 'VMNetwork', 'VLAN': 'yes', >> >> >>>>>>> >>> >>> > >> 'HOTPLUG': >> >> >>>>>>> >>> >>> > >> 'no', >> >> >>>>>>> >>> >>> > >> 'MTU': >> >> >>>>>>> >>> >>> > >> '1500', 'NM_CONTROLLED': 'no', 'DEVICE': >> 'bond0.36', >> >> >>>>>>> >>> >>> > >> 'ONBOOT': >> >> >>>>>>> >>> >>> > >> 'no'}, >> >> >>>>>>> >>> >>> > >> 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'], >> >> >>>>>>> >>> >>> > >> 'vlanid': >> >> >>>>>>> >>> >>> > >> 36, >> >> >>>>>>> >>> >>> > >> 'mtu': >> >> >>>>>>> >>> >>> > >> '1500', >> >> >>>>>>> >>> >>> > >> 'netmask': '', 'ipv4addrs': []}, 'bond1.100': >> >> >>>>>>> >>> >>> > >> {'iface': >> >> >>>>>>> >>> >>> > >> 'bond1', >> >> >>>>>>> >>> >>> > >> 'addr': >> >> >>>>>>> >>> >>> > >> '', 'cfg': {'BRIDGE': 'Internal', 'VLAN': 'yes', >> >> >>>>>>> >>> >>> > >> 'HOTPLUG': >> >> >>>>>>> >>> >>> > >> 'no', >> >> >>>>>>> >>> >>> > >> 'MTU': >> >> >>>>>>> >>> >>> > >> '9000', 'NM_CONTROLLED': 'no', 'DEVICE': >> 'bond1.100', >> >> >>>>>>> >>> >>> > >> 'ONBOOT': >> >> >>>>>>> >>> >>> > >> 'no'}, >> >> >>>>>>> >>> >>> > >> 'ipv6addrs': ['fe80::210:18ff:fecd:daac/64'], >> >> >>>>>>> >>> >>> > >> 'vlanid': >> >> >>>>>>> >>> >>> > >> 100, >> >> >>>>>>> >>> >>> > >> 'mtu': >> >> >>>>>>> >>> >>> > >> '9000', >> >> >>>>>>> >>> >>> > >> 'netmask': '', 'ipv4addrs': []}}, 'cpuCores': >> '12', >> >> >>>>>>> >>> >>> > >> 'kvmEnabled': >> >> >>>>>>> >>> >>> > >> 'true', >> >> >>>>>>> >>> >>> > >> 'guestOverhead': '65', 'cpuThreads': '24', >> >> >>>>>>> >>> >>> > >> 'emulatedMachines': >> >> >>>>>>> >>> >>> > >> [u'rhel6.5.0', u'pc', u'rhel6.4.0', u'rhel6.3.0', >> >> >>>>>>> >>> >>> > >> u'rhel6.2.0', >> >> >>>>>>> >>> >>> > >> u'rhel6.1.0', u'rhel6.0.0', u'rhel5.5.0', >> >> >>>>>>> >>> >>> > >> u'rhel5.4.4', >> >> >>>>>>> >>> >>> > >> u'rhel5.4.0'], >> >> >>>>>>> >>> >>> > >> 'operatingSystem': {'release': >> '5.el6.centos.11.1', >> >> >>>>>>> >>> >>> > >> 'version': >> >> >>>>>>> >>> >>> > >> '6', >> >> >>>>>>> >>> >>> > >> 'name': >> >> >>>>>>> >>> >>> > >> 'RHEL'}, 'lastClient': '10.10.10.2'}} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,620::BindingXMLRPC::1132::vds::(wrapper) >> >> >>>>>>> >>> >>> > >> client >> >> >>>>>>> >>> >>> > >> [10.10.10.2]::call >> >> >>>>>>> >>> >>> > >> getHardwareInfo with () {} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,621::BindingXMLRPC::1139::vds::(wrapper) >> >> >>>>>>> >>> >>> > >> return >> >> >>>>>>> >>> >>> > >> getHardwareInfo >> >> >>>>>>> >>> >>> > >> with {'status': {'message': 'Done', 'code': 0}, >> >> >>>>>>> >>> >>> > >> 'info': >> >> >>>>>>> >>> >>> > >> {'systemProductName': 'CS24-TY', >> >> >>>>>>> >>> >>> > >> 'systemSerialNumber': >> >> >>>>>>> >>> >>> > >> '7LWSPN1', >> >> >>>>>>> >>> >>> > >> 'systemFamily': 'Server', 'systemVersion': 'A00', >> >> >>>>>>> >>> >>> > >> 'systemUUID': >> >> >>>>>>> >>> >>> > >> '44454c4c-4c00-1057-8053-b7c04f504e31', >> >> >>>>>>> >>> >>> > >> 'systemManufacturer': >> >> >>>>>>> >>> >>> > >> 'Dell'}} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:41,733::BindingXMLRPC::1132::vds::(wrapper) >> >> >>>>>>> >>> >>> > >> client >> >> >>>>>>> >>> >>> > >> [10.10.10.2]::call >> >> >>>>>>> >>> >>> > >> hostsList with () {} flowID [222e8036] >> >> >>>>>>> >>> >>> > >> Thread-13::ERROR::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:44,753::BindingXMLRPC::1148::vds::(wrapper) >> >> >>>>>>> >>> >>> > >> vdsm >> >> >>>>>>> >>> >>> > >> exception >> >> >>>>>>> >>> >>> > >> occured >> >> >>>>>>> >>> >>> > >> Traceback (most recent call last): >> >> >>>>>>> >>> >>> > >> File "/usr/share/vdsm/rpc/BindingXMLRPC.py", >> line >> >> >>>>>>> >>> >>> > >> 1135, >> >> >>>>>>> >>> >>> > >> in >> >> >>>>>>> >>> >>> > >> wrapper >> >> >>>>>>> >>> >>> > >> res = f(*args, **kwargs) >> >> >>>>>>> >>> >>> > >> File "/usr/share/vdsm/gluster/api.py", line 54, >> in >> >> >>>>>>> >>> >>> > >> wrapper >> >> >>>>>>> >>> >>> > >> rv = func(*args, **kwargs) >> >> >>>>>>> >>> >>> > >> File "/usr/share/vdsm/gluster/api.py", line >> 251, in >> >> >>>>>>> >>> >>> > >> hostsList >> >> >>>>>>> >>> >>> > >> return {'hosts': >> >> >>>>>>> >>> >>> > >> self.svdsmProxy.glusterPeerStatus()} >> >> >>>>>>> >>> >>> > >> File "/usr/share/vdsm/supervdsm.py", line 50, in >> >> >>>>>>> >>> >>> > >> __call__ >> >> >>>>>>> >>> >>> > >> return callMethod() >> >> >>>>>>> >>> >>> > >> File "/usr/share/vdsm/supervdsm.py", line 48, in >> >> >>>>>>> >>> >>> > >> <lambda> >> >> >>>>>>> >>> >>> > >> **kwargs) >> >> >>>>>>> >>> >>> > >> File "<string>", line 2, in glusterPeerStatus >> >> >>>>>>> >>> >>> > >> File >> >> >>>>>>> >>> >>> > >> >> "/usr/lib64/python2.6/multiprocessing/managers.py", >> >> >>>>>>> >>> >>> > >> line >> >> >>>>>>> >>> >>> > >> 740, >> >> >>>>>>> >>> >>> > >> in >> >> >>>>>>> >>> >>> > >> _callmethod >> >> >>>>>>> >>> >>> > >> raise convert_to_error(kind, result) >> >> >>>>>>> >>> >>> > >> GlusterCmdExecFailedException: Command execution >> >> >>>>>>> >>> >>> > >> failed >> >> >>>>>>> >>> >>> > >> error: Connection failed. Please check if gluster >> >> >>>>>>> >>> >>> > >> daemon >> >> >>>>>>> >>> >>> > >> is >> >> >>>>>>> >>> >>> > >> operational. >> >> >>>>>>> >>> >>> > >> return code: 1 >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:50,949::task::595::Storage.TaskManager.Task::(_updateState) >> >> >>>>>>> >>> >>> > >> >> Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving >> >> >>>>>>> >>> >>> > >> from >> >> >>>>>>> >>> >>> > >> state >> >> >>>>>>> >>> >>> > >> init >> >> >>>>>>> >>> >>> > >> -> >> >> >>>>>>> >>> >>> > >> state preparing >> >> >>>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:50,950::logUtils::44::dispatcher::(wrapper) >> Run >> >> >>>>>>> >>> >>> > >> and >> >> >>>>>>> >>> >>> > >> protect: >> >> >>>>>>> >>> >>> > >> repoStats(options=None) >> >> >>>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24 >> >> >>>>>>> >>> >>> > >> 21:41:50,950::logUtils::47::dispatcher::(wrapper) >> Run >> >> >>>>>>> >>> >>> > >> and >> >> >>>>>>> >>> >>> > >> protect: >> >> >>>>>>> >>> >>> > >> repoStats, Return response: {} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:50,950::task::1191::Storage.TaskManager.Task::(prepare) >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> Task=`c9042986-c978-4b08-adb2-616f5299e115`::finished: {} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:50,950::task::595::Storage.TaskManager.Task::(_updateState) >> >> >>>>>>> >>> >>> > >> >> Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving >> >> >>>>>>> >>> >>> > >> from >> >> >>>>>>> >>> >>> > >> state >> >> >>>>>>> >>> >>> > >> preparing >> >> >>>>>>> >>> >>> > >> -> >> >> >>>>>>> >>> >>> > >> state finished >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:50,951::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) >> >> >>>>>>> >>> >>> > >> Owner.releaseAll requests {} resources {} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:50,951::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) >> >> >>>>>>> >>> >>> > >> Owner.cancelAll requests {} >> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> 21:41:50,951::task::993::Storage.TaskManager.Task::(_decref) >> >> >>>>>>> >>> >>> > >> Task=`c9042986-c978-4b08-adb2-616f5299e115`::ref 0 >> >> >>>>>>> >>> >>> > >> aborting >> >> >>>>>>> >>> >>> > >> False >> >> >>>>>>> >>> >>> > >> ------------------------------- >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> [root at compute4 ~]# service glusterd status >> >> >>>>>>> >>> >>> > >> glusterd is stopped >> >> >>>>>>> >>> >>> > >> [root at compute4 ~]# chkconfig --list | grep >> glusterd >> >> >>>>>>> >>> >>> > >> glusterd 0:off 1:off 2:on 3:on >> 4:on >> >> >>>>>>> >>> >>> > >> 5:on >> >> >>>>>>> >>> >>> > >> 6:off >> >> >>>>>>> >>> >>> > >> [root at compute4 ~]# >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> Thanks, >> >> >>>>>>> >>> >>> > >> Punit >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> On Mon, Nov 24, 2014 at 6:36 PM, Kanagaraj >> >> >>>>>>> >>> >>> > >> <kmayilsa at redhat.com> >> >> >>>>>>> >>> >>> > >> wrote: >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >>> Can you send the corresponding error in vdsm.log >> >> >>>>>>> >>> >>> > >>> from >> >> >>>>>>> >>> >>> > >>> the >> >> >>>>>>> >>> >>> > >>> host? >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> Also check if glusterd service is running. >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> Thanks, >> >> >>>>>>> >>> >>> > >>> Kanagaraj >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> On 11/24/2014 03:39 PM, Punit Dambiwal wrote: >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> Hi, >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> After reboot my Hypervisior host can not >> activate >> >> >>>>>>> >>> >>> > >>> again >> >> >>>>>>> >>> >>> > >>> in the >> >> >>>>>>> >>> >>> > >>> cluster >> >> >>>>>>> >>> >>> > >>> and failed with the following error :- >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> Gluster command [<UNKNOWN>] failed on server... >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> Engine logs :- >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:28,397 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-64) START, >> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand(HostName = Compute4, >> >> >>>>>>> >>> >>> > >>> HostId >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id: >> >> >>>>>>> >>> >>> > >>> 5f251c90 >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:30,609 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-64) FINISH, >> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand, >> >> >>>>>>> >>> >>> > >>> return: >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> {26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity at d95203e0 >> }, >> >> >>>>>>> >>> >>> > >>> log id: 5f251c90 >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,768 INFO >> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >> >> >>>>>>> >>> >>> > >>> (ajp--127.0.0.1-8702-8) >> >> >>>>>>> >>> >>> > >>> [287d570d] Lock Acquired to object EngineLock >> >> >>>>>>> >>> >>> > >>> [exclusiveLocks>> >> >>>>>>> >>> >>> > >>> key: >> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a value: VDS >> >> >>>>>>> >>> >>> > >>> , sharedLocks= ] >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,795 INFO >> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] >> >> >>>>>>> >>> >>> > >>> Running >> >> >>>>>>> >>> >>> > >>> command: >> >> >>>>>>> >>> >>> > >>> ActivateVdsCommand internal: false. Entities >> >> >>>>>>> >>> >>> > >>> affected : >> >> >>>>>>> >>> >>> > >>> ID: >> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: >> VDSAction >> >> >>>>>>> >>> >>> > >>> group >> >> >>>>>>> >>> >>> > >>> MANIPULATE_HOST >> >> >>>>>>> >>> >>> > >>> with role type ADMIN >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,796 INFO >> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] >> >> >>>>>>> >>> >>> > >>> Before >> >> >>>>>>> >>> >>> > >>> acquiring >> >> >>>>>>> >>> >>> > >>> lock in >> >> >>>>>>> >>> >>> > >>> order to prevent monitoring for host Compute5 >> from >> >> >>>>>>> >>> >>> > >>> data-center >> >> >>>>>>> >>> >>> > >>> SV_WTC >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,797 INFO >> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] >> Lock >> >> >>>>>>> >>> >>> > >>> acquired, >> >> >>>>>>> >>> >>> > >>> from >> >> >>>>>>> >>> >>> > >>> now a >> >> >>>>>>> >>> >>> > >>> monitoring of host will be skipped for host >> Compute5 >> >> >>>>>>> >>> >>> > >>> from >> >> >>>>>>> >>> >>> > >>> data-center >> >> >>>>>>> >>> >>> > >>> SV_WTC >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,817 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] >> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] >> >> >>>>>>> >>> >>> > >>> START, >> >> >>>>>>> >>> >>> > >>> SetVdsStatusVDSCommand(HostName = Compute5, >> HostId >> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a, >> >> >>>>>>> >>> >>> > >>> status=Unassigned, >> >> >>>>>>> >>> >>> > >>> nonOperationalReason=NONE, >> >> >>>>>>> >>> >>> > >>> stopSpmFailureLogged=false), >> >> >>>>>>> >>> >>> > >>> log id: >> >> >>>>>>> >>> >>> > >>> 1cbc7311 >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,820 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] >> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] >> >> >>>>>>> >>> >>> > >>> FINISH, >> >> >>>>>>> >>> >>> > >>> SetVdsStatusVDSCommand, log id: 1cbc7311 >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:34,086 INFO >> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) Activate >> >> >>>>>>> >>> >>> > >>> finished. >> >> >>>>>>> >>> >>> > >>> Lock >> >> >>>>>>> >>> >>> > >>> released. >> >> >>>>>>> >>> >>> > >>> Monitoring can run now for host Compute5 from >> >> >>>>>>> >>> >>> > >>> data-center >> >> >>>>>>> >>> >>> > >>> SV_WTC >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:34,088 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) Correlation >> ID: >> >> >>>>>>> >>> >>> > >>> 287d570d, >> >> >>>>>>> >>> >>> > >>> Job >> >> >>>>>>> >>> >>> > >>> ID: >> >> >>>>>>> >>> >>> > >>> 5ef8e4d6-b2bc-469e-8e81-7ef74b2a001a, Call Stack: >> >> >>>>>>> >>> >>> > >>> null, >> >> >>>>>>> >>> >>> > >>> Custom >> >> >>>>>>> >>> >>> > >>> Event ID: >> >> >>>>>>> >>> >>> > >>> -1, Message: Host Compute5 was activated by >> admin. >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:34,090 INFO >> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) Lock freed to >> >> >>>>>>> >>> >>> > >>> object >> >> >>>>>>> >>> >>> > >>> EngineLock >> >> >>>>>>> >>> >>> > >>> [exclusiveLocks= key: >> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a >> >> >>>>>>> >>> >>> > >>> value: >> >> >>>>>>> >>> >>> > >>> VDS >> >> >>>>>>> >>> >>> > >>> , sharedLocks= ] >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:35,792 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-55) [3706e836] >> START, >> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand(HostName = Compute4, >> >> >>>>>>> >>> >>> > >>> HostId >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id: >> >> >>>>>>> >>> >>> > >>> 48a0c832 >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,064 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) START, >> >> >>>>>>> >>> >>> > >>> GetHardwareInfoVDSCommand(HostName = Compute5, >> >> >>>>>>> >>> >>> > >>> HostId >> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a, >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> vds=Host[Compute5,0bf6b00f-7947-4411-b55a-cc5eea2b381a]), log >> >> >>>>>>> >>> >>> > >>> id: >> >> >>>>>>> >>> >>> > >>> 6d560cc2 >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,074 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) FINISH, >> >> >>>>>>> >>> >>> > >>> GetHardwareInfoVDSCommand, log >> >> >>>>>>> >>> >>> > >>> id: 6d560cc2 >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,093 WARN >> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.vdsbroker.VdsManager] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) Host Compute5 >> is >> >> >>>>>>> >>> >>> > >>> running >> >> >>>>>>> >>> >>> > >>> with >> >> >>>>>>> >>> >>> > >>> disabled >> >> >>>>>>> >>> >>> > >>> SELinux. >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,127 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf] >> >> >>>>>>> >>> >>> > >>> Running >> >> >>>>>>> >>> >>> > >>> command: >> >> >>>>>>> >>> >>> > >>> HandleVdsCpuFlagsOrClusterChangedCommand >> internal: >> >> >>>>>>> >>> >>> > >>> true. >> >> >>>>>>> >>> >>> > >>> Entities >> >> >>>>>>> >>> >>> > >>> affected >> >> >>>>>>> >>> >>> > >>> : ID: 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: >> >> >>>>>>> >>> >>> > >>> VDS >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,147 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf] >> START, >> >> >>>>>>> >>> >>> > >>> GlusterServersListVDSCommand(HostName = Compute5, >> >> >>>>>>> >>> >>> > >>> HostId >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a), log id: >> >> >>>>>>> >>> >>> > >>> 4faed87 >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,164 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf] >> >> >>>>>>> >>> >>> > >>> FINISH, >> >> >>>>>>> >>> >>> > >>> GlusterServersListVDSCommand, log id: 4faed87 >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,189 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.bll.SetNonOperationalVdsCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] >> >> >>>>>>> >>> >>> > >>> Running >> >> >>>>>>> >>> >>> > >>> command: >> >> >>>>>>> >>> >>> > >>> SetNonOperationalVdsCommand internal: true. >> Entities >> >> >>>>>>> >>> >>> > >>> affected : >> >> >>>>>>> >>> >>> > >>> ID: >> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,206 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] >> START, >> >> >>>>>>> >>> >>> > >>> SetVdsStatusVDSCommand(HostName = Compute5, >> HostId >> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a, >> >> >>>>>>> >>> >>> > >>> status=NonOperational, >> >> >>>>>>> >>> >>> > >>> nonOperationalReason=GLUSTER_COMMAND_FAILED, >> >> >>>>>>> >>> >>> > >>> stopSpmFailureLogged=false), >> >> >>>>>>> >>> >>> > >>> log id: fed5617 >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,209 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] >> >> >>>>>>> >>> >>> > >>> FINISH, >> >> >>>>>>> >>> >>> > >>> SetVdsStatusVDSCommand, log id: fed5617 >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,223 ERROR >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] >> >> >>>>>>> >>> >>> > >>> Correlation ID: >> >> >>>>>>> >>> >>> > >>> 4a84c4e5, >> >> >>>>>>> >>> >>> > >>> Job >> >> >>>>>>> >>> >>> > >>> ID: 4bfd4a6d-c3ef-468f-a40e-a3a6ca13011b, Call >> >> >>>>>>> >>> >>> > >>> Stack: >> >> >>>>>>> >>> >>> > >>> null, >> >> >>>>>>> >>> >>> > >>> Custom >> >> >>>>>>> >>> >>> > >>> Event >> >> >>>>>>> >>> >>> > >>> ID: -1, Message: Gluster command [<UNKNOWN>] >> failed >> >> >>>>>>> >>> >>> > >>> on >> >> >>>>>>> >>> >>> > >>> server >> >> >>>>>>> >>> >>> > >>> Compute5. >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,243 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] >> >> >>>>>>> >>> >>> > >>> Correlation ID: >> >> >>>>>>> >>> >>> > >>> null, >> >> >>>>>>> >>> >>> > >>> Call >> >> >>>>>>> >>> >>> > >>> Stack: null, Custom Event ID: -1, Message: >> Status of >> >> >>>>>>> >>> >>> > >>> host >> >> >>>>>>> >>> >>> > >>> Compute5 >> >> >>>>>>> >>> >>> > >>> was >> >> >>>>>>> >>> >>> > >>> set >> >> >>>>>>> >>> >>> > >>> to NonOperational. >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,272 INFO >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.bll.HandleVdsVersionCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [a0c8a7f] >> Running >> >> >>>>>>> >>> >>> > >>> command: >> >> >>>>>>> >>> >>> > >>> HandleVdsVersionCommand internal: true. Entities >> >> >>>>>>> >>> >>> > >>> affected : >> >> >>>>>>> >>> >>> > >>> ID: >> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,274 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [a0c8a7f] Host >> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a : Compute5 >> is >> >> >>>>>>> >>> >>> > >>> already in >> >> >>>>>>> >>> >>> > >>> NonOperational status for reason >> >> >>>>>>> >>> >>> > >>> GLUSTER_COMMAND_FAILED. >> >> >>>>>>> >>> >>> > >>> SetNonOperationalVds command is skipped. >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:38,065 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-55) [3706e836] >> >> >>>>>>> >>> >>> > >>> FINISH, >> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand, return: >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> {26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity at 4e72a1b1 >> }, >> >> >>>>>>> >>> >>> > >>> log id: 48a0c832 >> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:43,243 INFO >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-35) START, >> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand(HostName = Compute4, >> >> >>>>>>> >>> >>> > >>> HostId >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id: >> >> >>>>>>> >>> >>> > >>> 3ce13ebc >> >> >>>>>>> >>> >>> > >>> ^C >> >> >>>>>>> >>> >>> > >>> [root at ccr01 ~]# >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> Thanks, >> >> >>>>>>> >>> >>> > >>> Punit >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> _______________________________________________ >> >> >>>>>>> >>> >>> > >>> Users mailing >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> listUsers at ovirt.orghttp:// >> lists.ovirt.org/mailman/listinfo/users >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >>> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > >> >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > > >> >> >>>>>>> >>> >>> > >> >> >>>>>>> >>> >> >> >> >>>>>>> >>> >> >> >> >>>>>>> >>> > >> >> >>>>>>> >>> > >> >> >>>>>>> >>> > _______________________________________________ >> >> >>>>>>> >>> > Gluster-users mailing list >> >> >>>>>>> >>> > Gluster-users at gluster.org >> >> >>>>>>> >>> > >> >> >>>>>>> >>> > >> http://supercolony.gluster.org/mailman/listinfo/gluster-users >> >> >>>>>>> >> >> >> >>>>>>> >> >> >> >>>>>>> > >> >> >>>>>> >> >> >>>>>> >> >> >>>> >> >> >> >> > >> > >> > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20141207/d76090a7/attachment.html>
Punit Dambiwal
2014-Dec-09 09:54 UTC
[Gluster-users] [ovirt-users] Gluster command [<UNKNOWN>] failed on server...
Dear Kaushal, I tried various method...but still the same error...it seems it's gluster bug..is there any body can suggest work-around here ?? Thanks, Punit On Sun, Dec 7, 2014 at 8:40 PM, Punit Dambiwal <hypunit at gmail.com> wrote:> Hi Kaushal, > > Still the same error...even try your suggested workaround :- > > ------------------- > Can you replace 'Before=network-online.target' with > 'Wants=network-online.target' and try the boot again? This should > force the network to be online before starting GlusterD. > ------------------- > > Thanks, > Punit > > On Sat, Dec 6, 2014 at 11:44 AM, Punit Dambiwal <hypunit at gmail.com> wrote: > >> Hi Kaushal, >> >> I already have all the hosts entry in the /etc/hosts for the easy >> resolution....i will try your method in the glusterd.services and check and >> let you know....weather problem solve or not..... >> >> On Fri, Dec 5, 2014 at 9:50 PM, Kaushal M <kshlmster at gmail.com> wrote: >> >>> Can you replace 'Before=network-online.target' with >>> 'Wants=network-online.target' and try the boot again? This should >>> force the network to be online before starting GlusterD. >>> >>> If even that fails, you could try adding an entry into /etc/hosts with >>> the hostname of the system. This should prevent any more failures. >>> >>> I still don't believe it's a problem with Gluster. Gluster uses apis >>> provided by the system to perform name resolution. These definitely >>> work correctly because you can start GlusterD later. Since the >>> resolution failure only happens during boot, it points to system or >>> network setup issues during boot. To me it seems like the network >>> isn't completely setup at that point of time. >>> >>> ~kaushal >>> >>> On Fri, Dec 5, 2014 at 12:47 PM, Punit Dambiwal <hypunit at gmail.com> >>> wrote: >>> > Hi Kaushal, >>> > >>> > It seems it's bug in glusterfs 3.6....even i manage my systemd to >>> start the >>> > network service before glusterd...but it's still fail... >>> > >>> > --------------- >>> > [Unit] >>> > Description=GlusterFS, a clustered file-system server >>> > After=network.target rpcbind.service >>> > Before=network-online.target >>> > >>> > [Service] >>> > Type=forking >>> > PIDFile=/var/run/glusterd.pid >>> > LimitNOFILE=65536 >>> > ExecStartPre=/etc/rc.d/init.d/network start >>> > ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid >>> > KillMode=process >>> > >>> > [Install] >>> > WantedBy=multi-user.target >>> > ---------------- >>> > >>> > Thanks, >>> > Punit >>> > >>> > On Wed, Dec 3, 2014 at 8:56 PM, Kaushal M <kshlmster at gmail.com> wrote: >>> >> >>> >> I just remembered this. >>> >> >>> >> There was another user having a similar issue of GlusterD failing to >>> >> start on the mailing list a while back. The cause of his problem was >>> >> the way his network was brought up. >>> >> IIRC, he was using a static network configuration. The problem >>> >> vanished when he began using dhcp. Or it might have been he was using >>> >> dhcp.service and it got solved after switching to NetworkManager. >>> >> >>> >> This could be one more thing you could look at. >>> >> >>> >> I'll try to find the mail thread to see if it was the same problem as >>> you. >>> >> >>> >> ~kaushal >>> >> >>> >> On Wed, Dec 3, 2014 at 6:22 PM, Kaushal M <kshlmster at gmail.com> >>> wrote: >>> >> > I don't know much about how the network target is brought up in >>> >> > CentOS7, but I'll try as much as I can. >>> >> > >>> >> > It seems to me that, after the network has been brought up and by >>> the >>> >> > time GlusterD is started, >>> >> > a. The machine hasn't yet recieved it's hostname, or >>> >> > b. It hasn't yet registered with the name server. >>> >> > >>> >> > This is causing name resolution failures. >>> >> > >>> >> > I don't know if the network target could come up without the machine >>> >> > getting its hostname, so I'm pretty sure it's not a. >>> >> > >>> >> > So it seems to be b. But these kind of signing in happens only in >>> DDNS >>> >> > systems, which doesn't seem to be the case for you. >>> >> > >>> >> > Both of these reasons might be wrong (most likely wrong). You'd do >>> >> > good if you could ask for help from someone with more experience in >>> >> > systemd + networking. >>> >> > >>> >> > ~kaushal >>> >> > >>> >> > On Wed, Dec 3, 2014 at 10:54 AM, Punit Dambiwal <hypunit at gmail.com> >>> >> > wrote: >>> >> >> Hi Kaushal, >>> >> >> >>> >> >> This is the host...which i rebooted...would you mind to let me >>> know how >>> >> >> i >>> >> >> can make the glusterd sevice come up after network...i am using >>> >> >> centos7...if >>> >> >> network is the issue... >>> >> >> >>> >> >> On Wed, Dec 3, 2014 at 11:54 AM, Kaushal M <kshlmster at gmail.com> >>> wrote: >>> >> >>> >>> >> >>> This peer cannot be identified. >>> >> >>> >>> >> >>> " [2014-12-03 02:29:25.998153] D >>> >> >>> [glusterd-peer-utils.c:121:glusterd_peerinfo_find_by_hostname] >>> >> >>> 0-management: >>> >> >>> Unable to find friend: cpu05.zne01.hkg1.ovt.36stack.com" >>> >> >>> >>> >> >>> I don't know why this address is not being resolved during boot >>> time. >>> >> >>> If >>> >> >>> this is a valid peer, the the only reason I can think of this >>> that the >>> >> >>> network is not up. >>> >> >>> >>> >> >>> If you had previously detached the peer forcefully, the that could >>> >> >>> have >>> >> >>> left stale entries in some volumes. In this case as well, GlusterD >>> >> >>> will fail >>> >> >>> to identify the peer. >>> >> >>> >>> >> >>> Do either of these reasons seem a possibility to you? >>> >> >>> >>> >> >>> On Dec 3, 2014 8:07 AM, "Punit Dambiwal" <hypunit at gmail.com> >>> wrote: >>> >> >>>> >>> >> >>>> Hi Kaushal, >>> >> >>>> >>> >> >>>> Please find the logs here :- http://ur1.ca/iyoe5 and >>> >> >>>> http://ur1.ca/iyoed >>> >> >>>> >>> >> >>>> On Tue, Dec 2, 2014 at 10:43 PM, Kaushal M <kshlmster at gmail.com> >>> >> >>>> wrote: >>> >> >>>>> >>> >> >>>>> Hey Punit, >>> >> >>>>> In the logs you've provided, GlusterD appears to be running >>> >> >>>>> correctly. >>> >> >>>>> Could you provide the logs for the time period when GlusterD >>> >> >>>>> attempts to >>> >> >>>>> start but fails. >>> >> >>>>> >>> >> >>>>> ~kaushal >>> >> >>>>> >>> >> >>>>> On Dec 2, 2014 8:03 PM, "Punit Dambiwal" <hypunit at gmail.com> >>> wrote: >>> >> >>>>>> >>> >> >>>>>> Hi Kaushal, >>> >> >>>>>> >>> >> >>>>>> Please find the logs here :- http://ur1.ca/iyhs5 and >>> >> >>>>>> http://ur1.ca/iyhue >>> >> >>>>>> >>> >> >>>>>> Thanks, >>> >> >>>>>> punit >>> >> >>>>>> >>> >> >>>>>> >>> >> >>>>>> On Tue, Dec 2, 2014 at 12:00 PM, Kaushal M < >>> kshlmster at gmail.com> >>> >> >>>>>> wrote: >>> >> >>>>>>> >>> >> >>>>>>> Hey Punit, >>> >> >>>>>>> Could you start Glusterd in debug mode and provide the logs >>> here? >>> >> >>>>>>> To start it in debug mode, append '-LDEBUG' to the ExecStart >>> line >>> >> >>>>>>> in >>> >> >>>>>>> the service file. >>> >> >>>>>>> >>> >> >>>>>>> ~kaushal >>> >> >>>>>>> >>> >> >>>>>>> On Mon, Dec 1, 2014 at 9:05 AM, Punit Dambiwal < >>> hypunit at gmail.com> >>> >> >>>>>>> wrote: >>> >> >>>>>>> > Hi, >>> >> >>>>>>> > >>> >> >>>>>>> > Can Any body help me on this ?? >>> >> >>>>>>> > >>> >> >>>>>>> > On Thu, Nov 27, 2014 at 9:29 AM, Punit Dambiwal >>> >> >>>>>>> > <hypunit at gmail.com> >>> >> >>>>>>> > wrote: >>> >> >>>>>>> >> >>> >> >>>>>>> >> Hi Kaushal, >>> >> >>>>>>> >> >>> >> >>>>>>> >> Thanks for the detailed reply....let me explain my setup >>> first >>> >> >>>>>>> >> :- >>> >> >>>>>>> >> >>> >> >>>>>>> >> 1. Ovirt Engine >>> >> >>>>>>> >> 2. 4* host as well as storage machine (Host and gluster >>> >> >>>>>>> >> combined) >>> >> >>>>>>> >> 3. Every host has 24 bricks... >>> >> >>>>>>> >> >>> >> >>>>>>> >> Now whenever the host machine reboot...it can come up but >>> can >>> >> >>>>>>> >> not >>> >> >>>>>>> >> join the >>> >> >>>>>>> >> cluster again and through the following error "Gluster >>> command >>> >> >>>>>>> >> [<UNKNOWN>] >>> >> >>>>>>> >> failed on server.." >>> >> >>>>>>> >> >>> >> >>>>>>> >> Please check my comment in line :- >>> >> >>>>>>> >> >>> >> >>>>>>> >> 1. Use the same string for doing the peer probe and for the >>> >> >>>>>>> >> brick >>> >> >>>>>>> >> address >>> >> >>>>>>> >> during volume create/add-brick. Ideally, we suggest you use >>> >> >>>>>>> >> properly >>> >> >>>>>>> >> resolvable FQDNs everywhere. If that is not possible, then >>> use >>> >> >>>>>>> >> only >>> >> >>>>>>> >> IP >>> >> >>>>>>> >> addresses. Try to avoid short names. >>> >> >>>>>>> >> --------------- >>> >> >>>>>>> >> [root at cpu05 ~]# gluster peer status >>> >> >>>>>>> >> Number of Peers: 3 >>> >> >>>>>>> >> >>> >> >>>>>>> >> Hostname: cpu03.stack.com >>> >> >>>>>>> >> Uuid: 5729b8c4-e80d-4353-b456-6f467bddbdfb >>> >> >>>>>>> >> State: Peer in Cluster (Connected) >>> >> >>>>>>> >> >>> >> >>>>>>> >> Hostname: cpu04.stack.com >>> >> >>>>>>> >> Uuid: d272b790-c4b2-4bed-ba68-793656e6d7b0 >>> >> >>>>>>> >> State: Peer in Cluster (Connected) >>> >> >>>>>>> >> Other names: >>> >> >>>>>>> >> 10.10.0.8 >>> >> >>>>>>> >> >>> >> >>>>>>> >> Hostname: cpu02.stack.com >>> >> >>>>>>> >> Uuid: 8d8a7041-950e-40d0-85f9-58d14340ca25 >>> >> >>>>>>> >> State: Peer in Cluster (Connected) >>> >> >>>>>>> >> [root at cpu05 ~]# >>> >> >>>>>>> >> ---------------- >>> >> >>>>>>> >> 2. During boot up, make sure to launch glusterd only after >>> the >>> >> >>>>>>> >> network is >>> >> >>>>>>> >> up. This will allow the new peer identification mechanism >>> to do >>> >> >>>>>>> >> its >>> >> >>>>>>> >> job correctly. >>> >> >>>>>>> >> >> I think the service itself doing the same job.... >>> >> >>>>>>> >> >>> >> >>>>>>> >> [root at cpu05 ~]# cat >>> /usr/lib/systemd/system/glusterd.service >>> >> >>>>>>> >> [Unit] >>> >> >>>>>>> >> Description=GlusterFS, a clustered file-system server >>> >> >>>>>>> >> After=network.target rpcbind.service >>> >> >>>>>>> >> Before=network-online.target >>> >> >>>>>>> >> >>> >> >>>>>>> >> [Service] >>> >> >>>>>>> >> Type=forking >>> >> >>>>>>> >> PIDFile=/var/run/glusterd.pid >>> >> >>>>>>> >> LimitNOFILE=65536 >>> >> >>>>>>> >> ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid >>> >> >>>>>>> >> KillMode=process >>> >> >>>>>>> >> >>> >> >>>>>>> >> [Install] >>> >> >>>>>>> >> WantedBy=multi-user.target >>> >> >>>>>>> >> [root at cpu05 ~]# >>> >> >>>>>>> >> -------------------- >>> >> >>>>>>> >> >>> >> >>>>>>> >> gluster logs :- >>> >> >>>>>>> >> >>> >> >>>>>>> >> [2014-11-24 09:22:22.147471] I [MSGID: 100030] >>> >> >>>>>>> >> [glusterfsd.c:2018:main] >>> >> >>>>>>> >> 0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd >>> >> >>>>>>> >> version >>> >> >>>>>>> >> 3.6.1 >>> >> >>>>>>> >> (args: /usr/sbin/glusterd -p /var/run/glusterd.pid) >>> >> >>>>>>> >> [2014-11-24 09:22:22.151565] I [glusterd.c:1214:init] >>> >> >>>>>>> >> 0-management: >>> >> >>>>>>> >> Maximum allowed open file descriptors set to 65536 >>> >> >>>>>>> >> [2014-11-24 09:22:22.151599] I [glusterd.c:1259:init] >>> >> >>>>>>> >> 0-management: >>> >> >>>>>>> >> Using >>> >> >>>>>>> >> /var/lib/glusterd as working directory >>> >> >>>>>>> >> [2014-11-24 09:22:22.155216] W >>> >> >>>>>>> >> [rdma.c:4195:__gf_rdma_ctx_create] >>> >> >>>>>>> >> 0-rpc-transport/rdma: rdma_cm event channel creation >>> failed (No >>> >> >>>>>>> >> such device) >>> >> >>>>>>> >> [2014-11-24 09:22:22.155264] E [rdma.c:4483:init] >>> >> >>>>>>> >> 0-rdma.management: >>> >> >>>>>>> >> Failed to initialize IB Device >>> >> >>>>>>> >> [2014-11-24 09:22:22.155285] E >>> >> >>>>>>> >> [rpc-transport.c:333:rpc_transport_load] >>> >> >>>>>>> >> 0-rpc-transport: 'rdma' initialization failed >>> >> >>>>>>> >> [2014-11-24 09:22:22.155354] W >>> >> >>>>>>> >> [rpcsvc.c:1524:rpcsvc_transport_create] >>> >> >>>>>>> >> 0-rpc-service: cannot create listener, initing the >>> transport >>> >> >>>>>>> >> failed >>> >> >>>>>>> >> [2014-11-24 09:22:22.156290] I >>> >> >>>>>>> >> [glusterd.c:413:glusterd_check_gsync_present] 0-glusterd: >>> >> >>>>>>> >> geo-replication >>> >> >>>>>>> >> module not installed in the system >>> >> >>>>>>> >> [2014-11-24 09:22:22.161318] I >>> >> >>>>>>> >> [glusterd-store.c:2043:glusterd_restore_op_version] >>> 0-glusterd: >>> >> >>>>>>> >> retrieved >>> >> >>>>>>> >> op-version: 30600 >>> >> >>>>>>> >> [2014-11-24 09:22:22.821800] I >>> >> >>>>>>> >> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo] >>> >> >>>>>>> >> 0-management: >>> >> >>>>>>> >> connect returned 0 >>> >> >>>>>>> >> [2014-11-24 09:22:22.825810] I >>> >> >>>>>>> >> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo] >>> >> >>>>>>> >> 0-management: >>> >> >>>>>>> >> connect returned 0 >>> >> >>>>>>> >> [2014-11-24 09:22:22.828705] I >>> >> >>>>>>> >> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo] >>> >> >>>>>>> >> 0-management: >>> >> >>>>>>> >> connect returned 0 >>> >> >>>>>>> >> [2014-11-24 09:22:22.828771] I >>> >> >>>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init] >>> >> >>>>>>> >> 0-management: setting frame-timeout to 600 >>> >> >>>>>>> >> [2014-11-24 09:22:22.832670] I >>> >> >>>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init] >>> >> >>>>>>> >> 0-management: setting frame-timeout to 600 >>> >> >>>>>>> >> [2014-11-24 09:22:22.835919] I >>> >> >>>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init] >>> >> >>>>>>> >> 0-management: setting frame-timeout to 600 >>> >> >>>>>>> >> [2014-11-24 09:22:22.840209] E >>> >> >>>>>>> >> [glusterd-store.c:4248:glusterd_resolve_all_bricks] >>> 0-glusterd: >>> >> >>>>>>> >> resolve >>> >> >>>>>>> >> brick failed in restore >>> >> >>>>>>> >> [2014-11-24 09:22:22.840233] E [xlator.c:425:xlator_init] >>> >> >>>>>>> >> 0-management: >>> >> >>>>>>> >> Initialization of volume 'management' failed, review your >>> >> >>>>>>> >> volfile >>> >> >>>>>>> >> again >>> >> >>>>>>> >> [2014-11-24 09:22:22.840245] E >>> >> >>>>>>> >> [graph.c:322:glusterfs_graph_init] >>> >> >>>>>>> >> 0-management: initializing translator failed >>> >> >>>>>>> >> [2014-11-24 09:22:22.840264] E >>> >> >>>>>>> >> [graph.c:525:glusterfs_graph_activate] >>> >> >>>>>>> >> 0-graph: init failed >>> >> >>>>>>> >> [2014-11-24 09:22:22.840754] W >>> >> >>>>>>> >> [glusterfsd.c:1194:cleanup_and_exit] >>> >> >>>>>>> >> (--> >>> >> >>>>>>> >> 0-: received signum (0), shutting down >>> >> >>>>>>> >> >>> >> >>>>>>> >> Thanks, >>> >> >>>>>>> >> Punit >>> >> >>>>>>> >> >>> >> >>>>>>> >> >>> >> >>>>>>> >> >>> >> >>>>>>> >> >>> >> >>>>>>> >> On Wed, Nov 26, 2014 at 7:14 PM, Kaushal M >>> >> >>>>>>> >> <kshlmster at gmail.com> >>> >> >>>>>>> >> wrote: >>> >> >>>>>>> >>> >>> >> >>>>>>> >>> Based on the logs I can guess that glusterd is being >>> started >>> >> >>>>>>> >>> before >>> >> >>>>>>> >>> the network has come up and that the addresses given to >>> bricks >>> >> >>>>>>> >>> do >>> >> >>>>>>> >>> not >>> >> >>>>>>> >>> directly match the addresses used in during peer probe. >>> >> >>>>>>> >>> >>> >> >>>>>>> >>> The gluster_after_reboot log has the line "[2014-11-25 >>> >> >>>>>>> >>> 06:46:09.972113] E >>> >> >>>>>>> >>> [glusterd-store.c:2632:glusterd_resolve_all_bricks] >>> >> >>>>>>> >>> 0-glusterd: resolve brick failed in restore". >>> >> >>>>>>> >>> >>> >> >>>>>>> >>> Brick resolution fails when glusterd cannot match the >>> address >>> >> >>>>>>> >>> for >>> >> >>>>>>> >>> the >>> >> >>>>>>> >>> brick, with one of the peers. Brick resolution happens in >>> two >>> >> >>>>>>> >>> phases, >>> >> >>>>>>> >>> 1. We first try to identify the peer by performing string >>> >> >>>>>>> >>> comparisions >>> >> >>>>>>> >>> with the brick address and the peer addresses (The peer >>> names >>> >> >>>>>>> >>> will >>> >> >>>>>>> >>> be >>> >> >>>>>>> >>> the names/addresses that were given when the peer was >>> probed). >>> >> >>>>>>> >>> 2. If we don't find a match from step 1, we will then >>> resolve >>> >> >>>>>>> >>> all >>> >> >>>>>>> >>> the >>> >> >>>>>>> >>> brick address and the peer addresses into addrinfo >>> structs, >>> >> >>>>>>> >>> and >>> >> >>>>>>> >>> then >>> >> >>>>>>> >>> compare these structs to find a match. This process should >>> >> >>>>>>> >>> generally >>> >> >>>>>>> >>> find a match if available. This will fail only if the >>> network >>> >> >>>>>>> >>> is >>> >> >>>>>>> >>> not >>> >> >>>>>>> >>> up yet as we cannot resolve addresses. >>> >> >>>>>>> >>> >>> >> >>>>>>> >>> The above steps are applicable only to glusterfs versions >>> >> >>>>>>> >>> >=3.6. >>> >> >>>>>>> >>> They >>> >> >>>>>>> >>> were introduced to reduce problems with peer >>> identification, >>> >> >>>>>>> >>> like >>> >> >>>>>>> >>> the >>> >> >>>>>>> >>> one you encountered >>> >> >>>>>>> >>> >>> >> >>>>>>> >>> Since both of the steps failed to find a match in one >>> run, but >>> >> >>>>>>> >>> succeeded later, we can come to the conclusion that, >>> >> >>>>>>> >>> a) the bricks don't have the exact same string used in >>> peer >>> >> >>>>>>> >>> probe >>> >> >>>>>>> >>> for >>> >> >>>>>>> >>> their addresses as step 1 failed, and >>> >> >>>>>>> >>> b) the network was not up in the initial run, as step 2 >>> failed >>> >> >>>>>>> >>> during >>> >> >>>>>>> >>> the initial run, but passed in the second run. >>> >> >>>>>>> >>> >>> >> >>>>>>> >>> Please let me know if my conclusion is correct. >>> >> >>>>>>> >>> >>> >> >>>>>>> >>> If it is, you can solve your problem in two ways. >>> >> >>>>>>> >>> 1. Use the same string for doing the peer probe and for >>> the >>> >> >>>>>>> >>> brick >>> >> >>>>>>> >>> address during volume create/add-brick. Ideally, we >>> suggest >>> >> >>>>>>> >>> you >>> >> >>>>>>> >>> use >>> >> >>>>>>> >>> properly resolvable FQDNs everywhere. If that is not >>> possible, >>> >> >>>>>>> >>> then >>> >> >>>>>>> >>> use only IP addresses. Try to avoid short names. >>> >> >>>>>>> >>> 2. During boot up, make sure to launch glusterd only >>> after the >>> >> >>>>>>> >>> network >>> >> >>>>>>> >>> is up. This will allow the new peer identification >>> mechanism >>> >> >>>>>>> >>> to do >>> >> >>>>>>> >>> its >>> >> >>>>>>> >>> job correctly. >>> >> >>>>>>> >>> >>> >> >>>>>>> >>> >>> >> >>>>>>> >>> If you have already followed these steps and yet still >>> hit the >>> >> >>>>>>> >>> problem, then please provide more information (setup, >>> logs, >>> >> >>>>>>> >>> etc.). >>> >> >>>>>>> >>> It >>> >> >>>>>>> >>> could be much different problem that you are facing. >>> >> >>>>>>> >>> >>> >> >>>>>>> >>> ~kaushal >>> >> >>>>>>> >>> >>> >> >>>>>>> >>> On Wed, Nov 26, 2014 at 4:01 PM, Punit Dambiwal >>> >> >>>>>>> >>> <hypunit at gmail.com> >>> >> >>>>>>> >>> wrote: >>> >> >>>>>>> >>> > Is there any one can help on this ?? >>> >> >>>>>>> >>> > >>> >> >>>>>>> >>> > Thanks, >>> >> >>>>>>> >>> > punit >>> >> >>>>>>> >>> > >>> >> >>>>>>> >>> > On Wed, Nov 26, 2014 at 9:42 AM, Punit Dambiwal >>> >> >>>>>>> >>> > <hypunit at gmail.com> >>> >> >>>>>>> >>> > wrote: >>> >> >>>>>>> >>> >> >>> >> >>>>>>> >>> >> Hi, >>> >> >>>>>>> >>> >> >>> >> >>>>>>> >>> >> My Glusterfs version is :- glusterfs-3.6.1-1.el7 >>> >> >>>>>>> >>> >> >>> >> >>>>>>> >>> >> On Wed, Nov 26, 2014 at 1:59 AM, Kanagaraj Mayilsamy >>> >> >>>>>>> >>> >> <kmayilsa at redhat.com> >>> >> >>>>>>> >>> >> wrote: >>> >> >>>>>>> >>> >>> >>> >> >>>>>>> >>> >>> [+Gluster-users at gluster.org] >>> >> >>>>>>> >>> >>> >>> >> >>>>>>> >>> >>> "Initialization of volume 'management' failed, review >>> your >>> >> >>>>>>> >>> >>> volfile >>> >> >>>>>>> >>> >>> again", glusterd throws this error when the service is >>> >> >>>>>>> >>> >>> started >>> >> >>>>>>> >>> >>> automatically >>> >> >>>>>>> >>> >>> after the reboot. But the service is successfully >>> started >>> >> >>>>>>> >>> >>> later >>> >> >>>>>>> >>> >>> manually by >>> >> >>>>>>> >>> >>> the user. >>> >> >>>>>>> >>> >>> >>> >> >>>>>>> >>> >>> can somebody from gluster-users please help on this? >>> >> >>>>>>> >>> >>> >>> >> >>>>>>> >>> >>> glusterfs version: 3.5.1 >>> >> >>>>>>> >>> >>> >>> >> >>>>>>> >>> >>> Thanks, >>> >> >>>>>>> >>> >>> Kanagaraj >>> >> >>>>>>> >>> >>> >>> >> >>>>>>> >>> >>> ----- Original Message ----- >>> >> >>>>>>> >>> >>> > From: "Punit Dambiwal" <hypunit at gmail.com> >>> >> >>>>>>> >>> >>> > To: "Kanagaraj" <kmayilsa at redhat.com> >>> >> >>>>>>> >>> >>> > Cc: users at ovirt.org >>> >> >>>>>>> >>> >>> > Sent: Tuesday, November 25, 2014 7:24:45 PM >>> >> >>>>>>> >>> >>> > Subject: Re: [ovirt-users] Gluster command >>> [<UNKNOWN>] >>> >> >>>>>>> >>> >>> > failed on >>> >> >>>>>>> >>> >>> > server... >>> >> >>>>>>> >>> >>> > >>> >> >>>>>>> >>> >>> > Hi Kanagraj, >>> >> >>>>>>> >>> >>> > >>> >> >>>>>>> >>> >>> > Please check the attached log files....i didn't >>> find any >>> >> >>>>>>> >>> >>> > thing >>> >> >>>>>>> >>> >>> > special.... >>> >> >>>>>>> >>> >>> > >>> >> >>>>>>> >>> >>> > On Tue, Nov 25, 2014 at 12:12 PM, Kanagaraj >>> >> >>>>>>> >>> >>> > <kmayilsa at redhat.com> >>> >> >>>>>>> >>> >>> > wrote: >>> >> >>>>>>> >>> >>> > >>> >> >>>>>>> >>> >>> > > Do you see any errors in >>> >> >>>>>>> >>> >>> > > /var/log/glusterfs/etc-glusterfs-glusterd.vol.log >>> or >>> >> >>>>>>> >>> >>> > > vdsm.log >>> >> >>>>>>> >>> >>> > > when >>> >> >>>>>>> >>> >>> > > the >>> >> >>>>>>> >>> >>> > > service is trying to start automatically after the >>> >> >>>>>>> >>> >>> > > reboot? >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > > Thanks, >>> >> >>>>>>> >>> >>> > > Kanagaraj >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > > On 11/24/2014 08:13 PM, Punit Dambiwal wrote: >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > > Hi Kanagaraj, >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > > Yes...once i will start the gluster service and >>> then >>> >> >>>>>>> >>> >>> > > vdsmd >>> >> >>>>>>> >>> >>> > > ...the >>> >> >>>>>>> >>> >>> > > host >>> >> >>>>>>> >>> >>> > > can connect to cluster...but the question is why >>> it's >>> >> >>>>>>> >>> >>> > > not >>> >> >>>>>>> >>> >>> > > started >>> >> >>>>>>> >>> >>> > > even it >>> >> >>>>>>> >>> >>> > > has chkconfig enabled... >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > > I have tested it in two host cluster >>> >> >>>>>>> >>> >>> > > environment...(Centos 6.6 >>> >> >>>>>>> >>> >>> > > and >>> >> >>>>>>> >>> >>> > > centos 7.0) on both hypervisior cluster..it's >>> failed >>> >> >>>>>>> >>> >>> > > to >>> >> >>>>>>> >>> >>> > > reconnect >>> >> >>>>>>> >>> >>> > > in >>> >> >>>>>>> >>> >>> > > to >>> >> >>>>>>> >>> >>> > > cluster after reboot.... >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > > In both the environment glusterd enabled for next >>> >> >>>>>>> >>> >>> > > boot....but >>> >> >>>>>>> >>> >>> > > it's >>> >> >>>>>>> >>> >>> > > failed with the same error....seems it's bug in >>> either >>> >> >>>>>>> >>> >>> > > gluster or >>> >> >>>>>>> >>> >>> > > Ovirt ?? >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > > Please help me to find the workaround here if >>> can not >>> >> >>>>>>> >>> >>> > > resolve >>> >> >>>>>>> >>> >>> > > it...as >>> >> >>>>>>> >>> >>> > > without this the Host machine can not connect >>> after >>> >> >>>>>>> >>> >>> > > reboot....that >>> >> >>>>>>> >>> >>> > > means >>> >> >>>>>>> >>> >>> > > engine will consider it as down and every time >>> need to >>> >> >>>>>>> >>> >>> > > manually >>> >> >>>>>>> >>> >>> > > start >>> >> >>>>>>> >>> >>> > > the >>> >> >>>>>>> >>> >>> > > gluster service and vdsmd... ?? >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > > Thanks, >>> >> >>>>>>> >>> >>> > > Punit >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > > On Mon, Nov 24, 2014 at 10:20 PM, Kanagaraj >>> >> >>>>>>> >>> >>> > > <kmayilsa at redhat.com> >>> >> >>>>>>> >>> >>> > > wrote: >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > >> From vdsm.log "error: Connection failed. Please >>> >> >>>>>>> >>> >>> > >> check if >>> >> >>>>>>> >>> >>> > >> gluster >>> >> >>>>>>> >>> >>> > >> daemon >>> >> >>>>>>> >>> >>> > >> is operational." >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> Starting glusterd service should fix this issue. >>> >> >>>>>>> >>> >>> > >> 'service >>> >> >>>>>>> >>> >>> > >> glusterd >>> >> >>>>>>> >>> >>> > >> start' >>> >> >>>>>>> >>> >>> > >> But i am wondering why the glusterd was not >>> started >>> >> >>>>>>> >>> >>> > >> automatically >>> >> >>>>>>> >>> >>> > >> after >>> >> >>>>>>> >>> >>> > >> the reboot. >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> Thanks, >>> >> >>>>>>> >>> >>> > >> Kanagaraj >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> On 11/24/2014 07:18 PM, Punit Dambiwal wrote: >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> Hi Kanagaraj, >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> Please find the attached VDSM logs :- >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> ---------------- >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:17,182::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) >>> >> >>>>>>> >>> >>> > >> Owner.cancelAll requests {} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:17,182::task::993::Storage.TaskManager.Task::(_decref) >>> >> >>>>>>> >>> >>> > >> Task=`1691d409-9b27-4585-8281-5ec26154367a`::ref >>> 0 >>> >> >>>>>>> >>> >>> > >> aborting >>> >> >>>>>>> >>> >>> > >> False >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:32,393::task::595::Storage.TaskManager.Task::(_updateState) >>> >> >>>>>>> >>> >>> > >> >>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving >>> >> >>>>>>> >>> >>> > >> from >>> >> >>>>>>> >>> >>> > >> state >>> >> >>>>>>> >>> >>> > >> init >>> >> >>>>>>> >>> >>> > >> -> >>> >> >>>>>>> >>> >>> > >> state preparing >>> >> >>>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> 21:41:32,393::logUtils::44::dispatcher::(wrapper) Run >>> >> >>>>>>> >>> >>> > >> and >>> >> >>>>>>> >>> >>> > >> protect: >>> >> >>>>>>> >>> >>> > >> repoStats(options=None) >>> >> >>>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> 21:41:32,393::logUtils::47::dispatcher::(wrapper) Run >>> >> >>>>>>> >>> >>> > >> and >>> >> >>>>>>> >>> >>> > >> protect: >>> >> >>>>>>> >>> >>> > >> repoStats, Return response: {} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:32,393::task::1191::Storage.TaskManager.Task::(prepare) >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::finished: {} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:32,394::task::595::Storage.TaskManager.Task::(_updateState) >>> >> >>>>>>> >>> >>> > >> >>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving >>> >> >>>>>>> >>> >>> > >> from >>> >> >>>>>>> >>> >>> > >> state >>> >> >>>>>>> >>> >>> > >> preparing >>> >> >>>>>>> >>> >>> > >> -> >>> >> >>>>>>> >>> >>> > >> state finished >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:32,394::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) >>> >> >>>>>>> >>> >>> > >> Owner.releaseAll requests {} resources {} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:32,394::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) >>> >> >>>>>>> >>> >>> > >> Owner.cancelAll requests {} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:32,394::task::993::Storage.TaskManager.Task::(_decref) >>> >> >>>>>>> >>> >>> > >> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::ref >>> 0 >>> >> >>>>>>> >>> >>> > >> aborting >>> >> >>>>>>> >>> >>> > >> False >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,550::BindingXMLRPC::1132::vds::(wrapper) >>> >> >>>>>>> >>> >>> > >> client >>> >> >>>>>>> >>> >>> > >> [10.10.10.2]::call >>> >> >>>>>>> >>> >>> > >> getCapabilities with () {} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,553::utils::738::root::(execCmd) >>> >> >>>>>>> >>> >>> > >> /sbin/ip route show to 0.0.0.0/0 table all (cwd >>> None) >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,560::utils::758::root::(execCmd) >>> >> >>>>>>> >>> >>> > >> SUCCESS: <err> = ''; <rc> = 0 >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,588::caps::728::root::(_getKeyPackages) >>> rpm >>> >> >>>>>>> >>> >>> > >> package >>> >> >>>>>>> >>> >>> > >> ('gluster-swift',) not found >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,592::caps::728::root::(_getKeyPackages) >>> rpm >>> >> >>>>>>> >>> >>> > >> package >>> >> >>>>>>> >>> >>> > >> ('gluster-swift-object',) not found >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,593::caps::728::root::(_getKeyPackages) >>> rpm >>> >> >>>>>>> >>> >>> > >> package >>> >> >>>>>>> >>> >>> > >> ('gluster-swift-plugin',) not found >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,598::caps::728::root::(_getKeyPackages) >>> rpm >>> >> >>>>>>> >>> >>> > >> package >>> >> >>>>>>> >>> >>> > >> ('gluster-swift-account',) not found >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,598::caps::728::root::(_getKeyPackages) >>> rpm >>> >> >>>>>>> >>> >>> > >> package >>> >> >>>>>>> >>> >>> > >> ('gluster-swift-proxy',) not found >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,598::caps::728::root::(_getKeyPackages) >>> rpm >>> >> >>>>>>> >>> >>> > >> package >>> >> >>>>>>> >>> >>> > >> ('gluster-swift-doc',) not found >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,599::caps::728::root::(_getKeyPackages) >>> rpm >>> >> >>>>>>> >>> >>> > >> package >>> >> >>>>>>> >>> >>> > >> ('gluster-swift-container',) not found >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,599::caps::728::root::(_getKeyPackages) >>> rpm >>> >> >>>>>>> >>> >>> > >> package >>> >> >>>>>>> >>> >>> > >> ('glusterfs-geo-replication',) not found >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,600::caps::646::root::(get) >>> >> >>>>>>> >>> >>> > >> VirtioRNG DISABLED: libvirt version >>> 0.10.2-29.el6_5.9 >>> >> >>>>>>> >>> >>> > >> required >>> >> >>>>>>> >>> >>> > >> >>>> >> >>>>>>> >>> >>> > >> 0.10.2-31 >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,603::BindingXMLRPC::1139::vds::(wrapper) >>> >> >>>>>>> >>> >>> > >> return >>> >> >>>>>>> >>> >>> > >> getCapabilities >>> >> >>>>>>> >>> >>> > >> with {'status': {'message': 'Done', 'code': 0}, >>> >> >>>>>>> >>> >>> > >> 'info': >>> >> >>>>>>> >>> >>> > >> {'HBAInventory': >>> >> >>>>>>> >>> >>> > >> {'iSCSI': [{'InitiatorName': >>> >> >>>>>>> >>> >>> > >> 'iqn.1994-05.com.redhat:32151ce183c8'}], >>> >> >>>>>>> >>> >>> > >> 'FC': >>> >> >>>>>>> >>> >>> > >> []}, 'packages2': {'kernel': {'release': >>> >> >>>>>>> >>> >>> > >> '431.el6.x86_64', >>> >> >>>>>>> >>> >>> > >> 'buildtime': >>> >> >>>>>>> >>> >>> > >> 1385061309.0, 'version': '2.6.32'}, >>> 'glusterfs-rdma': >>> >> >>>>>>> >>> >>> > >> {'release': >>> >> >>>>>>> >>> >>> > >> '1.el6', >>> >> >>>>>>> >>> >>> > >> 'buildtime': 1403622628L, 'version': '3.5.1'}, >>> >> >>>>>>> >>> >>> > >> 'glusterfs-fuse': >>> >> >>>>>>> >>> >>> > >> {'release': '1.el6', 'buildtime': 1403622628L, >>> >> >>>>>>> >>> >>> > >> 'version': >>> >> >>>>>>> >>> >>> > >> '3.5.1'}, >>> >> >>>>>>> >>> >>> > >> 'spice-server': {'release': '6.el6_5.2', >>> 'buildtime': >>> >> >>>>>>> >>> >>> > >> 1402324637L, >>> >> >>>>>>> >>> >>> > >> 'version': '0.12.4'}, 'vdsm': {'release': >>> >> >>>>>>> >>> >>> > >> '1.gitdb83943.el6', >>> >> >>>>>>> >>> >>> > >> 'buildtime': >>> >> >>>>>>> >>> >>> > >> 1412784567L, 'version': '4.16.7'}, 'qemu-kvm': >>> >> >>>>>>> >>> >>> > >> {'release': >>> >> >>>>>>> >>> >>> > >> '2.415.el6_5.10', 'buildtime': 1402435700L, >>> >> >>>>>>> >>> >>> > >> 'version': >>> >> >>>>>>> >>> >>> > >> '0.12.1.2'}, >>> >> >>>>>>> >>> >>> > >> 'qemu-img': {'release': '2.415.el6_5.10', >>> >> >>>>>>> >>> >>> > >> 'buildtime': >>> >> >>>>>>> >>> >>> > >> 1402435700L, >>> >> >>>>>>> >>> >>> > >> 'version': '0.12.1.2'}, 'libvirt': {'release': >>> >> >>>>>>> >>> >>> > >> '29.el6_5.9', >>> >> >>>>>>> >>> >>> > >> 'buildtime': >>> >> >>>>>>> >>> >>> > >> 1402404612L, 'version': '0.10.2'}, 'glusterfs': >>> >> >>>>>>> >>> >>> > >> {'release': >>> >> >>>>>>> >>> >>> > >> '1.el6', >>> >> >>>>>>> >>> >>> > >> 'buildtime': 1403622628L, 'version': '3.5.1'}, >>> 'mom': >>> >> >>>>>>> >>> >>> > >> {'release': >>> >> >>>>>>> >>> >>> > >> '2.el6', >>> >> >>>>>>> >>> >>> > >> 'buildtime': 1403794344L, 'version': '0.4.1'}, >>> >> >>>>>>> >>> >>> > >> 'glusterfs-server': >>> >> >>>>>>> >>> >>> > >> {'release': '1.el6', 'buildtime': 1403622628L, >>> >> >>>>>>> >>> >>> > >> 'version': >>> >> >>>>>>> >>> >>> > >> '3.5.1'}}, >>> >> >>>>>>> >>> >>> > >> 'numaNodeDistance': {'1': [20, 10], '0': [10, >>> 20]}, >>> >> >>>>>>> >>> >>> > >> 'cpuModel': >>> >> >>>>>>> >>> >>> > >> 'Intel(R) >>> >> >>>>>>> >>> >>> > >> Xeon(R) CPU X5650 @ 2.67GHz', >>> 'liveMerge': >>> >> >>>>>>> >>> >>> > >> 'false', >>> >> >>>>>>> >>> >>> > >> 'hooks': >>> >> >>>>>>> >>> >>> > >> {}, >>> >> >>>>>>> >>> >>> > >> 'cpuSockets': '2', 'vmTypes': ['kvm'], 'selinux': >>> >> >>>>>>> >>> >>> > >> {'mode': '1'}, >>> >> >>>>>>> >>> >>> > >> 'kdumpStatus': 0, 'supportedProtocols': ['2.2', >>> >> >>>>>>> >>> >>> > >> '2.3'], >>> >> >>>>>>> >>> >>> > >> 'networks': >>> >> >>>>>>> >>> >>> > >> {'ovirtmgmt': {'iface': u'bond0.10', 'addr': >>> >> >>>>>>> >>> >>> > >> '43.252.176.16', >>> >> >>>>>>> >>> >>> > >> 'bridged': >>> >> >>>>>>> >>> >>> > >> False, 'ipv6addrs': >>> ['fe80::62eb:69ff:fe20:b46c/64'], >>> >> >>>>>>> >>> >>> > >> 'mtu': >>> >> >>>>>>> >>> >>> > >> '1500', >>> >> >>>>>>> >>> >>> > >> 'bootproto4': 'none', 'netmask': '255.255.255.0', >>> >> >>>>>>> >>> >>> > >> 'ipv4addrs': >>> >> >>>>>>> >>> >>> > >> [' >>> >> >>>>>>> >>> >>> > >> 43.252.176.16/24' <http://43.252.176.16/24%27>], >>> >> >>>>>>> >>> >>> > >> 'interface': >>> >> >>>>>>> >>> >>> > >> u'bond0.10', 'ipv6gateway': '::', 'gateway': >>> >> >>>>>>> >>> >>> > >> '43.25.17.1'}, >>> >> >>>>>>> >>> >>> > >> 'Internal': >>> >> >>>>>>> >>> >>> > >> {'iface': 'Internal', 'addr': '', 'cfg': >>> {'DEFROUTE': >>> >> >>>>>>> >>> >>> > >> 'no', >>> >> >>>>>>> >>> >>> > >> 'HOTPLUG': >>> >> >>>>>>> >>> >>> > >> 'no', 'MTU': '9000', 'DELAY': '0', >>> 'NM_CONTROLLED': >>> >> >>>>>>> >>> >>> > >> 'no', >>> >> >>>>>>> >>> >>> > >> 'BOOTPROTO': >>> >> >>>>>>> >>> >>> > >> 'none', 'STP': 'off', 'DEVICE': 'Internal', >>> 'TYPE': >>> >> >>>>>>> >>> >>> > >> 'Bridge', >>> >> >>>>>>> >>> >>> > >> 'ONBOOT': >>> >> >>>>>>> >>> >>> > >> 'no'}, 'bridged': True, 'ipv6addrs': >>> >> >>>>>>> >>> >>> > >> ['fe80::210:18ff:fecd:daac/64'], >>> >> >>>>>>> >>> >>> > >> 'gateway': '', 'bootproto4': 'none', 'netmask': >>> '', >>> >> >>>>>>> >>> >>> > >> 'stp': >>> >> >>>>>>> >>> >>> > >> 'off', >>> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [], 'mtu': '9000', 'ipv6gateway': >>> '::', >>> >> >>>>>>> >>> >>> > >> 'ports': >>> >> >>>>>>> >>> >>> > >> ['bond1.100']}, 'storage': {'iface': u'bond1', >>> >> >>>>>>> >>> >>> > >> 'addr': >>> >> >>>>>>> >>> >>> > >> '10.10.10.6', >>> >> >>>>>>> >>> >>> > >> 'bridged': False, 'ipv6addrs': >>> >> >>>>>>> >>> >>> > >> ['fe80::210:18ff:fecd:daac/64'], >>> >> >>>>>>> >>> >>> > >> 'mtu': >>> >> >>>>>>> >>> >>> > >> '9000', 'bootproto4': 'none', 'netmask': >>> >> >>>>>>> >>> >>> > >> '255.255.255.0', >>> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [' >>> >> >>>>>>> >>> >>> > >> 10.10.10.6/24' <http://10.10.10.6/24%27>], >>> >> >>>>>>> >>> >>> > >> 'interface': >>> >> >>>>>>> >>> >>> > >> u'bond1', >>> >> >>>>>>> >>> >>> > >> 'ipv6gateway': '::', 'gateway': ''}, 'VMNetwork': >>> >> >>>>>>> >>> >>> > >> {'iface': >>> >> >>>>>>> >>> >>> > >> 'VMNetwork', >>> >> >>>>>>> >>> >>> > >> 'addr': '', 'cfg': {'DEFROUTE': 'no', 'HOTPLUG': >>> >> >>>>>>> >>> >>> > >> 'no', >>> >> >>>>>>> >>> >>> > >> 'MTU': >>> >> >>>>>>> >>> >>> > >> '1500', >>> >> >>>>>>> >>> >>> > >> 'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO': >>> >> >>>>>>> >>> >>> > >> 'none', >>> >> >>>>>>> >>> >>> > >> 'STP': >>> >> >>>>>>> >>> >>> > >> 'off', >>> >> >>>>>>> >>> >>> > >> 'DEVICE': 'VMNetwork', 'TYPE': 'Bridge', >>> 'ONBOOT': >>> >> >>>>>>> >>> >>> > >> 'no'}, >>> >> >>>>>>> >>> >>> > >> 'bridged': >>> >> >>>>>>> >>> >>> > >> True, >>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'], >>> >> >>>>>>> >>> >>> > >> 'gateway': >>> >> >>>>>>> >>> >>> > >> '', >>> >> >>>>>>> >>> >>> > >> 'bootproto4': >>> >> >>>>>>> >>> >>> > >> 'none', 'netmask': '', 'stp': 'off', >>> 'ipv4addrs': [], >>> >> >>>>>>> >>> >>> > >> 'mtu': >>> >> >>>>>>> >>> >>> > >> '1500', >>> >> >>>>>>> >>> >>> > >> 'ipv6gateway': '::', 'ports': ['bond0.36']}}, >>> >> >>>>>>> >>> >>> > >> 'bridges': >>> >> >>>>>>> >>> >>> > >> {'Internal': >>> >> >>>>>>> >>> >>> > >> {'addr': '', 'cfg': {'DEFROUTE': 'no', 'HOTPLUG': >>> >> >>>>>>> >>> >>> > >> 'no', >>> >> >>>>>>> >>> >>> > >> 'MTU': >>> >> >>>>>>> >>> >>> > >> '9000', >>> >> >>>>>>> >>> >>> > >> 'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO': >>> >> >>>>>>> >>> >>> > >> 'none', >>> >> >>>>>>> >>> >>> > >> 'STP': >>> >> >>>>>>> >>> >>> > >> 'off', >>> >> >>>>>>> >>> >>> > >> 'DEVICE': 'Internal', 'TYPE': 'Bridge', 'ONBOOT': >>> >> >>>>>>> >>> >>> > >> 'no'}, >>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': >>> >> >>>>>>> >>> >>> > >> ['fe80::210:18ff:fecd:daac/64'], 'mtu': '9000', >>> >> >>>>>>> >>> >>> > >> 'netmask': '', >>> >> >>>>>>> >>> >>> > >> 'stp': >>> >> >>>>>>> >>> >>> > >> 'off', 'ipv4addrs': [], 'ipv6gateway': '::', >>> >> >>>>>>> >>> >>> > >> 'gateway': >>> >> >>>>>>> >>> >>> > >> '', >>> >> >>>>>>> >>> >>> > >> 'opts': >>> >> >>>>>>> >>> >>> > >> {'topology_change_detected': '0', >>> >> >>>>>>> >>> >>> > >> 'multicast_last_member_count': >>> >> >>>>>>> >>> >>> > >> '2', >>> >> >>>>>>> >>> >>> > >> 'hash_elasticity': '4', >>> >> >>>>>>> >>> >>> > >> 'multicast_query_response_interval': >>> >> >>>>>>> >>> >>> > >> '999', >>> >> >>>>>>> >>> >>> > >> 'multicast_snooping': '1', >>> >> >>>>>>> >>> >>> > >> 'multicast_startup_query_interval': >>> >> >>>>>>> >>> >>> > >> '3124', >>> >> >>>>>>> >>> >>> > >> 'hello_timer': '31', >>> 'multicast_querier_interval': >>> >> >>>>>>> >>> >>> > >> '25496', >>> >> >>>>>>> >>> >>> > >> 'max_age': >>> >> >>>>>>> >>> >>> > >> '1999', 'hash_max': '512', 'stp_state': '0', >>> >> >>>>>>> >>> >>> > >> 'root_id': >>> >> >>>>>>> >>> >>> > >> '8000.001018cddaac', 'priority': '32768', >>> >> >>>>>>> >>> >>> > >> 'multicast_membership_interval': >>> >> >>>>>>> >>> >>> > >> '25996', 'root_path_cost': '0', 'root_port': '0', >>> >> >>>>>>> >>> >>> > >> 'multicast_querier': >>> >> >>>>>>> >>> >>> > >> '0', >>> >> >>>>>>> >>> >>> > >> 'multicast_startup_query_count': '2', >>> 'hello_time': >>> >> >>>>>>> >>> >>> > >> '199', >>> >> >>>>>>> >>> >>> > >> 'topology_change': '0', 'bridge_id': >>> >> >>>>>>> >>> >>> > >> '8000.001018cddaac', >>> >> >>>>>>> >>> >>> > >> 'topology_change_timer': '0', 'ageing_time': >>> '29995', >>> >> >>>>>>> >>> >>> > >> 'gc_timer': >>> >> >>>>>>> >>> >>> > >> '31', >>> >> >>>>>>> >>> >>> > >> 'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0', >>> >> >>>>>>> >>> >>> > >> 'multicast_query_interval': '12498', >>> >> >>>>>>> >>> >>> > >> 'multicast_last_member_interval': >>> >> >>>>>>> >>> >>> > >> '99', 'multicast_router': '1', 'forward_delay': >>> '0'}, >>> >> >>>>>>> >>> >>> > >> 'ports': >>> >> >>>>>>> >>> >>> > >> ['bond1.100']}, 'VMNetwork': {'addr': '', 'cfg': >>> >> >>>>>>> >>> >>> > >> {'DEFROUTE': >>> >> >>>>>>> >>> >>> > >> 'no', >>> >> >>>>>>> >>> >>> > >> 'HOTPLUG': 'no', 'MTU': '1500', 'DELAY': '0', >>> >> >>>>>>> >>> >>> > >> 'NM_CONTROLLED': >>> >> >>>>>>> >>> >>> > >> 'no', >>> >> >>>>>>> >>> >>> > >> 'BOOTPROTO': 'none', 'STP': 'off', 'DEVICE': >>> >> >>>>>>> >>> >>> > >> 'VMNetwork', >>> >> >>>>>>> >>> >>> > >> 'TYPE': >>> >> >>>>>>> >>> >>> > >> 'Bridge', >>> >> >>>>>>> >>> >>> > >> 'ONBOOT': 'no'}, 'ipv6addrs': >>> >> >>>>>>> >>> >>> > >> ['fe80::62eb:69ff:fe20:b46c/64'], >>> >> >>>>>>> >>> >>> > >> 'mtu': >>> >> >>>>>>> >>> >>> > >> '1500', 'netmask': '', 'stp': 'off', >>> 'ipv4addrs': [], >>> >> >>>>>>> >>> >>> > >> 'ipv6gateway': >>> >> >>>>>>> >>> >>> > >> '::', >>> >> >>>>>>> >>> >>> > >> 'gateway': '', 'opts': >>> {'topology_change_detected': >>> >> >>>>>>> >>> >>> > >> '0', >>> >> >>>>>>> >>> >>> > >> 'multicast_last_member_count': '2', >>> >> >>>>>>> >>> >>> > >> 'hash_elasticity': >>> >> >>>>>>> >>> >>> > >> '4', >>> >> >>>>>>> >>> >>> > >> 'multicast_query_response_interval': '999', >>> >> >>>>>>> >>> >>> > >> 'multicast_snooping': >>> >> >>>>>>> >>> >>> > >> '1', >>> >> >>>>>>> >>> >>> > >> 'multicast_startup_query_interval': '3124', >>> >> >>>>>>> >>> >>> > >> 'hello_timer': >>> >> >>>>>>> >>> >>> > >> '131', >>> >> >>>>>>> >>> >>> > >> 'multicast_querier_interval': '25496', 'max_age': >>> >> >>>>>>> >>> >>> > >> '1999', >>> >> >>>>>>> >>> >>> > >> 'hash_max': >>> >> >>>>>>> >>> >>> > >> '512', 'stp_state': '0', 'root_id': >>> >> >>>>>>> >>> >>> > >> '8000.60eb6920b46c', >>> >> >>>>>>> >>> >>> > >> 'priority': >>> >> >>>>>>> >>> >>> > >> '32768', 'multicast_membership_interval': >>> '25996', >>> >> >>>>>>> >>> >>> > >> 'root_path_cost': >>> >> >>>>>>> >>> >>> > >> '0', >>> >> >>>>>>> >>> >>> > >> 'root_port': '0', 'multicast_querier': '0', >>> >> >>>>>>> >>> >>> > >> 'multicast_startup_query_count': '2', >>> 'hello_time': >>> >> >>>>>>> >>> >>> > >> '199', >>> >> >>>>>>> >>> >>> > >> 'topology_change': '0', 'bridge_id': >>> >> >>>>>>> >>> >>> > >> '8000.60eb6920b46c', >>> >> >>>>>>> >>> >>> > >> 'topology_change_timer': '0', 'ageing_time': >>> '29995', >>> >> >>>>>>> >>> >>> > >> 'gc_timer': >>> >> >>>>>>> >>> >>> > >> '31', >>> >> >>>>>>> >>> >>> > >> 'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0', >>> >> >>>>>>> >>> >>> > >> 'multicast_query_interval': '12498', >>> >> >>>>>>> >>> >>> > >> 'multicast_last_member_interval': >>> >> >>>>>>> >>> >>> > >> '99', 'multicast_router': '1', 'forward_delay': >>> '0'}, >>> >> >>>>>>> >>> >>> > >> 'ports': >>> >> >>>>>>> >>> >>> > >> ['bond0.36']}}, 'uuid': >>> >> >>>>>>> >>> >>> > >> '44454C4C-4C00-1057-8053-B7C04F504E31', >>> >> >>>>>>> >>> >>> > >> 'lastClientIface': 'bond1', 'nics': {'eth3': >>> >> >>>>>>> >>> >>> > >> {'permhwaddr': >>> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ae', 'addr': '', 'cfg': {'SLAVE': >>> >> >>>>>>> >>> >>> > >> 'yes', >>> >> >>>>>>> >>> >>> > >> 'NM_CONTROLLED': >>> >> >>>>>>> >>> >>> > >> 'no', 'MTU': '9000', 'HWADDR': >>> '00:10:18:cd:da:ae', >>> >> >>>>>>> >>> >>> > >> 'MASTER': >>> >> >>>>>>> >>> >>> > >> 'bond1', >>> >> >>>>>>> >>> >>> > >> 'DEVICE': 'eth3', 'ONBOOT': 'no'}, 'ipv6addrs': >>> [], >>> >> >>>>>>> >>> >>> > >> 'mtu': >>> >> >>>>>>> >>> >>> > >> '9000', >>> >> >>>>>>> >>> >>> > >> 'netmask': '', 'ipv4addrs': [], 'hwaddr': >>> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ac', >>> >> >>>>>>> >>> >>> > >> 'speed': >>> >> >>>>>>> >>> >>> > >> 1000}, 'eth2': {'permhwaddr': >>> '00:10:18:cd:da:ac', >>> >> >>>>>>> >>> >>> > >> 'addr': '', >>> >> >>>>>>> >>> >>> > >> 'cfg': >>> >> >>>>>>> >>> >>> > >> {'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU': >>> >> >>>>>>> >>> >>> > >> '9000', >>> >> >>>>>>> >>> >>> > >> 'HWADDR': >>> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ac', 'MASTER': 'bond1', 'DEVICE': >>> >> >>>>>>> >>> >>> > >> 'eth2', >>> >> >>>>>>> >>> >>> > >> 'ONBOOT': >>> >> >>>>>>> >>> >>> > >> 'no'}, >>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': [], 'mtu': '9000', 'netmask': '', >>> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [], >>> >> >>>>>>> >>> >>> > >> 'hwaddr': >>> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ac', 'speed': 1000}, 'eth1': >>> >> >>>>>>> >>> >>> > >> {'permhwaddr': >>> >> >>>>>>> >>> >>> > >> '60:eb:69:20:b4:6d', 'addr': '', 'cfg': {'SLAVE': >>> >> >>>>>>> >>> >>> > >> 'yes', >>> >> >>>>>>> >>> >>> > >> 'NM_CONTROLLED': >>> >> >>>>>>> >>> >>> > >> 'no', 'MTU': '1500', 'HWADDR': >>> '60:eb:69:20:b4:6d', >>> >> >>>>>>> >>> >>> > >> 'MASTER': >>> >> >>>>>>> >>> >>> > >> 'bond0', >>> >> >>>>>>> >>> >>> > >> 'DEVICE': 'eth1', 'ONBOOT': 'yes'}, 'ipv6addrs': >>> [], >>> >> >>>>>>> >>> >>> > >> 'mtu': >>> >> >>>>>>> >>> >>> > >> '1500', >>> >> >>>>>>> >>> >>> > >> 'netmask': '', 'ipv4addrs': [], 'hwaddr': >>> >> >>>>>>> >>> >>> > >> '60:eb:69:20:b4:6c', >>> >> >>>>>>> >>> >>> > >> 'speed': >>> >> >>>>>>> >>> >>> > >> 1000}, 'eth0': {'permhwaddr': >>> '60:eb:69:20:b4:6c', >>> >> >>>>>>> >>> >>> > >> 'addr': '', >>> >> >>>>>>> >>> >>> > >> 'cfg': >>> >> >>>>>>> >>> >>> > >> {'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU': >>> >> >>>>>>> >>> >>> > >> '1500', >>> >> >>>>>>> >>> >>> > >> 'HWADDR': >>> >> >>>>>>> >>> >>> > >> '60:eb:69:20:b4:6c', 'MASTER': 'bond0', 'DEVICE': >>> >> >>>>>>> >>> >>> > >> 'eth0', >>> >> >>>>>>> >>> >>> > >> 'ONBOOT': >>> >> >>>>>>> >>> >>> > >> 'yes'}, >>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': [], 'mtu': '1500', 'netmask': '', >>> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [], >>> >> >>>>>>> >>> >>> > >> 'hwaddr': >>> >> >>>>>>> >>> >>> > >> '60:eb:69:20:b4:6c', 'speed': 1000}}, >>> >> >>>>>>> >>> >>> > >> 'software_revision': '1', >>> >> >>>>>>> >>> >>> > >> 'clusterLevels': ['3.0', '3.1', '3.2', '3.3', >>> '3.4', >>> >> >>>>>>> >>> >>> > >> '3.5'], >>> >> >>>>>>> >>> >>> > >> 'cpuFlags': >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> u'fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,pdpe1gb,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,xtopology,nonstop_tsc,pni,pclmulqdq,dtes64,monitor,ds_cpl,vmx,smx,est,tm2,ssse3,cx16,xtpr,pdcm,pcid,dca,sse4_1,sse4_2,popcnt,aes,lahf_lm,tpr_shadow,vnmi,flexpriority,ept,vpid,model_Nehalem,model_Conroe,model_coreduo,model_core2duo,model_Penryn,model_Westmere,model_n270', >>> >> >>>>>>> >>> >>> > >> 'ISCSIInitiatorName': >>> >> >>>>>>> >>> >>> > >> 'iqn.1994-05.com.redhat:32151ce183c8', >>> >> >>>>>>> >>> >>> > >> 'netConfigDirty': 'False', 'supportedENGINEs': >>> >> >>>>>>> >>> >>> > >> ['3.0', >>> >> >>>>>>> >>> >>> > >> '3.1', >>> >> >>>>>>> >>> >>> > >> '3.2', >>> >> >>>>>>> >>> >>> > >> '3.3', >>> >> >>>>>>> >>> >>> > >> '3.4', '3.5'], 'autoNumaBalancing': 2, >>> 'reservedMem': >>> >> >>>>>>> >>> >>> > >> '321', >>> >> >>>>>>> >>> >>> > >> 'bondings': >>> >> >>>>>>> >>> >>> > >> {'bond4': {'addr': '', 'cfg': {}, 'mtu': '1500', >>> >> >>>>>>> >>> >>> > >> 'netmask': '', >>> >> >>>>>>> >>> >>> > >> 'slaves': >>> >> >>>>>>> >>> >>> > >> [], 'hwaddr': '00:00:00:00:00:00'}, 'bond0': >>> {'addr': >>> >> >>>>>>> >>> >>> > >> '', >>> >> >>>>>>> >>> >>> > >> 'cfg': >>> >> >>>>>>> >>> >>> > >> {'HOTPLUG': 'no', 'MTU': '1500', 'NM_CONTROLLED': >>> >> >>>>>>> >>> >>> > >> 'no', >>> >> >>>>>>> >>> >>> > >> 'BONDING_OPTS': >>> >> >>>>>>> >>> >>> > >> 'mode=4 miimon=100', 'DEVICE': 'bond0', 'ONBOOT': >>> >> >>>>>>> >>> >>> > >> 'yes'}, >>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': >>> >> >>>>>>> >>> >>> > >> ['fe80::62eb:69ff:fe20:b46c/64'], 'mtu': '1500', >>> >> >>>>>>> >>> >>> > >> 'netmask': '', >>> >> >>>>>>> >>> >>> > >> 'ipv4addrs': [], 'hwaddr': '60:eb:69:20:b4:6c', >>> >> >>>>>>> >>> >>> > >> 'slaves': >>> >> >>>>>>> >>> >>> > >> ['eth0', >>> >> >>>>>>> >>> >>> > >> 'eth1'], >>> >> >>>>>>> >>> >>> > >> 'opts': {'miimon': '100', 'mode': '4'}}, 'bond1': >>> >> >>>>>>> >>> >>> > >> {'addr': >>> >> >>>>>>> >>> >>> > >> '10.10.10.6', >>> >> >>>>>>> >>> >>> > >> 'cfg': {'DEFROUTE': 'no', 'IPADDR': '10.10.10.6', >>> >> >>>>>>> >>> >>> > >> 'HOTPLUG': >>> >> >>>>>>> >>> >>> > >> 'no', >>> >> >>>>>>> >>> >>> > >> 'MTU': >>> >> >>>>>>> >>> >>> > >> '9000', 'NM_CONTROLLED': 'no', 'NETMASK': >>> >> >>>>>>> >>> >>> > >> '255.255.255.0', >>> >> >>>>>>> >>> >>> > >> 'BOOTPROTO': >>> >> >>>>>>> >>> >>> > >> 'none', 'BONDING_OPTS': 'mode=4 miimon=100', >>> >> >>>>>>> >>> >>> > >> 'DEVICE': >>> >> >>>>>>> >>> >>> > >> 'bond1', >>> >> >>>>>>> >>> >>> > >> 'ONBOOT': >>> >> >>>>>>> >>> >>> > >> 'no'}, 'ipv6addrs': >>> ['fe80::210:18ff:fecd:daac/64'], >>> >> >>>>>>> >>> >>> > >> 'mtu': >>> >> >>>>>>> >>> >>> > >> '9000', >>> >> >>>>>>> >>> >>> > >> 'netmask': '255.255.255.0', 'ipv4addrs': >>> >> >>>>>>> >>> >>> > >> ['10.10.10.6/24' >>> >> >>>>>>> >>> >>> > >> <http://10.10.10.6/24%27>], 'hwaddr': >>> >> >>>>>>> >>> >>> > >> '00:10:18:cd:da:ac', >>> >> >>>>>>> >>> >>> > >> 'slaves': >>> >> >>>>>>> >>> >>> > >> ['eth2', 'eth3'], 'opts': {'miimon': '100', >>> 'mode': >>> >> >>>>>>> >>> >>> > >> '4'}}, >>> >> >>>>>>> >>> >>> > >> 'bond2': >>> >> >>>>>>> >>> >>> > >> {'addr': '', 'cfg': {}, 'mtu': '1500', >>> 'netmask': '', >>> >> >>>>>>> >>> >>> > >> 'slaves': >>> >> >>>>>>> >>> >>> > >> [], >>> >> >>>>>>> >>> >>> > >> 'hwaddr': '00:00:00:00:00:00'}, 'bond3': >>> {'addr': '', >>> >> >>>>>>> >>> >>> > >> 'cfg': {}, >>> >> >>>>>>> >>> >>> > >> 'mtu': >>> >> >>>>>>> >>> >>> > >> '1500', 'netmask': '', 'slaves': [], 'hwaddr': >>> >> >>>>>>> >>> >>> > >> '00:00:00:00:00:00'}}, >>> >> >>>>>>> >>> >>> > >> 'software_version': '4.16', 'memSize': '24019', >>> >> >>>>>>> >>> >>> > >> 'cpuSpeed': >>> >> >>>>>>> >>> >>> > >> '2667.000', >>> >> >>>>>>> >>> >>> > >> 'numaNodes': {u'1': {'totalMemory': '12288', >>> 'cpus': >>> >> >>>>>>> >>> >>> > >> [6, >>> >> >>>>>>> >>> >>> > >> 7, 8, >>> >> >>>>>>> >>> >>> > >> 9, >>> >> >>>>>>> >>> >>> > >> 10, 11, >>> >> >>>>>>> >>> >>> > >> 18, 19, 20, 21, 22, 23]}, u'0': {'totalMemory': >>> >> >>>>>>> >>> >>> > >> '12278', >>> >> >>>>>>> >>> >>> > >> 'cpus': >>> >> >>>>>>> >>> >>> > >> [0, >>> >> >>>>>>> >>> >>> > >> 1, 2, >>> >> >>>>>>> >>> >>> > >> 3, 4, 5, 12, 13, 14, 15, 16, 17]}}, >>> 'version_name': >>> >> >>>>>>> >>> >>> > >> 'Snow >>> >> >>>>>>> >>> >>> > >> Man', >>> >> >>>>>>> >>> >>> > >> 'vlans': >>> >> >>>>>>> >>> >>> > >> {'bond0.10': {'iface': 'bond0', 'addr': >>> >> >>>>>>> >>> >>> > >> '43.25.17.16', >>> >> >>>>>>> >>> >>> > >> 'cfg': >>> >> >>>>>>> >>> >>> > >> {'DEFROUTE': >>> >> >>>>>>> >>> >>> > >> 'yes', 'VLAN': 'yes', 'IPADDR': '43.25.17.16', >>> >> >>>>>>> >>> >>> > >> 'HOTPLUG': >>> >> >>>>>>> >>> >>> > >> 'no', >>> >> >>>>>>> >>> >>> > >> 'GATEWAY': >>> >> >>>>>>> >>> >>> > >> '43.25.17.1', 'NM_CONTROLLED': 'no', 'NETMASK': >>> >> >>>>>>> >>> >>> > >> '255.255.255.0', >>> >> >>>>>>> >>> >>> > >> 'BOOTPROTO': 'none', 'DEVICE': 'bond0.10', 'MTU': >>> >> >>>>>>> >>> >>> > >> '1500', >>> >> >>>>>>> >>> >>> > >> 'ONBOOT': >>> >> >>>>>>> >>> >>> > >> 'yes'}, >>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'], >>> >> >>>>>>> >>> >>> > >> 'vlanid': >>> >> >>>>>>> >>> >>> > >> 10, >>> >> >>>>>>> >>> >>> > >> 'mtu': >>> >> >>>>>>> >>> >>> > >> '1500', >>> >> >>>>>>> >>> >>> > >> 'netmask': '255.255.255.0', 'ipv4addrs': >>> >> >>>>>>> >>> >>> > >> ['43.25.17.16/24'] >>> >> >>>>>>> >>> >>> > >> <http://43.25.17.16/24%27%5D>}, 'bond0.36': >>> {'iface': >>> >> >>>>>>> >>> >>> > >> 'bond0', >>> >> >>>>>>> >>> >>> > >> 'addr': >>> >> >>>>>>> >>> >>> > >> '', 'cfg': {'BRIDGE': 'VMNetwork', 'VLAN': 'yes', >>> >> >>>>>>> >>> >>> > >> 'HOTPLUG': >>> >> >>>>>>> >>> >>> > >> 'no', >>> >> >>>>>>> >>> >>> > >> 'MTU': >>> >> >>>>>>> >>> >>> > >> '1500', 'NM_CONTROLLED': 'no', 'DEVICE': >>> 'bond0.36', >>> >> >>>>>>> >>> >>> > >> 'ONBOOT': >>> >> >>>>>>> >>> >>> > >> 'no'}, >>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'], >>> >> >>>>>>> >>> >>> > >> 'vlanid': >>> >> >>>>>>> >>> >>> > >> 36, >>> >> >>>>>>> >>> >>> > >> 'mtu': >>> >> >>>>>>> >>> >>> > >> '1500', >>> >> >>>>>>> >>> >>> > >> 'netmask': '', 'ipv4addrs': []}, 'bond1.100': >>> >> >>>>>>> >>> >>> > >> {'iface': >>> >> >>>>>>> >>> >>> > >> 'bond1', >>> >> >>>>>>> >>> >>> > >> 'addr': >>> >> >>>>>>> >>> >>> > >> '', 'cfg': {'BRIDGE': 'Internal', 'VLAN': 'yes', >>> >> >>>>>>> >>> >>> > >> 'HOTPLUG': >>> >> >>>>>>> >>> >>> > >> 'no', >>> >> >>>>>>> >>> >>> > >> 'MTU': >>> >> >>>>>>> >>> >>> > >> '9000', 'NM_CONTROLLED': 'no', 'DEVICE': >>> 'bond1.100', >>> >> >>>>>>> >>> >>> > >> 'ONBOOT': >>> >> >>>>>>> >>> >>> > >> 'no'}, >>> >> >>>>>>> >>> >>> > >> 'ipv6addrs': ['fe80::210:18ff:fecd:daac/64'], >>> >> >>>>>>> >>> >>> > >> 'vlanid': >>> >> >>>>>>> >>> >>> > >> 100, >>> >> >>>>>>> >>> >>> > >> 'mtu': >>> >> >>>>>>> >>> >>> > >> '9000', >>> >> >>>>>>> >>> >>> > >> 'netmask': '', 'ipv4addrs': []}}, 'cpuCores': >>> '12', >>> >> >>>>>>> >>> >>> > >> 'kvmEnabled': >>> >> >>>>>>> >>> >>> > >> 'true', >>> >> >>>>>>> >>> >>> > >> 'guestOverhead': '65', 'cpuThreads': '24', >>> >> >>>>>>> >>> >>> > >> 'emulatedMachines': >>> >> >>>>>>> >>> >>> > >> [u'rhel6.5.0', u'pc', u'rhel6.4.0', u'rhel6.3.0', >>> >> >>>>>>> >>> >>> > >> u'rhel6.2.0', >>> >> >>>>>>> >>> >>> > >> u'rhel6.1.0', u'rhel6.0.0', u'rhel5.5.0', >>> >> >>>>>>> >>> >>> > >> u'rhel5.4.4', >>> >> >>>>>>> >>> >>> > >> u'rhel5.4.0'], >>> >> >>>>>>> >>> >>> > >> 'operatingSystem': {'release': >>> '5.el6.centos.11.1', >>> >> >>>>>>> >>> >>> > >> 'version': >>> >> >>>>>>> >>> >>> > >> '6', >>> >> >>>>>>> >>> >>> > >> 'name': >>> >> >>>>>>> >>> >>> > >> 'RHEL'}, 'lastClient': '10.10.10.2'}} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,620::BindingXMLRPC::1132::vds::(wrapper) >>> >> >>>>>>> >>> >>> > >> client >>> >> >>>>>>> >>> >>> > >> [10.10.10.2]::call >>> >> >>>>>>> >>> >>> > >> getHardwareInfo with () {} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,621::BindingXMLRPC::1139::vds::(wrapper) >>> >> >>>>>>> >>> >>> > >> return >>> >> >>>>>>> >>> >>> > >> getHardwareInfo >>> >> >>>>>>> >>> >>> > >> with {'status': {'message': 'Done', 'code': 0}, >>> >> >>>>>>> >>> >>> > >> 'info': >>> >> >>>>>>> >>> >>> > >> {'systemProductName': 'CS24-TY', >>> >> >>>>>>> >>> >>> > >> 'systemSerialNumber': >>> >> >>>>>>> >>> >>> > >> '7LWSPN1', >>> >> >>>>>>> >>> >>> > >> 'systemFamily': 'Server', 'systemVersion': 'A00', >>> >> >>>>>>> >>> >>> > >> 'systemUUID': >>> >> >>>>>>> >>> >>> > >> '44454c4c-4c00-1057-8053-b7c04f504e31', >>> >> >>>>>>> >>> >>> > >> 'systemManufacturer': >>> >> >>>>>>> >>> >>> > >> 'Dell'}} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:41,733::BindingXMLRPC::1132::vds::(wrapper) >>> >> >>>>>>> >>> >>> > >> client >>> >> >>>>>>> >>> >>> > >> [10.10.10.2]::call >>> >> >>>>>>> >>> >>> > >> hostsList with () {} flowID [222e8036] >>> >> >>>>>>> >>> >>> > >> Thread-13::ERROR::2014-11-24 >>> >> >>>>>>> >>> >>> > >> 21:41:44,753::BindingXMLRPC::1148::vds::(wrapper) >>> >> >>>>>>> >>> >>> > >> vdsm >>> >> >>>>>>> >>> >>> > >> exception >>> >> >>>>>>> >>> >>> > >> occured >>> >> >>>>>>> >>> >>> > >> Traceback (most recent call last): >>> >> >>>>>>> >>> >>> > >> File "/usr/share/vdsm/rpc/BindingXMLRPC.py", >>> line >>> >> >>>>>>> >>> >>> > >> 1135, >>> >> >>>>>>> >>> >>> > >> in >>> >> >>>>>>> >>> >>> > >> wrapper >>> >> >>>>>>> >>> >>> > >> res = f(*args, **kwargs) >>> >> >>>>>>> >>> >>> > >> File "/usr/share/vdsm/gluster/api.py", line >>> 54, in >>> >> >>>>>>> >>> >>> > >> wrapper >>> >> >>>>>>> >>> >>> > >> rv = func(*args, **kwargs) >>> >> >>>>>>> >>> >>> > >> File "/usr/share/vdsm/gluster/api.py", line >>> 251, in >>> >> >>>>>>> >>> >>> > >> hostsList >>> >> >>>>>>> >>> >>> > >> return {'hosts': >>> >> >>>>>>> >>> >>> > >> self.svdsmProxy.glusterPeerStatus()} >>> >> >>>>>>> >>> >>> > >> File "/usr/share/vdsm/supervdsm.py", line 50, >>> in >>> >> >>>>>>> >>> >>> > >> __call__ >>> >> >>>>>>> >>> >>> > >> return callMethod() >>> >> >>>>>>> >>> >>> > >> File "/usr/share/vdsm/supervdsm.py", line 48, >>> in >>> >> >>>>>>> >>> >>> > >> <lambda> >>> >> >>>>>>> >>> >>> > >> **kwargs) >>> >> >>>>>>> >>> >>> > >> File "<string>", line 2, in glusterPeerStatus >>> >> >>>>>>> >>> >>> > >> File >>> >> >>>>>>> >>> >>> > >> >>> "/usr/lib64/python2.6/multiprocessing/managers.py", >>> >> >>>>>>> >>> >>> > >> line >>> >> >>>>>>> >>> >>> > >> 740, >>> >> >>>>>>> >>> >>> > >> in >>> >> >>>>>>> >>> >>> > >> _callmethod >>> >> >>>>>>> >>> >>> > >> raise convert_to_error(kind, result) >>> >> >>>>>>> >>> >>> > >> GlusterCmdExecFailedException: Command execution >>> >> >>>>>>> >>> >>> > >> failed >>> >> >>>>>>> >>> >>> > >> error: Connection failed. Please check if gluster >>> >> >>>>>>> >>> >>> > >> daemon >>> >> >>>>>>> >>> >>> > >> is >>> >> >>>>>>> >>> >>> > >> operational. >>> >> >>>>>>> >>> >>> > >> return code: 1 >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:50,949::task::595::Storage.TaskManager.Task::(_updateState) >>> >> >>>>>>> >>> >>> > >> >>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving >>> >> >>>>>>> >>> >>> > >> from >>> >> >>>>>>> >>> >>> > >> state >>> >> >>>>>>> >>> >>> > >> init >>> >> >>>>>>> >>> >>> > >> -> >>> >> >>>>>>> >>> >>> > >> state preparing >>> >> >>>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> 21:41:50,950::logUtils::44::dispatcher::(wrapper) Run >>> >> >>>>>>> >>> >>> > >> and >>> >> >>>>>>> >>> >>> > >> protect: >>> >> >>>>>>> >>> >>> > >> repoStats(options=None) >>> >> >>>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> 21:41:50,950::logUtils::47::dispatcher::(wrapper) Run >>> >> >>>>>>> >>> >>> > >> and >>> >> >>>>>>> >>> >>> > >> protect: >>> >> >>>>>>> >>> >>> > >> repoStats, Return response: {} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:50,950::task::1191::Storage.TaskManager.Task::(prepare) >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::finished: {} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:50,950::task::595::Storage.TaskManager.Task::(_updateState) >>> >> >>>>>>> >>> >>> > >> >>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving >>> >> >>>>>>> >>> >>> > >> from >>> >> >>>>>>> >>> >>> > >> state >>> >> >>>>>>> >>> >>> > >> preparing >>> >> >>>>>>> >>> >>> > >> -> >>> >> >>>>>>> >>> >>> > >> state finished >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:50,951::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) >>> >> >>>>>>> >>> >>> > >> Owner.releaseAll requests {} resources {} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:50,951::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) >>> >> >>>>>>> >>> >>> > >> Owner.cancelAll requests {} >>> >> >>>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24 >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> 21:41:50,951::task::993::Storage.TaskManager.Task::(_decref) >>> >> >>>>>>> >>> >>> > >> Task=`c9042986-c978-4b08-adb2-616f5299e115`::ref >>> 0 >>> >> >>>>>>> >>> >>> > >> aborting >>> >> >>>>>>> >>> >>> > >> False >>> >> >>>>>>> >>> >>> > >> ------------------------------- >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> [root at compute4 ~]# service glusterd status >>> >> >>>>>>> >>> >>> > >> glusterd is stopped >>> >> >>>>>>> >>> >>> > >> [root at compute4 ~]# chkconfig --list | grep >>> glusterd >>> >> >>>>>>> >>> >>> > >> glusterd 0:off 1:off 2:on 3:on >>> 4:on >>> >> >>>>>>> >>> >>> > >> 5:on >>> >> >>>>>>> >>> >>> > >> 6:off >>> >> >>>>>>> >>> >>> > >> [root at compute4 ~]# >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> Thanks, >>> >> >>>>>>> >>> >>> > >> Punit >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> On Mon, Nov 24, 2014 at 6:36 PM, Kanagaraj >>> >> >>>>>>> >>> >>> > >> <kmayilsa at redhat.com> >>> >> >>>>>>> >>> >>> > >> wrote: >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >>> Can you send the corresponding error in >>> vdsm.log >>> >> >>>>>>> >>> >>> > >>> from >>> >> >>>>>>> >>> >>> > >>> the >>> >> >>>>>>> >>> >>> > >>> host? >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> Also check if glusterd service is running. >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> Thanks, >>> >> >>>>>>> >>> >>> > >>> Kanagaraj >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> On 11/24/2014 03:39 PM, Punit Dambiwal wrote: >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> Hi, >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> After reboot my Hypervisior host can not >>> activate >>> >> >>>>>>> >>> >>> > >>> again >>> >> >>>>>>> >>> >>> > >>> in the >>> >> >>>>>>> >>> >>> > >>> cluster >>> >> >>>>>>> >>> >>> > >>> and failed with the following error :- >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> Gluster command [<UNKNOWN>] failed on server... >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> Engine logs :- >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:28,397 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-64) START, >>> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand(HostName >>> Compute4, >>> >> >>>>>>> >>> >>> > >>> HostId >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id: >>> >> >>>>>>> >>> >>> > >>> 5f251c90 >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:30,609 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-64) FINISH, >>> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand, >>> >> >>>>>>> >>> >>> > >>> return: >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> {26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity at d95203e0 >>> }, >>> >> >>>>>>> >>> >>> > >>> log id: 5f251c90 >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,768 INFO >>> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >>> >> >>>>>>> >>> >>> > >>> (ajp--127.0.0.1-8702-8) >>> >> >>>>>>> >>> >>> > >>> [287d570d] Lock Acquired to object EngineLock >>> >> >>>>>>> >>> >>> > >>> [exclusiveLocks>>> >> >>>>>>> >>> >>> > >>> key: >>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a value: VDS >>> >> >>>>>>> >>> >>> > >>> , sharedLocks= ] >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,795 INFO >>> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] >>> >> >>>>>>> >>> >>> > >>> Running >>> >> >>>>>>> >>> >>> > >>> command: >>> >> >>>>>>> >>> >>> > >>> ActivateVdsCommand internal: false. Entities >>> >> >>>>>>> >>> >>> > >>> affected : >>> >> >>>>>>> >>> >>> > >>> ID: >>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: >>> VDSAction >>> >> >>>>>>> >>> >>> > >>> group >>> >> >>>>>>> >>> >>> > >>> MANIPULATE_HOST >>> >> >>>>>>> >>> >>> > >>> with role type ADMIN >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,796 INFO >>> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] >>> >> >>>>>>> >>> >>> > >>> Before >>> >> >>>>>>> >>> >>> > >>> acquiring >>> >> >>>>>>> >>> >>> > >>> lock in >>> >> >>>>>>> >>> >>> > >>> order to prevent monitoring for host Compute5 >>> from >>> >> >>>>>>> >>> >>> > >>> data-center >>> >> >>>>>>> >>> >>> > >>> SV_WTC >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,797 INFO >>> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] >>> Lock >>> >> >>>>>>> >>> >>> > >>> acquired, >>> >> >>>>>>> >>> >>> > >>> from >>> >> >>>>>>> >>> >>> > >>> now a >>> >> >>>>>>> >>> >>> > >>> monitoring of host will be skipped for host >>> Compute5 >>> >> >>>>>>> >>> >>> > >>> from >>> >> >>>>>>> >>> >>> > >>> data-center >>> >> >>>>>>> >>> >>> > >>> SV_WTC >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,817 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] >>> >> >>>>>>> >>> >>> > >>> START, >>> >> >>>>>>> >>> >>> > >>> SetVdsStatusVDSCommand(HostName = Compute5, >>> HostId >>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a, >>> >> >>>>>>> >>> >>> > >>> status=Unassigned, >>> >> >>>>>>> >>> >>> > >>> nonOperationalReason=NONE, >>> >> >>>>>>> >>> >>> > >>> stopSpmFailureLogged=false), >>> >> >>>>>>> >>> >>> > >>> log id: >>> >> >>>>>>> >>> >>> > >>> 1cbc7311 >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:33,820 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] >>> >> >>>>>>> >>> >>> > >>> FINISH, >>> >> >>>>>>> >>> >>> > >>> SetVdsStatusVDSCommand, log id: 1cbc7311 >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:34,086 INFO >>> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) Activate >>> >> >>>>>>> >>> >>> > >>> finished. >>> >> >>>>>>> >>> >>> > >>> Lock >>> >> >>>>>>> >>> >>> > >>> released. >>> >> >>>>>>> >>> >>> > >>> Monitoring can run now for host Compute5 from >>> >> >>>>>>> >>> >>> > >>> data-center >>> >> >>>>>>> >>> >>> > >>> SV_WTC >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:34,088 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) Correlation >>> ID: >>> >> >>>>>>> >>> >>> > >>> 287d570d, >>> >> >>>>>>> >>> >>> > >>> Job >>> >> >>>>>>> >>> >>> > >>> ID: >>> >> >>>>>>> >>> >>> > >>> 5ef8e4d6-b2bc-469e-8e81-7ef74b2a001a, Call >>> Stack: >>> >> >>>>>>> >>> >>> > >>> null, >>> >> >>>>>>> >>> >>> > >>> Custom >>> >> >>>>>>> >>> >>> > >>> Event ID: >>> >> >>>>>>> >>> >>> > >>> -1, Message: Host Compute5 was activated by >>> admin. >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:34,090 INFO >>> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.bll.ActivateVdsCommand] >>> >> >>>>>>> >>> >>> > >>> (org.ovirt.thread.pool-8-thread-45) Lock freed >>> to >>> >> >>>>>>> >>> >>> > >>> object >>> >> >>>>>>> >>> >>> > >>> EngineLock >>> >> >>>>>>> >>> >>> > >>> [exclusiveLocks= key: >>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a >>> >> >>>>>>> >>> >>> > >>> value: >>> >> >>>>>>> >>> >>> > >>> VDS >>> >> >>>>>>> >>> >>> > >>> , sharedLocks= ] >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:35,792 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-55) [3706e836] >>> START, >>> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand(HostName >>> Compute4, >>> >> >>>>>>> >>> >>> > >>> HostId >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id: >>> >> >>>>>>> >>> >>> > >>> 48a0c832 >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,064 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) START, >>> >> >>>>>>> >>> >>> > >>> GetHardwareInfoVDSCommand(HostName = Compute5, >>> >> >>>>>>> >>> >>> > >>> HostId >>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a, >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> vds=Host[Compute5,0bf6b00f-7947-4411-b55a-cc5eea2b381a]), log >>> >> >>>>>>> >>> >>> > >>> id: >>> >> >>>>>>> >>> >>> > >>> 6d560cc2 >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,074 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) FINISH, >>> >> >>>>>>> >>> >>> > >>> GetHardwareInfoVDSCommand, log >>> >> >>>>>>> >>> >>> > >>> id: 6d560cc2 >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,093 WARN >>> >> >>>>>>> >>> >>> > >>> [org.ovirt.engine.core.vdsbroker.VdsManager] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) Host >>> Compute5 is >>> >> >>>>>>> >>> >>> > >>> running >>> >> >>>>>>> >>> >>> > >>> with >>> >> >>>>>>> >>> >>> > >>> disabled >>> >> >>>>>>> >>> >>> > >>> SELinux. >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,127 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf] >>> >> >>>>>>> >>> >>> > >>> Running >>> >> >>>>>>> >>> >>> > >>> command: >>> >> >>>>>>> >>> >>> > >>> HandleVdsCpuFlagsOrClusterChangedCommand >>> internal: >>> >> >>>>>>> >>> >>> > >>> true. >>> >> >>>>>>> >>> >>> > >>> Entities >>> >> >>>>>>> >>> >>> > >>> affected >>> >> >>>>>>> >>> >>> > >>> : ID: 0bf6b00f-7947-4411-b55a-cc5eea2b381a >>> Type: >>> >> >>>>>>> >>> >>> > >>> VDS >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,147 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf] >>> START, >>> >> >>>>>>> >>> >>> > >>> GlusterServersListVDSCommand(HostName >>> Compute5, >>> >> >>>>>>> >>> >>> > >>> HostId >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a), log id: >>> >> >>>>>>> >>> >>> > >>> 4faed87 >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,164 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf] >>> >> >>>>>>> >>> >>> > >>> FINISH, >>> >> >>>>>>> >>> >>> > >>> GlusterServersListVDSCommand, log id: 4faed87 >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,189 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.bll.SetNonOperationalVdsCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] >>> >> >>>>>>> >>> >>> > >>> Running >>> >> >>>>>>> >>> >>> > >>> command: >>> >> >>>>>>> >>> >>> > >>> SetNonOperationalVdsCommand internal: true. >>> Entities >>> >> >>>>>>> >>> >>> > >>> affected : >>> >> >>>>>>> >>> >>> > >>> ID: >>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,206 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] >>> START, >>> >> >>>>>>> >>> >>> > >>> SetVdsStatusVDSCommand(HostName = Compute5, >>> HostId >>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a, >>> >> >>>>>>> >>> >>> > >>> status=NonOperational, >>> >> >>>>>>> >>> >>> > >>> nonOperationalReason=GLUSTER_COMMAND_FAILED, >>> >> >>>>>>> >>> >>> > >>> stopSpmFailureLogged=false), >>> >> >>>>>>> >>> >>> > >>> log id: fed5617 >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,209 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] >>> >> >>>>>>> >>> >>> > >>> FINISH, >>> >> >>>>>>> >>> >>> > >>> SetVdsStatusVDSCommand, log id: fed5617 >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,223 ERROR >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] >>> >> >>>>>>> >>> >>> > >>> Correlation ID: >>> >> >>>>>>> >>> >>> > >>> 4a84c4e5, >>> >> >>>>>>> >>> >>> > >>> Job >>> >> >>>>>>> >>> >>> > >>> ID: 4bfd4a6d-c3ef-468f-a40e-a3a6ca13011b, Call >>> >> >>>>>>> >>> >>> > >>> Stack: >>> >> >>>>>>> >>> >>> > >>> null, >>> >> >>>>>>> >>> >>> > >>> Custom >>> >> >>>>>>> >>> >>> > >>> Event >>> >> >>>>>>> >>> >>> > >>> ID: -1, Message: Gluster command [<UNKNOWN>] >>> failed >>> >> >>>>>>> >>> >>> > >>> on >>> >> >>>>>>> >>> >>> > >>> server >>> >> >>>>>>> >>> >>> > >>> Compute5. >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,243 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] >>> >> >>>>>>> >>> >>> > >>> Correlation ID: >>> >> >>>>>>> >>> >>> > >>> null, >>> >> >>>>>>> >>> >>> > >>> Call >>> >> >>>>>>> >>> >>> > >>> Stack: null, Custom Event ID: -1, Message: >>> Status of >>> >> >>>>>>> >>> >>> > >>> host >>> >> >>>>>>> >>> >>> > >>> Compute5 >>> >> >>>>>>> >>> >>> > >>> was >>> >> >>>>>>> >>> >>> > >>> set >>> >> >>>>>>> >>> >>> > >>> to NonOperational. >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,272 INFO >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.bll.HandleVdsVersionCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [a0c8a7f] >>> Running >>> >> >>>>>>> >>> >>> > >>> command: >>> >> >>>>>>> >>> >>> > >>> HandleVdsVersionCommand internal: true. Entities >>> >> >>>>>>> >>> >>> > >>> affected : >>> >> >>>>>>> >>> >>> > >>> ID: >>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:37,274 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-69) [a0c8a7f] >>> Host >>> >> >>>>>>> >>> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a : Compute5 >>> is >>> >> >>>>>>> >>> >>> > >>> already in >>> >> >>>>>>> >>> >>> > >>> NonOperational status for reason >>> >> >>>>>>> >>> >>> > >>> GLUSTER_COMMAND_FAILED. >>> >> >>>>>>> >>> >>> > >>> SetNonOperationalVds command is skipped. >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:38,065 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-55) [3706e836] >>> >> >>>>>>> >>> >>> > >>> FINISH, >>> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand, return: >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> {26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity at 4e72a1b1 >>> }, >>> >> >>>>>>> >>> >>> > >>> log id: 48a0c832 >>> >> >>>>>>> >>> >>> > >>> 2014-11-24 18:05:43,243 INFO >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>> >> >>>>>>> >>> >>> > >>> (DefaultQuartzScheduler_Worker-35) START, >>> >> >>>>>>> >>> >>> > >>> GlusterVolumesListVDSCommand(HostName >>> Compute4, >>> >> >>>>>>> >>> >>> > >>> HostId >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id: >>> >> >>>>>>> >>> >>> > >>> 3ce13ebc >>> >> >>>>>>> >>> >>> > >>> ^C >>> >> >>>>>>> >>> >>> > >>> [root at ccr01 ~]# >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> Thanks, >>> >> >>>>>>> >>> >>> > >>> Punit >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> _______________________________________________ >>> >> >>>>>>> >>> >>> > >>> Users mailing >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> listUsers at ovirt.orghttp:// >>> lists.ovirt.org/mailman/listinfo/users >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >>> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > >> >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > > >>> >> >>>>>>> >>> >>> > >>> >> >>>>>>> >>> >> >>> >> >>>>>>> >>> >> >>> >> >>>>>>> >>> > >>> >> >>>>>>> >>> > >>> >> >>>>>>> >>> > _______________________________________________ >>> >> >>>>>>> >>> > Gluster-users mailing list >>> >> >>>>>>> >>> > Gluster-users at gluster.org >>> >> >>>>>>> >>> > >>> >> >>>>>>> >>> > >>> http://supercolony.gluster.org/mailman/listinfo/gluster-users >>> >> >>>>>>> >> >>> >> >>>>>>> >> >>> >> >>>>>>> > >>> >> >>>>>> >>> >> >>>>>> >>> >> >>>> >>> >> >> >>> > >>> > >>> >> >> >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20141209/671c550f/attachment.html>