Punit Dambiwal
2015-Jan-13 02:18 UTC
[Gluster-users] Failed to create volume in OVirt with gluster
Hi, Please find the more details on this ....can anybody from gluster will help me here :- Gluster CLI Logs :- /var/log/glusterfs/cli.log [2015-01-13 02:06:23.071969] T [cli.c:264:cli_rpc_notify] 0-glusterfs: got RPC_CLNT_CONNECT [2015-01-13 02:06:23.072012] T [cli-quotad-client.c:94:cli_quotad_notify] 0-glusterfs: got RPC_CLNT_CONNECT [2015-01-13 02:06:23.072024] I [socket.c:2344:socket_event_handler] 0-transport: disconnecting now [2015-01-13 02:06:23.072055] T [cli-quotad-client.c:100:cli_quotad_notify] 0-glusterfs: got RPC_CLNT_DISCONNECT [2015-01-13 02:06:23.072131] T [rpc-clnt.c:1381:rpc_clnt_record] 0-glusterfs: Auth Info: pid: 0, uid: 0, gid: 0, owner: [2015-01-13 02:06:23.072176] T [rpc-clnt.c:1238:rpc_clnt_record_build_header] 0-rpc-clnt: Request fraglen 128, payload: 64, rpc hdr: 64 [2015-01-13 02:06:23.072572] T [socket.c:2863:socket_connect] (--> /usr/lib64/libglusterfs.so.0(_gf_log_callingfn+0x1e0)[0x7fed02f15420] (--> /usr/lib64/glusterfs/3.6.1/rpc-transport/socket.so(+0x7293)[0x7fed001a4293] (--> /usr/lib64/libgfrpc.so.0(rpc_clnt_submit+0x468)[0x7fed0266df98] (--> /usr/sbin/gluster(cli_submit_request+0xdb)[0x40a9bb] (--> /usr/sbin/gluster(cli_cmd_submit+0x8e)[0x40b7be] ))))) 0-glusterfs: connect () called on transport already connected [2015-01-13 02:06:23.072616] T [rpc-clnt.c:1573:rpc_clnt_submit] 0-rpc-clnt: submitted request (XID: 0x1 Program: Gluster CLI, ProgVers: 2, Proc: 27) to rpc-transport (glusterfs) [2015-01-13 02:06:23.072633] D [rpc-clnt-ping.c:231:rpc_clnt_start_ping] 0-glusterfs: ping timeout is 0, returning [2015-01-13 02:06:23.075930] T [rpc-clnt.c:660:rpc_clnt_reply_init] 0-glusterfs: received rpc message (RPC XID: 0x1 Program: Gluster CLI, ProgVers: 2, Proc: 27) from rpc-transport (glusterfs) [2015-01-13 02:06:23.075976] D [cli-rpc-ops.c:6548:gf_cli_status_cbk] 0-cli: Received response to status cmd [2015-01-13 02:06:23.076025] D [cli-cmd.c:384:cli_cmd_submit] 0-cli: Returning 0 [2015-01-13 02:06:23.076049] D [cli-rpc-ops.c:6811:gf_cli_status_volume] 0-cli: Returning: 0 [2015-01-13 02:06:23.076192] D [cli-xml-output.c:84:cli_begin_xml_output] 0-cli: Returning 0 [2015-01-13 02:06:23.076244] D [cli-xml-output.c:131:cli_xml_output_common] 0-cli: Returning 0 [2015-01-13 02:06:23.076256] D [cli-xml-output.c:1375:cli_xml_output_vol_status_begin] 0-cli: Returning 0 [2015-01-13 02:06:23.076437] D [cli-xml-output.c:108:cli_end_xml_output] 0-cli: Returning 0 [2015-01-13 02:06:23.076459] D [cli-xml-output.c:1398:cli_xml_output_vol_status_end] 0-cli: Returning 0 [2015-01-13 02:06:23.076490] I [input.c:36:cli_batch] 0-: Exiting with: 0 Command log :- /var/log/glusterfs/.cmd_log_history Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. [2015-01-13 01:10:35.836676] : volume status all tasks : FAILED : Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. [2015-01-13 01:16:25.956514] : volume status all tasks : FAILED : Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. [2015-01-13 01:17:36.977833] : volume status all tasks : FAILED : Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. [2015-01-13 01:21:07.048053] : volume status all tasks : FAILED : Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. [2015-01-13 01:26:57.168661] : volume status all tasks : FAILED : Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. [2015-01-13 01:28:07.194428] : volume status all tasks : FAILED : Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. [2015-01-13 01:30:27.256667] : volume status vol01 : FAILED : Locking failed on cpu02.zne01.hkg1.stack.com. Please check log file for details. Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for details. Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for details. [2015-01-13 01:34:58.350748] : volume status all tasks : FAILED : Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. [2015-01-13 01:36:08.375326] : volume status all tasks : FAILED : Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. [2015-01-13 01:36:08.386470] : volume status vol01 : FAILED : Locking failed on cpu02.zne01.hkg1.stack.com. Please check log file for details. Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for details. Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for details. [2015-01-13 01:42:59.524215] : volume stop vol01 : FAILED : Locking failed on cpu02.zne01.hkg1.stack.com. Please check log file for details. Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for details. Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for details. [2015-01-13 01:45:10.550659] : volume status all tasks : FAILED : Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. Staging failed on 00000000-0000-0000-0000-000000000000. Please check log file for details. [2015-01-13 01:46:10.656802] : volume status all tasks : SUCCESS [2015-01-13 01:51:02.796031] : volume status all tasks : SUCCESS [2015-01-13 01:52:02.897804] : volume status all tasks : SUCCESS [2015-01-13 01:55:25.841070] : system:: uuid get : SUCCESS [2015-01-13 01:55:26.752084] : system:: uuid get : SUCCESS [2015-01-13 01:55:32.499049] : system:: uuid get : SUCCESS [2015-01-13 01:55:38.716907] : system:: uuid get : SUCCESS [2015-01-13 01:56:52.905899] : volume status all tasks : SUCCESS [2015-01-13 01:58:53.109613] : volume status all tasks : SUCCESS [2015-01-13 02:03:26.769430] : system:: uuid get : SUCCESS [2015-01-13 02:04:22.859213] : volume status all tasks : SUCCESS [2015-01-13 02:05:22.970393] : volume status all tasks : SUCCESS [2015-01-13 02:06:23.075823] : volume status all tasks : SUCCESS On Mon, Jan 12, 2015 at 10:53 PM, Kanagaraj Mayilsamy <kmayilsa at redhat.com> wrote:> I can see the failures in glusterd log. > > Can someone from glusterfs dev pls help on this? > > Thanks, > Kanagaraj > > ----- Original Message ----- > > From: "Punit Dambiwal" <hypunit at gmail.com> > > To: "Kanagaraj" <kmayilsa at redhat.com> > > Cc: "Martin Pavl?k" <mpavlik at redhat.com>, "Vijay Bellur" < > vbellur at redhat.com>, "Kaushal M" <kshlmster at gmail.com>, > > users at ovirt.org, gluster-users at gluster.org > > Sent: Monday, January 12, 2015 3:36:43 PM > > Subject: Re: Failed to create volume in OVirt with gluster > > > > Hi Kanagaraj, > > > > Please find the logs from here :- http://ur1.ca/jeszc > > > > [image: Inline image 1] > > > > [image: Inline image 2] > > > > On Mon, Jan 12, 2015 at 1:02 PM, Kanagaraj <kmayilsa at redhat.com> wrote: > > > > > Looks like there are some failures in gluster. > > > Can you send the log output from glusterd log file from the relevant > hosts? > > > > > > Thanks, > > > Kanagaraj > > > > > > > > > On 01/12/2015 10:24 AM, Punit Dambiwal wrote: > > > > > > Hi, > > > > > > Is there any one from gluster can help me here :- > > > > > > Engine logs :- > > > > > > 2015-01-12 12:50:33,841 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:34,725 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:36,824 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:36,853 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:36,866 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:37,751 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:39,849 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:39,878 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:39,890 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:40,776 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:42,878 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:42,903 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:42,916 INFO > > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock > > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > > > value: GLUSTER > > > , sharedLocks= ] > > > 2015-01-12 12:50:43,771 INFO > > > > [org.ovirt.engine.core.vdsbroker.gluster.CreateGlusterVolumeVDSCommand] > > > (ajp--127.0.0.1-8702-1) [330ace48] FINISH, > CreateGlusterVolumeVDSCommand, > > > log id: 303e70a4 > > > 2015-01-12 12:50:43,780 ERROR > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > > > (ajp--127.0.0.1-8702-1) [330ace48] Correlation ID: 330ace48, Job ID: > > > 896a69b3-a678-40a7-bceb-3635e4062aa0, Call Stack: null, Custom Event > ID: > > > -1, Message: Creation of Gluster Volume vol01 failed. > > > 2015-01-12 12:50:43,785 INFO > > > [org.ovirt.engine.core.bll.gluster.CreateGlusterVolumeCommand] > > > (ajp--127.0.0.1-8702-1) [330ace48] Lock freed to object EngineLock > > > [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 value: > GLUSTER > > > , sharedLocks= ] > > > > > > [image: Inline image 2] > > > > > > > > > On Sun, Jan 11, 2015 at 6:48 PM, Martin Pavl?k <mpavlik at redhat.com> > wrote: > > > > > >> Hi Punit, > > >> > > >> unfortunately I?am not that good with the gluster, I was just > following > > >> the obvious clue from the log. Could you try on the nodes if the > packages > > >> are even available for installation > > >> > > >> yum install gluster-swift gluster-swift-object gluster-swift-plugin > > >> gluster-swift-account > > >> gluster-swift-proxy gluster-swift-doc gluster-swift-container > > >> glusterfs-geo-replication > > >> > > >> if not you could try to get them in official gluster repo. > > >> > http://download.gluster.org/pub/gluster/glusterfs/LATEST/CentOS/glusterfs-epel.repo > > >> > > >> HTH > > >> > > >> M. > > >> > > >> > > >> > > >> > > >> On 10 Jan 2015, at 04:35, Punit Dambiwal <hypunit at gmail.com> wrote: > > >> > > >> Hi Martin, > > >> > > >> I installed gluster from ovirt repo....is it require to install those > > >> packages manually ?? > > >> > > >> On Fri, Jan 9, 2015 at 7:19 PM, Martin Pavl?k <mpavlik at redhat.com> > wrote: > > >> > > >>> Hi Punit, > > >>> > > >>> can you verify that nodes contain cluster packages from the following > > >>> log? > > >>> > > >>> Thread-14::DEBUG::2015-01-09 > > >>> 18:06:28,823::caps::716::root::(_getKeyPackages) rpm package > > >>> ('gluster-swift',) not found > > >>> Thread-14::DEBUG::2015-01-09 > > >>> 18:06:28,825::caps::716::root::(_getKeyPackages) rpm package > > >>> ('gluster-swift-object',) not found > > >>> Thread-14::DEBUG::2015-01-09 > > >>> 18:06:28,826::caps::716::root::(_getKeyPackages) rpm package > > >>> ('gluster-swift-plugin',) not found > > >>> Thread-14::DEBUG::2015-01-09 > > >>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package > > >>> ('gluster-swift-account',) not found > > >>> Thread-14::DEBUG::2015-01-09 > > >>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package > > >>> ('gluster-swift-proxy',) not found > > >>> Thread-14::DEBUG::2015-01-09 > > >>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package > > >>> ('gluster-swift-doc',) not found > > >>> Thread-14::DEBUG::2015-01-09 > > >>> 18:06:28,830::caps::716::root::(_getKeyPackages) rpm package > > >>> ('gluster-swift-container',) not found > > >>> Thread-14::DEBUG::2015-01-09 > > >>> 18:06:28,830::caps::716::root::(_getKeyPackages) rpm package > > >>> ('glusterfs-geo-replication',) not found > > >>> > > >>> > > >>> M. > > >>> > > >>> On 09 Jan 2015, at 11:13, Punit Dambiwal <hypunit at gmail.com> > wrote: > > >>> > > >>> Hi Kanagaraj, > > >>> > > >>> Please find the attached logs :- > > >>> > > >>> Engine Logs :- http://ur1.ca/jdopt > > >>> VDSM Logs :- http://ur1.ca/jdoq9 > > >>> > > >>> > > >>> > > >>> On Thu, Jan 8, 2015 at 6:05 PM, Kanagaraj <kmayilsa at redhat.com> > wrote: > > >>> > > >>>> Do you see any errors in the UI? > > >>>> > > >>>> Also please provide the engine.log and vdsm.log when the failure > > >>>> occured. > > >>>> > > >>>> Thanks, > > >>>> Kanagaraj > > >>>> > > >>>> > > >>>> On 01/08/2015 02:25 PM, Punit Dambiwal wrote: > > >>>> > > >>>> Hi Martin, > > >>>> > > >>>> The steps are below :- > > >>>> > > >>>> 1. Step the ovirt engine on the one server... > > >>>> 2. Installed centos 7 on 4 host node servers.. > > >>>> 3. I am using host node (compute+storage)....now i have added all 4 > > >>>> nodes to engine... > > >>>> 4. Create the gluster volume from GUI... > > >>>> > > >>>> Network :- > > >>>> eth0 :- public network (1G) > > >>>> eth1+eth2=bond0= VM public network (1G) > > >>>> eth3+eth4=bond1=ovirtmgmt+storage (10G private network) > > >>>> > > >>>> every hostnode has 24 bricks=24*4(distributed replicated) > > >>>> > > >>>> Thanks, > > >>>> Punit > > >>>> > > >>>> > > >>>> On Thu, Jan 8, 2015 at 3:20 PM, Martin Pavl?k <mpavlik at redhat.com> > > >>>> wrote: > > >>>> > > >>>>> Hi Punit, > > >>>>> > > >>>>> can you please provide also errors from /var/log/vdsm/vdsm.log and > > >>>>> /var/log/vdsm/vdsmd.log > > >>>>> > > >>>>> it would be really helpful if you provided exact steps how to > > >>>>> reproduce the problem. > > >>>>> > > >>>>> regards > > >>>>> > > >>>>> Martin Pavlik - rhev QE > > >>>>> > On 08 Jan 2015, at 03:06, Punit Dambiwal <hypunit at gmail.com> > wrote: > > >>>>> > > > >>>>> > Hi, > > >>>>> > > > >>>>> > I try to add gluster volume but it failed... > > >>>>> > > > >>>>> > Ovirt :- 3.5 > > >>>>> > VDSM :- vdsm-4.16.7-1.gitdb83943.el7 > > >>>>> > KVM :- 1.5.3 - 60.el7_0.2 > > >>>>> > libvirt-1.1.1-29.el7_0.4 > > >>>>> > Glusterfs :- glusterfs-3.5.3-1.el7 > > >>>>> > > > >>>>> > Engine Logs :- > > >>>>> > > > >>>>> > 2015-01-08 09:57:52,569 INFO > > >>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > >>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait > lock > > >>>>> EngineLock [exclusiveLocks= key: > 00000001-0001-0001-0001-000000000300 > > >>>>> value: GLUSTER > > >>>>> > , sharedLocks= ] > > >>>>> > 2015-01-08 09:57:52,609 INFO > > >>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > >>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait > lock > > >>>>> EngineLock [exclusiveLocks= key: > 00000001-0001-0001-0001-000000000300 > > >>>>> value: GLUSTER > > >>>>> > , sharedLocks= ] > > >>>>> > 2015-01-08 09:57:55,582 INFO > > >>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > >>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait > lock > > >>>>> EngineLock [exclusiveLocks= key: > 00000001-0001-0001-0001-000000000300 > > >>>>> value: GLUSTER > > >>>>> > , sharedLocks= ] > > >>>>> > 2015-01-08 09:57:55,591 INFO > > >>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > >>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait > lock > > >>>>> EngineLock [exclusiveLocks= key: > 00000001-0001-0001-0001-000000000300 > > >>>>> value: GLUSTER > > >>>>> > , sharedLocks= ] > > >>>>> > 2015-01-08 09:57:55,596 INFO > > >>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > >>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait > lock > > >>>>> EngineLock [exclusiveLocks= key: > 00000001-0001-0001-0001-000000000300 > > >>>>> value: GLUSTER > > >>>>> > , sharedLocks= ] > > >>>>> > 2015-01-08 09:57:55,633 INFO > > >>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > > >>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait > lock > > >>>>> EngineLock [exclusiveLocks= key: > 00000001-0001-0001-0001-000000000300 > > >>>>> value: GLUSTER > > >>>>> > , sharedLocks= ] > > >>>>> > ^C > > >>>>> > > > >>>>> > > > >>>>> > > >>>>> > > >>>> > > >>>> > > >>> <216 09-Jan-15.jpg><217 09-Jan-15.jpg> > > >>> > > >>> > > >>> > > >> > > >> > > > > > > > > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150113/aad716a6/attachment.html>
Atin Mukherjee
2015-Jan-13 04:37 UTC
[Gluster-users] Failed to create volume in OVirt with gluster
Punit, cli log wouldn't help much here. To debug this issue further can you please let us know the following: 1. gluster peer status output 2. gluster volume status output 3. gluster --version output. 4. Which command got failed 5. glusterd log file of all the nodes ~Atin On 01/13/2015 07:48 AM, Punit Dambiwal wrote:> Hi, > > Please find the more details on this ....can anybody from gluster will help > me here :- > > > Gluster CLI Logs :- /var/log/glusterfs/cli.log > > [2015-01-13 02:06:23.071969] T [cli.c:264:cli_rpc_notify] 0-glusterfs: got > RPC_CLNT_CONNECT > [2015-01-13 02:06:23.072012] T [cli-quotad-client.c:94:cli_quotad_notify] > 0-glusterfs: got RPC_CLNT_CONNECT > [2015-01-13 02:06:23.072024] I [socket.c:2344:socket_event_handler] > 0-transport: disconnecting now > [2015-01-13 02:06:23.072055] T [cli-quotad-client.c:100:cli_quotad_notify] > 0-glusterfs: got RPC_CLNT_DISCONNECT > [2015-01-13 02:06:23.072131] T [rpc-clnt.c:1381:rpc_clnt_record] > 0-glusterfs: Auth Info: pid: 0, uid: 0, gid: 0, owner: > [2015-01-13 02:06:23.072176] T > [rpc-clnt.c:1238:rpc_clnt_record_build_header] 0-rpc-clnt: Request fraglen > 128, payload: 64, rpc hdr: 64 > [2015-01-13 02:06:23.072572] T [socket.c:2863:socket_connect] (--> > /usr/lib64/libglusterfs.so.0(_gf_log_callingfn+0x1e0)[0x7fed02f15420] (--> > /usr/lib64/glusterfs/3.6.1/rpc-transport/socket.so(+0x7293)[0x7fed001a4293] > (--> /usr/lib64/libgfrpc.so.0(rpc_clnt_submit+0x468)[0x7fed0266df98] (--> > /usr/sbin/gluster(cli_submit_request+0xdb)[0x40a9bb] (--> > /usr/sbin/gluster(cli_cmd_submit+0x8e)[0x40b7be] ))))) 0-glusterfs: connect > () called on transport already connected > [2015-01-13 02:06:23.072616] T [rpc-clnt.c:1573:rpc_clnt_submit] > 0-rpc-clnt: submitted request (XID: 0x1 Program: Gluster CLI, ProgVers: 2, > Proc: 27) to rpc-transport (glusterfs) > [2015-01-13 02:06:23.072633] D [rpc-clnt-ping.c:231:rpc_clnt_start_ping] > 0-glusterfs: ping timeout is 0, returning > [2015-01-13 02:06:23.075930] T [rpc-clnt.c:660:rpc_clnt_reply_init] > 0-glusterfs: received rpc message (RPC XID: 0x1 Program: Gluster CLI, > ProgVers: 2, Proc: 27) from rpc-transport (glusterfs) > [2015-01-13 02:06:23.075976] D [cli-rpc-ops.c:6548:gf_cli_status_cbk] > 0-cli: Received response to status cmd > [2015-01-13 02:06:23.076025] D [cli-cmd.c:384:cli_cmd_submit] 0-cli: > Returning 0 > [2015-01-13 02:06:23.076049] D [cli-rpc-ops.c:6811:gf_cli_status_volume] > 0-cli: Returning: 0 > [2015-01-13 02:06:23.076192] D [cli-xml-output.c:84:cli_begin_xml_output] > 0-cli: Returning 0 > [2015-01-13 02:06:23.076244] D [cli-xml-output.c:131:cli_xml_output_common] > 0-cli: Returning 0 > [2015-01-13 02:06:23.076256] D > [cli-xml-output.c:1375:cli_xml_output_vol_status_begin] 0-cli: Returning 0 > [2015-01-13 02:06:23.076437] D [cli-xml-output.c:108:cli_end_xml_output] > 0-cli: Returning 0 > [2015-01-13 02:06:23.076459] D > [cli-xml-output.c:1398:cli_xml_output_vol_status_end] 0-cli: Returning 0 > [2015-01-13 02:06:23.076490] I [input.c:36:cli_batch] 0-: Exiting with: 0 > > Command log :- /var/log/glusterfs/.cmd_log_history > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > [2015-01-13 01:10:35.836676] : volume status all tasks : FAILED : Staging > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > [2015-01-13 01:16:25.956514] : volume status all tasks : FAILED : Staging > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > [2015-01-13 01:17:36.977833] : volume status all tasks : FAILED : Staging > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > [2015-01-13 01:21:07.048053] : volume status all tasks : FAILED : Staging > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > [2015-01-13 01:26:57.168661] : volume status all tasks : FAILED : Staging > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > [2015-01-13 01:28:07.194428] : volume status all tasks : FAILED : Staging > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > [2015-01-13 01:30:27.256667] : volume status vol01 : FAILED : Locking > failed on cpu02.zne01.hkg1.stack.com. Please check log file for details. > Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for > details. > Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for > details. > [2015-01-13 01:34:58.350748] : volume status all tasks : FAILED : Staging > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > [2015-01-13 01:36:08.375326] : volume status all tasks : FAILED : Staging > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > [2015-01-13 01:36:08.386470] : volume status vol01 : FAILED : Locking > failed on cpu02.zne01.hkg1.stack.com. Please check log file for details. > Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for > details. > Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for > details. > [2015-01-13 01:42:59.524215] : volume stop vol01 : FAILED : Locking failed > on cpu02.zne01.hkg1.stack.com. Please check log file for details. > Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for > details. > Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for > details. > [2015-01-13 01:45:10.550659] : volume status all tasks : FAILED : Staging > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > file for details. > [2015-01-13 01:46:10.656802] : volume status all tasks : SUCCESS > [2015-01-13 01:51:02.796031] : volume status all tasks : SUCCESS > [2015-01-13 01:52:02.897804] : volume status all tasks : SUCCESS > [2015-01-13 01:55:25.841070] : system:: uuid get : SUCCESS > [2015-01-13 01:55:26.752084] : system:: uuid get : SUCCESS > [2015-01-13 01:55:32.499049] : system:: uuid get : SUCCESS > [2015-01-13 01:55:38.716907] : system:: uuid get : SUCCESS > [2015-01-13 01:56:52.905899] : volume status all tasks : SUCCESS > [2015-01-13 01:58:53.109613] : volume status all tasks : SUCCESS > [2015-01-13 02:03:26.769430] : system:: uuid get : SUCCESS > [2015-01-13 02:04:22.859213] : volume status all tasks : SUCCESS > [2015-01-13 02:05:22.970393] : volume status all tasks : SUCCESS > [2015-01-13 02:06:23.075823] : volume status all tasks : SUCCESS > > > On Mon, Jan 12, 2015 at 10:53 PM, Kanagaraj Mayilsamy <kmayilsa at redhat.com> > wrote: > >> I can see the failures in glusterd log. >> >> Can someone from glusterfs dev pls help on this? >> >> Thanks, >> Kanagaraj >> >> ----- Original Message ----- >>> From: "Punit Dambiwal" <hypunit at gmail.com> >>> To: "Kanagaraj" <kmayilsa at redhat.com> >>> Cc: "Martin Pavl?k" <mpavlik at redhat.com>, "Vijay Bellur" < >> vbellur at redhat.com>, "Kaushal M" <kshlmster at gmail.com>, >>> users at ovirt.org, gluster-users at gluster.org >>> Sent: Monday, January 12, 2015 3:36:43 PM >>> Subject: Re: Failed to create volume in OVirt with gluster >>> >>> Hi Kanagaraj, >>> >>> Please find the logs from here :- http://ur1.ca/jeszc >>> >>> [image: Inline image 1] >>> >>> [image: Inline image 2] >>> >>> On Mon, Jan 12, 2015 at 1:02 PM, Kanagaraj <kmayilsa at redhat.com> wrote: >>> >>>> Looks like there are some failures in gluster. >>>> Can you send the log output from glusterd log file from the relevant >> hosts? >>>> >>>> Thanks, >>>> Kanagaraj >>>> >>>> >>>> On 01/12/2015 10:24 AM, Punit Dambiwal wrote: >>>> >>>> Hi, >>>> >>>> Is there any one from gluster can help me here :- >>>> >>>> Engine logs :- >>>> >>>> 2015-01-12 12:50:33,841 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:34,725 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:36,824 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:36,853 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:36,866 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:37,751 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:39,849 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:39,878 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:39,890 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:40,776 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:42,878 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:42,903 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:42,916 INFO >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait lock >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>> value: GLUSTER >>>> , sharedLocks= ] >>>> 2015-01-12 12:50:43,771 INFO >>>> >> [org.ovirt.engine.core.vdsbroker.gluster.CreateGlusterVolumeVDSCommand] >>>> (ajp--127.0.0.1-8702-1) [330ace48] FINISH, >> CreateGlusterVolumeVDSCommand, >>>> log id: 303e70a4 >>>> 2015-01-12 12:50:43,780 ERROR >>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>>> (ajp--127.0.0.1-8702-1) [330ace48] Correlation ID: 330ace48, Job ID: >>>> 896a69b3-a678-40a7-bceb-3635e4062aa0, Call Stack: null, Custom Event >> ID: >>>> -1, Message: Creation of Gluster Volume vol01 failed. >>>> 2015-01-12 12:50:43,785 INFO >>>> [org.ovirt.engine.core.bll.gluster.CreateGlusterVolumeCommand] >>>> (ajp--127.0.0.1-8702-1) [330ace48] Lock freed to object EngineLock >>>> [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 value: >> GLUSTER >>>> , sharedLocks= ] >>>> >>>> [image: Inline image 2] >>>> >>>> >>>> On Sun, Jan 11, 2015 at 6:48 PM, Martin Pavl?k <mpavlik at redhat.com> >> wrote: >>>> >>>>> Hi Punit, >>>>> >>>>> unfortunately I?am not that good with the gluster, I was just >> following >>>>> the obvious clue from the log. Could you try on the nodes if the >> packages >>>>> are even available for installation >>>>> >>>>> yum install gluster-swift gluster-swift-object gluster-swift-plugin >>>>> gluster-swift-account >>>>> gluster-swift-proxy gluster-swift-doc gluster-swift-container >>>>> glusterfs-geo-replication >>>>> >>>>> if not you could try to get them in official gluster repo. >>>>> >> http://download.gluster.org/pub/gluster/glusterfs/LATEST/CentOS/glusterfs-epel.repo >>>>> >>>>> HTH >>>>> >>>>> M. >>>>> >>>>> >>>>> >>>>> >>>>> On 10 Jan 2015, at 04:35, Punit Dambiwal <hypunit at gmail.com> wrote: >>>>> >>>>> Hi Martin, >>>>> >>>>> I installed gluster from ovirt repo....is it require to install those >>>>> packages manually ?? >>>>> >>>>> On Fri, Jan 9, 2015 at 7:19 PM, Martin Pavl?k <mpavlik at redhat.com> >> wrote: >>>>> >>>>>> Hi Punit, >>>>>> >>>>>> can you verify that nodes contain cluster packages from the following >>>>>> log? >>>>>> >>>>>> Thread-14::DEBUG::2015-01-09 >>>>>> 18:06:28,823::caps::716::root::(_getKeyPackages) rpm package >>>>>> ('gluster-swift',) not found >>>>>> Thread-14::DEBUG::2015-01-09 >>>>>> 18:06:28,825::caps::716::root::(_getKeyPackages) rpm package >>>>>> ('gluster-swift-object',) not found >>>>>> Thread-14::DEBUG::2015-01-09 >>>>>> 18:06:28,826::caps::716::root::(_getKeyPackages) rpm package >>>>>> ('gluster-swift-plugin',) not found >>>>>> Thread-14::DEBUG::2015-01-09 >>>>>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package >>>>>> ('gluster-swift-account',) not found >>>>>> Thread-14::DEBUG::2015-01-09 >>>>>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package >>>>>> ('gluster-swift-proxy',) not found >>>>>> Thread-14::DEBUG::2015-01-09 >>>>>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package >>>>>> ('gluster-swift-doc',) not found >>>>>> Thread-14::DEBUG::2015-01-09 >>>>>> 18:06:28,830::caps::716::root::(_getKeyPackages) rpm package >>>>>> ('gluster-swift-container',) not found >>>>>> Thread-14::DEBUG::2015-01-09 >>>>>> 18:06:28,830::caps::716::root::(_getKeyPackages) rpm package >>>>>> ('glusterfs-geo-replication',) not found >>>>>> >>>>>> >>>>>> M. >>>>>> >>>>>> On 09 Jan 2015, at 11:13, Punit Dambiwal <hypunit at gmail.com> >> wrote: >>>>>> >>>>>> Hi Kanagaraj, >>>>>> >>>>>> Please find the attached logs :- >>>>>> >>>>>> Engine Logs :- http://ur1.ca/jdopt >>>>>> VDSM Logs :- http://ur1.ca/jdoq9 >>>>>> >>>>>> >>>>>> >>>>>> On Thu, Jan 8, 2015 at 6:05 PM, Kanagaraj <kmayilsa at redhat.com> >> wrote: >>>>>> >>>>>>> Do you see any errors in the UI? >>>>>>> >>>>>>> Also please provide the engine.log and vdsm.log when the failure >>>>>>> occured. >>>>>>> >>>>>>> Thanks, >>>>>>> Kanagaraj >>>>>>> >>>>>>> >>>>>>> On 01/08/2015 02:25 PM, Punit Dambiwal wrote: >>>>>>> >>>>>>> Hi Martin, >>>>>>> >>>>>>> The steps are below :- >>>>>>> >>>>>>> 1. Step the ovirt engine on the one server... >>>>>>> 2. Installed centos 7 on 4 host node servers.. >>>>>>> 3. I am using host node (compute+storage)....now i have added all 4 >>>>>>> nodes to engine... >>>>>>> 4. Create the gluster volume from GUI... >>>>>>> >>>>>>> Network :- >>>>>>> eth0 :- public network (1G) >>>>>>> eth1+eth2=bond0= VM public network (1G) >>>>>>> eth3+eth4=bond1=ovirtmgmt+storage (10G private network) >>>>>>> >>>>>>> every hostnode has 24 bricks=24*4(distributed replicated) >>>>>>> >>>>>>> Thanks, >>>>>>> Punit >>>>>>> >>>>>>> >>>>>>> On Thu, Jan 8, 2015 at 3:20 PM, Martin Pavl?k <mpavlik at redhat.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Hi Punit, >>>>>>>> >>>>>>>> can you please provide also errors from /var/log/vdsm/vdsm.log and >>>>>>>> /var/log/vdsm/vdsmd.log >>>>>>>> >>>>>>>> it would be really helpful if you provided exact steps how to >>>>>>>> reproduce the problem. >>>>>>>> >>>>>>>> regards >>>>>>>> >>>>>>>> Martin Pavlik - rhev QE >>>>>>>> > On 08 Jan 2015, at 03:06, Punit Dambiwal <hypunit at gmail.com> >> wrote: >>>>>>>>> >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> I try to add gluster volume but it failed... >>>>>>>>> >>>>>>>>> Ovirt :- 3.5 >>>>>>>>> VDSM :- vdsm-4.16.7-1.gitdb83943.el7 >>>>>>>>> KVM :- 1.5.3 - 60.el7_0.2 >>>>>>>>> libvirt-1.1.1-29.el7_0.4 >>>>>>>>> Glusterfs :- glusterfs-3.5.3-1.el7 >>>>>>>>> >>>>>>>>> Engine Logs :- >>>>>>>>> >>>>>>>>> 2015-01-08 09:57:52,569 INFO >>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait >> lock >>>>>>>> EngineLock [exclusiveLocks= key: >> 00000001-0001-0001-0001-000000000300 >>>>>>>> value: GLUSTER >>>>>>>>> , sharedLocks= ] >>>>>>>>> 2015-01-08 09:57:52,609 INFO >>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait >> lock >>>>>>>> EngineLock [exclusiveLocks= key: >> 00000001-0001-0001-0001-000000000300 >>>>>>>> value: GLUSTER >>>>>>>>> , sharedLocks= ] >>>>>>>>> 2015-01-08 09:57:55,582 INFO >>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait >> lock >>>>>>>> EngineLock [exclusiveLocks= key: >> 00000001-0001-0001-0001-000000000300 >>>>>>>> value: GLUSTER >>>>>>>>> , sharedLocks= ] >>>>>>>>> 2015-01-08 09:57:55,591 INFO >>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait >> lock >>>>>>>> EngineLock [exclusiveLocks= key: >> 00000001-0001-0001-0001-000000000300 >>>>>>>> value: GLUSTER >>>>>>>>> , sharedLocks= ] >>>>>>>>> 2015-01-08 09:57:55,596 INFO >>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait >> lock >>>>>>>> EngineLock [exclusiveLocks= key: >> 00000001-0001-0001-0001-000000000300 >>>>>>>> value: GLUSTER >>>>>>>>> , sharedLocks= ] >>>>>>>>> 2015-01-08 09:57:55,633 INFO >>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait >> lock >>>>>>>> EngineLock [exclusiveLocks= key: >> 00000001-0001-0001-0001-000000000300 >>>>>>>> value: GLUSTER >>>>>>>>> , sharedLocks= ] >>>>>>>>> ^C >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>> <216 09-Jan-15.jpg><217 09-Jan-15.jpg> >>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>> >>>> >>> >> > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://www.gluster.org/mailman/listinfo/gluster-users >