Punit Dambiwal
2015-Jan-13 06:42 UTC
[Gluster-users] Failed to create volume in OVirt with gluster
Hi Atin, Please find the output from here :- http://ur1.ca/jf4bs On Tue, Jan 13, 2015 at 12:37 PM, Atin Mukherjee <amukherj at redhat.com> wrote:> Punit, > > cli log wouldn't help much here. To debug this issue further can you > please let us know the following: > > 1. gluster peer status output > 2. gluster volume status output > 3. gluster --version output. > 4. Which command got failed > 5. glusterd log file of all the nodes > > ~Atin > > > On 01/13/2015 07:48 AM, Punit Dambiwal wrote: > > Hi, > > > > Please find the more details on this ....can anybody from gluster will > help > > me here :- > > > > > > Gluster CLI Logs :- /var/log/glusterfs/cli.log > > > > [2015-01-13 02:06:23.071969] T [cli.c:264:cli_rpc_notify] 0-glusterfs: > got > > RPC_CLNT_CONNECT > > [2015-01-13 02:06:23.072012] T [cli-quotad-client.c:94:cli_quotad_notify] > > 0-glusterfs: got RPC_CLNT_CONNECT > > [2015-01-13 02:06:23.072024] I [socket.c:2344:socket_event_handler] > > 0-transport: disconnecting now > > [2015-01-13 02:06:23.072055] T > [cli-quotad-client.c:100:cli_quotad_notify] > > 0-glusterfs: got RPC_CLNT_DISCONNECT > > [2015-01-13 02:06:23.072131] T [rpc-clnt.c:1381:rpc_clnt_record] > > 0-glusterfs: Auth Info: pid: 0, uid: 0, gid: 0, owner: > > [2015-01-13 02:06:23.072176] T > > [rpc-clnt.c:1238:rpc_clnt_record_build_header] 0-rpc-clnt: Request > fraglen > > 128, payload: 64, rpc hdr: 64 > > [2015-01-13 02:06:23.072572] T [socket.c:2863:socket_connect] (--> > > /usr/lib64/libglusterfs.so.0(_gf_log_callingfn+0x1e0)[0x7fed02f15420] > (--> > > > /usr/lib64/glusterfs/3.6.1/rpc-transport/socket.so(+0x7293)[0x7fed001a4293] > > (--> /usr/lib64/libgfrpc.so.0(rpc_clnt_submit+0x468)[0x7fed0266df98] (--> > > /usr/sbin/gluster(cli_submit_request+0xdb)[0x40a9bb] (--> > > /usr/sbin/gluster(cli_cmd_submit+0x8e)[0x40b7be] ))))) 0-glusterfs: > connect > > () called on transport already connected > > [2015-01-13 02:06:23.072616] T [rpc-clnt.c:1573:rpc_clnt_submit] > > 0-rpc-clnt: submitted request (XID: 0x1 Program: Gluster CLI, ProgVers: > 2, > > Proc: 27) to rpc-transport (glusterfs) > > [2015-01-13 02:06:23.072633] D [rpc-clnt-ping.c:231:rpc_clnt_start_ping] > > 0-glusterfs: ping timeout is 0, returning > > [2015-01-13 02:06:23.075930] T [rpc-clnt.c:660:rpc_clnt_reply_init] > > 0-glusterfs: received rpc message (RPC XID: 0x1 Program: Gluster CLI, > > ProgVers: 2, Proc: 27) from rpc-transport (glusterfs) > > [2015-01-13 02:06:23.075976] D [cli-rpc-ops.c:6548:gf_cli_status_cbk] > > 0-cli: Received response to status cmd > > [2015-01-13 02:06:23.076025] D [cli-cmd.c:384:cli_cmd_submit] 0-cli: > > Returning 0 > > [2015-01-13 02:06:23.076049] D [cli-rpc-ops.c:6811:gf_cli_status_volume] > > 0-cli: Returning: 0 > > [2015-01-13 02:06:23.076192] D [cli-xml-output.c:84:cli_begin_xml_output] > > 0-cli: Returning 0 > > [2015-01-13 02:06:23.076244] D > [cli-xml-output.c:131:cli_xml_output_common] > > 0-cli: Returning 0 > > [2015-01-13 02:06:23.076256] D > > [cli-xml-output.c:1375:cli_xml_output_vol_status_begin] 0-cli: Returning > 0 > > [2015-01-13 02:06:23.076437] D [cli-xml-output.c:108:cli_end_xml_output] > > 0-cli: Returning 0 > > [2015-01-13 02:06:23.076459] D > > [cli-xml-output.c:1398:cli_xml_output_vol_status_end] 0-cli: Returning 0 > > [2015-01-13 02:06:23.076490] I [input.c:36:cli_batch] 0-: Exiting with: 0 > > > > Command log :- /var/log/glusterfs/.cmd_log_history > > > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > [2015-01-13 01:10:35.836676] : volume status all tasks : FAILED : > Staging > > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > > details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > [2015-01-13 01:16:25.956514] : volume status all tasks : FAILED : > Staging > > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > > details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > [2015-01-13 01:17:36.977833] : volume status all tasks : FAILED : > Staging > > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > > details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > [2015-01-13 01:21:07.048053] : volume status all tasks : FAILED : > Staging > > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > > details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > [2015-01-13 01:26:57.168661] : volume status all tasks : FAILED : > Staging > > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > > details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > [2015-01-13 01:28:07.194428] : volume status all tasks : FAILED : > Staging > > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > > details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > [2015-01-13 01:30:27.256667] : volume status vol01 : FAILED : Locking > > failed on cpu02.zne01.hkg1.stack.com. Please check log file for details. > > Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for > > details. > > Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for > > details. > > [2015-01-13 01:34:58.350748] : volume status all tasks : FAILED : > Staging > > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > > details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > [2015-01-13 01:36:08.375326] : volume status all tasks : FAILED : > Staging > > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > > details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > [2015-01-13 01:36:08.386470] : volume status vol01 : FAILED : Locking > > failed on cpu02.zne01.hkg1.stack.com. Please check log file for details. > > Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for > > details. > > Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for > > details. > > [2015-01-13 01:42:59.524215] : volume stop vol01 : FAILED : Locking > failed > > on cpu02.zne01.hkg1.stack.com. Please check log file for details. > > Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for > > details. > > Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for > > details. > > [2015-01-13 01:45:10.550659] : volume status all tasks : FAILED : > Staging > > failed on 00000000-0000-0000-0000-000000000000. Please check log file for > > details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > Staging failed on 00000000-0000-0000-0000-000000000000. Please check log > > file for details. > > [2015-01-13 01:46:10.656802] : volume status all tasks : SUCCESS > > [2015-01-13 01:51:02.796031] : volume status all tasks : SUCCESS > > [2015-01-13 01:52:02.897804] : volume status all tasks : SUCCESS > > [2015-01-13 01:55:25.841070] : system:: uuid get : SUCCESS > > [2015-01-13 01:55:26.752084] : system:: uuid get : SUCCESS > > [2015-01-13 01:55:32.499049] : system:: uuid get : SUCCESS > > [2015-01-13 01:55:38.716907] : system:: uuid get : SUCCESS > > [2015-01-13 01:56:52.905899] : volume status all tasks : SUCCESS > > [2015-01-13 01:58:53.109613] : volume status all tasks : SUCCESS > > [2015-01-13 02:03:26.769430] : system:: uuid get : SUCCESS > > [2015-01-13 02:04:22.859213] : volume status all tasks : SUCCESS > > [2015-01-13 02:05:22.970393] : volume status all tasks : SUCCESS > > [2015-01-13 02:06:23.075823] : volume status all tasks : SUCCESS > > > > > > On Mon, Jan 12, 2015 at 10:53 PM, Kanagaraj Mayilsamy < > kmayilsa at redhat.com> > > wrote: > > > >> I can see the failures in glusterd log. > >> > >> Can someone from glusterfs dev pls help on this? > >> > >> Thanks, > >> Kanagaraj > >> > >> ----- Original Message ----- > >>> From: "Punit Dambiwal" <hypunit at gmail.com> > >>> To: "Kanagaraj" <kmayilsa at redhat.com> > >>> Cc: "Martin Pavl?k" <mpavlik at redhat.com>, "Vijay Bellur" < > >> vbellur at redhat.com>, "Kaushal M" <kshlmster at gmail.com>, > >>> users at ovirt.org, gluster-users at gluster.org > >>> Sent: Monday, January 12, 2015 3:36:43 PM > >>> Subject: Re: Failed to create volume in OVirt with gluster > >>> > >>> Hi Kanagaraj, > >>> > >>> Please find the logs from here :- http://ur1.ca/jeszc > >>> > >>> [image: Inline image 1] > >>> > >>> [image: Inline image 2] > >>> > >>> On Mon, Jan 12, 2015 at 1:02 PM, Kanagaraj <kmayilsa at redhat.com> > wrote: > >>> > >>>> Looks like there are some failures in gluster. > >>>> Can you send the log output from glusterd log file from the relevant > >> hosts? > >>>> > >>>> Thanks, > >>>> Kanagaraj > >>>> > >>>> > >>>> On 01/12/2015 10:24 AM, Punit Dambiwal wrote: > >>>> > >>>> Hi, > >>>> > >>>> Is there any one from gluster can help me here :- > >>>> > >>>> Engine logs :- > >>>> > >>>> 2015-01-12 12:50:33,841 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:34,725 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:36,824 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:36,853 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:36,866 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:37,751 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:39,849 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:39,878 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:39,890 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:40,776 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:42,878 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:42,903 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:42,916 INFO > >>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait > lock > >>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 > >>>> value: GLUSTER > >>>> , sharedLocks= ] > >>>> 2015-01-12 12:50:43,771 INFO > >>>> > >> [org.ovirt.engine.core.vdsbroker.gluster.CreateGlusterVolumeVDSCommand] > >>>> (ajp--127.0.0.1-8702-1) [330ace48] FINISH, > >> CreateGlusterVolumeVDSCommand, > >>>> log id: 303e70a4 > >>>> 2015-01-12 12:50:43,780 ERROR > >>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > >>>> (ajp--127.0.0.1-8702-1) [330ace48] Correlation ID: 330ace48, Job ID: > >>>> 896a69b3-a678-40a7-bceb-3635e4062aa0, Call Stack: null, Custom Event > >> ID: > >>>> -1, Message: Creation of Gluster Volume vol01 failed. > >>>> 2015-01-12 12:50:43,785 INFO > >>>> [org.ovirt.engine.core.bll.gluster.CreateGlusterVolumeCommand] > >>>> (ajp--127.0.0.1-8702-1) [330ace48] Lock freed to object EngineLock > >>>> [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 value: > >> GLUSTER > >>>> , sharedLocks= ] > >>>> > >>>> [image: Inline image 2] > >>>> > >>>> > >>>> On Sun, Jan 11, 2015 at 6:48 PM, Martin Pavl?k <mpavlik at redhat.com> > >> wrote: > >>>> > >>>>> Hi Punit, > >>>>> > >>>>> unfortunately I?am not that good with the gluster, I was just > >> following > >>>>> the obvious clue from the log. Could you try on the nodes if the > >> packages > >>>>> are even available for installation > >>>>> > >>>>> yum install gluster-swift gluster-swift-object gluster-swift-plugin > >>>>> gluster-swift-account > >>>>> gluster-swift-proxy gluster-swift-doc gluster-swift-container > >>>>> glusterfs-geo-replication > >>>>> > >>>>> if not you could try to get them in official gluster repo. > >>>>> > >> > http://download.gluster.org/pub/gluster/glusterfs/LATEST/CentOS/glusterfs-epel.repo > >>>>> > >>>>> HTH > >>>>> > >>>>> M. > >>>>> > >>>>> > >>>>> > >>>>> > >>>>> On 10 Jan 2015, at 04:35, Punit Dambiwal <hypunit at gmail.com> > wrote: > >>>>> > >>>>> Hi Martin, > >>>>> > >>>>> I installed gluster from ovirt repo....is it require to install > those > >>>>> packages manually ?? > >>>>> > >>>>> On Fri, Jan 9, 2015 at 7:19 PM, Martin Pavl?k <mpavlik at redhat.com> > >> wrote: > >>>>> > >>>>>> Hi Punit, > >>>>>> > >>>>>> can you verify that nodes contain cluster packages from the > following > >>>>>> log? > >>>>>> > >>>>>> Thread-14::DEBUG::2015-01-09 > >>>>>> 18:06:28,823::caps::716::root::(_getKeyPackages) rpm package > >>>>>> ('gluster-swift',) not found > >>>>>> Thread-14::DEBUG::2015-01-09 > >>>>>> 18:06:28,825::caps::716::root::(_getKeyPackages) rpm package > >>>>>> ('gluster-swift-object',) not found > >>>>>> Thread-14::DEBUG::2015-01-09 > >>>>>> 18:06:28,826::caps::716::root::(_getKeyPackages) rpm package > >>>>>> ('gluster-swift-plugin',) not found > >>>>>> Thread-14::DEBUG::2015-01-09 > >>>>>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package > >>>>>> ('gluster-swift-account',) not found > >>>>>> Thread-14::DEBUG::2015-01-09 > >>>>>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package > >>>>>> ('gluster-swift-proxy',) not found > >>>>>> Thread-14::DEBUG::2015-01-09 > >>>>>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package > >>>>>> ('gluster-swift-doc',) not found > >>>>>> Thread-14::DEBUG::2015-01-09 > >>>>>> 18:06:28,830::caps::716::root::(_getKeyPackages) rpm package > >>>>>> ('gluster-swift-container',) not found > >>>>>> Thread-14::DEBUG::2015-01-09 > >>>>>> 18:06:28,830::caps::716::root::(_getKeyPackages) rpm package > >>>>>> ('glusterfs-geo-replication',) not found > >>>>>> > >>>>>> > >>>>>> M. > >>>>>> > >>>>>> On 09 Jan 2015, at 11:13, Punit Dambiwal <hypunit at gmail.com> > >> wrote: > >>>>>> > >>>>>> Hi Kanagaraj, > >>>>>> > >>>>>> Please find the attached logs :- > >>>>>> > >>>>>> Engine Logs :- http://ur1.ca/jdopt > >>>>>> VDSM Logs :- http://ur1.ca/jdoq9 > >>>>>> > >>>>>> > >>>>>> > >>>>>> On Thu, Jan 8, 2015 at 6:05 PM, Kanagaraj <kmayilsa at redhat.com> > >> wrote: > >>>>>> > >>>>>>> Do you see any errors in the UI? > >>>>>>> > >>>>>>> Also please provide the engine.log and vdsm.log when the failure > >>>>>>> occured. > >>>>>>> > >>>>>>> Thanks, > >>>>>>> Kanagaraj > >>>>>>> > >>>>>>> > >>>>>>> On 01/08/2015 02:25 PM, Punit Dambiwal wrote: > >>>>>>> > >>>>>>> Hi Martin, > >>>>>>> > >>>>>>> The steps are below :- > >>>>>>> > >>>>>>> 1. Step the ovirt engine on the one server... > >>>>>>> 2. Installed centos 7 on 4 host node servers.. > >>>>>>> 3. I am using host node (compute+storage)....now i have added all 4 > >>>>>>> nodes to engine... > >>>>>>> 4. Create the gluster volume from GUI... > >>>>>>> > >>>>>>> Network :- > >>>>>>> eth0 :- public network (1G) > >>>>>>> eth1+eth2=bond0= VM public network (1G) > >>>>>>> eth3+eth4=bond1=ovirtmgmt+storage (10G private network) > >>>>>>> > >>>>>>> every hostnode has 24 bricks=24*4(distributed replicated) > >>>>>>> > >>>>>>> Thanks, > >>>>>>> Punit > >>>>>>> > >>>>>>> > >>>>>>> On Thu, Jan 8, 2015 at 3:20 PM, Martin Pavl?k <mpavlik at redhat.com> > >>>>>>> wrote: > >>>>>>> > >>>>>>>> Hi Punit, > >>>>>>>> > >>>>>>>> can you please provide also errors from /var/log/vdsm/vdsm.log and > >>>>>>>> /var/log/vdsm/vdsmd.log > >>>>>>>> > >>>>>>>> it would be really helpful if you provided exact steps how to > >>>>>>>> reproduce the problem. > >>>>>>>> > >>>>>>>> regards > >>>>>>>> > >>>>>>>> Martin Pavlik - rhev QE > >>>>>>>> > On 08 Jan 2015, at 03:06, Punit Dambiwal <hypunit at gmail.com> > >> wrote: > >>>>>>>>> > >>>>>>>>> Hi, > >>>>>>>>> > >>>>>>>>> I try to add gluster volume but it failed... > >>>>>>>>> > >>>>>>>>> Ovirt :- 3.5 > >>>>>>>>> VDSM :- vdsm-4.16.7-1.gitdb83943.el7 > >>>>>>>>> KVM :- 1.5.3 - 60.el7_0.2 > >>>>>>>>> libvirt-1.1.1-29.el7_0.4 > >>>>>>>>> Glusterfs :- glusterfs-3.5.3-1.el7 > >>>>>>>>> > >>>>>>>>> Engine Logs :- > >>>>>>>>> > >>>>>>>>> 2015-01-08 09:57:52,569 INFO > >>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait > >> lock > >>>>>>>> EngineLock [exclusiveLocks= key: > >> 00000001-0001-0001-0001-000000000300 > >>>>>>>> value: GLUSTER > >>>>>>>>> , sharedLocks= ] > >>>>>>>>> 2015-01-08 09:57:52,609 INFO > >>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait > >> lock > >>>>>>>> EngineLock [exclusiveLocks= key: > >> 00000001-0001-0001-0001-000000000300 > >>>>>>>> value: GLUSTER > >>>>>>>>> , sharedLocks= ] > >>>>>>>>> 2015-01-08 09:57:55,582 INFO > >>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait > >> lock > >>>>>>>> EngineLock [exclusiveLocks= key: > >> 00000001-0001-0001-0001-000000000300 > >>>>>>>> value: GLUSTER > >>>>>>>>> , sharedLocks= ] > >>>>>>>>> 2015-01-08 09:57:55,591 INFO > >>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait > >> lock > >>>>>>>> EngineLock [exclusiveLocks= key: > >> 00000001-0001-0001-0001-000000000300 > >>>>>>>> value: GLUSTER > >>>>>>>>> , sharedLocks= ] > >>>>>>>>> 2015-01-08 09:57:55,596 INFO > >>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait > >> lock > >>>>>>>> EngineLock [exclusiveLocks= key: > >> 00000001-0001-0001-0001-000000000300 > >>>>>>>> value: GLUSTER > >>>>>>>>> , sharedLocks= ] > >>>>>>>>> 2015-01-08 09:57:55,633 INFO > >>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > >>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait > >> lock > >>>>>>>> EngineLock [exclusiveLocks= key: > >> 00000001-0001-0001-0001-000000000300 > >>>>>>>> value: GLUSTER > >>>>>>>>> , sharedLocks= ] > >>>>>>>>> ^C > >>>>>>>>> > >>>>>>>>> > >>>>>>>> > >>>>>>>> > >>>>>>> > >>>>>>> > >>>>>> <216 09-Jan-15.jpg><217 09-Jan-15.jpg> > >>>>>> > >>>>>> > >>>>>> > >>>>> > >>>>> > >>>> > >>>> > >>> > >> > > > > > > > > _______________________________________________ > > Gluster-users mailing list > > Gluster-users at gluster.org > > http://www.gluster.org/mailman/listinfo/gluster-users > > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150113/a19b2868/attachment.html>
Atin Mukherjee
2015-Jan-13 07:00 UTC
[Gluster-users] Failed to create volume in OVirt with gluster
On 01/13/2015 12:12 PM, Punit Dambiwal wrote:> Hi Atin, > > Please find the output from here :- http://ur1.ca/jf4bs >Looks like http://review.gluster.org/#/c/9269/ should solve this issue. Please note this patch has not been taken in 3.6 release. Would you be able to apply this patch on the source and re-test? ~Atin> On Tue, Jan 13, 2015 at 12:37 PM, Atin Mukherjee <amukherj at redhat.com> > wrote: > >> Punit, >> >> cli log wouldn't help much here. To debug this issue further can you >> please let us know the following: >> >> 1. gluster peer status output >> 2. gluster volume status output >> 3. gluster --version output. >> 4. Which command got failed >> 5. glusterd log file of all the nodes >> >> ~Atin >> >> >> On 01/13/2015 07:48 AM, Punit Dambiwal wrote: >>> Hi, >>> >>> Please find the more details on this ....can anybody from gluster will >> help >>> me here :- >>> >>> >>> Gluster CLI Logs :- /var/log/glusterfs/cli.log >>> >>> [2015-01-13 02:06:23.071969] T [cli.c:264:cli_rpc_notify] 0-glusterfs: >> got >>> RPC_CLNT_CONNECT >>> [2015-01-13 02:06:23.072012] T [cli-quotad-client.c:94:cli_quotad_notify] >>> 0-glusterfs: got RPC_CLNT_CONNECT >>> [2015-01-13 02:06:23.072024] I [socket.c:2344:socket_event_handler] >>> 0-transport: disconnecting now >>> [2015-01-13 02:06:23.072055] T >> [cli-quotad-client.c:100:cli_quotad_notify] >>> 0-glusterfs: got RPC_CLNT_DISCONNECT >>> [2015-01-13 02:06:23.072131] T [rpc-clnt.c:1381:rpc_clnt_record] >>> 0-glusterfs: Auth Info: pid: 0, uid: 0, gid: 0, owner: >>> [2015-01-13 02:06:23.072176] T >>> [rpc-clnt.c:1238:rpc_clnt_record_build_header] 0-rpc-clnt: Request >> fraglen >>> 128, payload: 64, rpc hdr: 64 >>> [2015-01-13 02:06:23.072572] T [socket.c:2863:socket_connect] (--> >>> /usr/lib64/libglusterfs.so.0(_gf_log_callingfn+0x1e0)[0x7fed02f15420] >> (--> >>> >> /usr/lib64/glusterfs/3.6.1/rpc-transport/socket.so(+0x7293)[0x7fed001a4293] >>> (--> /usr/lib64/libgfrpc.so.0(rpc_clnt_submit+0x468)[0x7fed0266df98] (--> >>> /usr/sbin/gluster(cli_submit_request+0xdb)[0x40a9bb] (--> >>> /usr/sbin/gluster(cli_cmd_submit+0x8e)[0x40b7be] ))))) 0-glusterfs: >> connect >>> () called on transport already connected >>> [2015-01-13 02:06:23.072616] T [rpc-clnt.c:1573:rpc_clnt_submit] >>> 0-rpc-clnt: submitted request (XID: 0x1 Program: Gluster CLI, ProgVers: >> 2, >>> Proc: 27) to rpc-transport (glusterfs) >>> [2015-01-13 02:06:23.072633] D [rpc-clnt-ping.c:231:rpc_clnt_start_ping] >>> 0-glusterfs: ping timeout is 0, returning >>> [2015-01-13 02:06:23.075930] T [rpc-clnt.c:660:rpc_clnt_reply_init] >>> 0-glusterfs: received rpc message (RPC XID: 0x1 Program: Gluster CLI, >>> ProgVers: 2, Proc: 27) from rpc-transport (glusterfs) >>> [2015-01-13 02:06:23.075976] D [cli-rpc-ops.c:6548:gf_cli_status_cbk] >>> 0-cli: Received response to status cmd >>> [2015-01-13 02:06:23.076025] D [cli-cmd.c:384:cli_cmd_submit] 0-cli: >>> Returning 0 >>> [2015-01-13 02:06:23.076049] D [cli-rpc-ops.c:6811:gf_cli_status_volume] >>> 0-cli: Returning: 0 >>> [2015-01-13 02:06:23.076192] D [cli-xml-output.c:84:cli_begin_xml_output] >>> 0-cli: Returning 0 >>> [2015-01-13 02:06:23.076244] D >> [cli-xml-output.c:131:cli_xml_output_common] >>> 0-cli: Returning 0 >>> [2015-01-13 02:06:23.076256] D >>> [cli-xml-output.c:1375:cli_xml_output_vol_status_begin] 0-cli: Returning >> 0 >>> [2015-01-13 02:06:23.076437] D [cli-xml-output.c:108:cli_end_xml_output] >>> 0-cli: Returning 0 >>> [2015-01-13 02:06:23.076459] D >>> [cli-xml-output.c:1398:cli_xml_output_vol_status_end] 0-cli: Returning 0 >>> [2015-01-13 02:06:23.076490] I [input.c:36:cli_batch] 0-: Exiting with: 0 >>> >>> Command log :- /var/log/glusterfs/.cmd_log_history >>> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> [2015-01-13 01:10:35.836676] : volume status all tasks : FAILED : >> Staging >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file for >>> details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> [2015-01-13 01:16:25.956514] : volume status all tasks : FAILED : >> Staging >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file for >>> details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> [2015-01-13 01:17:36.977833] : volume status all tasks : FAILED : >> Staging >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file for >>> details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> [2015-01-13 01:21:07.048053] : volume status all tasks : FAILED : >> Staging >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file for >>> details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> [2015-01-13 01:26:57.168661] : volume status all tasks : FAILED : >> Staging >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file for >>> details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> [2015-01-13 01:28:07.194428] : volume status all tasks : FAILED : >> Staging >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file for >>> details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> [2015-01-13 01:30:27.256667] : volume status vol01 : FAILED : Locking >>> failed on cpu02.zne01.hkg1.stack.com. Please check log file for details. >>> Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for >>> details. >>> Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for >>> details. >>> [2015-01-13 01:34:58.350748] : volume status all tasks : FAILED : >> Staging >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file for >>> details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> [2015-01-13 01:36:08.375326] : volume status all tasks : FAILED : >> Staging >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file for >>> details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> [2015-01-13 01:36:08.386470] : volume status vol01 : FAILED : Locking >>> failed on cpu02.zne01.hkg1.stack.com. Please check log file for details. >>> Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for >>> details. >>> Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for >>> details. >>> [2015-01-13 01:42:59.524215] : volume stop vol01 : FAILED : Locking >> failed >>> on cpu02.zne01.hkg1.stack.com. Please check log file for details. >>> Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for >>> details. >>> Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for >>> details. >>> [2015-01-13 01:45:10.550659] : volume status all tasks : FAILED : >> Staging >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file for >>> details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >>> file for details. >>> [2015-01-13 01:46:10.656802] : volume status all tasks : SUCCESS >>> [2015-01-13 01:51:02.796031] : volume status all tasks : SUCCESS >>> [2015-01-13 01:52:02.897804] : volume status all tasks : SUCCESS >>> [2015-01-13 01:55:25.841070] : system:: uuid get : SUCCESS >>> [2015-01-13 01:55:26.752084] : system:: uuid get : SUCCESS >>> [2015-01-13 01:55:32.499049] : system:: uuid get : SUCCESS >>> [2015-01-13 01:55:38.716907] : system:: uuid get : SUCCESS >>> [2015-01-13 01:56:52.905899] : volume status all tasks : SUCCESS >>> [2015-01-13 01:58:53.109613] : volume status all tasks : SUCCESS >>> [2015-01-13 02:03:26.769430] : system:: uuid get : SUCCESS >>> [2015-01-13 02:04:22.859213] : volume status all tasks : SUCCESS >>> [2015-01-13 02:05:22.970393] : volume status all tasks : SUCCESS >>> [2015-01-13 02:06:23.075823] : volume status all tasks : SUCCESS >>> >>> >>> On Mon, Jan 12, 2015 at 10:53 PM, Kanagaraj Mayilsamy < >> kmayilsa at redhat.com> >>> wrote: >>> >>>> I can see the failures in glusterd log. >>>> >>>> Can someone from glusterfs dev pls help on this? >>>> >>>> Thanks, >>>> Kanagaraj >>>> >>>> ----- Original Message ----- >>>>> From: "Punit Dambiwal" <hypunit at gmail.com> >>>>> To: "Kanagaraj" <kmayilsa at redhat.com> >>>>> Cc: "Martin Pavl?k" <mpavlik at redhat.com>, "Vijay Bellur" < >>>> vbellur at redhat.com>, "Kaushal M" <kshlmster at gmail.com>, >>>>> users at ovirt.org, gluster-users at gluster.org >>>>> Sent: Monday, January 12, 2015 3:36:43 PM >>>>> Subject: Re: Failed to create volume in OVirt with gluster >>>>> >>>>> Hi Kanagaraj, >>>>> >>>>> Please find the logs from here :- http://ur1.ca/jeszc >>>>> >>>>> [image: Inline image 1] >>>>> >>>>> [image: Inline image 2] >>>>> >>>>> On Mon, Jan 12, 2015 at 1:02 PM, Kanagaraj <kmayilsa at redhat.com> >> wrote: >>>>> >>>>>> Looks like there are some failures in gluster. >>>>>> Can you send the log output from glusterd log file from the relevant >>>> hosts? >>>>>> >>>>>> Thanks, >>>>>> Kanagaraj >>>>>> >>>>>> >>>>>> On 01/12/2015 10:24 AM, Punit Dambiwal wrote: >>>>>> >>>>>> Hi, >>>>>> >>>>>> Is there any one from gluster can help me here :- >>>>>> >>>>>> Engine logs :- >>>>>> >>>>>> 2015-01-12 12:50:33,841 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:34,725 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:36,824 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:36,853 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:36,866 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:37,751 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:39,849 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:39,878 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:39,890 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:40,776 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:42,878 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:42,903 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:42,916 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >>>>>> EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >>>>>> value: GLUSTER >>>>>> , sharedLocks= ] >>>>>> 2015-01-12 12:50:43,771 INFO >>>>>> >>>> [org.ovirt.engine.core.vdsbroker.gluster.CreateGlusterVolumeVDSCommand] >>>>>> (ajp--127.0.0.1-8702-1) [330ace48] FINISH, >>>> CreateGlusterVolumeVDSCommand, >>>>>> log id: 303e70a4 >>>>>> 2015-01-12 12:50:43,780 ERROR >>>>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>>>>> (ajp--127.0.0.1-8702-1) [330ace48] Correlation ID: 330ace48, Job ID: >>>>>> 896a69b3-a678-40a7-bceb-3635e4062aa0, Call Stack: null, Custom Event >>>> ID: >>>>>> -1, Message: Creation of Gluster Volume vol01 failed. >>>>>> 2015-01-12 12:50:43,785 INFO >>>>>> [org.ovirt.engine.core.bll.gluster.CreateGlusterVolumeCommand] >>>>>> (ajp--127.0.0.1-8702-1) [330ace48] Lock freed to object EngineLock >>>>>> [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 value: >>>> GLUSTER >>>>>> , sharedLocks= ] >>>>>> >>>>>> [image: Inline image 2] >>>>>> >>>>>> >>>>>> On Sun, Jan 11, 2015 at 6:48 PM, Martin Pavl?k <mpavlik at redhat.com> >>>> wrote: >>>>>> >>>>>>> Hi Punit, >>>>>>> >>>>>>> unfortunately I?am not that good with the gluster, I was just >>>> following >>>>>>> the obvious clue from the log. Could you try on the nodes if the >>>> packages >>>>>>> are even available for installation >>>>>>> >>>>>>> yum install gluster-swift gluster-swift-object gluster-swift-plugin >>>>>>> gluster-swift-account >>>>>>> gluster-swift-proxy gluster-swift-doc gluster-swift-container >>>>>>> glusterfs-geo-replication >>>>>>> >>>>>>> if not you could try to get them in official gluster repo. >>>>>>> >>>> >> http://download.gluster.org/pub/gluster/glusterfs/LATEST/CentOS/glusterfs-epel.repo >>>>>>> >>>>>>> HTH >>>>>>> >>>>>>> M. >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> On 10 Jan 2015, at 04:35, Punit Dambiwal <hypunit at gmail.com> >> wrote: >>>>>>> >>>>>>> Hi Martin, >>>>>>> >>>>>>> I installed gluster from ovirt repo....is it require to install >> those >>>>>>> packages manually ?? >>>>>>> >>>>>>> On Fri, Jan 9, 2015 at 7:19 PM, Martin Pavl?k <mpavlik at redhat.com> >>>> wrote: >>>>>>> >>>>>>>> Hi Punit, >>>>>>>> >>>>>>>> can you verify that nodes contain cluster packages from the >> following >>>>>>>> log? >>>>>>>> >>>>>>>> Thread-14::DEBUG::2015-01-09 >>>>>>>> 18:06:28,823::caps::716::root::(_getKeyPackages) rpm package >>>>>>>> ('gluster-swift',) not found >>>>>>>> Thread-14::DEBUG::2015-01-09 >>>>>>>> 18:06:28,825::caps::716::root::(_getKeyPackages) rpm package >>>>>>>> ('gluster-swift-object',) not found >>>>>>>> Thread-14::DEBUG::2015-01-09 >>>>>>>> 18:06:28,826::caps::716::root::(_getKeyPackages) rpm package >>>>>>>> ('gluster-swift-plugin',) not found >>>>>>>> Thread-14::DEBUG::2015-01-09 >>>>>>>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package >>>>>>>> ('gluster-swift-account',) not found >>>>>>>> Thread-14::DEBUG::2015-01-09 >>>>>>>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package >>>>>>>> ('gluster-swift-proxy',) not found >>>>>>>> Thread-14::DEBUG::2015-01-09 >>>>>>>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package >>>>>>>> ('gluster-swift-doc',) not found >>>>>>>> Thread-14::DEBUG::2015-01-09 >>>>>>>> 18:06:28,830::caps::716::root::(_getKeyPackages) rpm package >>>>>>>> ('gluster-swift-container',) not found >>>>>>>> Thread-14::DEBUG::2015-01-09 >>>>>>>> 18:06:28,830::caps::716::root::(_getKeyPackages) rpm package >>>>>>>> ('glusterfs-geo-replication',) not found >>>>>>>> >>>>>>>> >>>>>>>> M. >>>>>>>> >>>>>>>> On 09 Jan 2015, at 11:13, Punit Dambiwal <hypunit at gmail.com> >>>> wrote: >>>>>>>> >>>>>>>> Hi Kanagaraj, >>>>>>>> >>>>>>>> Please find the attached logs :- >>>>>>>> >>>>>>>> Engine Logs :- http://ur1.ca/jdopt >>>>>>>> VDSM Logs :- http://ur1.ca/jdoq9 >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Thu, Jan 8, 2015 at 6:05 PM, Kanagaraj <kmayilsa at redhat.com> >>>> wrote: >>>>>>>> >>>>>>>>> Do you see any errors in the UI? >>>>>>>>> >>>>>>>>> Also please provide the engine.log and vdsm.log when the failure >>>>>>>>> occured. >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> Kanagaraj >>>>>>>>> >>>>>>>>> >>>>>>>>> On 01/08/2015 02:25 PM, Punit Dambiwal wrote: >>>>>>>>> >>>>>>>>> Hi Martin, >>>>>>>>> >>>>>>>>> The steps are below :- >>>>>>>>> >>>>>>>>> 1. Step the ovirt engine on the one server... >>>>>>>>> 2. Installed centos 7 on 4 host node servers.. >>>>>>>>> 3. I am using host node (compute+storage)....now i have added all 4 >>>>>>>>> nodes to engine... >>>>>>>>> 4. Create the gluster volume from GUI... >>>>>>>>> >>>>>>>>> Network :- >>>>>>>>> eth0 :- public network (1G) >>>>>>>>> eth1+eth2=bond0= VM public network (1G) >>>>>>>>> eth3+eth4=bond1=ovirtmgmt+storage (10G private network) >>>>>>>>> >>>>>>>>> every hostnode has 24 bricks=24*4(distributed replicated) >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> Punit >>>>>>>>> >>>>>>>>> >>>>>>>>> On Thu, Jan 8, 2015 at 3:20 PM, Martin Pavl?k <mpavlik at redhat.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Hi Punit, >>>>>>>>>> >>>>>>>>>> can you please provide also errors from /var/log/vdsm/vdsm.log and >>>>>>>>>> /var/log/vdsm/vdsmd.log >>>>>>>>>> >>>>>>>>>> it would be really helpful if you provided exact steps how to >>>>>>>>>> reproduce the problem. >>>>>>>>>> >>>>>>>>>> regards >>>>>>>>>> >>>>>>>>>> Martin Pavlik - rhev QE >>>>>>>>>> > On 08 Jan 2015, at 03:06, Punit Dambiwal <hypunit at gmail.com> >>>> wrote: >>>>>>>>>>> >>>>>>>>>>> Hi, >>>>>>>>>>> >>>>>>>>>>> I try to add gluster volume but it failed... >>>>>>>>>>> >>>>>>>>>>> Ovirt :- 3.5 >>>>>>>>>>> VDSM :- vdsm-4.16.7-1.gitdb83943.el7 >>>>>>>>>>> KVM :- 1.5.3 - 60.el7_0.2 >>>>>>>>>>> libvirt-1.1.1-29.el7_0.4 >>>>>>>>>>> Glusterfs :- glusterfs-3.5.3-1.el7 >>>>>>>>>>> >>>>>>>>>>> Engine Logs :- >>>>>>>>>>> >>>>>>>>>>> 2015-01-08 09:57:52,569 INFO >>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait >>>> lock >>>>>>>>>> EngineLock [exclusiveLocks= key: >>>> 00000001-0001-0001-0001-000000000300 >>>>>>>>>> value: GLUSTER >>>>>>>>>>> , sharedLocks= ] >>>>>>>>>>> 2015-01-08 09:57:52,609 INFO >>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait >>>> lock >>>>>>>>>> EngineLock [exclusiveLocks= key: >>>> 00000001-0001-0001-0001-000000000300 >>>>>>>>>> value: GLUSTER >>>>>>>>>>> , sharedLocks= ] >>>>>>>>>>> 2015-01-08 09:57:55,582 INFO >>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait >>>> lock >>>>>>>>>> EngineLock [exclusiveLocks= key: >>>> 00000001-0001-0001-0001-000000000300 >>>>>>>>>> value: GLUSTER >>>>>>>>>>> , sharedLocks= ] >>>>>>>>>>> 2015-01-08 09:57:55,591 INFO >>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait >>>> lock >>>>>>>>>> EngineLock [exclusiveLocks= key: >>>> 00000001-0001-0001-0001-000000000300 >>>>>>>>>> value: GLUSTER >>>>>>>>>>> , sharedLocks= ] >>>>>>>>>>> 2015-01-08 09:57:55,596 INFO >>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait >>>> lock >>>>>>>>>> EngineLock [exclusiveLocks= key: >>>> 00000001-0001-0001-0001-000000000300 >>>>>>>>>> value: GLUSTER >>>>>>>>>>> , sharedLocks= ] >>>>>>>>>>> 2015-01-08 09:57:55,633 INFO >>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and wait >>>> lock >>>>>>>>>> EngineLock [exclusiveLocks= key: >>>> 00000001-0001-0001-0001-000000000300 >>>>>>>>>> value: GLUSTER >>>>>>>>>>> , sharedLocks= ] >>>>>>>>>>> ^C >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> <216 09-Jan-15.jpg><217 09-Jan-15.jpg> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>> >>>> >>> >>> >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://www.gluster.org/mailman/listinfo/gluster-users >>> >> >