On Tue, Mar 1, 2016 at 5:07 PM, paf1 at email.cz <paf1 at email.cz> wrote:> Hello, can anybody explain this error no.13 ( open file ) in sanlock.log > . >This is EACCES Can you share the outoput of: ls -lh /rhev/data-center/mnt/<server>:<_path>/<sd_uuid>/dom_md> > The size of "ids" file is zero (0) >This is how we create the ids file when initializing it. But then we use sanlock to initialize the ids file, and it should be 1MiB after that. Is this ids files created by vdsm, or one you created yourself?> 2016-02-28 03:25:46+0100 269626 [1951]: open error -13 > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids > 2016-02-28 03:25:46+0100 269626 [1951]: s187985 open_disk > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids > error -13 > 2016-02-28 03:25:56+0100 269636 [11304]: s187992 lockspace > 7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0 > > If the main problem is about zero file size, can I regenerate this file > online securely , with no VM dependence ???? >Yes, I think I already referred to the instructions how to do that in a previous mail.> > dist = RHEL - 7 - 2.1511 > kernel = 3.10.0 - 327.10.1.el7.x86_64 > KVM = 2.3.0 - 29.1.el7 > libvirt = libvirt-1.2.17-13.el7_2.3 > vdsm = vdsm-4.16.30-0.el7 > GlusterFS = glusterfs-3.7.8-1.el7 > > > regs. > Pavel > > _______________________________________________ > Users mailing list > Users at ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160301/e1ff7f47/attachment.html>
paf1 at email.cz
2016-Mar-01 20:51 UTC
[Gluster-users] [ovirt-users] open error -13 = sanlock
HI, requested output: # ls -lh /rhev/data-center/mnt/glusterSD/localhost:*/*/dom_md /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md: total 2,1M -rw-rw---- 1 vdsm kvm 1,0M 1. b?e 21.28 ids -rw-rw---- 1 vdsm kvm 16M 7. lis 22.16 inbox -rw-rw---- 1 vdsm kvm 2,0M 7. lis 22.17 leases -rw-r--r-- 1 vdsm kvm 335 7. lis 22.17 metadata -rw-rw---- 1 vdsm kvm 16M 7. lis 22.16 outbox /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P1/553d9b92-e4a0-4042-a579-4cabeb55ded4/dom_md: total 1,1M -rw-r--r-- 1 vdsm kvm 0 24. ?no 07.41 ids -rw-rw---- 1 vdsm kvm 16M 7. lis 00.14 inbox -rw-rw---- 1 vdsm kvm 2,0M 7. lis 03.56 leases -rw-r--r-- 1 vdsm kvm 333 7. lis 03.56 metadata -rw-rw---- 1 vdsm kvm 16M 7. lis 00.14 outbox /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md: total 1,1M -rw-r--r-- 1 vdsm kvm 0 24. ?no 07.43 ids -rw-rw---- 1 vdsm kvm 16M 7. lis 00.15 inbox -rw-rw---- 1 vdsm kvm 2,0M 7. lis 22.14 leases -rw-r--r-- 1 vdsm kvm 333 7. lis 22.14 metadata -rw-rw---- 1 vdsm kvm 16M 7. lis 00.15 outbox /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P3/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md: total 1,1M -rw-r--r-- 1 vdsm kvm 0 24. ?no 07.43 ids -rw-rw---- 1 vdsm kvm 16M 23. ?no 22.51 inbox -rw-rw---- 1 vdsm kvm 2,0M 23. ?no 23.12 leases -rw-r--r-- 1 vdsm kvm 998 25. ?no 00.35 metadata -rw-rw---- 1 vdsm kvm 16M 7. lis 00.16 outbox /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md: total 1,1M -rw-r--r-- 1 vdsm kvm 0 24. ?no 07.44 ids -rw-rw---- 1 vdsm kvm 16M 7. lis 00.17 inbox -rw-rw---- 1 vdsm kvm 2,0M 7. lis 00.18 leases -rw-r--r-- 1 vdsm kvm 333 7. lis 00.18 metadata -rw-rw---- 1 vdsm kvm 16M 7. lis 00.17 outbox /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P1/42d710a9-b844-43dc-be41-77002d1cd553/dom_md: total 1,1M -rw-rw-r-- 1 vdsm kvm 0 24. ?no 07.32 ids -rw-rw---- 1 vdsm kvm 16M 7. lis 22.18 inbox -rw-rw---- 1 vdsm kvm 2,0M 7. lis 22.18 leases -rw-r--r-- 1 vdsm kvm 333 7. lis 22.18 metadata -rw-rw---- 1 vdsm kvm 16M 7. lis 22.18 outbox /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md: total 3,0M -rw-rw-r-- 1 vdsm kvm 1,0M 1. b?e 21.28 ids -rw-rw---- 1 vdsm kvm 16M 25. ?no 00.42 inbox -rw-rw---- 1 vdsm kvm 2,0M 25. ?no 00.44 leases -rw-r--r-- 1 vdsm kvm 997 24. ?no 02.46 metadata -rw-rw---- 1 vdsm kvm 16M 25. ?no 00.44 outbox /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P3/ef010d08-aed1-41c4-ba9a-e6d9bdecb4b4/dom_md: total 2,1M -rw-r--r-- 1 vdsm kvm 0 24. ?no 07.34 ids -rw-rw---- 1 vdsm kvm 16M 23. ?no 22.35 inbox -rw-rw---- 1 vdsm kvm 2,0M 23. ?no 22.38 leases -rw-r--r-- 1 vdsm kvm 1,1K 24. ?no 19.07 metadata -rw-rw---- 1 vdsm kvm 16M 23. ?no 22.27 outbox /rhev/data-center/mnt/glusterSD/localhost:_2KVM12__P4/300e9ac8-3c2f-4703-9bb1-1df2130c7c97/dom_md: total 3,0M -rw-rw-r-- 1 vdsm kvm 1,0M 1. b?e 21.28 ids -rw-rw-r-- 1 vdsm kvm 16M 6. lis 23.50 inbox -rw-rw-r-- 1 vdsm kvm 2,0M 6. lis 23.51 leases -rw-rw-r-- 1 vdsm kvm 734 7. lis 02.13 metadata -rw-rw-r-- 1 vdsm kvm 16M 6. lis 16.55 outbox /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P5/1ca56b45-701e-4c22-9f59-3aebea4d8477/dom_md: total 1,1M -rw-rw-r-- 1 vdsm kvm 0 24. ?no 07.35 ids -rw-rw-r-- 1 vdsm kvm 16M 24. ?no 01.06 inbox -rw-rw-r-- 1 vdsm kvm 2,0M 24. ?no 02.44 leases -rw-r--r-- 1 vdsm kvm 998 24. ?no 19.07 metadata -rw-rw-r-- 1 vdsm kvm 16M 7. lis 22.20 outbox The ids files was generated by "touch" command after deleting them due "sanlock locking hang" gluster crash & reboot I expected that they will be filled automaticaly after gluster reboot ( the shadow copy from ".gluster " directory was deleted & created empty too ) OK, it looks that sanlock can't work with empty file or rewrite them . Am I right ?? The last point - about "ids" workaround - this is offline version = VMs have to be moved out from for continual running with maintenance volume mode But this is not acceptable in current situation, so the question again, is it safe to do it online ?? ( YES / NO ) regs. Pavel On 1.3.2016 18:38, Nir Soffer wrote:> On Tue, Mar 1, 2016 at 5:07 PM, paf1 at email.cz <mailto:paf1 at email.cz> > <paf1 at email.cz <mailto:paf1 at email.cz>> wrote: > > Hello, can anybody explain this error no.13 ( open file ) in > sanlock.log . > > > This is EACCES > > Can you share the outoput of: > > ls -lh /rhev/data-center/mnt/<server>:<_path>/<sd_uuid>/dom_md > > > The size of "ids" file is zero (0) > > > This is how we create the ids file when initializing it. > > But then we use sanlock to initialize the ids file, and it should be > 1MiB after that. > > Is this ids files created by vdsm, or one you created yourself? > > 2016-02-28 03:25:46+0100 269626 [1951]: open error -13 > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids > 2016-02-28 03:25:46+0100 269626 [1951]: s187985 open_disk > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids > error -13 > 2016-02-28 03:25:56+0100 269636 [11304]: s187992 lockspace > 7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0 > > If the main problem is about zero file size, can I regenerate > this file online securely , with no VM dependence ???? > > > Yes, I think I already referred to the instructions how to do that in > a previous mail. > > > > dist = RHEL - 7 - 2.1511 > kernel = 3.10.0 - 327.10.1.el7.x86_64 > KVM = 2.3.0 - 29.1.el7 > libvirt = libvirt-1.2.17-13.el7_2.3 > vdsm = vdsm-4.16.30-0.el7 > GlusterFS = glusterfs-3.7.8-1.el7 > > > regs. > Pavel > > _______________________________________________ > Users mailing list > Users at ovirt.org <mailto:Users at ovirt.org> > http://lists.ovirt.org/mailman/listinfo/users > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160301/02c83ca7/attachment.html>