Pranith Kumar Karampuri
2015-Jun-02 13:12 UTC
[Gluster-users] gfapi access not working with 3.7.0
On 05/31/2015 01:02 AM, Alessandro De Salvo wrote:> Thanks again Pranith!Unfortunately the fixes missed the window for 3.7.1. These fixes will be available in the next release. Pranith> > Alessandro > >> Il giorno 30/mag/2015, alle ore 03:16, Pranith Kumar Karampuri <pkarampu at redhat.com> ha scritto: >> >> Alessandro, >> Same issue as the bug you talked about in "gluster volume heal info" thread. http://review.gluster.org/11002 should address this (Not the same fix you patched for glfsheal). I will backport this one to 3.7.1 as well. >> >> Pranith >> On 05/30/2015 12:23 AM, Alessandro De Salvo wrote: >>> Hi, >>> I'm trying to access a volume using gfapi and gluster 3.7.0. This was >>> working with 3.6.3, but not working anymore after the upgrade. >>> The volume has snapshots enabled, and it's configured in the following >>> way: >>> >>> # gluster volume info adsnet-vm-01 >>> Volume Name: adsnet-vm-01 >>> Type: Replicate >>> Volume ID: f8f615df-3dde-4ea6-9bdb-29a1706e864c >>> Status: Started >>> Number of Bricks: 1 x 2 = 2 >>> Transport-type: tcp >>> Bricks: >>> Brick1: gwads02.sta.adsnet.it:/gluster/vm01/data >>> Brick2: gwads03.sta.adsnet.it:/gluster/vm01/data >>> Options Reconfigured: >>> server.allow-insecure: on >>> features.file-snapshot: on >>> features.barrier: disable >>> nfs.disable: true >>> >>> Also, my /etc/glusterfs/glusterd.vol has the needed option: >>> >>> # cat /etc/glusterfs/glusterd.vol >>> # This file is managed by puppet, do not change >>> volume management >>> type mgmt/glusterd >>> option working-directory /var/lib/glusterd >>> option transport-type socket,rdma >>> option transport.socket.keepalive-time 10 >>> option transport.socket.keepalive-interval 2 >>> option transport.socket.read-fail-log off >>> option ping-timeout 30 >>> option rpc-auth-allow-insecure on >>> # option base-port 49152 >>> end-volume >>> >>> However, when I try for example to access an image via qemu-img it >>> segfaults: >>> >>> # qemu-img info >>> gluster://gwads03.sta.adsnet.it/adsnet-vm-01/images/foreman7.vm.adsnet.it.qcow2 >>> [2015-05-29 18:39:41.436951] E [MSGID: 108006] >>> [afr-common.c:3919:afr_notify] 0-adsnet-vm-01-replicate-0: All >>> subvolumes are down. Going offline until atleast one of them comes back >>> up. >>> [2015-05-29 18:39:41.438234] E [rpc-transport.c:512:rpc_transport_unref] >>> (--> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x186)[0x7fc3851caf16] >>> (--> /lib64/libgfrpc.so.0(rpc_transport_unref+0xa3)[0x7fc387c855a3] >>> (--> /lib64/libgfrpc.so.0(rpc_clnt_unref+0x5c)[0x7fc387c888ec] >>> (--> /lib64/libglusterfs.so.0(+0x21791)[0x7fc3851c7791] >>> (--> /lib64/libglusterfs.so.0(+0x21725)[0x7fc3851c7725] ))))) >>> 0-rpc_transport: invalid argument: this >>> [2015-05-29 18:39:41.438484] E [rpc-transport.c:512:rpc_transport_unref] >>> (--> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x186)[0x7fc3851caf16] >>> (--> /lib64/libgfrpc.so.0(rpc_transport_unref+0xa3)[0x7fc387c855a3] >>> (--> /lib64/libgfrpc.so.0(rpc_clnt_unref+0x5c)[0x7fc387c888ec] >>> (--> /lib64/libglusterfs.so.0(+0x21791)[0x7fc3851c7791] >>> (--> /lib64/libglusterfs.so.0(+0x21725)[0x7fc3851c7725] ))))) >>> 0-rpc_transport: invalid argument: this >>> Segmentation fault (core dumped) >>> >>> The volume is fine: >>> >>> # gluster volume status adsnet-vm-01 >>> Status of volume: adsnet-vm-01 >>> Gluster process TCP Port RDMA Port Online >>> Pid >>> ------------------------------------------------------------------------------ >>> Brick gwads02.sta.adsnet.it:/gluster/vm01/d >>> ata 49159 0 Y >>> 27878 >>> Brick gwads03.sta.adsnet.it:/gluster/vm01/d >>> ata 49159 0 Y >>> 24638 >>> Self-heal Daemon on localhost N/A N/A Y >>> 28031 >>> Self-heal Daemon on gwads03.sta.adsnet.it N/A N/A Y >>> 24667 >>> Task Status of Volume adsnet-vm-01 >>> ------------------------------------------------------------------------------ >>> There are no active volume tasks >>> >>> >>> Running with the debugger I see the following: >>> >>> (gdb) r >>> Starting program: /usr/bin/qemu-img info >>> gluster://gwads03.sta.adsnet.it/adsnet-vm-01/images/foreman7.vm.adsnet.it.qcow2 >>> [Thread debugging using libthread_db enabled] >>> Using host libthread_db library "/lib64/libthread_db.so.1". >>> [New Thread 0x7ffff176a700 (LWP 30027)] >>> [New Thread 0x7ffff0f69700 (LWP 30028)] >>> [New Thread 0x7fffe99ab700 (LWP 30029)] >>> [New Thread 0x7fffe8fa7700 (LWP 30030)] >>> [New Thread 0x7fffe3fff700 (LWP 30031)] >>> [New Thread 0x7fffdbfff700 (LWP 30032)] >>> [New Thread 0x7fffdb2dd700 (LWP 30033)] >>> [2015-05-29 18:51:25.656014] E [MSGID: 108006] >>> [afr-common.c:3919:afr_notify] 0-adsnet-vm-01-replicate-0: All >>> subvolumes are down. Going offline until atleast one of them comes back >>> up. >>> [2015-05-29 18:51:25.657338] E [rpc-transport.c:512:rpc_transport_unref] >>> (--> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x186)[0x7ffff48bcf16] >>> (--> /lib64/libgfrpc.so.0(rpc_transport_unref+0xa3)[0x7ffff73775a3] >>> (--> /lib64/libgfrpc.so.0(rpc_clnt_unref+0x5c)[0x7ffff737a8ec] >>> (--> /lib64/libglusterfs.so.0(+0x21791)[0x7ffff48b9791] >>> (--> /lib64/libglusterfs.so.0(+0x21725)[0x7ffff48b9725] ))))) >>> 0-rpc_transport: invalid argument: this >>> [2015-05-29 18:51:25.657619] E [rpc-transport.c:512:rpc_transport_unref] >>> (--> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x186)[0x7ffff48bcf16] >>> (--> /lib64/libgfrpc.so.0(rpc_transport_unref+0xa3)[0x7ffff73775a3] >>> (--> /lib64/libgfrpc.so.0(rpc_clnt_unref+0x5c)[0x7ffff737a8ec] >>> (--> /lib64/libglusterfs.so.0(+0x21791)[0x7ffff48b9791] >>> (--> /lib64/libglusterfs.so.0(+0x21725)[0x7ffff48b9725] ))))) >>> 0-rpc_transport: invalid argument: this >>> >>> Program received signal SIGSEGV, Segmentation fault. >>> inode_unref (inode=0x7fffd975c06c) at inode.c:499 >>> 499 table = inode->table; >>> (gdb) bt >>> #0 inode_unref (inode=0x7fffd975c06c) at inode.c:499 >>> #1 0x00007fffe14d5a61 in fini (this=<optimized out>) at >>> qemu-block.c:1092 >>> #2 0x00007ffff48b9791 in xlator_fini_rec (xl=0x7fffdc00b520) at >>> xlator.c:463 >>> #3 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc00c940) at >>> xlator.c:453 >>> #4 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc00dcf0) at >>> xlator.c:453 >>> #5 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc00f0a0) at >>> xlator.c:453 >>> #6 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc010470) at >>> xlator.c:453 >>> #7 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc011820) at >>> xlator.c:453 >>> #8 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc012bd0) at >>> xlator.c:453 >>> #9 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc014020) at >>> xlator.c:453 >>> #10 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc0154b0) at >>> xlator.c:453 >>> #11 0x00007ffff48baeea in xlator_tree_fini (xl=<optimized out>) at >>> xlator.c:545 >>> #12 0x00007ffff48f6b25 in glusterfs_graph_deactivate (graph=<optimized >>> out>) at graph.c:340 >>> #13 0x00007ffff758de3c in pub_glfs_fini (fs=0x555555c13cd0) at >>> glfs.c:1155 >>> #14 0x0000555555568f49 in bdrv_close (bs=bs at entry=0x555555c10790) at >>> block.c:1522 >>> #15 0x0000555555568e08 in bdrv_delete (bs=0x555555c10790) at >>> block.c:1749 >>> #16 bdrv_unref (bs=0x555555c10790) at block.c:5121 >>> #17 0x0000555555568fd3 in bdrv_close (bs=bs at entry=0x555555c0b520) at >>> block.c:1544 >>> #18 0x0000555555568e08 in bdrv_delete (bs=0x555555c0b520) at >>> block.c:1749 >>> #19 bdrv_unref (bs=bs at entry=0x555555c0b520) at block.c:5121 >>> #20 0x00005555555b75d6 in collect_image_info_list (chain=false, >>> fmt=<optimized out>, filename=<optimized out>) >>> at qemu-img.c:1820 >>> #21 img_info (argc=<optimized out>, argv=<optimized out>) at >>> qemu-img.c:1897 >>> #22 0x00007ffff5004af5 in __libc_start_main (main=0x555555561740 <main>, >>> argc=3, ubp_av=0x7fffffffe398, >>> init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized >>> out>, stack_end=0x7fffffffe388) >>> at libc-start.c:274 >>> #23 0x000055555556185d in _start () >>> >>> >>> It seems very similar to what I reported this morning about the healing, >>> still a problem with fini, but I'm not sure it's the real cause. >>> Any help? >>> Many thanks! >>> >>> Alessandro >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://www.gluster.org/mailman/listinfo/gluster-users
Alessandro De Salvo
2015-Jun-02 13:22 UTC
[Gluster-users] gfapi access not working with 3.7.0
OK, Thanks Pranith. Do you have a timeline for that? Cheers, Alessandro> Il giorno 02/giu/2015, alle ore 15:12, Pranith Kumar Karampuri <pkarampu at redhat.com> ha scritto: > > > > On 05/31/2015 01:02 AM, Alessandro De Salvo wrote: >> Thanks again Pranith! > Unfortunately the fixes missed the window for 3.7.1. These fixes will be available in the next release. > > Pranith >> >> Alessandro >> >>> Il giorno 30/mag/2015, alle ore 03:16, Pranith Kumar Karampuri <pkarampu at redhat.com> ha scritto: >>> >>> Alessandro, >>> Same issue as the bug you talked about in "gluster volume heal info" thread. http://review.gluster.org/11002 should address this (Not the same fix you patched for glfsheal). I will backport this one to 3.7.1 as well. >>> >>> Pranith >>> On 05/30/2015 12:23 AM, Alessandro De Salvo wrote: >>>> Hi, >>>> I'm trying to access a volume using gfapi and gluster 3.7.0. This was >>>> working with 3.6.3, but not working anymore after the upgrade. >>>> The volume has snapshots enabled, and it's configured in the following >>>> way: >>>> >>>> # gluster volume info adsnet-vm-01 >>>> Volume Name: adsnet-vm-01 >>>> Type: Replicate >>>> Volume ID: f8f615df-3dde-4ea6-9bdb-29a1706e864c >>>> Status: Started >>>> Number of Bricks: 1 x 2 = 2 >>>> Transport-type: tcp >>>> Bricks: >>>> Brick1: gwads02.sta.adsnet.it:/gluster/vm01/data >>>> Brick2: gwads03.sta.adsnet.it:/gluster/vm01/data >>>> Options Reconfigured: >>>> server.allow-insecure: on >>>> features.file-snapshot: on >>>> features.barrier: disable >>>> nfs.disable: true >>>> >>>> Also, my /etc/glusterfs/glusterd.vol has the needed option: >>>> >>>> # cat /etc/glusterfs/glusterd.vol >>>> # This file is managed by puppet, do not change >>>> volume management >>>> type mgmt/glusterd >>>> option working-directory /var/lib/glusterd >>>> option transport-type socket,rdma >>>> option transport.socket.keepalive-time 10 >>>> option transport.socket.keepalive-interval 2 >>>> option transport.socket.read-fail-log off >>>> option ping-timeout 30 >>>> option rpc-auth-allow-insecure on >>>> # option base-port 49152 >>>> end-volume >>>> >>>> However, when I try for example to access an image via qemu-img it >>>> segfaults: >>>> >>>> # qemu-img info >>>> gluster://gwads03.sta.adsnet.it/adsnet-vm-01/images/foreman7.vm.adsnet.it.qcow2 >>>> [2015-05-29 18:39:41.436951] E [MSGID: 108006] >>>> [afr-common.c:3919:afr_notify] 0-adsnet-vm-01-replicate-0: All >>>> subvolumes are down. Going offline until atleast one of them comes back >>>> up. >>>> [2015-05-29 18:39:41.438234] E [rpc-transport.c:512:rpc_transport_unref] >>>> (--> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x186)[0x7fc3851caf16] >>>> (--> /lib64/libgfrpc.so.0(rpc_transport_unref+0xa3)[0x7fc387c855a3] >>>> (--> /lib64/libgfrpc.so.0(rpc_clnt_unref+0x5c)[0x7fc387c888ec] >>>> (--> /lib64/libglusterfs.so.0(+0x21791)[0x7fc3851c7791] >>>> (--> /lib64/libglusterfs.so.0(+0x21725)[0x7fc3851c7725] ))))) >>>> 0-rpc_transport: invalid argument: this >>>> [2015-05-29 18:39:41.438484] E [rpc-transport.c:512:rpc_transport_unref] >>>> (--> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x186)[0x7fc3851caf16] >>>> (--> /lib64/libgfrpc.so.0(rpc_transport_unref+0xa3)[0x7fc387c855a3] >>>> (--> /lib64/libgfrpc.so.0(rpc_clnt_unref+0x5c)[0x7fc387c888ec] >>>> (--> /lib64/libglusterfs.so.0(+0x21791)[0x7fc3851c7791] >>>> (--> /lib64/libglusterfs.so.0(+0x21725)[0x7fc3851c7725] ))))) >>>> 0-rpc_transport: invalid argument: this >>>> Segmentation fault (core dumped) >>>> >>>> The volume is fine: >>>> >>>> # gluster volume status adsnet-vm-01 >>>> Status of volume: adsnet-vm-01 >>>> Gluster process TCP Port RDMA Port Online >>>> Pid >>>> ------------------------------------------------------------------------------ >>>> Brick gwads02.sta.adsnet.it:/gluster/vm01/d >>>> ata 49159 0 Y >>>> 27878 >>>> Brick gwads03.sta.adsnet.it:/gluster/vm01/d >>>> ata 49159 0 Y >>>> 24638 >>>> Self-heal Daemon on localhost N/A N/A Y >>>> 28031 >>>> Self-heal Daemon on gwads03.sta.adsnet.it N/A N/A Y >>>> 24667 >>>> Task Status of Volume adsnet-vm-01 >>>> ------------------------------------------------------------------------------ >>>> There are no active volume tasks >>>> >>>> >>>> Running with the debugger I see the following: >>>> >>>> (gdb) r >>>> Starting program: /usr/bin/qemu-img info >>>> gluster://gwads03.sta.adsnet.it/adsnet-vm-01/images/foreman7.vm.adsnet.it.qcow2 >>>> [Thread debugging using libthread_db enabled] >>>> Using host libthread_db library "/lib64/libthread_db.so.1". >>>> [New Thread 0x7ffff176a700 (LWP 30027)] >>>> [New Thread 0x7ffff0f69700 (LWP 30028)] >>>> [New Thread 0x7fffe99ab700 (LWP 30029)] >>>> [New Thread 0x7fffe8fa7700 (LWP 30030)] >>>> [New Thread 0x7fffe3fff700 (LWP 30031)] >>>> [New Thread 0x7fffdbfff700 (LWP 30032)] >>>> [New Thread 0x7fffdb2dd700 (LWP 30033)] >>>> [2015-05-29 18:51:25.656014] E [MSGID: 108006] >>>> [afr-common.c:3919:afr_notify] 0-adsnet-vm-01-replicate-0: All >>>> subvolumes are down. Going offline until atleast one of them comes back >>>> up. >>>> [2015-05-29 18:51:25.657338] E [rpc-transport.c:512:rpc_transport_unref] >>>> (--> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x186)[0x7ffff48bcf16] >>>> (--> /lib64/libgfrpc.so.0(rpc_transport_unref+0xa3)[0x7ffff73775a3] >>>> (--> /lib64/libgfrpc.so.0(rpc_clnt_unref+0x5c)[0x7ffff737a8ec] >>>> (--> /lib64/libglusterfs.so.0(+0x21791)[0x7ffff48b9791] >>>> (--> /lib64/libglusterfs.so.0(+0x21725)[0x7ffff48b9725] ))))) >>>> 0-rpc_transport: invalid argument: this >>>> [2015-05-29 18:51:25.657619] E [rpc-transport.c:512:rpc_transport_unref] >>>> (--> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x186)[0x7ffff48bcf16] >>>> (--> /lib64/libgfrpc.so.0(rpc_transport_unref+0xa3)[0x7ffff73775a3] >>>> (--> /lib64/libgfrpc.so.0(rpc_clnt_unref+0x5c)[0x7ffff737a8ec] >>>> (--> /lib64/libglusterfs.so.0(+0x21791)[0x7ffff48b9791] >>>> (--> /lib64/libglusterfs.so.0(+0x21725)[0x7ffff48b9725] ))))) >>>> 0-rpc_transport: invalid argument: this >>>> >>>> Program received signal SIGSEGV, Segmentation fault. >>>> inode_unref (inode=0x7fffd975c06c) at inode.c:499 >>>> 499 table = inode->table; >>>> (gdb) bt >>>> #0 inode_unref (inode=0x7fffd975c06c) at inode.c:499 >>>> #1 0x00007fffe14d5a61 in fini (this=<optimized out>) at >>>> qemu-block.c:1092 >>>> #2 0x00007ffff48b9791 in xlator_fini_rec (xl=0x7fffdc00b520) at >>>> xlator.c:463 >>>> #3 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc00c940) at >>>> xlator.c:453 >>>> #4 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc00dcf0) at >>>> xlator.c:453 >>>> #5 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc00f0a0) at >>>> xlator.c:453 >>>> #6 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc010470) at >>>> xlator.c:453 >>>> #7 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc011820) at >>>> xlator.c:453 >>>> #8 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc012bd0) at >>>> xlator.c:453 >>>> #9 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc014020) at >>>> xlator.c:453 >>>> #10 0x00007ffff48b9725 in xlator_fini_rec (xl=0x7fffdc0154b0) at >>>> xlator.c:453 >>>> #11 0x00007ffff48baeea in xlator_tree_fini (xl=<optimized out>) at >>>> xlator.c:545 >>>> #12 0x00007ffff48f6b25 in glusterfs_graph_deactivate (graph=<optimized >>>> out>) at graph.c:340 >>>> #13 0x00007ffff758de3c in pub_glfs_fini (fs=0x555555c13cd0) at >>>> glfs.c:1155 >>>> #14 0x0000555555568f49 in bdrv_close (bs=bs at entry=0x555555c10790) at >>>> block.c:1522 >>>> #15 0x0000555555568e08 in bdrv_delete (bs=0x555555c10790) at >>>> block.c:1749 >>>> #16 bdrv_unref (bs=0x555555c10790) at block.c:5121 >>>> #17 0x0000555555568fd3 in bdrv_close (bs=bs at entry=0x555555c0b520) at >>>> block.c:1544 >>>> #18 0x0000555555568e08 in bdrv_delete (bs=0x555555c0b520) at >>>> block.c:1749 >>>> #19 bdrv_unref (bs=bs at entry=0x555555c0b520) at block.c:5121 >>>> #20 0x00005555555b75d6 in collect_image_info_list (chain=false, >>>> fmt=<optimized out>, filename=<optimized out>) >>>> at qemu-img.c:1820 >>>> #21 img_info (argc=<optimized out>, argv=<optimized out>) at >>>> qemu-img.c:1897 >>>> #22 0x00007ffff5004af5 in __libc_start_main (main=0x555555561740 <main>, >>>> argc=3, ubp_av=0x7fffffffe398, >>>> init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized >>>> out>, stack_end=0x7fffffffe388) >>>> at libc-start.c:274 >>>> #23 0x000055555556185d in _start () >>>> >>>> >>>> It seems very similar to what I reported this morning about the healing, >>>> still a problem with fini, but I'm not sure it's the real cause. >>>> Any help? >>>> Many thanks! >>>> >>>> Alessandro >>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> http://www.gluster.org/mailman/listinfo/gluster-users >-------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 1770 bytes Desc: not available URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150602/db520ff1/attachment.p7s>