Hello When i start a virtual machine,which image was stored in Glusterfs , i met a problem like this : 009-04-22 13:26:58 W [afr.c:609:afr_open] afr1: returning EIO, file has to be manually corrected in backend 2009-04-22 13:26:58 W [stripe.c:1869:stripe_open_cbk] bricks: afr1 returned error Input/output error 2009-04-22 13:26:58 E [fuse-bridge.c:662:fuse_fd_cbk] glusterfs-fuse: 50149: OPEN() /rhel5-210/disk0 => -1 (Input/output error) 2009-04-22 13:26:58 W [afr.c:609:afr_open] afr1: returning EIO, file has to be manually corrected in backend 2009-04-22 13:26:58 W [stripe.c:1869:stripe_open_cbk] bricks: afr1 returned error Input/output error 2009-04-22 13:26:58 E [fuse-bridge.c:662:fuse_fd_cbk] glusterfs-fuse: 50150: OPEN() /rhel5-210/disk0 => -1 (Input/output error) 2009-04-22 13:26:58 W [afr.c:609:afr_open] afr1: returning EIO, file has to be manually corrected in backend 2009-04-22 13:26:58 W [stripe.c:1869:stripe_open_cbk] bricks: afr1 returned error Input/output error 2009-04-22 13:26:58 E [fuse-bridge.c:662:fuse_fd_cbk] glusterfs-fuse: 50151: OPEN() /rhel5-210/disk0 => -1 (Input/output error) 2009-04-22 13:30:03 E [afr-self-heal-data.c:778:afr_sh_data_fix] afr1: Unable to resolve conflicting data of /rhel5-210/disk0. Pleas e resolve manually by deleting the file /rhel5-210/disk0 from all but the preferred subvolume. Please consider 'option favorite-chil d <>' so i delete the films in one side of afr mode , the log was : 2009-04-22 13:30:56 W [afr-self-heal-common.c:871:sh_missing_entries_lookup_cbk] afr1: path /rhel5-210/disk0 on subvolume client4 => -1 (No such file or directory) 2009-04-22 13:30:56 W [afr-self-heal-common.c:871:sh_missing_entries_lookup_cbk] afr3: path /rhel5-210/disk0 on subvolume client6 => -1 (No such file or directory) 2009-04-22 13:30:56 W [afr-self-heal-common.c:871:sh_missing_entries_lookup_cbk] afr2: path /rhel5-210/disk0 on subvolume client5 => -1 (No such file or directory) 2009-04-22 13:30:56 W [afr-self-heal-data.c:615:afr_sh_data_open_cbk] afr1: sourcing file /rhel5-210/disk0 from client1 to other sin ks 2009-04-22 13:30:56 W [afr-self-heal-data.c:615:afr_sh_data_open_cbk] afr3: sourcing file /rhel5-210/disk0 from client3 to other sin ks 2009-04-22 13:30:56 W [afr-self-heal-data.c:615:afr_sh_data_open_cbk] afr2: sourcing file /rhel5-210/disk0 from client2 to other sin ks At last i found the delete side , file was larger than other side ,like that was 3.1G and this is 8.0G which was whole image size . Why the data was not the same with afr ? My configure is this : volume client1 ##### type protocol/client option transport-type tcp option remote-host 172.20.92.184 # IP address of the remote brick option remote-port 6997 option transport-timeout 10 # seconds to wait for a reply option remote-subvolume brick # name of the remote volume end-volume volume client2 ##### type protocol/client option transport-type tcp option remote-host 172.20.92.78 # IP address of the remote brick option remote-port 6997 option transport-timeout 10 # seconds to wait for a reply option remote-subvolume brick # name of the remote volume end-volume volume client3 ##### type protocol/client option transport-type tcp option remote-host 172.20.92.190 # IP address of the remote brick option remote-port 6997 option transport-timeout 10 # seconds to wait for a reply option remote-subvolume brick # name of the remote volume end-volume volume client4 ##### type protocol/client option transport-type tcp option remote-host 172.20.92.153 # IP address of the remote brick option remote-port 6997 option transport-timeout 10 # seconds to wait for a reply option remote-subvolume brick # name of the remote volume end-volume volume client5 ##### type protocol/client option transport-type tcp option remote-host 172.20.93.13 # IP address of the remote brick option remote-port 6997 option transport-timeout 10 # seconds to wait for a reply option remote-subvolume brick # name of the remote volume end-volume volume client6 ##### type protocol/client option transport-type tcp option remote-host 172.20.92.209 # IP address of the remote brick option remote-port 6997 option transport-timeout 10 # seconds to wait for a reply option remote-subvolume brick # name of the remote volume end-volume volume afr1 ##### type cluster/afr subvolumes client1 client4 end-volume volume afr2 ##### type cluster/afr subvolumes client2 client5 end-volume volume afr3 ##### type cluster/afr subvolumes client3 client6 end-volume volume bricks ##### type cluster/stripe option block-size 256MB subvolumes afr1 afr2 afr3 end-volume 2009-04-22 eagleeyes -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20090422/496cc529/attachment.html>