My ongoing saga in trying to get unify to work consistently across all my files - does anyone know what the best way to debug the file accesses going on on my bricks would be? Running glusterfs in debug mode doesn't appear to yield any useful information - what's concerning me is I can't tell whether gluster tries to access the files on the brick and fails (and hence doesn't show them up in the unify view) and if that's the case, at what level is it failing? - Will -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20081008/c7de3fcd/attachment.html>
Will, What problem are you facing? If you could tell that we can fix it. The problem can be as simple as incorrect spec file. Regards Krishna On Wed, Oct 8, 2008 at 1:24 PM, Will Rouesnel <electricitylikesme at hotmail.com> wrote:> My ongoing saga in trying to get unify to work consistently across all my > files - does anyone know what the best way to debug the file accesses going > on on my bricks would be? Running glusterfs in debug mode doesn't appear to > yield any useful information - what's concerning me is I can't tell whether > gluster tries to access the files on the brick and fails (and hence doesn't > show them up in the unify view) and if that's the case, at what level is it > failing? > > - Will > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > >
I have 2 servers that AFR eachother. one of them suffered a drive failure and is being rebuilt. the question is. What will happen if I just mount the empty drive back as the AFR node. will it just start grabbing the data from the other server (which is exactly what I want), OR will it start deleting the data from the other server (which is terribly bad). Another thought was to, on the current working server with good data, disable the remote afr node from it''s config (so it''s only using AFR on itself), and then leave the other machines config as is, and turn it on. This way I can be sure that the node with the data wont go nuts and start deleting, but that updates to it will get replicated to the other machine. this particular set is running 1.4pre5 if that changes the answer. Thanks, Keith
I have 2 servers that AFR eachother. one of them suffered a drive failure and is being rebuilt. the question is. What will happen if I just mount the empty drive back as the AFR node. will it just start grabbing the data from the other server (which is exactly what I want), OR will it start deleting the data from the other server (which is terribly bad). Another thought was to, on the current working server with good data, disable the remote afr node from it''s config (so it''s only using AFR on itself), and then leave the other machines config as is, and turn it on. This way I can be sure that the node with the data wont go nuts and start deleting, but that updates to it will get replicated to the other machine. this particular set is running 1.4pre5 if that changes the answer. Thanks, Keith
I have 2 servers that AFR eachother. one of them suffered a drive failure and is being rebuilt. the question is. What will happen if I just mount the empty drive back as the AFR node. will it just start grabbing the data from the other server (which is exactly what I want), OR will it start deleting the data from the other server (which is terribly bad). Another thought was to, on the current working server with good data, disable the remote afr node from it''s config (so it''s only using AFR on itself), and then leave the other machines config as is, and turn it on. This way I can be sure that the node with the data wont go nuts and start deleting, but that updates to it will get replicated to the other machine. this particular set is running 1.4pre5 if that changes the answer. Thanks, Keith
no one answered me so I''ll just report my findings: Server1 and ServerB full of data.. Gluster 1.4pre5. FedoraCore9 Disk on serverb crashes, loose everything. re-install copied over my AFR config from server1 change IP addresses as appropriate. the OS filesystems are in /gluster/home mounted gluster filesystem is /home on server1, I changed the AFR config to only list itself (because I wasn''t sure if this would work and I didn''t want afr to go delete everything instead of copy it over). so I mount /home on serverB while tailing the gluster log. the interesting thing, it created the directories in /home and many produced i/o errors the log showed many entries such as this: 2008-10-10 04:45:49 C [posix.c:2756:ensure_file_type] home1: entry /gluster/home/freeform/access-logs is a different type of file than expected (one for each directory under /home) so, that access-logs file is a symlink to somewhere else. there were also entries for any other symlinks in the first level directory. For any of these links which pointed to directories which didn''t exist, it produced an error and the directory wasn''t available. strange.. but how to fix.. here''s what I did. if I then unmount /home rm -f /gluster/home/*/access-log mkdir (link destination for access-log) mount /home there were a couple directories still inaccessible, it turned out they had bad symlinks as well. did the same thing and now it''s plugging along using the find /home/XXXX -type f -print0 | xargs -0 head -c1 > /dev/null to auto-heal them and it seems to be going just fine. once that''s finished, I''ll re-add serverB to server1''s AFR config and I presume it''ll be fine. I just found it odd that these missing symlink destinations would cause such a problem. anyway, it was a minor irritation and overall the auto-healing once going has been a lifesaver. Keith At 06:21 AM 10/9/2008, Keith Freedman wrote:>I have 2 servers that AFR eachother. > >one of them suffered a drive failure and is being rebuilt. > >the question is. What will happen if I just mount the empty drive >back as the AFR node. > >will it just start grabbing the data from the other server (which is >exactly what I want), OR >will it start deleting the data from the other server (which is terribly bad). > >Another thought was to, on the current working server with good data, >disable the remote afr node from it''s config (so it''s only using AFR >on itself), and then leave the other machines config as is, and turn it on. >This way I can be sure that the node with the data wont go nuts and >start deleting, but that updates to it will get replicated to the >other machine. > >this particular set is running 1.4pre5 if that changes the answer. > >Thanks, >Keith > > >_______________________________________________ >Gluster-users mailing list >Gluster-users at gluster.org >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
no one answered me so I''ll just report my findings: Server1 and ServerB full of data.. Gluster 1.4pre5. FedoraCore9 Disk on serverb crashes, loose everything. re-install copied over my AFR config from server1 change IP addresses as appropriate. the OS filesystems are in /gluster/home mounted gluster filesystem is /home on server1, I changed the AFR config to only list itself (because I wasn''t sure if this would work and I didn''t want afr to go delete everything instead of copy it over). so I mount /home on serverB while tailing the gluster log. the interesting thing, it created the directories in /home and many produced i/o errors the log showed many entries such as this: 2008-10-10 04:45:49 C [posix.c:2756:ensure_file_type] home1: entry /gluster/home/freeform/access-logs is a different type of file than expected (one for each directory under /home) so, that access-logs file is a symlink to somewhere else. there were also entries for any other symlinks in the first level directory. For any of these links which pointed to directories which didn''t exist, it produced an error and the directory wasn''t available. strange.. but how to fix.. here''s what I did. if I then unmount /home rm -f /gluster/home/*/access-log mkdir (link destination for access-log) mount /home there were a couple directories still inaccessible, it turned out they had bad symlinks as well. did the same thing and now it''s plugging along using the find /home/XXXX -type f -print0 | xargs -0 head -c1 > /dev/null to auto-heal them and it seems to be going just fine. once that''s finished, I''ll re-add serverB to server1''s AFR config and I presume it''ll be fine. I just found it odd that these missing symlink destinations would cause such a problem. anyway, it was a minor irritation and overall the auto-healing once going has been a lifesaver. Keith At 06:21 AM 10/9/2008, Keith Freedman wrote:>I have 2 servers that AFR eachother. > >one of them suffered a drive failure and is being rebuilt. > >the question is. What will happen if I just mount the empty drive >back as the AFR node. > >will it just start grabbing the data from the other server (which is >exactly what I want), OR >will it start deleting the data from the other server (which is terribly bad). > >Another thought was to, on the current working server with good data, >disable the remote afr node from it''s config (so it''s only using AFR >on itself), and then leave the other machines config as is, and turn it on. >This way I can be sure that the node with the data wont go nuts and >start deleting, but that updates to it will get replicated to the >other machine. > >this particular set is running 1.4pre5 if that changes the answer. > >Thanks, >Keith > > >_______________________________________________ >Gluster-users mailing list >Gluster-users at gluster.org >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
no one answered me so I''ll just report my findings: Server1 and ServerB full of data.. Gluster 1.4pre5. FedoraCore9 Disk on serverb crashes, loose everything. re-install copied over my AFR config from server1 change IP addresses as appropriate. the OS filesystems are in /gluster/home mounted gluster filesystem is /home on server1, I changed the AFR config to only list itself (because I wasn''t sure if this would work and I didn''t want afr to go delete everything instead of copy it over). so I mount /home on serverB while tailing the gluster log. the interesting thing, it created the directories in /home and many produced i/o errors the log showed many entries such as this: 2008-10-10 04:45:49 C [posix.c:2756:ensure_file_type] home1: entry /gluster/home/freeform/access-logs is a different type of file than expected (one for each directory under /home) so, that access-logs file is a symlink to somewhere else. there were also entries for any other symlinks in the first level directory. For any of these links which pointed to directories which didn''t exist, it produced an error and the directory wasn''t available. strange.. but how to fix.. here''s what I did. if I then unmount /home rm -f /gluster/home/*/access-log mkdir (link destination for access-log) mount /home there were a couple directories still inaccessible, it turned out they had bad symlinks as well. did the same thing and now it''s plugging along using the find /home/XXXX -type f -print0 | xargs -0 head -c1 > /dev/null to auto-heal them and it seems to be going just fine. once that''s finished, I''ll re-add serverB to server1''s AFR config and I presume it''ll be fine. I just found it odd that these missing symlink destinations would cause such a problem. anyway, it was a minor irritation and overall the auto-healing once going has been a lifesaver. Keith At 06:21 AM 10/9/2008, Keith Freedman wrote:>I have 2 servers that AFR eachother. > >one of them suffered a drive failure and is being rebuilt. > >the question is. What will happen if I just mount the empty drive >back as the AFR node. > >will it just start grabbing the data from the other server (which is >exactly what I want), OR >will it start deleting the data from the other server (which is terribly bad). > >Another thought was to, on the current working server with good data, >disable the remote afr node from it''s config (so it''s only using AFR >on itself), and then leave the other machines config as is, and turn it on. >This way I can be sure that the node with the data wont go nuts and >start deleting, but that updates to it will get replicated to the >other machine. > >this particular set is running 1.4pre5 if that changes the answer. > >Thanks, >Keith > > >_______________________________________________ >Gluster-users mailing list >Gluster-users at gluster.org >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
just a quick status for anyone who cares. once the find seemed to work properly for a few directories I remounted both servers with the proper AFR confi and they seem to be working just fine. auto-healing serverB as appropriate and sever1 getting updated when a file on serverb is updated. :) At 03:22 AM 10/10/2008, Keith Freedman wrote:>no one answered me so I''ll just report my findings: > >Server1 and ServerB full of data.. Gluster 1.4pre5. FedoraCore9 >Disk on serverb crashes, loose everything. re-install >copied over my AFR config from server1 change IP addresses as appropriate. > >using the find /home/XXXX -type f -print0 | xargs -0 head -c1 > /dev/null >to auto-heal them and it seems to be going just fine. > >once that''s finished, I''ll re-add serverB to server1''s AFR config and >I presume it''ll be fine. > >anyway, it was a minor irritation and overall the auto-healing once >going has been a lifesaver. > >Keith > >At 06:21 AM 10/9/2008, Keith Freedman wrote: > >I have 2 servers that AFR eachother. > > > >one of them suffered a drive failure and is being rebuilt. > > > >the question is. What will happen if I just mount the empty drive > >back as the AFR node. > > > >will it just start grabbing the data from the other server (which is > >exactly what I want), OR > >will it start deleting the data from the other server (which is > terribly bad). > > > >Another thought was to, on the current working server with good data, > >disable the remote afr node from it''s config (so it''s only using AFR > >on itself), and then leave the other machines config as is, and turn it on. > >This way I can be sure that the node with the data wont go nuts and > >start deleting, but that updates to it will get replicated to the > >other machine. > > > >this particular set is running 1.4pre5 if that changes the answer. > > > >Thanks, > >Keith > > > > > >_______________________________________________ > >Gluster-users mailing list > >Gluster-users at gluster.org > >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > > >_______________________________________________ >Gluster-users mailing list >Gluster-users at gluster.org >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
just a quick status for anyone who cares. once the find seemed to work properly for a few directories I remounted both servers with the proper AFR confi and they seem to be working just fine. auto-healing serverB as appropriate and sever1 getting updated when a file on serverb is updated. :) At 03:22 AM 10/10/2008, Keith Freedman wrote:>no one answered me so I''ll just report my findings: > >Server1 and ServerB full of data.. Gluster 1.4pre5. FedoraCore9 >Disk on serverb crashes, loose everything. re-install >copied over my AFR config from server1 change IP addresses as appropriate. > >using the find /home/XXXX -type f -print0 | xargs -0 head -c1 > /dev/null >to auto-heal them and it seems to be going just fine. > >once that''s finished, I''ll re-add serverB to server1''s AFR config and >I presume it''ll be fine. > >anyway, it was a minor irritation and overall the auto-healing once >going has been a lifesaver. > >Keith > >At 06:21 AM 10/9/2008, Keith Freedman wrote: > >I have 2 servers that AFR eachother. > > > >one of them suffered a drive failure and is being rebuilt. > > > >the question is. What will happen if I just mount the empty drive > >back as the AFR node. > > > >will it just start grabbing the data from the other server (which is > >exactly what I want), OR > >will it start deleting the data from the other server (which is > terribly bad). > > > >Another thought was to, on the current working server with good data, > >disable the remote afr node from it''s config (so it''s only using AFR > >on itself), and then leave the other machines config as is, and turn it on. > >This way I can be sure that the node with the data wont go nuts and > >start deleting, but that updates to it will get replicated to the > >other machine. > > > >this particular set is running 1.4pre5 if that changes the answer. > > > >Thanks, > >Keith > > > > > >_______________________________________________ > >Gluster-users mailing list > >Gluster-users at gluster.org > >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > > >_______________________________________________ >Gluster-users mailing list >Gluster-users at gluster.org >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
just a quick status for anyone who cares. once the find seemed to work properly for a few directories I remounted both servers with the proper AFR confi and they seem to be working just fine. auto-healing serverB as appropriate and sever1 getting updated when a file on serverb is updated. :) At 03:22 AM 10/10/2008, Keith Freedman wrote:>no one answered me so I''ll just report my findings: > >Server1 and ServerB full of data.. Gluster 1.4pre5. FedoraCore9 >Disk on serverb crashes, loose everything. re-install >copied over my AFR config from server1 change IP addresses as appropriate. > >using the find /home/XXXX -type f -print0 | xargs -0 head -c1 > /dev/null >to auto-heal them and it seems to be going just fine. > >once that''s finished, I''ll re-add serverB to server1''s AFR config and >I presume it''ll be fine. > >anyway, it was a minor irritation and overall the auto-healing once >going has been a lifesaver. > >Keith > >At 06:21 AM 10/9/2008, Keith Freedman wrote: > >I have 2 servers that AFR eachother. > > > >one of them suffered a drive failure and is being rebuilt. > > > >the question is. What will happen if I just mount the empty drive > >back as the AFR node. > > > >will it just start grabbing the data from the other server (which is > >exactly what I want), OR > >will it start deleting the data from the other server (which is > terribly bad). > > > >Another thought was to, on the current working server with good data, > >disable the remote afr node from it''s config (so it''s only using AFR > >on itself), and then leave the other machines config as is, and turn it on. > >This way I can be sure that the node with the data wont go nuts and > >start deleting, but that updates to it will get replicated to the > >other machine. > > > >this particular set is running 1.4pre5 if that changes the answer. > > > >Thanks, > >Keith > > > > > >_______________________________________________ > >Gluster-users mailing list > >Gluster-users at gluster.org > >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > > >_______________________________________________ >Gluster-users mailing list >Gluster-users at gluster.org >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
Keith, It was difficult to follow your mail. What is your setup? I understand that when destination of symlinks don't exist it causes problems during selfheal, is that right? Do you have server side afr or client side? Krishna On Fri, Oct 10, 2008 at 4:16 PM, Keith Freedman <freedman at freeformit.com> wrote:> just a quick status for anyone who cares. > > once the find seemed to work properly for a few directories I > remounted both servers with the proper AFR confi and they seem to be > working just fine. auto-healing serverB as appropriate and sever1 > getting updated when a file on serverb is updated. > > :) > > > At 03:22 AM 10/10/2008, Keith Freedman wrote: >>no one answered me so I'll just report my findings: >> >>Server1 and ServerB full of data.. Gluster 1.4pre5. FedoraCore9 >>Disk on serverb crashes, loose everything. re-install >>copied over my AFR config from server1 change IP addresses as appropriate. >> >>using the find /home/XXXX -type f -print0 | xargs -0 head -c1 > /dev/null >>to auto-heal them and it seems to be going just fine. >> >>once that's finished, I'll re-add serverB to server1's AFR config and >>I presume it'll be fine. >> >>anyway, it was a minor irritation and overall the auto-healing once >>going has been a lifesaver. >> >>Keith >> >>At 06:21 AM 10/9/2008, Keith Freedman wrote: >> >I have 2 servers that AFR eachother. >> > >> >one of them suffered a drive failure and is being rebuilt. >> > >> >the question is. What will happen if I just mount the empty drive >> >back as the AFR node. >> > >> >will it just start grabbing the data from the other server (which is >> >exactly what I want), OR >> >will it start deleting the data from the other server (which is >> terribly bad). >> > >> >Another thought was to, on the current working server with good data, >> >disable the remote afr node from it's config (so it's only using AFR >> >on itself), and then leave the other machines config as is, and turn it on. >> >This way I can be sure that the node with the data wont go nuts and >> >start deleting, but that updates to it will get replicated to the >> >other machine. >> > >> >this particular set is running 1.4pre5 if that changes the answer. >> > >> >Thanks, >> >Keith >> > >> > >> >_______________________________________________ >> >Gluster-users mailing list >> >Gluster-users at gluster.org >> >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users >> >> >>_______________________________________________ >>Gluster-users mailing list >>Gluster-users at gluster.org >>http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users >
2 servers which AFR eachother as server & client. I was pretty happy with it, I think if it could just not care about those silly simlink issues it''d be great. nice to know I can just add a brick and it''ll just work. I didn''t mind the bit of manual intervention since it was minimal and ultimate got the job done. but here''s my config to help you understand. the configs on both servers are identical. volume home1 type storage/posix # POSIX FS translator option directory /gluster/home # Export this directory end-volume volume posix-locks-home1 type features/posix-locks option mandatory on subvolumes home1 end-volume ## Reference volume "home2" from remote server volume home2 type protocol/client # POSIX FS translator option transport-type tcp/client option remote-host ##.##.##.## # IP address of remote host option remote-subvolume posix-locks-home1 # use home1 on remote host option transport-timeout 10 # value in seconds; it should be set rel atively low end-volume volume server type protocol/server option transport-type tcp/server # For TCP/IP transport subvolumes posix-locks-home1 option auth.addr.posix-locks-home1.allow ##.##.##.##,127.0.0.1 # Allow access to "home1" volume end-volume ### Create automatic file replication volume home type cluster/afr option read-subvolume posix-locks-home1 subvolumes posix-locks-home1 home2 # subvolumes posix-locks-home1 end-volume At 09:19 PM 10/10/2008, Krishna Srinivas wrote:>Keith, > >It was difficult to follow your mail. What is your setup? I understand >that when destination of symlinks don''t exist it causes problems >during selfheal, is that right? Do you have server side afr or client >side? > >Krishna > >On Fri, Oct 10, 2008 at 4:16 PM, Keith Freedman ><freedman at freeformit.com> wrote: > > just a quick status for anyone who cares. > > > > once the find seemed to work properly for a few directories I > > remounted both servers with the proper AFR confi and they seem to be > > working just fine. auto-healing serverB as appropriate and sever1 > > getting updated when a file on serverb is updated. > > > > :) > > > > > > At 03:22 AM 10/10/2008, Keith Freedman wrote: > >>no one answered me so I''ll just report my findings: > >> > >>Server1 and ServerB full of data.. Gluster 1.4pre5. FedoraCore9 > >>Disk on serverb crashes, loose everything. re-install > >>copied over my AFR config from server1 change IP addresses as appropriate. > >> > >>using the find /home/XXXX -type f -print0 | xargs -0 head -c1 > /dev/null > >>to auto-heal them and it seems to be going just fine. > >> > >>once that''s finished, I''ll re-add serverB to server1''s AFR config and > >>I presume it''ll be fine. > >> > >>anyway, it was a minor irritation and overall the auto-healing once > >>going has been a lifesaver. > >> > >>Keith > >> > >>At 06:21 AM 10/9/2008, Keith Freedman wrote: > >> >I have 2 servers that AFR eachother. > >> > > >> >one of them suffered a drive failure and is being rebuilt. > >> > > >> >the question is. What will happen if I just mount the empty drive > >> >back as the AFR node. > >> > > >> >will it just start grabbing the data from the other server (which is > >> >exactly what I want), OR > >> >will it start deleting the data from the other server (which is > >> terribly bad). > >> > > >> >Another thought was to, on the current working server with good data, > >> >disable the remote afr node from it''s config (so it''s only using AFR > >> >on itself), and then leave the other machines config as is, and > turn it on. > >> >This way I can be sure that the node with the data wont go nuts and > >> >start deleting, but that updates to it will get replicated to the > >> >other machine. > >> > > >> >this particular set is running 1.4pre5 if that changes the answer. > >> > > >> >Thanks, > >> >Keith > >> > > >> > > >> >_______________________________________________ > >> >Gluster-users mailing list > >> >Gluster-users at gluster.org > >> >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > >> > >> > >>_______________________________________________ > >>Gluster-users mailing list > >>Gluster-users at gluster.org > >>http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > > > > > > _______________________________________________ > > Gluster-users mailing list > > Gluster-users at gluster.org > > http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > >
2 servers which AFR eachother as server & client. I was pretty happy with it, I think if it could just not care about those silly simlink issues it''d be great. nice to know I can just add a brick and it''ll just work. I didn''t mind the bit of manual intervention since it was minimal and ultimate got the job done. but here''s my config to help you understand. the configs on both servers are identical. volume home1 type storage/posix # POSIX FS translator option directory /gluster/home # Export this directory end-volume volume posix-locks-home1 type features/posix-locks option mandatory on subvolumes home1 end-volume ## Reference volume "home2" from remote server volume home2 type protocol/client # POSIX FS translator option transport-type tcp/client option remote-host ##.##.##.## # IP address of remote host option remote-subvolume posix-locks-home1 # use home1 on remote host option transport-timeout 10 # value in seconds; it should be set rel atively low end-volume volume server type protocol/server option transport-type tcp/server # For TCP/IP transport subvolumes posix-locks-home1 option auth.addr.posix-locks-home1.allow ##.##.##.##,127.0.0.1 # Allow access to "home1" volume end-volume ### Create automatic file replication volume home type cluster/afr option read-subvolume posix-locks-home1 subvolumes posix-locks-home1 home2 # subvolumes posix-locks-home1 end-volume At 09:19 PM 10/10/2008, Krishna Srinivas wrote:>Keith, > >It was difficult to follow your mail. What is your setup? I understand >that when destination of symlinks don''t exist it causes problems >during selfheal, is that right? Do you have server side afr or client >side? > >Krishna > >On Fri, Oct 10, 2008 at 4:16 PM, Keith Freedman ><freedman at freeformit.com> wrote: > > just a quick status for anyone who cares. > > > > once the find seemed to work properly for a few directories I > > remounted both servers with the proper AFR confi and they seem to be > > working just fine. auto-healing serverB as appropriate and sever1 > > getting updated when a file on serverb is updated. > > > > :) > > > > > > At 03:22 AM 10/10/2008, Keith Freedman wrote: > >>no one answered me so I''ll just report my findings: > >> > >>Server1 and ServerB full of data.. Gluster 1.4pre5. FedoraCore9 > >>Disk on serverb crashes, loose everything. re-install > >>copied over my AFR config from server1 change IP addresses as appropriate. > >> > >>using the find /home/XXXX -type f -print0 | xargs -0 head -c1 > /dev/null > >>to auto-heal them and it seems to be going just fine. > >> > >>once that''s finished, I''ll re-add serverB to server1''s AFR config and > >>I presume it''ll be fine. > >> > >>anyway, it was a minor irritation and overall the auto-healing once > >>going has been a lifesaver. > >> > >>Keith > >> > >>At 06:21 AM 10/9/2008, Keith Freedman wrote: > >> >I have 2 servers that AFR eachother. > >> > > >> >one of them suffered a drive failure and is being rebuilt. > >> > > >> >the question is. What will happen if I just mount the empty drive > >> >back as the AFR node. > >> > > >> >will it just start grabbing the data from the other server (which is > >> >exactly what I want), OR > >> >will it start deleting the data from the other server (which is > >> terribly bad). > >> > > >> >Another thought was to, on the current working server with good data, > >> >disable the remote afr node from it''s config (so it''s only using AFR > >> >on itself), and then leave the other machines config as is, and > turn it on. > >> >This way I can be sure that the node with the data wont go nuts and > >> >start deleting, but that updates to it will get replicated to the > >> >other machine. > >> > > >> >this particular set is running 1.4pre5 if that changes the answer. > >> > > >> >Thanks, > >> >Keith > >> > > >> > > >> >_______________________________________________ > >> >Gluster-users mailing list > >> >Gluster-users at gluster.org > >> >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > >> > >> > >>_______________________________________________ > >>Gluster-users mailing list > >>Gluster-users at gluster.org > >>http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > > > > > > _______________________________________________ > > Gluster-users mailing list > > Gluster-users at gluster.org > > http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > >
2 servers which AFR eachother as server & client. I was pretty happy with it, I think if it could just not care about those silly simlink issues it''d be great. nice to know I can just add a brick and it''ll just work. I didn''t mind the bit of manual intervention since it was minimal and ultimate got the job done. but here''s my config to help you understand. the configs on both servers are identical. volume home1 type storage/posix # POSIX FS translator option directory /gluster/home # Export this directory end-volume volume posix-locks-home1 type features/posix-locks option mandatory on subvolumes home1 end-volume ## Reference volume "home2" from remote server volume home2 type protocol/client # POSIX FS translator option transport-type tcp/client option remote-host ##.##.##.## # IP address of remote host option remote-subvolume posix-locks-home1 # use home1 on remote host option transport-timeout 10 # value in seconds; it should be set rel atively low end-volume volume server type protocol/server option transport-type tcp/server # For TCP/IP transport subvolumes posix-locks-home1 option auth.addr.posix-locks-home1.allow ##.##.##.##,127.0.0.1 # Allow access to "home1" volume end-volume ### Create automatic file replication volume home type cluster/afr option read-subvolume posix-locks-home1 subvolumes posix-locks-home1 home2 # subvolumes posix-locks-home1 end-volume At 09:19 PM 10/10/2008, Krishna Srinivas wrote:>Keith, > >It was difficult to follow your mail. What is your setup? I understand >that when destination of symlinks don''t exist it causes problems >during selfheal, is that right? Do you have server side afr or client >side? > >Krishna > >On Fri, Oct 10, 2008 at 4:16 PM, Keith Freedman ><freedman at freeformit.com> wrote: > > just a quick status for anyone who cares. > > > > once the find seemed to work properly for a few directories I > > remounted both servers with the proper AFR confi and they seem to be > > working just fine. auto-healing serverB as appropriate and sever1 > > getting updated when a file on serverb is updated. > > > > :) > > > > > > At 03:22 AM 10/10/2008, Keith Freedman wrote: > >>no one answered me so I''ll just report my findings: > >> > >>Server1 and ServerB full of data.. Gluster 1.4pre5. FedoraCore9 > >>Disk on serverb crashes, loose everything. re-install > >>copied over my AFR config from server1 change IP addresses as appropriate. > >> > >>using the find /home/XXXX -type f -print0 | xargs -0 head -c1 > /dev/null > >>to auto-heal them and it seems to be going just fine. > >> > >>once that''s finished, I''ll re-add serverB to server1''s AFR config and > >>I presume it''ll be fine. > >> > >>anyway, it was a minor irritation and overall the auto-healing once > >>going has been a lifesaver. > >> > >>Keith > >> > >>At 06:21 AM 10/9/2008, Keith Freedman wrote: > >> >I have 2 servers that AFR eachother. > >> > > >> >one of them suffered a drive failure and is being rebuilt. > >> > > >> >the question is. What will happen if I just mount the empty drive > >> >back as the AFR node. > >> > > >> >will it just start grabbing the data from the other server (which is > >> >exactly what I want), OR > >> >will it start deleting the data from the other server (which is > >> terribly bad). > >> > > >> >Another thought was to, on the current working server with good data, > >> >disable the remote afr node from it''s config (so it''s only using AFR > >> >on itself), and then leave the other machines config as is, and > turn it on. > >> >This way I can be sure that the node with the data wont go nuts and > >> >start deleting, but that updates to it will get replicated to the > >> >other machine. > >> > > >> >this particular set is running 1.4pre5 if that changes the answer. > >> > > >> >Thanks, > >> >Keith > >> > > >> > > >> >_______________________________________________ > >> >Gluster-users mailing list > >> >Gluster-users at gluster.org > >> >http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > >> > >> > >>_______________________________________________ > >>Gluster-users mailing list > >>Gluster-users at gluster.org > >>http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > > > > > > _______________________________________________ > > Gluster-users mailing list > > Gluster-users at gluster.org > > http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > >