Hi, I had a 4 nodes in replication setup. After removing one of the nodes with: gluster> volume remove-brick images_1 replica 3 vmhost5:/exports/1 Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y Remove Brick commit force successful I started to see a high cpu usage on 2 of the remaining boxes, and this error in glustershd.log: [2013-04-18 12:40:47.160559] E [afr-self-heald.c:685:_link_inode_update_loc] 0-images_1-replicate-0: inode link failed on the inode (00000000-0000-0000-0000-000000000000) [2013-04-18 12:41:55.784510] I [afr-self-heald.c:1082:afr_dir_exclusive_crawl] 0-images_1-replicate-0: Another crawl is in progress for images_1-client-1 any ideas? -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130418/f23b2794/attachment.html>