We are getting the following messages. Jan 6 10:19:25 houided006 kernel: ocfs: Adding houided005 (node 0) to clustered device (8,208) Jan 6 10:19:52 houided006 kernel: ocfs: Removing houided005 (node 0) from clustered device (8,208) every few minutes. Any advice on how to troubleshoot this? David. __________________________________ Do you Yahoo!? Read only the mail you want - Yahoo! Mail SpamGuard. http://promotions.yahoo.com/new_mail
how many ocfs mounts do you have on that box? which kernel? ocfs version? David McWhinnie wrote:>We are getting the following messages. >Jan 6 10:19:25 houided006 kernel: ocfs: Adding >houided005 (node 0) to clustered device (8,208) >Jan 6 10:19:52 houided006 kernel: ocfs: Removing >houided005 (node 0) from clustered device (8,208) > >every few minutes. > >Any advice on how to troubleshoot this? > >David. > > > >__________________________________ >Do you Yahoo!? >Read only the mail you want - Yahoo! Mail SpamGuard. >http://promotions.yahoo.com/new_mail >_______________________________________________ >Ocfs-users mailing list >Ocfs-users@oss.oracle.com >http://oss.oracle.com/mailman/listinfo/ocfs-users > >
We have 16 OCFS mounts. We are running RedHat 2.1 OCFS version 1. Latest patch. Turns out a dd was being done without the o_direct option. So that could be the cause. Interesting this ls, df etc would all hang while the node was removed from the cluster. David --- Sunil Mushran <Sunil.Mushran@oracle.com> wrote:> how many ocfs mounts do you have on that box? > which kernel? > ocfs version? > > David McWhinnie wrote: > > >We are getting the following messages. > >Jan 6 10:19:25 houided006 kernel: ocfs: Adding > >houided005 (node 0) to clustered device (8,208) > >Jan 6 10:19:52 houided006 kernel: ocfs: Removing > >houided005 (node 0) from clustered device (8,208) > > > >every few minutes. > > > >Any advice on how to troubleshoot this? > > > >David. > > > > > > > >__________________________________ > >Do you Yahoo!? > >Read only the mail you want - Yahoo! Mail > SpamGuard. > >http://promotions.yahoo.com/new_mail > >_______________________________________________ > >Ocfs-users mailing list > >Ocfs-users@oss.oracle.com > >http://oss.oracle.com/mailman/listinfo/ocfs-users > > > > >__________________________________ Do you Yahoo!? Yahoo! Mail - Easier than ever with enhanced search. Learn more. http://info.mail.yahoo.com/mail_250
No, dd is mostly unrelated. What those messages indicate that the heartbeat thread is not keeping up, which is not good. Yes, all the processes requiring a cluster lock will hang as a node is being evicted from the cluster. That you are getting these errors with as low as 16 mounts is puzzling. When do these errors messages pop up? Any relationship with the load on the system? David McWhinnie wrote:>We have 16 OCFS mounts. >We are running RedHat 2.1 >OCFS version 1. Latest patch. > >Turns out a dd was being done without the o_direct >option. So that could be the cause. Interesting this >ls, df etc would all hang while the node was removed >from the cluster. > >David >--- Sunil Mushran <Sunil.Mushran@oracle.com> wrote: > > > >>how many ocfs mounts do you have on that box? >>which kernel? >>ocfs version? >> >>David McWhinnie wrote: >> >> >> >>>We are getting the following messages. >>>Jan 6 10:19:25 houided006 kernel: ocfs: Adding >>>houided005 (node 0) to clustered device (8,208) >>>Jan 6 10:19:52 houided006 kernel: ocfs: Removing >>>houided005 (node 0) from clustered device (8,208) >>> >>>every few minutes. >>> >>>Any advice on how to troubleshoot this? >>> >>>David. >>> >>> >>> >>>__________________________________ >>>Do you Yahoo!? >>>Read only the mail you want - Yahoo! Mail >>> >>> >>SpamGuard. >> >> >>>http://promotions.yahoo.com/new_mail >>>_______________________________________________ >>>Ocfs-users mailing list >>>Ocfs-users@oss.oracle.com >>>http://oss.oracle.com/mailman/listinfo/ocfs-users >>> >>> >>> >>> > > > > >__________________________________ >Do you Yahoo!? >Yahoo! Mail - Easier than ever with enhanced search. Learn more. >http://info.mail.yahoo.com/mail_250 > >