New MDS. Kernel 2.6.18-92.1.17.el5_lustre.1.8.0smp on a base RH5.3 system. Nehalem CPU. I''m stuck. Did I miss some step here? The filesystem is about 300G. Thanks, Bill ====================================================== I make the filesystem: [root at mds ~]# mkfs.lustre --fsname=lustre --mgs --mdt /dev/sdb1 Permanent disk data: Target: lustre-MDTffff Index: unassigned Lustre FS: lustre Mount type: ldiskfs Flags: 0x75 (MDT MGS needs_index first_time update ) Persistent mount opts: errors=remount-ro,iopen_nopriv,user_xattr Parameters: mdt.group_upcall=/usr/sbin/l_getgroups checking for existing Lustre data: not found device size = 285561MB 2 6 18 formatting backing filesystem ldiskfs on /dev/sdb1 target name lustre-MDTffff 4k blocks 0 options -J size=400 -i 4096 -I 512 -q -O dir_index,uninit_groups -F mkfs_cmd = mkfs.ext2 -j -b 4096 -L lustre-MDTffff -J size=400 -i 4096 -I 512 -q -O dir_index,uninit_groups -F /dev/sdb1 Writing CONFIGS/mountdata ==================================================Then try to mount it: [root at mds ~]# mount -t lustre /dev/sdb1 /MDT mount.lustre: mount /dev/sdb1 at /MDT failed: Operation not supported =================================================== The errors in /var/log/messages are: Jun 19 14:01:25 mds kernel: kjournald starting. Commit interval 5 seconds Jun 19 14:01:25 mds kernel: LDISKFS FS on sdb1, internal journal Jun 19 14:01:25 mds kernel: LDISKFS-fs: mounted filesystem with ordered data mode. Jun 19 14:01:25 mds kernel: kjournald starting. Commit interval 5 seconds Jun 19 14:01:25 mds kernel: LDISKFS FS on sdb1, internal journal Jun 19 14:01:25 mds kernel: LDISKFS-fs: mounted filesystem with ordered data mode. Jun 19 14:01:25 mds kernel: Lustre: MGS MGS started Jun 19 14:01:25 mds kernel: Lustre: MGC172.23.10.4 at tcp: Reactivating import Jun 19 14:01:25 mds kernel: Lustre: Setting parameter lustre-MDT0000.mdt.group_upcall in log lustre-MDT0000 Jun 19 14:01:25 mds kernel: Lustre: Enabling user_xattr Jun 19 14:01:25 mds kernel: Lustre: lustre-MDT0000: new disk, initializing Jun 19 14:01:25 mds kernel: Lustre: MDT lustre-MDT0000 now serving lustre-MDT0000_UUID (lustre-MDT0000/dc93cffc-68e5-8351-46eb-52215bd7a771) with recovery enabled Jun 19 14:01:25 mds kernel: Lustre: 3177:0:(lproc_mds.c:271:lprocfs_wr_group_upcall()) lustre-MDT0000: group upcall set to /usr/sbin/l_getgroups Jun 19 14:01:25 mds kernel: Lustre: lustre-MDT0000.mdt: set parameter group_upcall=/usr/sbin/l_getgroups Jun 19 14:01:25 mds kernel: Lustre: Server lustre-MDT0000 on device /dev/sdb1 has started Jun 19 14:01:25 mds kernel: SELinux: (dev lustre, type lustre) has no xattr support Jun 19 14:01:25 mds kernel: Lustre: Failing over lustre-MDT0000 Jun 19 14:01:25 mds kernel: Lustre: Skipped 1 previous similar message Jun 19 14:01:25 mds kernel: Lustre: *** setting obd lustre-MDT0000 device ''sdb1'' read-only *** Jun 19 14:01:25 mds kernel: Turning device sdb (0x800011) read-only Jun 19 14:01:25 mds kernel: Lustre: Failing over lustre-mdtlov Jun 19 14:01:25 mds kernel: Lustre: lustre-MDT0000: shutting down for failover; client state will be preserved. Jun 19 14:01:25 mds kernel: Lustre: MDT lustre-MDT0000 has stopped. Jun 19 14:01:25 mds kernel: LustreError: 3072:0:(ldlm_request.c:1043:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway Jun 19 14:01:25 mds kernel: LustreError: 3072:0:(ldlm_request.c:1632:ldlm_cli_cancel_list()) ldlm_cli_cancel_list: -108 Jun 19 14:01:25 mds kernel: Lustre: MGS has stopped. Jun 19 14:01:25 mds kernel: Removing read-only on unknown block (0x800011) Jun 19 14:01:25 mds kernel: Lustre: server umount lustre-MDT0000 complete [root at mds ~]# lctl network up LNET configured [root at mds ~]# lctl list_nids 172.23.10.4 at tcp
Just ran into that myself this morning - in my case, it was dealt with by disabling SELinux (setting it to disabled in /etc/sysconfig/selinux). I''ve only been doing testing on non-production systems though, so I didn''t look into what specifically SELinux was throwing a fit about/modification to SELinux configuration to fix it. -- -------------------------------------------------------------------------------------------------- Gary Gogick senior systems administrator | workhabit,inc. On Fri, Jun 19, 2009 at 3:06 PM, Bill Wichser <bill at princeton.edu> wrote:> New MDS. Kernel 2.6.18-92.1.17.el5_lustre.1.8.0smp on a base RH5.3 > system. Nehalem CPU. > > I''m stuck. Did I miss some step here? The filesystem is about 300G. > > Thanks, > Bill > > ======================================================> > I make the filesystem: > > [root at mds ~]# mkfs.lustre --fsname=lustre --mgs --mdt /dev/sdb1 > > Permanent disk data: > Target: lustre-MDTffff > Index: unassigned > Lustre FS: lustre > Mount type: ldiskfs > Flags: 0x75 > (MDT MGS needs_index first_time update ) > Persistent mount opts: errors=remount-ro,iopen_nopriv,user_xattr > Parameters: mdt.group_upcall=/usr/sbin/l_getgroups > > checking for existing Lustre data: not found > device size = 285561MB > 2 6 18 > formatting backing filesystem ldiskfs on /dev/sdb1 > target name lustre-MDTffff > 4k blocks 0 > options -J size=400 -i 4096 -I 512 -q -O > dir_index,uninit_groups -F > mkfs_cmd = mkfs.ext2 -j -b 4096 -L lustre-MDTffff -J size=400 -i 4096 > -I 512 -q -O dir_index,uninit_groups -F /dev/sdb1 > Writing CONFIGS/mountdata > > ==================================================> Then try to mount it: > > [root at mds ~]# mount -t lustre /dev/sdb1 /MDT > mount.lustre: mount /dev/sdb1 at /MDT failed: Operation not supported > > ===================================================> > The errors in /var/log/messages are: > Jun 19 14:01:25 mds kernel: kjournald starting. Commit interval 5 seconds > Jun 19 14:01:25 mds kernel: LDISKFS FS on sdb1, internal journal > Jun 19 14:01:25 mds kernel: LDISKFS-fs: mounted filesystem with ordered > data mode. > Jun 19 14:01:25 mds kernel: kjournald starting. Commit interval 5 seconds > Jun 19 14:01:25 mds kernel: LDISKFS FS on sdb1, internal journal > Jun 19 14:01:25 mds kernel: LDISKFS-fs: mounted filesystem with ordered > data mode. > Jun 19 14:01:25 mds kernel: Lustre: MGS MGS started > Jun 19 14:01:25 mds kernel: Lustre: MGC172.23.10.4 at tcp: Reactivating > import > Jun 19 14:01:25 mds kernel: Lustre: Setting parameter > lustre-MDT0000.mdt.group_upcall in log lustre-MDT0000 > Jun 19 14:01:25 mds kernel: Lustre: Enabling user_xattr > Jun 19 14:01:25 mds kernel: Lustre: lustre-MDT0000: new disk, initializing > Jun 19 14:01:25 mds kernel: Lustre: MDT lustre-MDT0000 now serving > lustre-MDT0000_UUID > (lustre-MDT0000/dc93cffc-68e5-8351-46eb-52215bd7a771) with recovery enabled > Jun 19 14:01:25 mds kernel: Lustre: > 3177:0:(lproc_mds.c:271:lprocfs_wr_group_upcall()) lustre-MDT0000: group > upcall set to /usr/sbin/l_getgroups > Jun 19 14:01:25 mds kernel: Lustre: lustre-MDT0000.mdt: set parameter > group_upcall=/usr/sbin/l_getgroups > Jun 19 14:01:25 mds kernel: Lustre: Server lustre-MDT0000 on device > /dev/sdb1 has started > Jun 19 14:01:25 mds kernel: SELinux: (dev lustre, type lustre) has no > xattr support > Jun 19 14:01:25 mds kernel: Lustre: Failing over lustre-MDT0000 > Jun 19 14:01:25 mds kernel: Lustre: Skipped 1 previous similar message > Jun 19 14:01:25 mds kernel: Lustre: *** setting obd lustre-MDT0000 > device ''sdb1'' read-only *** > Jun 19 14:01:25 mds kernel: Turning device sdb (0x800011) read-only > Jun 19 14:01:25 mds kernel: Lustre: Failing over lustre-mdtlov > Jun 19 14:01:25 mds kernel: Lustre: lustre-MDT0000: shutting down for > failover; client state will be preserved. > Jun 19 14:01:25 mds kernel: Lustre: MDT lustre-MDT0000 has stopped. > Jun 19 14:01:25 mds kernel: LustreError: > 3072:0:(ldlm_request.c:1043:ldlm_cli_cancel_req()) Got rc -108 from > cancel RPC: canceling anyway > Jun 19 14:01:25 mds kernel: LustreError: > 3072:0:(ldlm_request.c:1632:ldlm_cli_cancel_list()) > ldlm_cli_cancel_list: -108 > Jun 19 14:01:25 mds kernel: Lustre: MGS has stopped. > Jun 19 14:01:25 mds kernel: Removing read-only on unknown block (0x800011) > Jun 19 14:01:25 mds kernel: Lustre: server umount lustre-MDT0000 complete > > [root at mds ~]# lctl network up > LNET configured > [root at mds ~]# lctl list_nids > 172.23.10.4 at tcp > _______________________________________________ > Lustre-discuss mailing list > Lustre-discuss at lists.lustre.org > http://lists.lustre.org/mailman/listinfo/lustre-discuss >-------------- next part -------------- An HTML attachment was scrubbed... URL: http://lists.lustre.org/pipermail/lustre-discuss/attachments/20090619/5d4e4423/attachment.html
Even though it was set to permissive mode this was preventing the mount. Once I disabled it entirely, the mount proceeds fine. Thanks for finding this! Have a great weekend...Bill Gary Gogick wrote:> Just ran into that myself this morning - in my case, it was dealt with > by disabling SELinux (setting it to disabled in > /etc/sysconfig/selinux). I''ve only been doing testing on non-production > systems though, so I didn''t look into what specifically SELinux was > throwing a fit about/modification to SELinux configuration to fix it. > > -- > -------------------------------------------------------------------------------------------------- > Gary Gogick > senior systems administrator | workhabit,inc. > > > On Fri, Jun 19, 2009 at 3:06 PM, Bill Wichser <bill at princeton.edu > <mailto:bill at princeton.edu>> wrote: > > New MDS. Kernel 2.6.18-92.1.17.el5_lustre.1.8.0smp on a base RH5.3 > system. Nehalem CPU. > > I''m stuck. Did I miss some step here? The filesystem is about 300G. > > Thanks, > Bill > > ======================================================> > I make the filesystem: > > [root at mds ~]# mkfs.lustre --fsname=lustre --mgs --mdt /dev/sdb1 > > Permanent disk data: > Target: lustre-MDTffff > Index: unassigned > Lustre FS: lustre > Mount type: ldiskfs > Flags: 0x75 > (MDT MGS needs_index first_time update ) > Persistent mount opts: errors=remount-ro,iopen_nopriv,user_xattr > Parameters: mdt.group_upcall=/usr/sbin/l_getgroups > > checking for existing Lustre data: not found > device size = 285561MB > 2 6 18 > formatting backing filesystem ldiskfs on /dev/sdb1 > target name lustre-MDTffff > 4k blocks 0 > options -J size=400 -i 4096 -I 512 -q -O > dir_index,uninit_groups -F > mkfs_cmd = mkfs.ext2 -j -b 4096 -L lustre-MDTffff -J size=400 -i 4096 > -I 512 -q -O dir_index,uninit_groups -F /dev/sdb1 > Writing CONFIGS/mountdata > > ==================================================> Then try to mount it: > > [root at mds ~]# mount -t lustre /dev/sdb1 /MDT > mount.lustre: mount /dev/sdb1 at /MDT failed: Operation not supported > > ===================================================> > The errors in /var/log/messages are: > Jun 19 14:01:25 mds kernel: kjournald starting. Commit interval 5 > seconds > Jun 19 14:01:25 mds kernel: LDISKFS FS on sdb1, internal journal > Jun 19 14:01:25 mds kernel: LDISKFS-fs: mounted filesystem with ordered > data mode. > Jun 19 14:01:25 mds kernel: kjournald starting. Commit interval 5 > seconds > Jun 19 14:01:25 mds kernel: LDISKFS FS on sdb1, internal journal > Jun 19 14:01:25 mds kernel: LDISKFS-fs: mounted filesystem with ordered > data mode. > Jun 19 14:01:25 mds kernel: Lustre: MGS MGS started > Jun 19 14:01:25 mds kernel: Lustre: MGC172.23.10.4 at tcp: Reactivating > import > Jun 19 14:01:25 mds kernel: Lustre: Setting parameter > lustre-MDT0000.mdt.group_upcall in log lustre-MDT0000 > Jun 19 14:01:25 mds kernel: Lustre: Enabling user_xattr > Jun 19 14:01:25 mds kernel: Lustre: lustre-MDT0000: new disk, > initializing > Jun 19 14:01:25 mds kernel: Lustre: MDT lustre-MDT0000 now serving > lustre-MDT0000_UUID > (lustre-MDT0000/dc93cffc-68e5-8351-46eb-52215bd7a771) with recovery > enabled > Jun 19 14:01:25 mds kernel: Lustre: > 3177:0:(lproc_mds.c:271:lprocfs_wr_group_upcall()) lustre-MDT0000: group > upcall set to /usr/sbin/l_getgroups > Jun 19 14:01:25 mds kernel: Lustre: lustre-MDT0000.mdt: set parameter > group_upcall=/usr/sbin/l_getgroups > Jun 19 14:01:25 mds kernel: Lustre: Server lustre-MDT0000 on device > /dev/sdb1 has started > Jun 19 14:01:25 mds kernel: SELinux: (dev lustre, type lustre) has no > xattr support > Jun 19 14:01:25 mds kernel: Lustre: Failing over lustre-MDT0000 > Jun 19 14:01:25 mds kernel: Lustre: Skipped 1 previous similar message > Jun 19 14:01:25 mds kernel: Lustre: *** setting obd lustre-MDT0000 > device ''sdb1'' read-only *** > Jun 19 14:01:25 mds kernel: Turning device sdb (0x800011) read-only > Jun 19 14:01:25 mds kernel: Lustre: Failing over lustre-mdtlov > Jun 19 14:01:25 mds kernel: Lustre: lustre-MDT0000: shutting down for > failover; client state will be preserved. > Jun 19 14:01:25 mds kernel: Lustre: MDT lustre-MDT0000 has stopped. > Jun 19 14:01:25 mds kernel: LustreError: > 3072:0:(ldlm_request.c:1043:ldlm_cli_cancel_req()) Got rc -108 from > cancel RPC: canceling anyway > Jun 19 14:01:25 mds kernel: LustreError: > 3072:0:(ldlm_request.c:1632:ldlm_cli_cancel_list()) > ldlm_cli_cancel_list: -108 > Jun 19 14:01:25 mds kernel: Lustre: MGS has stopped. > Jun 19 14:01:25 mds kernel: Removing read-only on unknown block > (0x800011) > Jun 19 14:01:25 mds kernel: Lustre: server umount lustre-MDT0000 > complete > > [root at mds ~]# lctl network up > LNET configured > [root at mds ~]# lctl list_nids > 172.23.10.4 at tcp > _______________________________________________ > Lustre-discuss mailing list > Lustre-discuss at lists.lustre.org <mailto:Lustre-discuss at lists.lustre.org> > http://lists.lustre.org/mailman/listinfo/lustre-discuss > > >