Nirmal Seenu
2009-Aug-13 16:24 UTC
[Lustre-discuss] INFO: possible recursive locking detected
I am running lustre 1.8.0.1 on 2.6.22.19 kernel and I am getting the following Warning messages once on each lustre server when I mount the lustre partitions but everything seems to be working fine. Could someone please let me know if this a known problem that can be ignored. TIA Nirmal Aug 11 15:54:23 mdt1p kernel: Lustre: OBD class driver, http://www.lustre.org/ Aug 11 15:54:23 mdt1p kernel: Lustre: Lustre Version: 1.8.0.1 Aug 11 15:54:23 mdt1p kernel: Lustre: Build Version: 1.8.0.1-19691231180000-PRISTINE-.usr.src.linux-2.6.22.19-2.6.22.19 Aug 11 15:54:23 mdt1p kernel: Lustre: Added LNI 172.19.11.210 at tcp1 [8/256] Aug 11 15:54:23 mdt1p kernel: Lustre: Accept secure, port 988 Aug 11 15:54:23 mdt1p kernel: LustreError: 8541:0:(router_proc.c:1020:lnet_proc_init()) couldn''t create proc entry sys/lnet/stats Aug 11 15:54:24 mdt1p kernel: Lustre: Lustre Client File System; http://www.lustre.org/ Aug 11 15:54:24 mdt1p kernel: kjournald starting. Commit interval 5 seconds Aug 11 15:54:24 mdt1p kernel: LDISKFS FS on dm-1, internal journal Aug 11 15:54:24 mdt1p kernel: LDISKFS-fs: mounted filesystem with ordered data mode. Aug 11 15:54:24 mdt1p kernel: kjournald starting. Commit interval 5 seconds Aug 11 15:54:24 mdt1p kernel: LDISKFS FS on dm-1, internal journal Aug 11 15:54:24 mdt1p kernel: LDISKFS-fs: mounted filesystem with ordered data mode. Aug 11 15:54:25 mdt1p kernel: Lustre: MGS MGS started Aug 11 15:54:25 mdt1p kernel: Lustre: Server MGS on device /dev/mapper/mdt1_vol-mgs has started Aug 11 15:54:25 mdt1p kernel: Lustre: MGC172.19.11.210 at tcp1: Reactivating import Aug 11 15:54:28 mdt1p kernel: kjournald starting. Commit interval 5 seconds Aug 11 15:54:28 mdt1p kernel: LDISKFS FS on dm-2, internal journal Aug 11 15:54:28 mdt1p kernel: LDISKFS-fs: recovery complete. Aug 11 15:54:28 mdt1p kernel: LDISKFS-fs: mounted filesystem with ordered data mode. Aug 11 15:54:28 mdt1p kernel: kjournald starting. Commit interval 5 seconds Aug 11 15:54:28 mdt1p kernel: LDISKFS FS on dm-2, internal journal Aug 11 15:54:28 mdt1p kernel: LDISKFS-fs: mounted filesystem with ordered data mode. Aug 11 15:54:28 mdt1p kernel: Aug 11 15:54:28 mdt1p kernel: ============================================Aug 11 15:54:28 mdt1p kernel: [ INFO: possible recursive locking detected ] Aug 11 15:54:28 mdt1p kernel: 2.6.22.19 #1 Aug 11 15:54:28 mdt1p kernel: --------------------------------------------- Aug 11 15:54:28 mdt1p kernel: mount.lustre/8761 is trying to acquire lock: Aug 11 15:54:28 mdt1p kernel: (&inode->i_mutex){--..}, at: [<ffffffff80468a4b>] mutex_lock+0x25/0x29 Aug 11 15:54:28 mdt1p kernel: Aug 11 15:54:28 mdt1p kernel: but task is already holding lock: Aug 11 15:54:28 mdt1p kernel: (&inode->i_mutex){--..}, at: [<ffffffff80468a4b>] mutex_lock+0x25/0x29 Aug 11 15:54:28 mdt1p kernel: Aug 11 15:54:28 mdt1p kernel: other info that might help us debug this: Aug 11 15:54:28 mdt1p kernel: 2 locks held by mount.lustre/8761: Aug 11 15:54:28 mdt1p kernel: #0: (&type->s_umount_key#22){--..}, at: [<ffffffff8029c490>] sget+0x240/0x3b7 Aug 11 15:54:29 mdt1p kernel: #1: (&inode->i_mutex){--..}, at: [<ffffffff80468a4b>] mutex_lock+0x25/0x29 Aug 11 15:54:29 mdt1p kernel: Aug 11 15:54:29 mdt1p kernel: stack backtrace: Aug 11 15:54:29 mdt1p kernel: Aug 11 15:54:29 mdt1p kernel: Call Trace: Aug 11 15:54:29 mdt1p kernel: [<ffffffff802506fa>] __lock_acquire+0x162/0xbd8 Aug 11 15:54:29 mdt1p kernel: [<ffffffff80468a4b>] mutex_lock+0x25/0x29 Aug 11 15:54:29 mdt1p kernel: [<ffffffff802511ec>] lock_acquire+0x7c/0xa0 Aug 11 15:54:29 mdt1p kernel: [<ffffffff80468a4b>] mutex_lock+0x25/0x29 Aug 11 15:54:29 mdt1p kernel: [<ffffffff8046889c>] __mutex_lock_slowpath+0xef/0x279 Aug 11 15:54:29 mdt1p kernel: [<ffffffff80468a4b>] mutex_lock+0x25/0x29 Aug 11 15:54:29 mdt1p kernel: [<ffffffff802a32bd>] vfs_unlink+0x86/0x10e Aug 11 15:54:29 mdt1p kernel: [<ffffffff8846d248>] :obdclass:llog_lvfs_destroy+0x168/0x980 Aug 11 15:54:29 mdt1p kernel: [<ffffffff883f137b>] :libcfs:cfs_alloc+0x2b/0x60 Aug 11 15:54:29 mdt1p kernel: [<ffffffff884687d6>] :obdclass:llog_init_handle+0xf6/0x880 Aug 11 15:54:29 mdt1p kernel: [<ffffffff80294549>] __kmalloc+0x136/0x146 Aug 11 15:54:29 mdt1p kernel: [<ffffffff8875b6a3>] :mgc:mgc_process_log+0x1953/0x24f0 Aug 11 15:54:29 mdt1p kernel: [<ffffffff8875cae0>] :mgc:mgc_blocking_ast+0x0/0x4a0 Aug 11 15:54:29 mdt1p kernel: [<ffffffff88513830>] :ptlrpc:ldlm_completion_ast+0x0/0x830 Aug 11 15:54:29 mdt1p kernel: [<ffffffff887597c1>] :mgc:config_log_find+0xb1/0x360 Aug 11 15:54:29 mdt1p kernel: [<ffffffff8875fafa>] :mgc:mgc_process_config+0x8aa/0x1080 Aug 11 15:54:29 mdt1p kernel: [<ffffffff884a3a0a>] :obdclass:lustre_process_log+0x35a/0xed0 Aug 11 15:54:29 mdt1p kernel: [<ffffffff80294549>] __kmalloc+0x136/0x146 Aug 11 15:54:29 mdt1p kernel: [<ffffffff884a4641>] :obdclass:server_find_mount+0x51/0x1b0 Aug 11 15:54:29 mdt1p kernel: [<ffffffff884a9ffc>] :obdclass:server_start_targets+0x98c/0x19e0 Aug 11 15:54:29 mdt1p kernel: [<ffffffff884ae510>] :obdclass:server_fill_super+0x1530/0x2380 Aug 11 15:54:29 mdt1p kernel: [<ffffffff80294136>] cache_alloc_refill+0x77/0x210 Aug 11 15:54:29 mdt1p kernel: [<ffffffff883f137b>] :libcfs:cfs_alloc+0x2b/0x60 Aug 11 15:54:29 mdt1p kernel: [<ffffffff80292eb6>] poison_obj+0x27/0x32 Aug 11 15:54:29 mdt1p kernel: [<ffffffff883f137b>] :libcfs:cfs_alloc+0x2b/0x60 Aug 11 15:54:29 mdt1p kernel: [<ffffffff80292fd4>] cache_alloc_debugcheck_after+0x113/0x1c2 Aug 11 15:54:29 mdt1p kernel: [<ffffffff80294549>] __kmalloc+0x136/0x146 Aug 11 15:54:29 mdt1p kernel: [<ffffffff884b054d>] :obdclass:lustre_fill_super+0x11ed/0x1870 Aug 11 15:54:29 mdt1p kernel: [<ffffffff802afd71>] get_filesystem+0x1a/0x40 Aug 11 15:54:29 mdt1p kernel: [<ffffffff8029c5f5>] sget+0x3a5/0x3b7 Aug 11 15:54:29 mdt1p kernel: [<ffffffff8029bf25>] set_anon_super+0x0/0xb7 Aug 11 15:54:29 mdt1p kernel: [<ffffffff884af360>] :obdclass:lustre_fill_super+0x0/0x1870 Aug 11 15:54:29 mdt1p kernel: [<ffffffff8029d0c7>] get_sb_nodev+0x57/0x97 Aug 11 15:54:29 mdt1p kernel: [<ffffffff884a0e96>] :obdclass:lustre_get_sb+0x16/0x20 Aug 11 15:54:29 mdt1p kernel: [<ffffffff8029cb0f>] vfs_kern_mount+0x98/0x121 Aug 11 15:54:29 mdt1p kernel: [<ffffffff8029cbf1>] do_kern_mount+0x47/0xe2 Aug 11 15:54:29 mdt1p kernel: [<ffffffff802b1d84>] do_mount+0x6a1/0x714 Aug 11 15:54:29 mdt1p kernel: [<ffffffff802510dc>] __lock_acquire+0xb44/0xbd8 Aug 11 15:54:29 mdt1p kernel: [<ffffffff80250102>] trace_hardirqs_on+0x11c/0x147 Aug 11 15:54:29 mdt1p kernel: [<ffffffff802510dc>] __lock_acquire+0xb44/0xbd8 Aug 11 15:54:29 mdt1p kernel: [<ffffffff80314437>] __up_read+0x1a/0x83 Aug 11 15:54:29 mdt1p kernel: [<ffffffff8027545c>] get_page_from_freelist+0x2a2/0x5ac Aug 11 15:54:29 mdt1p kernel: [<ffffffff80250102>] trace_hardirqs_on+0x11c/0x147 Aug 11 15:54:29 mdt1p kernel: [<ffffffff8027546a>] get_page_from_freelist+0x2b0/0x5ac Aug 11 15:54:29 mdt1p kernel: [<ffffffff80292fd4>] cache_alloc_debugcheck_after+0x113/0x1c2 Aug 11 15:54:29 mdt1p kernel: [<ffffffff8028df39>] alloc_pages_current+0xa8/0xb0 Aug 11 15:54:29 mdt1p kernel: [<ffffffff802b1e80>] sys_mount+0x89/0xcb Aug 11 15:54:29 mdt1p kernel: [<ffffffff8020d043>] syscall_trace_enter+0x95/0x99 Aug 11 15:54:29 mdt1p kernel: [<ffffffff80209ec5>] tracesys+0xdc/0xe1 Aug 11 15:54:29 mdt1p kernel: Aug 11 15:54:29 mdt1p kernel: Lustre: Enabling user_xattr Aug 11 15:54:29 mdt1p kernel: Lustre: Server lqcdproj-MDT0000 on device /dev/mapper/mdt1_vol-mdt has started