similar to: oss umount hangs forever

Displaying 7 results from an estimated 7 matches similar to: "oss umount hangs forever"

2008 Apr 15
5
o2ib module prevents shutdown
Hello, Not sure if this is the right forum: I''m encountering difficulties with o2ib which prevents an LNET shutdown from proceeding: Unloading OpenIB kernel modules:NET: Unregistered protocal family 27 Failed to unload rdma_cm Failed to unload rdma_cm Failed to unload ib_cm Failed to unload ib_sa LustreError: 131-3: Received notification of device removal Please shutdown LNET
2001 Oct 07
2
"DRQ after issuing write" error
While writing a file to an ext3 filesystem, i get some drive errors on the console. I guess is a drive/ide related problem more that a ext3 one, but anyway the debug output is here if you want to take a look. And, by the way, what means that DRQ thing?? Oct 7 20:31:36 fargo kernel: (journal.c, 218): kjournald: kjournald wakes Oct 7 20:31:36 fargo kernel: (journal.c, 202): kjournald:
2010 Jul 07
0
How to evict a dead client?
Dear, everyone We have stuck with the problem that the OSS connect one dead client or one with changed IP address all the time until we reboot the dead client. From the OSS log message, we can get the information as follows: Jul 7 14:45:07 com01 kernel: Lustre: 12180:0:(socklnd_cb.cLustre: 12180:0:(socklnd_cb.c:915:ksocknal_launch_packet()) No usable routes to 12345-202.Lustre:
2008 Feb 12
0
Lustre-discuss Digest, Vol 25, Issue 17
Hi, i just want to know whether there are any alternative file systems for HP SFS. I heard that there is Cluster Gateway from Polyserve. Can anybody plz help me in finding more abt this Cluster Gateway. Thanks and Regards, Ashok Bharat -----Original Message----- From: lustre-discuss-bounces at lists.lustre.org on behalf of lustre-discuss-request at lists.lustre.org Sent: Tue 2/12/2008 3:18 AM
2010 Aug 04
6
[PATCH -v2 0/3] jbd2 scalability patches
This version fixes three bugs in the 2nd patch of this series that caused kernel BUG when the system was under race. We weren't accounting with t_oustanding_credits correctly, and there were race conditions caused by the fact the I had overlooked the fact that __jbd2_log_wait_for_space() and jbd2_get_transaction() requires j_state_lock to be write locked. Theodore Ts'o (3): jbd2: Use
2001 Jun 06
1
ext3-0.0.7a for 2.2.19 is released
Hi all, ext3-0.0.7a is now out, at the usual places: ftp.uk.linux.org:/pub/linux/sct/fs/jfs/ and propagating now from ftp.*.kernel.org:/pub/linux/kernel/people/sct/ext3/ This fixes one major bug in ext3 recovery for metadata-only journaling. Because recovery can also happen in e2fsck, users should also upgrade to at least e2fsprogs-1.21-WIP-0601 (also provided at the above ftp sites).
2008 Feb 04
32
Luster clients getting evicted
on our cluster that has been running lustre for about 1 month. I have 1 MDT/MGS and 1 OSS with 2 OST''s. Our cluster uses all Gige and has about 608 nodes 1854 cores. We have allot of jobs that die, and/or go into high IO wait, strace shows processes stuck in fstat(). The big problem is (i think) I would like some feedback on it that of these 608 nodes 209 of them have in dmesg