similar to: Assert in jbd-kernel.c

Displaying 20 results from an estimated 2000 matches similar to: "Assert in jbd-kernel.c"

2005 Jun 14
2
[2.6 patch] fs/jbd/: possible cleanups
This patch contains the following possible cleanups: - make needlessly global functions static - journal.c: remove the unused global function __journal_internal_check and move the check to journal_init - remove the following write-only global variable: - journal.c: current_journal - remove the following unneeded EXPORT_SYMBOL's: - journal.c: journal_check_used_features -
2005 Jul 19
1
[2.6 patch] fs/jbd/: cleanups
This patch contains the following cleanups: - make needlessly global functions static - journal.c: remove the unused global function __journal_internal_check and move the check to journal_init - remove the following write-only global variable: - journal.c: current_journal - remove the following unneeded EXPORT_SYMBOL: - journal.c: journal_recover Signed-off-by: Adrian Bunk
2005 Apr 22
2
[2.6 patch] fs/jbd/: possible cleanups
This patch contains the following possible cleanups: - make needlessly global functions static - #if 0 the following unused global functions: - journal.c: __journal_internal_check - journal.c: journal_ack_err - remove the following write-only global variable: - journal.c: current_journal - remove the following unneeded EXPORT_SYMBOL's: - journal.c: journal_check_used_features -
2009 Feb 04
1
Strange dmesg messages
Hi list, Something went wrong this morning and we have a node ( #0 ) reboot. Something blocked the NFS access from both nodes, one rebooted and the another we restarted the nfsd and it brought him back. Looking at node #0 - the one that rebooted - logs everything seems normal, but looking at the othere node dmesg's we saw this messages: First the o2net detected that node #0 was dead: (It
2006 Nov 28
4
how to prevent filesystem check
Hi all, I want to setup a RAID storage system, where i have two systems connected to it. the filesystems are mapped out to both connectors. I want the master host mount them read write, and the slave read only. in my fstab on the slave I have a line like the following: /dev/sdb1 /mount ext3 acl,noauto,user_xattr,nosuid,ro 0 0 so in man 5 fstab, it is written, that when the 6. field
2005 Sep 09
7
[PATCH 0/6] jbd cleanup
The following 6 patches cleanup the jbd code and kill about 200 lines. First of 4 patches can apply to 2.6.13-git8 and 2.6.13-mm2. The rest of them can apply to 2.6.13-mm2. fs/jbd/checkpoint.c | 179 +++++++++++-------------------------------- fs/jbd/commit.c | 101 ++++++++++-------------- fs/jbd/journal.c | 11 +- fs/jbd/revoke.c | 158
2002 May 31
2
PATCH for filesys corruption in ext3 with data=journal
Hi, as I mentioned in earlier mail to ext3-users I have been getting some corruption on an ext3 filesystem that has been serving NFS. I am now confident that I fully understand the problem and have a patch. It only affects data=journal mode and I wonder if it might also be the cause of the corruption noted by a number of people on linux-kernel. First I will explain the problem. Then display
2009 Sep 23
0
jbd/kjournald oops on 2.6.30.1
Hi, I am getting the following Oops on 2.6.30.1 kernel. The bad part is, it happens rarely (twice in last 1.5 months) and the system is pretty lightly loaded when this happens (no heavy file/disk io). Any insights or patches that I can try? (i searched lkml and ext3 lists but could not find any similar oops/reports). == Oops =================== BUG: unable to handle kernel NULL pointer
2002 Sep 06
1
kjournald & jbd
Hello everybody, Could someone please explain to me what is the difference between kjournald and jbd (precisely, what does each of them do?) Thank you, Alina _________________________________________________________________ Chat with friends online, try MSN Messenger: http://messenger.msn.com
2006 Aug 27
1
how can I get the 64bit JBD patch?
I see that ocfs2 has used 64bit JBD. I want to know where is the 64bit JBD patch? Whether the patch is developed by ocfs2 or not? Whether the patch is developed by JBD itself? Anyone can help me? -------------- next part -------------- An HTML attachment was scrubbed... URL: http://oss.oracle.com/pipermail/ocfs2-devel/attachments/20060828/98fdb6c1/attachment.html
2008 Mar 18
1
Problems patching fs/jbd/checkpoint.c in RHEL4 2.6.9-67.0.4 kernel
My manual patching of the rejects in checkpoint.c didn''t work out; a delete of 10,000 files caused a panic (in any ext fs, not just Lustre). In the new checkpoint.c, two routines expected by the patch no longer exist: __cleanup_transaction and __flush_buffer. I can avoid the panic if I omit (don''t try to manually patch) the following: Index: linux-2.6.9/fs/jbd/checkpoint.c
2007 Jan 29
1
seeking a developper documentation for jbd and ext3
Hi, I am a student in computer science and I develop a program that tries to explain other students the mecanisms of the ext3 filesystem : we show the content of each structure and explain what it means. But I was unable to find a developper documentation for the jounalizing functionality (jbd). Could you please tell me where can I find one ? ( in english or in french ) Also a documentation
2010 Aug 02
1
JBD: failed to read block at offset
Im getting errors like JBD: failed to read block at offset 4360 on a raid 5 parition , have run several fcsk on it.... are these Journal errors recoverable? I've done a physical scan of the hard disks with HP insight manager as well -- ------------------------------------------------------------------------------------------------------------------------------------- NOTICE: This message,
2001 Sep 22
7
FS locks
Hello, unfortunately I got some trouble using ext3 - the system hangs when working in a certain directory. Not completely, I can switch between consoles and reboot with SysRQ, but cannot do anything which requires disk IO: not start any applications, not umount, nothing. I could send the strace output, but I don't think this would help anyone. Sorry, but I don't have a clue where to start
2002 Dec 06
1
Assertion failure in do_get_write_access() at fs/jbd/transaction.c:746
I guess this could be filed under the 'shot myself in the foot' section, but I figured I'd report it anyways.. Kernel 2.5.50-bk6, x86, preempt, PIIX IDE, ext3 fs. I dd'ed the first part of a ext3 partition to tmp, then mounted it via loopback. (the image passed a fsck, except for warning about corrupt part. table or superblock) I tried to 'cd' into a directory within
2001 Sep 24
1
ext3-2.4-0.9.10
An ext3 patch against linux 2.4.10 is at http://www.uow.edu.au/~andrewm/linux/ext3/ This patch is *lightly tested* - ie, it boots and does stuff. The changes to ext3 are small, but the kernel which it patches has recently changed a lot. If you're cautious, please wait a couple of days. The patch retains the buffer-tracing code. This will soon be broken out into a separate patch to make
2001 Nov 01
3
ext3 partition still gets mounted as ext2 for me too!
Hi, I am a new ext3 user and I am having problems similar to what many people are facing on this list. I have done everything as the emails suggest. I am using SuSE 7.3 with the the 2.4.10 kernel and ext3 compiled as a module. Because it is 2.4.10 kernel, I created all my journals from a rescue disk and everything seems to be working right for the non-root filesystems. My fstab has all my
2007 Mar 08
4
ocfs2 cluster becomes unresponsive
We are running OCFS2 on SLES9 machines using a FC SAN. Without warning both nodes will become unresponsive. Can not access either machine via ssh or terminal (hangs after typing in username). However the machine still responds to pings. This continues until one node is rebooted, at which time the second node resumes normal operations. I am not entirely sure that this is an OCFS2 problem at all
2005 Nov 16
0
(large, external) data journal BUG (Assertion failure in __journal_drop_transaction() at fs/jbd/checkpoint.c:626: "transaction->t_forget == NULL")
Hi, A couple of our important servers, both running FC4 but one i386 and one x86_64, have been crashing recently. They both are running ext3 data=journal with large external journals and high commit intervals. Both machines use the gdth driver for their hardware RAID sets, if that's of any use. I think the hardware is good in both cases. I hope someone finds this data useful enough to be
2002 Jun 06
2
More ext3 fileserver woes ...
Well.... you might remember that I have had problems will my NFS fileserver that run ext3 with data=journal. The filesystem corruption now seems too be solved with the patch (plus amendment) that I posted, so I am happy about that... but there is more. I have known for a while that ext3 doesn't behave very well when the journal fills up. If it finds that the journal is full, and the