similar to: Kernel Panic, Server not coming back up

Displaying 20 results from an estimated 110 matches similar to: "Kernel Panic, Server not coming back up"

2009 Aug 02
1
Non sparse extend init issue
The patch was created against a 1.4 tree. However, it applies cleanly to mainline too. The patch has been lightly tested. I am running fill_verify_holes on a non sparse volume currently. Please review. Sunil
2009 Aug 03
1
Non sparse init fix v3
One line fix from Joel's version. Also, some comments removed. 18:58 <sunil> wc->w_first_new_cpos = 18:58 <sunil> - ocfs2_align_bytes_to_clusters(inode->i_sb, i_size_read(inode)); 18:58 <sunil> + ocfs2_clusters_for_bytes(inode->i_sb, i_size_read(inode));
2011 Aug 10
1
fsck hangs in Pass 0a
Hello list, I have a ~10TB ocfs2 filesystem in a 8-node cluster. This sits on a logical volume (I know lv is not cluster aware, but I make sure no one touches the lv, while the cluster is running). The LV consists of 5x2TB multipath devices. I recently had errors like this on some nodes: OCFS2: ERROR (device dm-7): ocfs2_check_group_descriptor: Group Descriptor # 0 has bad signature File
2009 Jul 29
1
ocfs2 quota qs.
Hi Jan, I am now reading quota support in ocfs2 and have some qs. Wish you can help me clarify it. Thanks. ocfs2_quota_write: In the comment, you said that "we know the transaction is already started", so it should be called within a transaction. But in this function, we call ocfs2_extend_no_holes and ocfs2_simple_size_update which will start another transaction. So we can survive
2009 Apr 07
0
[PATCH 1/1] OCFS2: use i_size_read() instead of direct accessing inode->i_size
use i_size_read() instead of accessing inode->i_size directly. Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com> -- diff --git a/fs/ocfs2/aops.c b/fs/ocfs2/aops.c index b2c52b3..4491851 100644 --- a/fs/ocfs2/aops.c +++ b/fs/ocfs2/aops.c @@ -1920,7 +1920,7 @@ int ocfs2_write_end_nolock(struct address_space *mapping, out_write_size: pos += copied; - if (pos >
2009 Mar 18
2
[RFC] metadata alloc fix in machines which has PAGE_SIZE > CLUSTER_SIZE
Hi Mark/Joel, I meet with some meta allocation bugs when I implement reflink these days. And after some investigation, I think we should have the same problem when we have PAGE_SIZE > CLUSTER_SIZE. So I create a scenario today in one ppc box and try. the box panic as I expected. ;) The scenario is that: Create a file with the disk layout like this(with bs=512, and cs=4K). debugfs: stat
2008 Sep 16
1
OCFS2 and Xen - aio error -14
I've configured a text Xen system using ocfs2 1.4.1 as the shared filesystem to enable migration. So far the file based xen disk sharing is working well. Yesterday I tried to do a "xm save vm /ocfs2/vm.save" and I keep getting the following message: __ocfs2_file_aio_write:2185 ERROR: status = -14 The file created is only 9113 bytes long where it should be about 2 GB i.e.
2009 Jul 13
1
[PATCH 1/1] adds mlogs to aops.c
this patch adds mlogs to apos to help tracing. Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com> --- fs/ocfs2/aops.c | 233 ++++++++++++++++++++++++++++++++++++++++++++----------- 1 files changed, 189 insertions(+), 44 deletions(-) diff --git a/fs/ocfs2/aops.c b/fs/ocfs2/aops.c index b2c52b3..b730010 100644 --- a/fs/ocfs2/aops.c +++ b/fs/ocfs2/aops.c @@ -90,7 +90,7 @@ static int
2009 Jul 21
1
(no subject)
>From c70adcaca99acf93bc00cf2edc4d549b83e2f95d Mon Sep 17 00:00:00 2001 From: Wengang Wang <wen.gang.wang at oracle.com> Date: Tue, 21 Jul 2009 10:52:52 +0800 Subject: [PATCH 1/1] ocfs2: adds mlogs to aops.c -V2 this patch adds some mlogs to apos.c helping tracing and narrowing down bugs. Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com> --- fs/ocfs2/aops.c | 242
2017 Jun 19
1
core dump on ocfs2
Hi everybody I'm finding on my server a lot of a lot of errors like this: Jun 19 14:22:45 posta2 kernel: [885017.412902] BUG: soft lockup - CPU#2 stuck for 22s! [dovecot-lda:11955] Jun 19 14:22:45 posta2 kernel: [885017.412906] Modules linked in: ocfs2(E) jbd2 quota_tree dm_service_time dm_multipath ocfs2_dlmfs(E) ocfs2_stack_o2cb(E) ocfs2_dlm( E) ocfs2_nodemanager(E)
2009 Jul 21
1
[PATCH 1/1] ocfs2: adds mlogs to aops.c -V2
this patch adds some mlogs to apos.c helping tracing and narrowing down bugs. Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com> --- fs/ocfs2/aops.c | 242 +++++++++++++++++++++++++++++++++++++++++++++---------- 1 files changed, 198 insertions(+), 44 deletions(-) diff --git a/fs/ocfs2/aops.c b/fs/ocfs2/aops.c index b2c52b3..4527f16 100644 --- a/fs/ocfs2/aops.c +++
2009 Jul 22
8
[PATCH 0/7] OCFS2 quota fixes (version 2)
Hi, here is the second version of OCFS2 quota fixes with Joel's comments fixed. Also I've added a patch defining counts of credits for quota operations as Joel asked. Honza
2013 Jun 19
1
[PATCH] fs/jbd2: t_updates should increase when start_this_handle() failed in jbd2__journal_restart()
jbd2_journal_restart() would restart a handle. In this function, it calls start_this_handle(). Before calling start_this_handle()?subtract 1 from transaction->t_updates. If start_this_handle() succeeds, transaction->t_updates increases by 1 in it. But if start_this_handle() fails, transaction->t_updates does not increase. So, when commit the handle's transaction in
2009 Jul 15
7
[PATCH 0/6] Quota fixes for 2.6.31-rc4
Hi, I did some more in-depth testing of OCFS2 quota code, especially with ECC feature enabled and spotted some problems. First four patches fix them. I need the fifth patch to be able to mount OCFS2 filesystem with 2.6.31-rc3. The sixth patch fixes a potential problem when quota syncing interval is updated while the cluster is running (which is not possible currently). Joel, could you please
2009 Feb 11
1
Possible lock inversion in directory locking
Hi, I've been playing lately with lockdep annotations of OCFS2. I seem to have most of the false positives sorted out and currently I hit the report below. I've analyzed that ocfs2_extend_dir() does first lock local alloc inode in ocfs2_reserve_clusters() and then acquires ip_alloc_sem from the directory. The usual ordering e.g. in ocfs2_write_begin() is to first acquire ip_alloc_sem
2009 Feb 26
3
[PATCH 0/3] ocfs2-1.4: Backport inode alloc from mainline.
Hi all, this patch set are the backport of inode alloc improvement from mainline to ocfs2-1.4. the patches are almost the same excpet one thing: Joel has added JBD2 support to ocfs2, so he has added "max_blocks" to alloc_context and add a new function "ocfs2_reserve_clusters_with_limit". We don't have that in ocfs2-1.4. So there are some great difference in patch 2.
2009 Mar 03
3
[PATCH 1/1] OCFS2: anti stale inode for nfs (V6)
For nfs exporting, ocfs2_get_dentry() returns the dentry for fh. ocfs2_get_dentry() may read from disk(when inode not in memory) without any cross cluster lock. this leads to load a stale inode. this patch fixes above problem. solution is that in case of inode is not in memory, we get the cluster lock(PR) of alloc inode where the inode in question is allocated from(this causes node on which
2013 Sep 03
1
[PATCH V3 RESENT] ocfs2: lighten up allocate transaction
Hi Andrew Morton, The patch "[patch 05/22] ocfs2: lighten up allocate transaction" in mm tree is not the newest one, please drop it from mm tree. This patch V3 is newest, please replace it. Thanks. The issue scenario is as following: When fallocating a very large disk space for a small file, __ocfs2_extend_allocation attempts to get a very large transaction. For some journal
2009 Mar 05
0
[PATCH 1/1] OCFS2: anti stale inode for nfs (V6.2)
#against V6, corrects some format problem pointed out by checkpatch.pl. For nfs exporting, ocfs2_get_dentry() returns the dentry for fh. ocfs2_get_dentry() may read from disk(when inode not in memory) without any cross cluster lock. this leads to load a stale inode. this patch fixes above problem. solution is that in case of inode is not in memory, we get the cluster lock(PR) of alloc inode
2009 Mar 06
0
[PATCH 1/1] OCFS2: anti stale inode for nfs (V6.3)
#against V6.2, add indentation. For nfs exporting, ocfs2_get_dentry() returns the dentry for fh. ocfs2_get_dentry() may read from disk(when inode not in memory) without any cross cluster lock. this leads to load a stale inode. this patch fixes above problem. solution is that in case of inode is not in memory, we get the cluster lock(PR) of alloc inode where the inode in question is allocated