similar to: [PATCH] ocfs2: Don't delete orphaned files if we are in the process of umount.

Displaying 20 results from an estimated 100 matches similar to: "[PATCH] ocfs2: Don't delete orphaned files if we are in the process of umount."

2010 Jan 29
0
[PATCH 3/3] ocfs2:freeze-thaw: make it work -v2
This patch adds freeze_fs()/unfreeze_fs() for ocfs2 so that it supports freeze/thaw. Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com> --- fs/ocfs2/dlmglue.c | 115 ++++++++++++++++++++++++++++++++++++++++++++++- fs/ocfs2/dlmglue.h | 2 + fs/ocfs2/journal.c | 1 + fs/ocfs2/ocfs2.h | 12 +++++ fs/ocfs2/super.c | 129 +++++++++++++++++++++++++++++++++++++++++++++++++++-
2009 Jul 20
1
[PATCH] ocfs2: flush dentry lock drop when sync ocfs2 volume.
In commit ea455f8ab68338ba69f5d3362b342c115bea8e13, we move the dentry lock put process into ocfs2_wq. This is OK for most case, but as for umount, it lead to at least 2 bugs. See http://oss.oracle.com/bugzilla/show_bug.cgi?id=1133 and http://oss.oracle.com/bugzilla/show_bug.cgi?id=1135. And it happens easily if we have opened a lot of inodes. For 1135, the reason is that during umount will call
2009 Jan 12
5
[PATCH 0/5] OCFS2 quota fixes
Hello, the following series of patches fixes some issues with OCFS2 quotas. The first patch modifies VFS quota locking, the next patch uses the fact to simplify OCFS2 quota locking and solves a few deadlock issues. The third and the fourth patches fix another possible deadlocks in OCFS2 quota code and the last patch is a minor cleanup. Honza
2015 Oct 14
1
Ocfs2-devel Digest, Vol 138, Issue 31 review
OCFS2 is often used in high-availaibility systems, This patch enhances robustness for the filesystem. but storage network is unstable?it still triggers a panic? such as ocfs2_start_trans -> __ocfs2_abort ->panic. The 's_mount_opt' should depend on the mount option set, If errors=continue is set, mark as a EIO error, change OCFS2_MOUNT_ERRORS_PANIC to OCFS2_MOUNT_ERRORS_CONT in
2015 Oct 14
1
Ocfs2-devel Digest, Vol 138, Issue 31 review
OCFS2 is often used in high-availaibility systems, This patch enhances robustness for the filesystem. but storage network is unstable?it still triggers a panic? such as ocfs2_start_trans -> __ocfs2_abort ->panic. The 's_mount_opt' should depend on the mount option set, If errors=continue is set, mark as a EIO error, change OCFS2_MOUNT_ERRORS_PANIC to OCFS2_MOUNT_ERRORS_CONT in
2023 May 04
1
[PATCH 2/2] ocfs2: add error handling path when jbd2 enter ABORT status
On Thu, May 04, 2023 at 05:41:29PM +0800, Joseph Qi wrote: > > > On 5/4/23 4:02 PM, Heming Zhao wrote: > > On Thu, May 04, 2023 at 03:34:49PM +0800, Joseph Qi wrote: > >> > >> > >> On 5/4/23 2:21 PM, Heming Zhao wrote: > >>> On Thu, May 04, 2023 at 10:27:46AM +0800, Joseph Qi wrote: > >>>> > >>>> >
2010 Jun 14
3
Diagnosing some OCFS2 error messages
Hello. I am experimenting with OCFS2 on Suse Linux Enterprise Server 11 Service Pack 1. I am performing various stress tests. My current exercise involves writing to files using a shared-writable mmap() from two nodes. (Each node mmaps and writes to different files; I am not trying to access the same file from multiple nodes.) Both nodes are logging messages like these: [94355.116255]
2023 Apr 30
3
[PATCH 2/2] ocfs2: add error handling path when jbd2 enter ABORT status
fstest generic cases 347 361 628 629 trigger a same issue: When jbd2 enter ABORT status, ocfs2 ignores it and keep going to commit journal. This commit gives ocfs2 ability to handle jbd2 ABORT case. Signed-off-by: Heming Zhao <heming.zhao at suse.com> --- fs/ocfs2/alloc.c | 10 ++++++---- fs/ocfs2/journal.c | 5 +++++ fs/ocfs2/localalloc.c | 3 +++ 3 files changed, 14
2009 Jan 12
1
Bug in inode deletion code leading to stale inodes
Hello, I've hit a bug in OCFS2 delete code which results in inodes being left on disk without any links to them. The workload triggering this creates directories on one node and deletes them on another node in the cluster. The inode is not deleted because both nodes bail out from ocfs2_delete_inode() with: Skipping delete of 100405 because it is in use on other nodes The scenario which I
2008 Sep 22
0
[PATCH 1/1] OCFS2: add validation in ocfs2_get_dentry()
we need do more checking for validation in ocfs2_get_dentry(). 1) OCFS2_INODE_DELETED flag 2) i_nlink the patch is against 1.4 git. Signed-off-by: Wengang wang <wen.gang.wang at oracle.com> -- diff --git a/fs/ocfs2/export.c b/fs/ocfs2/export.c index 649f3c8..36ccf27 100644 --- a/fs/ocfs2/export.c +++ b/fs/ocfs2/export.c @@ -49,6 +49,7 @@ static struct dentry *ocfs2_get_dentry(struct
2008 Sep 22
1
[PATCH 1/1] OCFS2: add nlink check in ocfs2_inode_revalidate()
nlink should be also checked in ocfs2_inode_revalidate(). before setting flag OCFS2_INODE_DELETED ip_flags (between unlink and delete vote), the nlink may be 0. the patch is against 1.4 git. 1.2 svn should has the same patch with different line number. Signed-off-by: Wengang wang <wen.gang.wang at oracle.com> -- diff --git a/fs/ocfs2/inode.c b/fs/ocfs2/inode.c index 591e693..6b5a83e
2010 Apr 29
2
Hardware error or ocfs2 error?
Hello, today I noticed the following on *only* one node: ----- cut here ----- Apr 29 11:01:18 node06 kernel: [2569440.616036] INFO: task ocfs2_wq:5214 blocked for more than 120 seconds. Apr 29 11:01:18 node06 kernel: [2569440.616056] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. Apr 29 11:01:18 node06 kernel: [2569440.616080] ocfs2_wq D
2009 Feb 03
1
[git patches] Ocfs2 and Configfs fixes
Please pull from 'upstream-linus' branch of git://git.kernel.org/pub/scm/linux/kernel/git/mfasheh/ocfs2.git upstream-linus to receive the following updates: fs/configfs/dir.c | 59 +++++++++++++++++++++++++++++++++++++++++++++++ fs/ocfs2/alloc.c | 3 ++ fs/ocfs2/dcache.c | 42 +++++++++++++++++++++++++++++++-- fs/ocfs2/dcache.h | 9 ++++++-
2011 May 13
1
答复: hi,if i can post ocfs2-dmesg to you?i have trouble on ocfs2
On 05/12/2011 06:50 PM, Longguang Yue wrote: > > Thank you first of all > > totally, there are 5 kinds of error occur. > > Spinlock leads to cpu lockup, o2net modules panic, kernel BUG at mm/slub.c:2969, BUG unable to handle kernel paging request at addr > > My environment: kernel-2.6.32.23 + xen + ocfs2(1.5) > > Thanks again. > This looks like the deadlock as
2010 Jul 16
0
[GIT PULL] Important ocfs2 fixes for 2.6.35
Linus, et al, Here are the major ocfs2 fixes for 2.6.35-rc5. A few more fixes have trickled in, but as these major fixes have been tested, I didn't want to delay them any longer. The first major fix you know about: the tail zeroing fix when extending files. It's been run under heavy testing for a week now, and we're quite happy. We should also no longer be touching any pages past
2011 Dec 20
1
OCFS2 problems when connectivity lost
Hello, We are having a problem with a 3-node cluster based on Pacemaker/Corosync with 2 primary DRBD+OCFS2 nodes and a quorum node. Nodes run on Debian Squeeze, all packages are from the stable branch except for Corosync (which is from backports for udpu functionality). Each node has a single network card. When the network is up, everything works without any problems, graceful shutdown of
2012 Jun 14
0
[ocfs2-announce] OCFS2 1.4.10-1 released
All, We are pleased to announce the release of OCFS2 1.4.10-1 and OCFS2 tools 1.6.3-2 for Oracle Linux 5 Update 7 and higher and Redhat Enterprise Linux 5 Update 7 and higher. Oracle's Unbreakable Linux Network users who are subscribing to the "OCFS2 1.4 packages for Enterprise Linux 5" channel can upgrade to this release by running up2date. Red Hat's Enterprise Linux 5
2011 Dec 20
8
ocfs2 - Kernel panic on many write/read from both
Sorry i don`t copy everything: TEST-MAIL1# echo "ls //orphan_dir:0000"|debugfs.ocfs2 /dev/dm-0|wc debugfs.ocfs2 1.6.4 5239722 26198604 246266859 TEST-MAIL1# echo "ls //orphan_dir:0001"|debugfs.ocfs2 /dev/dm-0|wc debugfs.ocfs2 1.6.4 6074335 30371669 285493670 TEST-MAIL2 ~ # echo "ls //orphan_dir:0000"|debugfs.ocfs2 /dev/dm-0|wc debugfs.ocfs2 1.6.4 5239722 26198604
2011 Apr 01
1
Node Recovery locks I/O in two-node OCFS2 cluster (DRBD 8.3.8 / Ubuntu 10.10)
I am running a two-node web cluster on OCFS2 via DRBD Primary/Primary (v8.3.8) and Pacemaker. Everything seems to be working great, except during testing of hard-boot scenarios. Whenever I hard-boot one of the nodes, the other node is successfully fenced and marked ?Outdated? * <resource minor="0" cs="WFConnection" ro1="Primary" ro2="Unknown"
2009 Feb 13
44
[PATCH 0/40] ocfs2: Detach ocfs2 metadata I/O from struct inode
The following series of patches attempts to detach metadata I/O from struct inode. They are currently tied together pretty tightly. Metadata reads happen via the ocfs2_read_blocks() functions, writes via both jbd2 and ocfs2_write_blocks(). - Each inode has a cache of associated metadata blocks stored on its ip_metadata_cache member. The ocfs2_read/write_blocks() functions take a struct