similar to: [PATCH] ocfs2: remove some pointless conditionals before kfree()

Displaying 20 results from an estimated 100 matches similar to: "[PATCH] ocfs2: remove some pointless conditionals before kfree()"

2009 May 01
0
[PATCH 1/3] OCFS2: speed up dlm_lockr_resouce hash_table lookups
use multiple pages for the hash table. mainline git commit: 03d864c02c3ea803b1718940ac6953a257182d7a Authored-by: Daniel Phillips <phillips at google.com> Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com> -- Index: ocfs2-1.2/fs/ocfs2/dlm/dlmdomain.c =================================================================== --- ocfs2-1.2/fs/ocfs2/dlm/dlmdomain.c (revision 1) +++
2009 Apr 22
1
[PATCH 1/1] OCFS2: fasten dlm_lock_resource hash_table lookups
#backporting the 3 patches at http://kernel.us.oracle.com/~smushran/srini/ to 1.2. enlarge hash_table capacity to fasten hash_table lookups. Signed-off-by: Wengang wang <wen.gang.wang at oracle.com> -- diff -up ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c.orig ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c --- ./svnocfs2-1.2/fs/ocfs2/dlm/dlmdebug.c.orig 2009-04-22 11:00:37.000000000 +0800 +++
2013 Mar 10
1
[PATCH] fs: fs2fs: Replaced calls to kmalloc and memcpy with kmemdup
Replaced calls to kmalloc followd by memcpy with a single call to kmemdup. This patch was found using coccicheck. Signed-off-by: Alexandru Gheorghiu <gheorghiuandru at gmail.com> --- fs/ocfs2/localalloc.c | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/fs/ocfs2/localalloc.c b/fs/ocfs2/localalloc.c index aebeacd..839183f 100644 --- a/fs/ocfs2/localalloc.c +++
2009 Jul 07
2
[PATCH 1/1] ocfs2-devel: trivial fix for s/migrate/migration/ in dlmrecovery.c, line 1121
in dlmrecovery.c:1121, replace 'migrate' to 'migration' to keep the consistency by comparing to other lines with the similar log info in the same file. Signed-off-by: Jeff Liu <jeff.liu at oracle.com> --- fs/ocfs2/dlm/dlmrecovery.c | 2 +- 1 files changed, 1 insertions(+), 1 deletions(-) diff --git a/fs/ocfs2/dlm/dlmrecovery.c b/fs/ocfs2/dlm/dlmrecovery.c index
2008 Jan 09
2
[PATCH 1/1] Clear joining_node no matter whether it is in the domain map or not.
Currently the process of dlm join contains 2 steps: query join and assert join. After query join, the joined node will set its joining_node. So if the joining node happens to panic before the 2nd step, the joined node will fail to clear its joining_node flag because that node isn't in the domain map. It at least cause 2 problems. 1. All the new join request will fail. So no new node can mount
2013 Aug 27
0
[patch 07/22] ocfs2: dlm_request_all_locks() should deal with the status sent from target node
From: Xue jiufei <xuejiufei at huawei.com> Subject: ocfs2: dlm_request_all_locks() should deal with the status sent from target node dlm_request_all_locks() should deal with the status sent from target node if DLM_LOCK_REQUEST_MSG is sent successfully, or recovery master will fall into endless loop, waiting for other nodes to send locks and DLM_RECO_DATA_DONE_MSG to me. NodeA
2013 Jul 26
0
[PATCH] ocfs2: dlm_request_all_locks() should deal with the status sent from target node
dlm_request_all_locks() should deal with the status sent from target node if DLM_LOCK_REQUEST_MSG is sent successfully, or recovery master will fall into endless loop, waiting for other nodes to send locks and DLM_RECO_DATA_DONE_MSG to me. NodeA NodeB selected as recovery master
2010 Aug 26
1
[PATCH 2/5] ocfs2/dlm: add lockres as parameter to dlm_new_lock()
Wether the dlm_lock needs to access lvb or not depends on dlm_lock_resource it belongs to. So a new parameter "struct dlm_lock_resource *res" is added to dlm_new_lock() so that we can know if we need to allocate lvb for the dlm_lock. And we have to make the lockres availale for calling dlm_new_lock(). Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com> ---
2010 Jul 16
0
[GIT PULL] Important ocfs2 fixes for 2.6.35
Linus, et al, Here are the major ocfs2 fixes for 2.6.35-rc5. A few more fixes have trickled in, but as these major fixes have been tested, I didn't want to delay them any longer. The first major fix you know about: the tail zeroing fix when extending files. It's been run under heavy testing for a week now, and we're quite happy. We should also no longer be touching any pages past
2010 May 20
0
[GIT PULL] ocfs2 updates for 2.6.35
Linus et al, Here are the ocfs2 updates for 2.6.35. There are two major changes. Mark added allocation reservations to our node-local allocators. This gets us much more contiguousness when many processes are growing files in parallel. On the other end of contiguousness, ocfs2 has had a major limitation since the beginning. When ocfs2 needs new inodes, it has always grown its metadata
2009 Aug 11
0
[GIT PULL] ocfs2 fixes for 2.6.31-rc5
Linus, et al, Here are all the outstanding ocfs2 fixes for 2.6.31. The major fixes are for aio write and for non-sparse file extending. There are also a couple of b-tree fixes as well as some for quotas. Please pull. Joel The following changes since commit 44b572809581d5a10dbe35aa6bf689f32b9c5ad6: Yinghai Lu (1): x86: don't clear nodes_states[N_NORMAL_MEMORY] when numa is not
2009 Feb 28
1
[PATCH 1/1] Patch to recover orphans in offline slots during recovery.
In the current recovery procedure, recovering node recovers orphans from the slot it is holding and the slots that dead nodes were holding. Other online nodes recover orphans from their slots. But orphans in offline slots are left un-recovered. This patch queues recovery_completion for offline slots. Signed-off-by: Srinivas Eeda <srinivas.eeda at oracle.com> --- fs/ocfs2/journal.c | 21
2007 May 17
1
[PATCH] ocfs: use list_for_each_entry where benefical
Signed-off-by: Christoph Hellwig <hch@lst.de> Index: linux-2.6/fs/ocfs2/cluster/tcp.c =================================================================== --- linux-2.6.orig/fs/ocfs2/cluster/tcp.c 2007-05-06 13:51:17.000000000 +0200 +++ linux-2.6/fs/ocfs2/cluster/tcp.c 2007-05-17 15:00:14.000000000 +0200 @@ -261,14 +261,12 @@ out: static void o2net_complete_nodes_nsw(struct o2net_node
2009 Apr 17
26
OCFS2 1.4: Patches backported from mainline
Please review the list of patches being applied to the ocfs2 1.4 tree. All patches list the mainline commit hash. Thanks Sunil
2008 Jan 25
0
[git patches] ocfs2 and configfs updates
These patches comprise the bulk of our outstanding changes for Ocfs2 and Configfs. Configfs gets some lockdep updates, and we remove it's EXPERIMENTAL tag as things have been quite stable for some time now. Ocfs2 gets fixes that made it in too late for 2.6.24, and several new features. The most interesting are listed below: We've been on the path to removing the "vote"
2010 Apr 14
2
[PATCH 1/2] ocfs2/dlm: Make o2dlm domain join/leave messages KERN_NOTICE
o2dlm join and leave messages are more than informational as they are required is debugging locking issues. This patch changes them from KERN_INFO to KERN_NOTICE. Signed-off-by: Sunil Mushran <sunil.mushran at oracle.com> --- fs/ocfs2/dlm/dlmdomain.c | 6 +++--- 1 files changed, 3 insertions(+), 3 deletions(-) diff --git a/fs/ocfs2/dlm/dlmdomain.c b/fs/ocfs2/dlm/dlmdomain.c index
2010 Jun 16
2
[PATCH] ocfs2/dlm: check dlm_state under spinlock
We should check dlm->dlm_state under dlm->spinlock though maybe in this case it doesn't hurt. Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com> --- fs/ocfs2/dlm/dlmdomain.c | 13 ++++++------- 1 files changed, 6 insertions(+), 7 deletions(-) diff --git a/fs/ocfs2/dlm/dlmdomain.c b/fs/ocfs2/dlm/dlmdomain.c index 6b5a492..ab82add 100644 --- a/fs/ocfs2/dlm/dlmdomain.c +++
2009 Mar 04
2
[PATCH 1/1] Patch to recover orphans in offline slots during recovery and mount
During recovery, a node recovers orphans in it's slot and the dead node(s). But if the dead nodes were holding orphans in offline slots, they will be left unrecovered. If the dead node is the last one to die and is holding orphans in other slots and is the first one to mount, then it only recovers it's own slot, which leaves orphans in offline slots. This patch queues complete_recovery
2009 Apr 15
1
I'd like to stop the OCSF2_disk_heart_beat.
This is Hide. This is the first contribution. I'd like to configure cluster file system by OCFS2. The number of nodes are two. CompactFlash card(CF) is used as shared DISK. Guarantees of the writing frequency to CF are up to 100,000 times. However, it exceed in two day, due to the disk_heart_beat executed every two seconds. I tried following procedures to solve that. It seems solved. 1.
2009 Mar 17
33
[git patches] Ocfs2 updates for 2.6.30
Hi, The following patches comprise the bulk of Ocfs2 updates for the 2.6.30 merge window. Aside from larger, more involved fixes, we're adding the following features, which I will describe in the order their patches are mailed. Sunil's exported some more state to our debugfs files, and consolidated some other aspects of our debugfs infrastructure. This will further aid us in debugging