Displaying 20 results from an estimated 2000 matches similar to: "Re: [Ocfs2-commits] manish commits r1580 - trunk/src"
2009 Jun 19
6
[PATCH 1/5] ocfs2: Pin journal head before accessing jh->b_committed_data
This patch adds jbd_lock_bh_state() and jbd_unlock_bh_state() around accessses
to jh->b_committed_data.
Fixes oss bugzilla#1131
http://oss.oracle.com/bugzilla/show_bug.cgi?id=1131
Signed-off-by: Sunil Mushran <sunil.mushran at oracle.com>
---
fs/ocfs2/suballoc.c | 28 ++++++++++++++++++++++++----
1 files changed, 24 insertions(+), 4 deletions(-)
diff --git a/fs/ocfs2/suballoc.c
2009 Mar 06
0
[PATCH 1/1] OCFS2: anti stale inode for nfs (V6.3)
#against V6.2, add indentation.
For nfs exporting, ocfs2_get_dentry() returns the dentry for fh.
ocfs2_get_dentry() may read from disk(when inode not in memory) without
any cross cluster lock. this leads to load a stale inode.
this patch fixes above problem.
solution is that in case of inode is not in memory, we get the cluster lock(PR)
of alloc inode where the inode in question is allocated
2009 Feb 27
2
[PATCH 1/1] OCFS2: anti stale inode for nfs (V5)
changes from v4:
1, let suballoc lock covers the checking of the group.
2, add/correct some log messages.
3, use ocfs2_read_group_descriptor() instead of diry reading the group.
Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com>
--
dlmglue.c | 45 ++++++++++++++++
dlmglue.h | 2
export.c | 77 +++++++++++++++++++++++++--
inode.c | 24 ++++++++
2009 Mar 05
0
[PATCH 1/1] OCFS2: anti stale inode for nfs (V6.2)
#against V6, corrects some format problem pointed out by checkpatch.pl.
For nfs exporting, ocfs2_get_dentry() returns the dentry for fh.
ocfs2_get_dentry() may read from disk(when inode not in memory) without
any cross cluster lock. this leads to load a stale inode.
this patch fixes above problem.
solution is that in case of inode is not in memory, we get the cluster lock(PR)
of alloc inode
2009 Mar 06
2
[PATCH 1/1] OCFS2: anti stale inode for nfs (for 1.4git)
Back porting from mainline.
For nfs exporting, ocfs2_get_dentry() returns the dentry for fh.
ocfs2_get_dentry() may read from disk(when inode not in memory) without
any cross cluster lock. this leads to load a stale inode.
this patch fixes above problem.
solution is that in case of inode is not in memory, we get the cluster lock(PR)
of alloc inode where the inode in question is allocated
2009 Feb 20
3
[PATCH 1/1] OCFS2: anti stale inode for nfs (V4)
changes from v3:
1, move codes that checks inode allocation bit to subfunction
ocfs2_test_inode_bit().
2, release the suballoc lock just after we get it. we should release it asap
and doing so doesn't affect functionility.
3, add inode alloc slot validation.
Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com>
--
dlmglue.c | 45 +++++++++++++++++
dlmglue.h | 2
2009 Mar 03
3
[PATCH 1/1] OCFS2: anti stale inode for nfs (V6)
For nfs exporting, ocfs2_get_dentry() returns the dentry for fh.
ocfs2_get_dentry() may read from disk(when inode not in memory) without
any cross cluster lock. this leads to load a stale inode.
this patch fixes above problem.
solution is that in case of inode is not in memory, we get the cluster lock(PR)
of alloc inode where the inode in question is allocated from(this causes node
on which
2006 Aug 15
0
[git patches] ocfs2 updates
This set of patches includes a few dlm related fixes from Kurt, and a small,
trivial cleanup by Adrian.
Also included are three disk allocation patches by me - two fixes and one
incremental improvement in our allocation strategy. These have been around
since early June, so I think they've had enough testing that they can go
upstream.
Please pull from 'upstream-linus' branch of
2009 Feb 17
1
[PATCH 1/1] OCFS2: anti stale inode for nfs (V3)
For nfs exporting, ocfs2_get_dentry() returns the dentry for fh.
ocfs2_get_dentry() may read from disk(when inode not in memory) without
any cross cluster lock. this leads to load a stale inode.
this patch fixes above problem.
solution is that in case of inode is not in memory, we get the cluster lock(PR)
of alloc inode where the inode in question is allocated from(this causes node
on which
2009 Apr 21
0
[PATCH] ocfs2: Fix some printk() warnings.
The old %llu vs u64 battle. Cast them correctly.
Signed-off-by: Joel Becker <joel.becker at oracle.com>
---
fs/ocfs2/export.c | 9 +++++----
fs/ocfs2/suballoc.c | 19 ++++++++++++-------
2 files changed, 17 insertions(+), 11 deletions(-)
diff --git a/fs/ocfs2/export.c b/fs/ocfs2/export.c
index de3da8e..15713cb 100644
--- a/fs/ocfs2/export.c
+++ b/fs/ocfs2/export.c
@@ -100,7 +100,8
2013 Feb 21
1
[PATCH] the ac->ac_allow_chain_relink=0 won't disable group relink
From: "Xiaowei.Hu" <xiaowei.hu at oracle.com>
ocfs2_block_group_alloc_discontig() disables chain relink by setting
ac->ac_allow_chain_relink = 0 because it grabs clusters from multiple
cluster groups. It doesn't keep the credits for all chain relink,but
ocfs2_claim_suballoc_bits overrides this in this call trace:
2005 May 17
0
Re: [Ocfs2-commits] manish commits r2251 - trunk/fs/ocfs2
On Mon, May 16, 2005 at 08:14:11PM -0500, svn-commits@oss.oracle.com wrote:
> Author: manish
> Signed-off-by: mfasheh
> Date: 2005-05-16 20:14:10 -0500 (Mon, 16 May 2005)
> New Revision: 2251
>
> Modified:
> trunk/fs/ocfs2/namei.c
> Log:
> Cosmetic coding style fix
While we're fixing at style issues, could you please remove all the
typedefs for struct types?
2010 Oct 18
1
yum list updates error (Manish Kathuria)
There was no proxy in the file /etc/yum.conf. I specified the proxy setting
and now "yum list updates" is working. A part of the output was:
Loaded plugins: fastestmirror
Loading mirror speeds from cached hostfile
* base: centos.01link.hk
* updates: centos.01link.hk
* addons: centos.01link.hk
* extras: centos.01link.hk
Updated Packages
ImageMagick.i386
2009 Feb 24
2
[PATCH 1/3] ocfs2: Optimize inode allocation by remembering last group.
In ocfs2, the inode block search looks for the "emptiest" inode
group to allocate from. So if an inode alloc file has many equally
(or almost equally) empty groups, new inodes will tend to get
spread out amongst them, which in turn can put them all over the
disk. This is undesirable because directory operations on conceptually
"nearby" inodes force a large number of seeks.
So
2009 Mar 18
3
[PATCH] ocfs2: Fix 2 warning during ocfs2 make.
fs/ocfs2/dir.c: In function ?ocfs2_extend_dir?:
fs/ocfs2/dir.c:2700: warning: ?ret? may be used uninitialized in this function
fs/ocfs2/suballoc.c: In function ?ocfs2_get_suballoc_slot_bit?:
fs/ocfs2/suballoc.c:2216: warning: comparison is always true due to limited range of data type
Signed-off-by: Tao Ma <tao.ma at oracle.com>
---
fs/ocfs2/dir.c | 2 +-
fs/ocfs2/suballoc.c |
2009 May 01
1
[GIT PULL] ocfs2 fixes
Linus, et al,
Here are the outstanding ocfs2 fixes we have for 2.6.30. The
important ones are journal credit fixes for the new indexed directories.
There's also MAINTAINERS update for my ocfs2 git tree. Please pull.
Joel
The following changes since commit 612392307cb09e49051225092cbbd7049bd8db93:
Ryusuke Konishi (1):
nilfs2: support nanosecond timestamp
are available in the
2007 Apr 13
2
Anaconda can't squeeze out the repomd.xml
Greetings.
There must be some minor changes to anaconda. I'm getting the error:
"Cannot open repomd.xml file...."
the file seems to be located in the repodata directory...
I'm using the following .cf taken directly from the CentOS 4.4 install :
install
url --url ftp://centos.westmancom.com/5.0/os/i386/
#cdrom
lang en_US.UTF-8
langsupport --default=en_US.UTF-8 en_US.UTF-8
2005 Jun 17
1
kickstart software raid on sata drives
This is what I have for configuring software raid on sata drives in my
kickstart config
<snip kickstart.cfg>
clearpart --initlabel --all
part swap --size=2048 --ondisk=sdb
part swap --size=2048 --ondisk=sda
part raid.01 --size=101 --ondisk=sda
part raid.02 --size=101 --ondisk=sdb
part raid.04 --size=1 --grow --ondisk=sdb
part raid.03 --size=1 --grow --ondisk=sda
raid / --fstype ext3
2010 Nov 18
1
kickstart raid disk partitioning
Hello.
A couple of years ago I installed two file-servers
using kickstart. The server has two 1TB sata disks
with two software raid1 partitions as follows:
# cat /proc/mdstat
Personalities : [raid1]
md1 : active raid1 sdb4[1] sda4[0]
933448704 blocks [2/2] [UU]
md0 : active raid1 sdb1[1] sda2[2](F)
40957568 blocks [2/1] [_U]
Now the drives are starting to be failing and next week
2006 Mar 02
0
Discrepancies in Anaconda-ks.cfg after kickstart
I am asking this here, that although a kickstart question, may have
something to do with the Centos install?
I took my Anaconda-ks.cfg from my system, turned it into a ks.cfg and
did the install. Everything SEEMS ok, but why?
First the partitioning information:
ks.cfg supplied:
clearpart --all --drives=hda
part /boot --fstype ext3 --size=100 --ondisk=hda
part / --fstype ext3 --start=14