similar to: Re: [Ocfs2-commits] manish commits r1580 - trunk/src

Displaying 20 results from an estimated 2000 matches similar to: "Re: [Ocfs2-commits] manish commits r1580 - trunk/src"

2009 Jun 19
6
[PATCH 1/5] ocfs2: Pin journal head before accessing jh->b_committed_data
This patch adds jbd_lock_bh_state() and jbd_unlock_bh_state() around accessses to jh->b_committed_data. Fixes oss bugzilla#1131 http://oss.oracle.com/bugzilla/show_bug.cgi?id=1131 Signed-off-by: Sunil Mushran <sunil.mushran at oracle.com> --- fs/ocfs2/suballoc.c | 28 ++++++++++++++++++++++++---- 1 files changed, 24 insertions(+), 4 deletions(-) diff --git a/fs/ocfs2/suballoc.c
2009 Mar 06
0
[PATCH 1/1] OCFS2: anti stale inode for nfs (V6.3)
#against V6.2, add indentation. For nfs exporting, ocfs2_get_dentry() returns the dentry for fh. ocfs2_get_dentry() may read from disk(when inode not in memory) without any cross cluster lock. this leads to load a stale inode. this patch fixes above problem. solution is that in case of inode is not in memory, we get the cluster lock(PR) of alloc inode where the inode in question is allocated
2009 Feb 27
2
[PATCH 1/1] OCFS2: anti stale inode for nfs (V5)
changes from v4: 1, let suballoc lock covers the checking of the group. 2, add/correct some log messages. 3, use ocfs2_read_group_descriptor() instead of diry reading the group. Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com> -- dlmglue.c | 45 ++++++++++++++++ dlmglue.h | 2 export.c | 77 +++++++++++++++++++++++++-- inode.c | 24 ++++++++
2009 Mar 05
0
[PATCH 1/1] OCFS2: anti stale inode for nfs (V6.2)
#against V6, corrects some format problem pointed out by checkpatch.pl. For nfs exporting, ocfs2_get_dentry() returns the dentry for fh. ocfs2_get_dentry() may read from disk(when inode not in memory) without any cross cluster lock. this leads to load a stale inode. this patch fixes above problem. solution is that in case of inode is not in memory, we get the cluster lock(PR) of alloc inode
2009 Mar 06
2
[PATCH 1/1] OCFS2: anti stale inode for nfs (for 1.4git)
Back porting from mainline. For nfs exporting, ocfs2_get_dentry() returns the dentry for fh. ocfs2_get_dentry() may read from disk(when inode not in memory) without any cross cluster lock. this leads to load a stale inode. this patch fixes above problem. solution is that in case of inode is not in memory, we get the cluster lock(PR) of alloc inode where the inode in question is allocated
2009 Feb 20
3
[PATCH 1/1] OCFS2: anti stale inode for nfs (V4)
changes from v3: 1, move codes that checks inode allocation bit to subfunction ocfs2_test_inode_bit(). 2, release the suballoc lock just after we get it. we should release it asap and doing so doesn't affect functionility. 3, add inode alloc slot validation. Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com> -- dlmglue.c | 45 +++++++++++++++++ dlmglue.h | 2
2009 Mar 03
3
[PATCH 1/1] OCFS2: anti stale inode for nfs (V6)
For nfs exporting, ocfs2_get_dentry() returns the dentry for fh. ocfs2_get_dentry() may read from disk(when inode not in memory) without any cross cluster lock. this leads to load a stale inode. this patch fixes above problem. solution is that in case of inode is not in memory, we get the cluster lock(PR) of alloc inode where the inode in question is allocated from(this causes node on which
2006 Aug 15
0
[git patches] ocfs2 updates
This set of patches includes a few dlm related fixes from Kurt, and a small, trivial cleanup by Adrian. Also included are three disk allocation patches by me - two fixes and one incremental improvement in our allocation strategy. These have been around since early June, so I think they've had enough testing that they can go upstream. Please pull from 'upstream-linus' branch of
2009 Feb 17
1
[PATCH 1/1] OCFS2: anti stale inode for nfs (V3)
For nfs exporting, ocfs2_get_dentry() returns the dentry for fh. ocfs2_get_dentry() may read from disk(when inode not in memory) without any cross cluster lock. this leads to load a stale inode. this patch fixes above problem. solution is that in case of inode is not in memory, we get the cluster lock(PR) of alloc inode where the inode in question is allocated from(this causes node on which
2009 Apr 21
0
[PATCH] ocfs2: Fix some printk() warnings.
The old %llu vs u64 battle. Cast them correctly. Signed-off-by: Joel Becker <joel.becker at oracle.com> --- fs/ocfs2/export.c | 9 +++++---- fs/ocfs2/suballoc.c | 19 ++++++++++++------- 2 files changed, 17 insertions(+), 11 deletions(-) diff --git a/fs/ocfs2/export.c b/fs/ocfs2/export.c index de3da8e..15713cb 100644 --- a/fs/ocfs2/export.c +++ b/fs/ocfs2/export.c @@ -100,7 +100,8
2013 Feb 21
1
[PATCH] the ac->ac_allow_chain_relink=0 won't disable group relink
From: "Xiaowei.Hu" <xiaowei.hu at oracle.com> ocfs2_block_group_alloc_discontig() disables chain relink by setting ac->ac_allow_chain_relink = 0 because it grabs clusters from multiple cluster groups. It doesn't keep the credits for all chain relink,but ocfs2_claim_suballoc_bits overrides this in this call trace:
2005 May 17
0
Re: [Ocfs2-commits] manish commits r2251 - trunk/fs/ocfs2
On Mon, May 16, 2005 at 08:14:11PM -0500, svn-commits@oss.oracle.com wrote: > Author: manish > Signed-off-by: mfasheh > Date: 2005-05-16 20:14:10 -0500 (Mon, 16 May 2005) > New Revision: 2251 > > Modified: > trunk/fs/ocfs2/namei.c > Log: > Cosmetic coding style fix While we're fixing at style issues, could you please remove all the typedefs for struct types?
2010 Oct 18
1
yum list updates error (Manish Kathuria)
There was no proxy in the file /etc/yum.conf. I specified the proxy setting and now "yum list updates" is working. A part of the output was: Loaded plugins: fastestmirror Loading mirror speeds from cached hostfile * base: centos.01link.hk * updates: centos.01link.hk * addons: centos.01link.hk * extras: centos.01link.hk Updated Packages ImageMagick.i386
2009 Feb 24
2
[PATCH 1/3] ocfs2: Optimize inode allocation by remembering last group.
In ocfs2, the inode block search looks for the "emptiest" inode group to allocate from. So if an inode alloc file has many equally (or almost equally) empty groups, new inodes will tend to get spread out amongst them, which in turn can put them all over the disk. This is undesirable because directory operations on conceptually "nearby" inodes force a large number of seeks. So
2009 Mar 18
3
[PATCH] ocfs2: Fix 2 warning during ocfs2 make.
fs/ocfs2/dir.c: In function ?ocfs2_extend_dir?: fs/ocfs2/dir.c:2700: warning: ?ret? may be used uninitialized in this function fs/ocfs2/suballoc.c: In function ?ocfs2_get_suballoc_slot_bit?: fs/ocfs2/suballoc.c:2216: warning: comparison is always true due to limited range of data type Signed-off-by: Tao Ma <tao.ma at oracle.com> --- fs/ocfs2/dir.c | 2 +- fs/ocfs2/suballoc.c |
2009 May 01
1
[GIT PULL] ocfs2 fixes
Linus, et al, Here are the outstanding ocfs2 fixes we have for 2.6.30. The important ones are journal credit fixes for the new indexed directories. There's also MAINTAINERS update for my ocfs2 git tree. Please pull. Joel The following changes since commit 612392307cb09e49051225092cbbd7049bd8db93: Ryusuke Konishi (1): nilfs2: support nanosecond timestamp are available in the
2007 Apr 13
2
Anaconda can't squeeze out the repomd.xml
Greetings. There must be some minor changes to anaconda. I'm getting the error: "Cannot open repomd.xml file...." the file seems to be located in the repodata directory... I'm using the following .cf taken directly from the CentOS 4.4 install : install url --url ftp://centos.westmancom.com/5.0/os/i386/ #cdrom lang en_US.UTF-8 langsupport --default=en_US.UTF-8 en_US.UTF-8
2005 Jun 17
1
kickstart software raid on sata drives
This is what I have for configuring software raid on sata drives in my kickstart config <snip kickstart.cfg> clearpart --initlabel --all part swap --size=2048 --ondisk=sdb part swap --size=2048 --ondisk=sda part raid.01 --size=101 --ondisk=sda part raid.02 --size=101 --ondisk=sdb part raid.04 --size=1 --grow --ondisk=sdb part raid.03 --size=1 --grow --ondisk=sda raid / --fstype ext3
2010 Nov 18
1
kickstart raid disk partitioning
Hello. A couple of years ago I installed two file-servers using kickstart. The server has two 1TB sata disks with two software raid1 partitions as follows: # cat /proc/mdstat Personalities : [raid1] md1 : active raid1 sdb4[1] sda4[0] 933448704 blocks [2/2] [UU] md0 : active raid1 sdb1[1] sda2[2](F) 40957568 blocks [2/1] [_U] Now the drives are starting to be failing and next week
2006 Mar 02
0
Discrepancies in Anaconda-ks.cfg after kickstart
I am asking this here, that although a kickstart question, may have something to do with the Centos install? I took my Anaconda-ks.cfg from my system, turned it into a ks.cfg and did the install. Everything SEEMS ok, but why? First the partitioning information: ks.cfg supplied: clearpart --all --drives=hda part /boot --fstype ext3 --size=100 --ondisk=hda part / --fstype ext3 --start=14