similar to: [PATCH] the ac->ac_allow_chain_relink=0 won't disable group relink

Displaying 16 results from an estimated 16 matches similar to: "[PATCH] the ac->ac_allow_chain_relink=0 won't disable group relink"

2006 Aug 15
0
[git patches] ocfs2 updates
This set of patches includes a few dlm related fixes from Kurt, and a small, trivial cleanup by Adrian. Also included are three disk allocation patches by me - two fixes and one incremental improvement in our allocation strategy. These have been around since early June, so I think they've had enough testing that they can go upstream. Please pull from 'upstream-linus' branch of
2010 Apr 05
1
Kernel Panic, Server not coming back up
I have a relatively new test environment setup that is a little different from your typical scenario. This is my first time using OCFS2, but I believe it should work the way I have it setup. All of this is setup on VMWare virtual hosts. I have two front-end web servers and one backend administrative server. They all share 2 virtual hard drives within VMware (independent, persistent, &
2017 Jun 19
1
core dump on ocfs2
Hi everybody I'm finding on my server a lot of a lot of errors like this: Jun 19 14:22:45 posta2 kernel: [885017.412902] BUG: soft lockup - CPU#2 stuck for 22s! [dovecot-lda:11955] Jun 19 14:22:45 posta2 kernel: [885017.412906] Modules linked in: ocfs2(E) jbd2 quota_tree dm_service_time dm_multipath ocfs2_dlmfs(E) ocfs2_stack_o2cb(E) ocfs2_dlm( E) ocfs2_nodemanager(E)
2011 Sep 02
5
Linux kernel crash due to ocfs2
Hello, we have a pair of IBM P570 servers running RHEL5.2 kernel 2.6.18-92.el5.ppc64 We have Oracle RAC on ocfs2 storage ocfs2 is 1.4.7-1 for the above kernel (downloaded from oracle oss site) Recently both servers have been crashing with the following error: Assertion failure in journal_dirty_metadata() at fs/jbd/transaction.c:1130: "handle->h_buffer_credits > 0" kernel BUG in
2009 Feb 26
3
[PATCH 0/3] ocfs2-1.4: Backport inode alloc from mainline.
Hi all, this patch set are the backport of inode alloc improvement from mainline to ocfs2-1.4. the patches are almost the same excpet one thing: Joel has added JBD2 support to ocfs2, so he has added "max_blocks" to alloc_context and add a new function "ocfs2_reserve_clusters_with_limit". We don't have that in ocfs2-1.4. So there are some great difference in patch 2.
2011 Aug 10
1
fsck hangs in Pass 0a
Hello list, I have a ~10TB ocfs2 filesystem in a 8-node cluster. This sits on a logical volume (I know lv is not cluster aware, but I make sure no one touches the lv, while the cluster is running). The LV consists of 5x2TB multipath devices. I recently had errors like this on some nodes: OCFS2: ERROR (device dm-7): ocfs2_check_group_descriptor: Group Descriptor # 0 has bad signature File
2010 May 20
0
[GIT PULL] ocfs2 updates for 2.6.35
Linus et al, Here are the ocfs2 updates for 2.6.35. There are two major changes. Mark added allocation reservations to our node-local allocators. This gets us much more contiguousness when many processes are growing files in parallel. On the other end of contiguousness, ocfs2 has had a major limitation since the beginning. When ocfs2 needs new inodes, it has always grown its metadata
2012 Jun 14
0
[ocfs2-announce] OCFS2 1.4.10-1 released
All, We are pleased to announce the release of OCFS2 1.4.10-1 and OCFS2 tools 1.6.3-2 for Oracle Linux 5 Update 7 and higher and Redhat Enterprise Linux 5 Update 7 and higher. Oracle's Unbreakable Linux Network users who are subscribing to the "OCFS2 1.4 packages for Enterprise Linux 5" channel can upgrade to this release by running up2date. Red Hat's Enterprise Linux 5
2007 Jun 23
1
[LLVMdev] on-the-fly recompilation/relinking
Hi Assuming I had an application stub based on (say) lli, and I'd like to allow a hypothetical IDE to recompile and relink a function: Is that supported in any way currently? Assuming actually patching all call sites to point to the newly generated function is possible, I guess what I need is to be able to either track or find all of the call sites. The trickiest bit seems to be if the
2018 Jul 23
2
Relinking (syscall-free) ELF executable into Mach-O and PE executables
Couldn’t you write the relocations to the ELF executable? I don’t know if current linkers have support for this, but it seems possible in theory to make a relinkable executable. If you want to do this with an already linked executable though, then yea this won’t be possible. On Fri, Jul 20, 2018 at 10:30 AM Reid Kleckner via llvm-dev < llvm-dev at lists.llvm.org> wrote: > Typically
2018 Jul 20
4
Relinking (syscall-free) ELF executable into Mach-O and PE executables
Hi, Let's suppose we have an ELF executable that doesn't issue any syscall (I mean, syscalls are issued from an external dynamic library, not from the executable, and we can ignore such dynamic library because we have the proper equivalent library with the proper syscalls in MacOS and Windows). So, the question: Is it "currently possible" (by "currently possible" I
2008 Sep 04
4
[PATCH 0/3] ocfs2: Switch over to JBD2.
ocfs2 currently uses the Journaled Block Device (JBD) for its journaling. This is a very stable and tested codebase. However, JBD is limited by architecture to 32bit block numbers. This means an ocfs2 filesystem is limited to 2^32 blocks. With a 4K blocksize, that's 16TB. People want larger volumes. Fortunately, there is now JBD2. JBD2 adds 64bit block number support and some other
2009 Feb 24
2
[PATCH 1/3] ocfs2: Optimize inode allocation by remembering last group.
In ocfs2, the inode block search looks for the "emptiest" inode group to allocate from. So if an inode alloc file has many equally (or almost equally) empty groups, new inodes will tend to get spread out amongst them, which in turn can put them all over the disk. This is undesirable because directory operations on conceptually "nearby" inodes force a large number of seeks. So
2009 Jan 15
5
[PATCH 0/3] ocfs2: Inode Allocation Strategy Improvement.v2
Changelog from V1 to V2: 1. Modify some codes according to Mark's advice. 2. Attach some test statistics in the commit log of patch 3 and in this e-mail also. See below. Hi all, In ocfs2, when we create a fresh file system and create inodes in it, they are contiguous and good for readdir+stat. While if we delete all the inodes and created again, the new inodes will get spread out and that
2009 Apr 17
26
OCFS2 1.4: Patches backported from mainline
Please review the list of patches being applied to the ocfs2 1.4 tree. All patches list the mainline commit hash. Thanks Sunil
2009 Mar 17
33
[git patches] Ocfs2 updates for 2.6.30
Hi, The following patches comprise the bulk of Ocfs2 updates for the 2.6.30 merge window. Aside from larger, more involved fixes, we're adding the following features, which I will describe in the order their patches are mailed. Sunil's exported some more state to our debugfs files, and consolidated some other aspects of our debugfs infrastructure. This will further aid us in debugging