similar to: debian ocfs2 debug fs_locks

Displaying 20 results from an estimated 4000 matches similar to: "debian ocfs2 debug fs_locks"

2014 May 06
0
poor write performance or locking issues with ocfs2
Hello all, I've got heavy troubles with my ocfs2 environment. Cluster filesystem worked fine for about 3-6 weeks after initial setup, but since 1 week performance issues occurs. I've already searched long time in google and on this mailing list but I wasn't able to found any solution. I've found a lot of posts with "same" problems but without the magic answer :-)
2014 Sep 10
1
How to unlock a bloked resource? Thanks
Hi All: As we test with two node in one OCFS2 cluster. The cluster is hang up may be for dead lock. We use the debugfs.ocfs tool founding that one resource is holding by one node who has it for long time and another node can still wait for the resource. So the cluster is hang up. debugfs.ocfs2 -R "fs_locks -B" /dev/dm-0 debugfs.ocfs2 -R "dlm_locks LOCKID_XXX" /dev/dm-0 How
2014 Sep 10
1
How to unlock a bloked resource? Thanks
Hi All: As we test with two node in one OCFS2 cluster. The cluster is hang up may be for dead lock. We use the debugfs.ocfs tool founding that one resource is holding by one node who has it for long time and another node can still wait for the resource. So the cluster is hang up. debugfs.ocfs2 -R "fs_locks -B" /dev/dm-0 debugfs.ocfs2 -R "dlm_locks LOCKID_XXX" /dev/dm-0 How
2018 Jan 15
1
Aw: Re: Lmtp Memory Limit
2016 Mar 08
0
OCFS2 showing "No space left on device" on a device with free space
Good day Guys We just had an issue on one of our clients servers, whereby the server was claiming there was no more space, but df showed there there plenty of space (even checked the inodes). We were really stummped, but after Googling around, the following command was needed to be run tunefs.ocfs2 --fs-features=discontig-bg /dev/drbd0 Here is the output before tunefs.ocfs2 was run user at
2014 Sep 26
2
One node hangs up issue requiring goog idea, thanks
Hi, all, As we use OCFS2, the network is not good. When the converting request message can?t send to the another node, there will be a node hangs up which will still waiting for the dlm. CAS2/logdir/var/log/syslog.1-6778-Sep 16 20:57:16 CAS2 kernel: [516366.623623] o2net: Connection to node CAS1 (num 1) at 10.172.254.1:7100 has been idle for 30.87 secs, shutting it down.
2020 Sep 16
1
another problem with 2.3.36.4 after update os
Hi A few days ago I upgraded debian8 to debian9 dovecot is from source # 2.2.36.4 (baf9232c1): /etc/dovecot/dovecot.conf # Pigeonhole version 0.4.24.2 (aaba65b7) # OS: Linux 4.9.0-13-amd64 x86_64 Debian 9.13 Today I get some times in logs: "dovecot-uidlist line 112: Invalid data" and I dont know why This is claster dovecot: dovecot1 - debian8 dovecot2 - debian8 dovecot3 - debian8
2018 Jan 14
2
Lmtp Memory Limit
Hi, i am using dovecot 2.2.33.2 on CentOS 7.4. Since i upgraded from CentOS 7.2. to CentOS 7.4. (without upgrading dovecot), my dovecot sieve-pipe scripts crash with Out of memory: Out of memory (allocated 262144) (tried to allocate 8793 bytes) There are some memory limits in dovecot or sieve? Can i change this value? Kernel limitks: [root at xxx software]# ulimit -a core file size
2006 Jul 04
2
ocfs2-tools-1.2.0-1 compile error
Hello, i want to install ocfs2-tools-1.2.0-1 on slackware 10.2 (up-to-date) with 2.6.17.2 kernel and i got error: ./configure checking build system type... i686-pc-linux-gnu checking host system type... i686-pc-linux-gnu checking for gcc... gcc checking for C compiler default output file name... a.out checking whether the C compiler works... yes checking whether we are cross compiling... no
2008 Sep 01
1
(no subject)
Hello, We just experienced a hang that looks superficially very similar to http://www.mail-archive.com/ocfs2-users at oss.oracle.com/msg02359.html There are 3 nodes in the cluster ocfs2-1.4.1 rhel 5.2. Versions, uname's in the attached text file which also includes fs_locks dumps and various other diagnostics. The lock up happened when we were restarting a java application that was
2006 Nov 21
2
Memory leak in ocfs2/dlm?
Hi! Seems we're facing some memory leak here. This is vanilla 2.6.19-rc6 on a x86_64 box, 4GB RAM. A simple `ls -Rn' on a filesystem with lots of files makes the box leak so much RAM that the OOM killer starts to kick in. With slab alloc debugging turned on, we see this: # mount; ls -Rn; wait some seconds; Ctrl-C [root@lnxp-1038:/backend1]$ cat /proc/slab_allocators | egrep
2011 Dec 20
8
ocfs2 - Kernel panic on many write/read from both
Sorry i don`t copy everything: TEST-MAIL1# echo "ls //orphan_dir:0000"|debugfs.ocfs2 /dev/dm-0|wc debugfs.ocfs2 1.6.4 5239722 26198604 246266859 TEST-MAIL1# echo "ls //orphan_dir:0001"|debugfs.ocfs2 /dev/dm-0|wc debugfs.ocfs2 1.6.4 6074335 30371669 285493670 TEST-MAIL2 ~ # echo "ls //orphan_dir:0000"|debugfs.ocfs2 /dev/dm-0|wc debugfs.ocfs2 1.6.4 5239722 26198604
2006 Aug 24
0
debugfs.ocfs2 usage message fix
Hi all, The help message in debufs.ocfs2 is not consist among ``extent'', ``findpath'', ``locate'', ``ncheck'' and ``help''. ``help'' show that extent accept a block number as its argument. ---8<--- extent <block#> Show extent block ---8<--- But ``extent'' says its argument is an inode number.
2017 Jan 24
2
[Release-testers] [cfe-dev] [4.0.0 Release] Relase Candidate 1 has been tagged
Hi, Looks ok for native MIPS, I have two failures on debian8: Failing Tests (2): XRay-x86_64-linux :: TestCases/Linux/argv0-log-file-name.cc XRay-x86_64-linux :: TestCases/Linux/fixedsize-logging.cc I'll investigate these failures. Otherwise looks ok. I've uploaded the binaries. 9d5a389c20eb5b3071e6a0504b7cf87d clang+llvm-4.0.0-rc1-mipsel-linux-gnu.tar.xz
2006 Nov 02
0
Re: [Ocfs2-tools-devel] Patch for debugfs.ocfs2 ``icheck'' command
http://oss.oracle.com/~smushran/debugfs.ocfs2-icheck-04a.patch The comments have been inlined in the patch. Summary; 1. Follow Tao's recommendation for checking against known holes. The check for the first cluster group can be done before hand, and for the remaining, it can be done as part of the group descriptor check. 2. It should allow users to provide multiple block numbers. See
2008 May 07
1
[PATCH]ocfs2-1.2: Add dput for uuid entry.
In ocfs2-1.2, when we mount a device, a debugfs dir will be created using its uuid. When 2 devices have the same uuid, after the 1st device is mounted, the 2nd one can't be mounted. This is OK. But the problem is that the dentry's reference is added. So when the 1st volume is umounted, none of these 2 volumes can be mounted then. So this fix solves this problem by dputting the
2010 Oct 22
0
[GIT PULL] ocfs2 changes for 2.6.37
Linus, et al, Here are the ocfs2 changes for 2.6.37. There are three major additions. Tao Ma has added readahead to our CoW operations. Sunil Mushran has added a global heartbeat mode, allowing one device heartbeat to support multiple ocfs2 mounts. Finally, Patrick J. LoPresti has done the final work to enable ocfs2 mounts on devices larger than 16TB. The ocfs2 disk format has always
2009 Apr 30
0
[PATCH] ocfs2: Add statistics for the checksum and ecc operations.
It would be nice to know how often we get checksum failures. Even better, how many of them we can fix with the single bit ecc. So, we add a statistics structure. The structure can be installed into debugfs wherever the user wants. For ocfs2, we'll put it in the superblock-specific debugfs directory and pass it down from our higher-level functions. The stats are only registered with
2018 Feb 05
0
releases.llvm.org randomly failing with Service Unavailable
I just ran into this again: $ wget http://releases.llvm.org/4.0.0/clang+llvm-4.0.0-x86_64-linux-gnu-debian8.tar.xz --2018-02-05 13:40:00-- http://releases.llvm.org/4.0.0/clang+llvm-4.0.0-x86_64-linux-gnu-debian8.tar.xz Resolving releases.llvm.org (releases.llvm.org)... 151.101.122.49 Connecting to releases.llvm.org (releases.llvm.org)|151.101.122.49|:80... connected. HTTP request sent, awaiting
2017 Dec 15
0
OCFS2 cluster debian8 / debian9
Hi, On 12/05/2017 11:19 PM, BASSAGET C?dric wrote: > Hello > Retried from scratch; and still have an error when trying to bring up > the second cluster : > > root at LAB-virtm6:/# o2cb register-cluster ocfs2new > o2cb: Internal logic failure while registering cluster 'ocfs2new' > > root at LAB-virtm6:/mnt/vol1_iscsi_san1# o2cb list-clusters > ocfs2 >