similar to: Re: [PATCH] open files in kjounald (2)

Displaying 20 results from an estimated 300 matches similar to: "Re: [PATCH] open files in kjounald (2)"

2002 Apr 22
3
[PATCH] open files in kjounald
Hello everybody! As I wrote in my mail the previous week ("BUG: 2.4.19pre1 & journal_thread & open filehandles") I followed the problem a little bit further. Here's my patch; I beg the ext3-maintainers (Stephen, Andreas, Andrew) to have a look at it and submit it to Marcelo and Linux for inclusion. (2.4 is for now for me more important than 2.5). Patch ----- On every
2002 Aug 22
1
kjounald takes too much cpu
Hi all, The problem I have now is similar to previous posts in this list before, but I couldn't find the solution. If the solution is there, please refer me to that, and I apologize for not being able to find it. I recently installed RH7.3, with ext3 fs, on several identical computers. The processor is AMD 1900+, the motherboard is ASUS A7V266-C, and the hard drive is a Maxtor 40GB, 7200
2002 Dec 07
6
kjournald using up majority cpu%.
We run several RH7.2/7.3 servers & recently 2 of them, although still working fine, have started to show kjournald as generally using over 50% cpu% on 'top' - virtually continuously. Free cpu% generally less than 25% now! Both machines are also using software RAID5 EIDE ....... otherwise are standard server installs. Any info on what kjournald is & why it should have recently
2023 May 22
1
[PATCH 1/3] signal: Don't always put SIGKILL in shared_pending
When get_pending detects the task has been marked to be killed we try to clean up the SIGKLL by doing a sigdelset and recalc_sigpending, but we still leave it in shared_pending. If the signal is being short circuit delivered there is no need to put in shared_pending so this adds a check in complete_signal. This patch was modified from Eric Biederman <ebiederm at xmission.com> original
2015 Feb 11
4
Fwd: samba & Oracle ACFS Issues
On Wed, Feb 11, 2015 at 10:27:50AM +0100, Nacho del Rey wrote: > Hi Volker and the list > > Yesterday the problem appeared again > > I could launch a strace command over a samba PID to see what was going on > and this is the result > > [PRO] [root at pf3il0024 ~]# strace -frp 57686 > Process 57686 attached - interrupt to quit > 0.000000 fcntl(14, F_SETLKW,
2014 Jun 09
36
[Bug 79823] New: Mozilla apps freeze on startup with nouveau-dri-10.2.1 libs on dual-screen
https://bugs.freedesktop.org/show_bug.cgi?id=79823 Priority: medium Bug ID: 79823 Assignee: nouveau at lists.freedesktop.org Summary: Mozilla apps freeze on startup with nouveau-dri-10.2.1 libs on dual-screen Severity: normal Classification: Unclassified OS: Linux (All) Reporter: natrio at
2012 Jul 17
1
Unmountable btrfs filesystem
Hi devs, I can''t mount my btrfs-based external disk. I don''t know what happened to the disk, but usually it gets disconnected (usb cable pulled out) buy my 8 months old daughter. Here''s the output from dmesg: [ 299.699543] parent transid verify failed on 528952573952 wanted 22726 found 22728 [ 299.700947] btrfs read error corrected: ino 1 off 528952573952 (dev
2023 Jun 13
1
Upssched 100% CPU after updating Debian 12
Hi, Thanks Jim! Here is more system information from the commands you mentioned. Kari root at fricka:~# lsof -p 1716171 lsof: WARNING: can't stat() fuse.gvfsd-fuse file system /run/user/1002/gvfs Output information may be incomplete. COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME upssched 1716171 root cwd DIR 8,2 4096 2 /
2023 Jun 13
1
Upssched 100% CPU after updating Debian 12
Hi, Thanks Jim! Here is more system information from the commands you mentioned. Kari root at fricka:~# lsof -p 1716171 lsof: WARNING: can't stat() fuse.gvfsd-fuse file system /run/user/1002/gvfs Output information may be incomplete. COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME upssched 1716171 root cwd DIR 8,2 4096 2 /
2006 Dec 15
12
Numbering of releases
Am I right in thinking that the numbering of the versions are in synch with the gcc version that one is recommened to build with? In other words "use gcc version 3.foo". -- John Maclean - 07739 171 531 MSc (DIC) _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
2023 Jun 13
2
Upssched 100% CPU after updating Debian 12
FWIW, I wonder if this is a fallout of PR #1274 : https://github.com/networkupstools/nut/pull/1274/files and specifically https://github.com/networkupstools/nut/commit/550064930e369fb3a322c3b28ffff8b4acf53532 which added a `sock_read()` loop continuation effectively if `pconf_char()` returned 0. Just grasping at straws here (since it is a change a couple of months before the release was cut),
2023 Jun 13
2
Upssched 100% CPU after updating Debian 12
FWIW, I wonder if this is a fallout of PR #1274 : https://github.com/networkupstools/nut/pull/1274/files and specifically https://github.com/networkupstools/nut/commit/550064930e369fb3a322c3b28ffff8b4acf53532 which added a `sock_read()` loop continuation effectively if `pconf_char()` returned 0. Just grasping at straws here (since it is a change a couple of months before the release was cut),
2006 Aug 26
3
Reproduced the Login process died too early issue.
Hi. I tried for some time to reproduce the "Login process died too early issue". It seems related to when the system run out of file descriptors, to force the issue i forwarded 3k of messages trough the server(using thunderbird). I also tried with my usual mail application - mail.app, but no luck there. Aug 26 18:46:36 soekris postfix/cleanup[14739]: fatal: accept
2014 Apr 07
3
Software RAID10 - which two disks can fail?
Hi All. I have a server which uses RAID10 made of 4 partitions for / and boots from it. It looks like so: mdadm -D /dev/md1 /dev/md1: Version : 00.90 Creation Time : Mon Apr 27 09:25:05 2009 Raid Level : raid10 Array Size : 973827968 (928.71 GiB 997.20 GB) Used Dev Size : 486913984 (464.36 GiB 498.60 GB) Raid Devices : 4 Total Devices : 4 Preferred Minor : 1
2023 Jun 13
3
Upssched 100% CPU after updating Debian 12
After launching the command several times, with debug (posted by new code in a new branch for the investigation) confirming that the same daemon handles operations from the new client instances, its strace now has numerous FDs to report after select() - so I guess it is a problem of detecting an exit of the counterpart. 0.000000 [D2] parse_at: is 'heartbeat at localhost' in AT
2023 Jun 13
3
Upssched 100% CPU after updating Debian 12
After launching the command several times, with debug (posted by new code in a new branch for the investigation) confirming that the same daemon handles operations from the new client instances, its strace now has numerous FDs to report after select() - so I guess it is a problem of detecting an exit of the counterpart. 0.000000 [D2] parse_at: is 'heartbeat at localhost' in AT
2015 May 16
5
[Bug 90482] New: Xorg take 100% CPU when using multiple independent screen configuration
https://bugs.freedesktop.org/show_bug.cgi?id=90482 Bug ID: 90482 Summary: Xorg take 100% CPU when using multiple independent screen configuration Product: xorg Version: unspecified Hardware: x86-64 (AMD64) OS: Linux (All) Status: NEW Severity: normal Priority: medium
2008 Apr 24
1
Utility tool for dm-ioband.
Hi everyone, I made a utility tool for dm-ioband version 0.0.4, named iobandctl. It enables you to easily apply I/O bandwidth control to an entire disk, and manage it. It helps you set the percentage of bandwidth to give each partition, and each user, process, group, or cgroup. (You are not able to use cgroup support yet, because the dm-ioband patch to enable cgroup support
2008 Apr 24
1
Utility tool for dm-ioband.
Hi everyone, I made a utility tool for dm-ioband version 0.0.4, named iobandctl. It enables you to easily apply I/O bandwidth control to an entire disk, and manage it. It helps you set the percentage of bandwidth to give each partition, and each user, process, group, or cgroup. (You are not able to use cgroup support yet, because the dm-ioband patch to enable cgroup support
2008 Apr 24
1
Utility tool for dm-ioband.
Hi everyone, I made a utility tool for dm-ioband version 0.0.4, named iobandctl. It enables you to easily apply I/O bandwidth control to an entire disk, and manage it. It helps you set the percentage of bandwidth to give each partition, and each user, process, group, or cgroup. (You are not able to use cgroup support yet, because the dm-ioband patch to enable cgroup support