similar to: ext3_free_blocks_sb: bit already cleared for block

Displaying 20 results from an estimated 10000 matches similar to: "ext3_free_blocks_sb: bit already cleared for block"

2007 Jan 11
4
Help understanding some benchmark results
G''day, all, So, I''ve decided to migrate my home server from Linux+swRAID+LVM to Solaris+ZFS, because it seems to hold much better promise for data integrity, which is my primary concern. However, naturally, I decided to do some benchmarks in the process, and I don''t understand why the results are what they are. I though I had a reasonable understanding of ZFS, but now
2009 Jun 24
3
Unexplained reboots in DRBD82 + OCFS2 setup
We're trying to setup a dual-primary DRBD environment, with a shared disk with either OCFS2 or GFS. The environment is a Centos 5.3 with DRBD82 (but also tried with DRBD83 from testing) . Setting up a single primary disk and running bonnie++ on it works. Setting up a dual-primary disk, only mounting it on one node (ext3) and running bonnie++ works When setting up ocfs2 on the /dev/drbd0
2005 Oct 31
4
Best mkfs.ext2 performance options on RAID5 in CentOS 4.2
I can't seem to get the read and write performance better than approximately 40MB/s on an ext2 file system. IMO, this is horrible performance for a 6-drive, hardware RAID 5 array. Please have a look at what I'm doing and let me know if anybody has any suggestions on how to improve the performance... System specs: ----------------- 2 x 2.8GHz Xeons 6GB RAM 1 3ware 9500S-12 2 x 6-drive,
2009 Jun 24
3
Unexplained reboots in DRBD82 + OCFS2 setup
We're trying to setup a dual-primary DRBD environment, with a shared disk with either OCFS2 or GFS. The environment is a Centos 5.3 with DRBD82 (but also tried with DRBD83 from testing) . Setting up a single primary disk and running bonnie++ on it works. Setting up a dual-primary disk, only mounting it on one node (ext3) and running bonnie++ works When setting up ocfs2 on the /dev/drbd0
2009 Jan 10
3
Poor RAID performance new Xeon server?
I have just purchased an HP ProLiant HP ML110 G5 server and install ed CentOS 5.2 x86_64 on it. It has the following spec: Intel(R) Xeon(R) CPU 3065 @ 2.33GHz 4GB ECC memory 4 x 250GB SATA hard disks running at 1.5GB/s Onboard RAID controller is enabled but at the moment I have used mdadm to configure the array. RAID bus controller: Intel Corporation 82801 SATA RAID Controller For a simple
2004 Sep 04
1
"bit already cleared" messages
There's been a lot of discussion in the past on these somewhat-mysterious "bit already cleared" messages appearing in the logs when using Ext3 file systems. Unfortunately, I didn't see a conclusion to any of these threads. We're still seeing these messages pop up occasionally and fsck's of the file system reveal lot of orphaned inodes and such. Anybody else see these
2001 Nov 26
1
EXT3 crash: ext3_free_blocks: bit already cleared
I had a failure over the weekend of my ext3 filesystems. Hardware: Dell 2550 (dual PIII 1 Ghz, 1GB ram, Acenic GigE card) Software: linux-2.4.14 ext3-2.4-0.9.15-2414 Trond's NFS patches (pathconf, tune,read,rpc_blk) The filesystem is RAID0 using LVM (1.0.1-rc4). The disks are connected using fiber channel (qlogic 2200, qlax00-4.27beta). I was generating high load on my nfs
2006 Oct 04
2
server disk subsystem benchmarks, bonnie++ and/or others?
Greetings I've searched to no avail so far... there is bound to be something more intelligible out there...??? I am playing with bonnie++ for the first time... May I please get some advise and list experience on using this or other disk subsystem benchmark programs properly with or without a GUI ? Test system in this case is a Compaq DL360 with 2 to 4 Gig DRAM and qty (2) 36Gig 10k drives
2012 Jan 04
9
Stress test zfs
Hi all, I''ve got a solaris 10 running 9/10 on a T3. It''s an oracle box with 128GB memory RIght now oracle . I''ve been trying to load test the box with bonnie++. I can seem to get 80 to 90 K writes, but can''t seem to get more than a couple K for writes. Any suggestions? Or should I take this to a bonnie++ mailing list? Any help is appreciated. I''m kinda
2007 Nov 26
4
Filesystem for Maildir
<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN"> <html> <head> </head> <body bgcolor="#ffffff" text="#000000"> <font size="-1"><font face="Verdana">Hi all,<br> <br> In last year, i had made some research and benchmarks based on CentOS 4 to know which filesystem is better for
2005 May 11
5
Xen reboots on dom-U disk stress...
Hi all I tried to run bonnie++ disk stresser in dom-U, who''s disk is backed with non-local (on nfs) loop-back file. The machine rebooted pretty quickly. So, how do I tell what''s barfing? Is it Xen? Is it dom-0 (nfs or loop-back)? I looked in dom-0''s /var/log/messages and didn''t see any obvious record of a dom-0 whoopsie (but is that the right place to look
2009 Dec 24
6
benchmark results
I've had the chance to use a testsystem here and couldn't resist running a few benchmark programs on them: bonnie++, tiobench, dbench and a few generic ones (cp/rm/tar/etc...) on ext{234}, btrfs, jfs, ufs, xfs, zfs. All with standard mkfs/mount options and +noatime for all of them. Here are the results, no graphs - sorry: http://nerdbynature.de/benchmarks/v40z/2009-12-22/ Reiserfs
2009 Dec 24
6
benchmark results
I've had the chance to use a testsystem here and couldn't resist running a few benchmark programs on them: bonnie++, tiobench, dbench and a few generic ones (cp/rm/tar/etc...) on ext{234}, btrfs, jfs, ufs, xfs, zfs. All with standard mkfs/mount options and +noatime for all of them. Here are the results, no graphs - sorry: http://nerdbynature.de/benchmarks/v40z/2009-12-22/ Reiserfs
2003 Sep 22
3
Fwd: privsep in ssh
It was suggested to me that I forward this message to you. ---------- Forwarded Message ---------- Subject: privsep in ssh Date: Fri, 19 Sep 2003 12:22 From: Russell Coker <russell at coker.com.au> To: SE Linux <selinux at tycho.nsa.gov> Cc: Colin Watson <cjwatson at debian.org> #ifdef DISABLE_FD_PASSING if (1) { #else if (authctxt->pw->pw_uid == 0 ||
2005 Jul 14
1
a comparison of ext3, jfs, and xfs on hardware raid
I'm setting up a new file server and I just can't seem to get the expected performance from ext3. Unfortunately I'm stuck with ext3 due to my use of Lustre. So I'm hoping you dear readers will send me some tips for increasing ext3 performance. The system is using an Areca hardware raid controller with 5 7200RPM SATA disks. The RAID controller has 128MB of cache and the disks
2008 Nov 13
7
Kernel oops when running bonnie++ on btrfs
I wanted to see how btrfs compares to other filesystems so I have been running bonnie++ on it. While the results are good(much faster then ext2) every once in awhile I get a kernel oops. I am testing on xubuntu 8.10 with the 2.6.27-7-686 kernel using the latest git sources. Most of the time the oops happens within 20min of running bonnie++ but sometimes it takes a few hours. This happens with and
2008 Feb 27
8
Physical disks + Software Raid + lvm in domU
Hi, I''m trying to setup my box as described below: - The Dom0 exports two disks as physical devices to the DomU - DomU use the two devices to assemble a software raid device (/dev/md0) - On the /dev/md0 I create a lvm volume group and lvm volumes on it. Everything seems to work fine if the lvm volumes on the DomU are lightly used. Under heavy load the DomU freeze up immediately. The
2007 Nov 08
1
XEN HVMs on LVM over iSCSI - test results, (crashes) and questions
Hi all, some results from my configuration. I''ve this configuration (I''m not interested in raw top performances but in reliability.. so, I can accept slow MB/sec and prefer to rely on a RAID6, for example): 1 Infortrend iSCSI Array A16E-G2130-4 with: - 1GB DDR cache - RAID6 - 7 x 500GB sataII Seagate ST3500630NS with 16mb (no budget for SAS) - one of the logic volumes (about
2001 Aug 26
2
ext2 vs. ext3
Hi, I just tested ext2 and ext3(2.4-0.9.6) with 2.4.8-ac8, using bonnie++-1.01. The hardware is an old Pentium 100 box with 48M RAM and a 16.6MB/s EIDE harddrive. The test results(see attachment) show ext3 is a lot slower than ext2 in most aspects. Does anybody get better results? Wenzhuo
2011 Nov 23
1
SSD diagnostics / test suite?
Hey folks, I looked back through the list archives and there are surprisingly few threads with "SSD" in the subject. In my new job I've been handed over a number of things that were outstanding with the previous Sys Admin, and one of them was an SSD that was suspect. I just plugged it into a diagnostics laptop to use the Linux Smart software to check it - and it all seems fine