similar to: ext2 vs. ext3

Displaying 20 results from an estimated 1000 matches similar to: "ext2 vs. ext3"

2005 Oct 31
4
Best mkfs.ext2 performance options on RAID5 in CentOS 4.2
I can't seem to get the read and write performance better than approximately 40MB/s on an ext2 file system. IMO, this is horrible performance for a 6-drive, hardware RAID 5 array. Please have a look at what I'm doing and let me know if anybody has any suggestions on how to improve the performance... System specs: ----------------- 2 x 2.8GHz Xeons 6GB RAM 1 3ware 9500S-12 2 x 6-drive,
2012 Mar 07
2
hardware issues? driver issues?
Got a bunch of servers from Penguin. Supermicro m/b's H8QG6. We put a 3tb drive in for additional workspace for the users, and some of them won't read, others will go for weeks, then spit out DRDY errors. lshw shows the controller as an ATI SB7x0/SB8x0/SB9x0 SATA. I did notice that it shows *-storage description: SATA controller product: SB7x0/SB8x0/SB9x0 SATA
2001 Sep 05
3
[e2fsprogs-1.24] "fsck -A -a" fails on reboot
Dear Ted, I upgraded e2fsprogs and util-linux to the latest versions, as per the instructions on the "ext3 for 2.4" page, to make the switching between ext2-only and ext3-enabled kernels seamless. Now that if the filesystems have not been unmounted cleanly, due to a power failure for example, "fsck -A -a" cannot continue after checking the root filesystem on reboot, issuing
2009 Jun 24
3
Unexplained reboots in DRBD82 + OCFS2 setup
We're trying to setup a dual-primary DRBD environment, with a shared disk with either OCFS2 or GFS. The environment is a Centos 5.3 with DRBD82 (but also tried with DRBD83 from testing) . Setting up a single primary disk and running bonnie++ on it works. Setting up a dual-primary disk, only mounting it on one node (ext3) and running bonnie++ works When setting up ocfs2 on the /dev/drbd0
2009 Jun 24
3
Unexplained reboots in DRBD82 + OCFS2 setup
We're trying to setup a dual-primary DRBD environment, with a shared disk with either OCFS2 or GFS. The environment is a Centos 5.3 with DRBD82 (but also tried with DRBD83 from testing) . Setting up a single primary disk and running bonnie++ on it works. Setting up a dual-primary disk, only mounting it on one node (ext3) and running bonnie++ works When setting up ocfs2 on the /dev/drbd0
2009 Jan 10
3
Poor RAID performance new Xeon server?
I have just purchased an HP ProLiant HP ML110 G5 server and install ed CentOS 5.2 x86_64 on it. It has the following spec: Intel(R) Xeon(R) CPU 3065 @ 2.33GHz 4GB ECC memory 4 x 250GB SATA hard disks running at 1.5GB/s Onboard RAID controller is enabled but at the moment I have used mdadm to configure the array. RAID bus controller: Intel Corporation 82801 SATA RAID Controller For a simple
2006 Oct 04
2
server disk subsystem benchmarks, bonnie++ and/or others?
Greetings I've searched to no avail so far... there is bound to be something more intelligible out there...??? I am playing with bonnie++ for the first time... May I please get some advise and list experience on using this or other disk subsystem benchmark programs properly with or without a GUI ? Test system in this case is a Compaq DL360 with 2 to 4 Gig DRAM and qty (2) 36Gig 10k drives
2007 Jan 11
4
Help understanding some benchmark results
G''day, all, So, I''ve decided to migrate my home server from Linux+swRAID+LVM to Solaris+ZFS, because it seems to hold much better promise for data integrity, which is my primary concern. However, naturally, I decided to do some benchmarks in the process, and I don''t understand why the results are what they are. I though I had a reasonable understanding of ZFS, but now
2012 Jan 04
9
Stress test zfs
Hi all, I''ve got a solaris 10 running 9/10 on a T3. It''s an oracle box with 128GB memory RIght now oracle . I''ve been trying to load test the box with bonnie++. I can seem to get 80 to 90 K writes, but can''t seem to get more than a couple K for writes. Any suggestions? Or should I take this to a bonnie++ mailing list? Any help is appreciated. I''m kinda
2007 Nov 26
4
Filesystem for Maildir
<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN"> <html> <head> </head> <body bgcolor="#ffffff" text="#000000"> <font size="-1"><font face="Verdana">Hi all,<br> <br> In last year, i had made some research and benchmarks based on CentOS 4 to know which filesystem is better for
2002 Jan 05
2
e2fsck: bad magic number in super-block
Hi, I'm on kernel 2.14.13-ac8 running Redhat 7.1. I've successfully converted all my file system to ext3 and is working fine for a couple of weeks. However, I did a silly thing when I e2fsck (version 1.23) an ext3 /home partiction without umounting it. Now, I believe my partiton super-block is corrupted and the system wasn't able to mount /home. I tried as suggested to do a e2fsck -b
2005 May 11
5
Xen reboots on dom-U disk stress...
Hi all I tried to run bonnie++ disk stresser in dom-U, who''s disk is backed with non-local (on nfs) loop-back file. The machine rebooted pretty quickly. So, how do I tell what''s barfing? Is it Xen? Is it dom-0 (nfs or loop-back)? I looked in dom-0''s /var/log/messages and didn''t see any obvious record of a dom-0 whoopsie (but is that the right place to look
2009 Dec 24
6
benchmark results
I've had the chance to use a testsystem here and couldn't resist running a few benchmark programs on them: bonnie++, tiobench, dbench and a few generic ones (cp/rm/tar/etc...) on ext{234}, btrfs, jfs, ufs, xfs, zfs. All with standard mkfs/mount options and +noatime for all of them. Here are the results, no graphs - sorry: http://nerdbynature.de/benchmarks/v40z/2009-12-22/ Reiserfs
2009 Dec 24
6
benchmark results
I've had the chance to use a testsystem here and couldn't resist running a few benchmark programs on them: bonnie++, tiobench, dbench and a few generic ones (cp/rm/tar/etc...) on ext{234}, btrfs, jfs, ufs, xfs, zfs. All with standard mkfs/mount options and +noatime for all of them. Here are the results, no graphs - sorry: http://nerdbynature.de/benchmarks/v40z/2009-12-22/ Reiserfs
2003 Sep 22
3
Fwd: privsep in ssh
It was suggested to me that I forward this message to you. ---------- Forwarded Message ---------- Subject: privsep in ssh Date: Fri, 19 Sep 2003 12:22 From: Russell Coker <russell at coker.com.au> To: SE Linux <selinux at tycho.nsa.gov> Cc: Colin Watson <cjwatson at debian.org> #ifdef DISABLE_FD_PASSING if (1) { #else if (authctxt->pw->pw_uid == 0 ||
2005 Jul 14
1
a comparison of ext3, jfs, and xfs on hardware raid
I'm setting up a new file server and I just can't seem to get the expected performance from ext3. Unfortunately I'm stuck with ext3 due to my use of Lustre. So I'm hoping you dear readers will send me some tips for increasing ext3 performance. The system is using an Areca hardware raid controller with 5 7200RPM SATA disks. The RAID controller has 128MB of cache and the disks
2007 Feb 20
1
Digital Audio (via EIDE ribbon cables) supported in Kernel in CentOS 5.0?
Hello: Will Digital Audio (via EIDE ribbon cables) be supported in the Kernel in CentOS 5.0? I'm now using CentOS 4.4 on a Dell Dimension 2400 box (Celeron 2.6 GHz CPU). This is a dual boot box, rarely booted into Windows XP, but when I do use Windows, and I place an Audio CD into the DVD drive or CD-RW drive, I can hear the music. Using CentOS 4.4, I can see that the software and hardware are
2008 Nov 13
7
Kernel oops when running bonnie++ on btrfs
I wanted to see how btrfs compares to other filesystems so I have been running bonnie++ on it. While the results are good(much faster then ext2) every once in awhile I get a kernel oops. I am testing on xubuntu 8.10 with the 2.6.27-7-686 kernel using the latest git sources. Most of the time the oops happens within 20min of running bonnie++ but sometimes it takes a few hours. This happens with and
2003 Apr 08
5
Status of SPARC64 port?
I was looking at the Sparc 64 4.7 or 4.8 port, wondering if its in any state to put on my Solaris v100. Just need the basics, eide, network, kbm,mouse, graphics (x). A bonus would be support for the lom (light out management) features and the flash card. Reason for change is I HOPE that the math (ssl) performance is better on FBSD, as well as I am just tired of keeping track of
2008 Feb 27
8
Physical disks + Software Raid + lvm in domU
Hi, I''m trying to setup my box as described below: - The Dom0 exports two disks as physical devices to the DomU - DomU use the two devices to assemble a software raid device (/dev/md0) - On the /dev/md0 I create a lvm volume group and lvm volumes on it. Everything seems to work fine if the lvm volumes on the DomU are lightly used. Under heavy load the DomU freeze up immediately. The