Hi forum,
I''m currently a little playing around with ZFS on my workstation.
I created a standard mirrored pool over 2 disk-slices.
# zpool status
 Pool: mypool
 Status: ONLINE
 scrub: Keine erforderlich
config:
        NAME          STATE     READ WRITE CKSUM
        mypool        ONLINE       0     0     0
          mirror        ONLINE       0     0     0
            c0t0d0s4  ONLINE       0     0     0
            c0t2d0s4  ONLINE       0     0     0
Then i created a ZFS with no extra options:
# zfs create mypool/zfs01
# zfs list
NAME                   USED  AVAIL  REFER  MOUNTPOINT
mypool                 106K  27,8G  25,5K  /mypool
mypool/zfs01          24,5K  27,8G  24,5K  /mypool/zfs01
When I now send a mkfile on the new FS, the performance of the whole system
breaks down near zero:
# mkfile 5g test
last pid: 25286;  load avg:  3.54,  2.28,  1.29;       up 0+01:44:26            
16:16:24
66 processes: 61 sleeping, 3 running, 1 zombie, 1 on cpu
CPU states:  0.0% idle,  2.1% user, 97.9% kernel,  0.0% iowait,  0.0% swap
Memory: 512M phys mem, 65M free mem, 2050M swap, 2050M free swap
   PID USERNAME LWP PRI NICE  SIZE   RES STATE    TIME    CPU COMMAND
 25285 root       1   8    4 1184K  752K run      0:09 66.28% mkfile
It seams that some kind of kernel activity while writing to ZFS blocks the
system.
Is this a known problem? Do you need additional information?
regards
Mathias
 
 
This message posted from opensolaris.org
Jürgen Keil
2006-Sep-15  15:07 UTC
[zfs-discuss] Re: [Blade 150] ZFS: extreme low performance
The disks in that Blade 100, are these IDE disks? The performance problem is probably bug 6421427: http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6421427 A fix for the issue was integrated into the Opensolaris 20060904 source drop (actually closed binary drop): http://dlc.sun.com/osol/on/downloads/20060904/on-changelog-20060904.html ... but has been removed in the next update: http://dlc.sun.com/osol/on/downloads/20060911/on-changelog-20060911.html This message posted from opensolaris.org