similar to: ZFS, COW, write(2), directIO...

Displaying 20 results from an estimated 10000 matches similar to: "ZFS, COW, write(2), directIO..."

2011 Feb 12
10
native ZFS on Linux
I see that Pinguy OS, an uber-Ubuntu o/s, includes native ZFS support. Any pointers to more info on this? -- David
2008 Jan 31
1
simulating directio on zfs?
The big problem that I have with non-directio is that buffering delays program execution. When reading/writing files that are many times larger than RAM without directio, it is very apparent that system response drops through the floor- it can take several minutes for an ssh login to prompt for a password. This is true both for UFS and ZFS. Repeat the exercise with directio on UFS and there is no
2008 Dec 15
15
Need Help Invalidating Uberblock
I have a ZFS pool that has been corrupted. The pool contains a single device which was actually a file on UFS. The machine was accidentally halted and now the pool is corrupt. There are (of course) no backups and I''ve been asked to recover the pool. The system panics when trying to do anything with the pool. root@:/$ zpool status panic[cpu1]/thread=fffffe8000758c80: assertion failed:
2009 Apr 20
6
simulating directio on zfs?
I had to let this go and get on with testing DB2 on Solaris. I had to abandon zfs on local discs in x64 Solaris 10 5/08. The situation was that: * DB2 buffer pools occupied up to 90% of 32GB RAM on each host * DB2 cached the entire database in its buffer pools o having the file system repeat this was not helpful * running high-load DB2 tests for 2 weeks showed 100%
2010 Aug 28
4
ufs root to zfs root liveupgrade?
hi all Try to learn how UFS root to ZFS root liveUG work. I download the vbox image of s10u8, it come up as UFS root. add a new disks (16GB) create zpool rpool run lucreate -n zfsroot -p rpool run luactivate zfsroot run lustatus it do show zfsroot will be active in next boot init 6 but it come up with UFS root, lustatus show ufsroot active zpool rpool is mounted but not used by boot Is this a
2010 Oct 08
3
ZFS equivalent of inotify
Is there a ZFS equivalent (or alternative) of inotify? You have some thing, which wants to be notified whenever a specific file or directory changes. For example, a "live" sync application of some kind... -------------- next part -------------- An HTML attachment was scrubbed... URL:
2008 Nov 08
9
How does zfs COW deal with ''..'' in brother directory?
Hi matt, I have some problems about understanding zfs COW implemention. Suppose b and c are both children dir of a, if c changes, there will be new versions of both a and c, namely c'' and a''. a a'' b c c'' Because ''..'' in b points to a before this change, shall we modify b to let ''..'' point to a''? If yes,
2007 Jun 09
41
zfs reports small st_size for directories?
Why does ZFS report such small directory sizes? For example, take a maildir directory with ten entries: total 2385 drwx------ 8 17121 vmail 10 Jun 8 23:50 . drwx--x--x 14 root root 14 May 12 2006 .. drwx------ 5 17121 vmail 5 May 25 18:16 .Trash drwx------ 5 17121 staff 6 Jun 9 00:01 .testing -rw------- 1 17121 staff 0 Jun
2010 Jan 27
13
zfs destroy hangs machine if snapshot exists- workaround found
Hi, I was suffering for weeks from the following problem: a zfs dataset contained an automatic snapshot (monthly) that used 2.8 TB of data. The dataset was deprecated, so I chose to destroy it after I had deleted some files; eventually it was completely blank besides the snapshot that still locked 2.8 TB on the pool. ''zfs destroy -r pool/dataset'' hung the machine within seconds
2007 Jun 19
38
ZFS Scalability/performance
Hello, I''m quite interested in ZFS, like everybody else I suppose, and am about to install FBSD with ZFS. On that note, i have a different first question to start with. I personally am a Linux fanboy, and would love to see/use ZFS on linux. I assume that I can use those ZFS disks later with any os that can work/recognizes ZFS correct? e.g. I can install/setup ZFS in FBSD, and later use
2007 Sep 11
13
ZFS but how?
Hi, I''m still debating wether I should use ZFS or not and how. Here is my scenario. I want to run a server with a lot of storage, that gets disks added/upgraded from time to time to expand space. I''d want to store large files on it, 15mb - 5gb per file, and they''d only need to be accessible via NFS/FTP and maybe CIFS. The machine I have dedicated for this job would be
2008 Dec 08
5
How to use mbuffer with zfs send/recv
>> How do i compile mbuffer for our system, Thanks to Mike Futerko for help with the compile, i now have it installed OK. >> and what syntax to i use to invoke it within the zfs send recv? Still looking for answers to this one? Any example syntax, gotchas etc would be much appreciated. -- Kind regards, Jules free. open. honest. love. kindness. generosity. energy. frenetic.
2006 Jun 27
28
Supporting ~10K users on ZFS
OK, I know that there''s been some discussion on this before, but I''m not sure that any specific advice came out of it. What would the advice be for supporting a largish number of users (10,000 say) on a system that supports ZFS? We currently use vxfs and assign a user quota, and backups are done via Legato Networker. >From what little I currently understand, the general
2008 Nov 22
6
RC1 Zfs writes a lot slower when running X
Hi, I have OpenSolaris on an Amd64 Asus-A8NE with 2gig of Rams and 4x320 gig sata drives in raidz1. With dd, I can write at quasi disk maximum speed of 80meg each for a total of 250meg/s if I have no Xsession at all (only console tty). But as soon as I have an Xsession running, the write speed drops to about 120MB/s. Its even worse if I have a VBoxHeadless running with an idle win2k3 inside. It
2007 Feb 27
16
understanding zfs/thunoer "bottlenecks"?
Currently I''m trying to figure out the best zfs layout for a thumper wrt. to read AND write performance. I did some simple mkfile 512G tests and found out, that per average ~ 500 MB/s seems to be the maximum on can reach (tried initial default setup, all 46 HDDs as R0, etc.). According to http://www.amd.com/us-en/assets/content_type/DownloadableAssets/ArchitectureWP_062806.pdf I would
2009 Apr 26
9
Peculiarities of COW over COW?
We run our IMAP spool on ZFS that''s derived from LUNs on a Netapp filer. There''s a great deal of churn in e-mail folders, with messages appearing and being deleted frequently. I know that ZFS uses copy-on- write, so that blocks in use are never overwritten, and that deleted blocks are added to a free list. This behavior would spread the free list all over the zpool. As well,
2008 Nov 19
7
Upgrading from a single disk.
Suppose I have a single ZFS pool on a single disk; I want to upgrade the system to use two different, larger disks and I want to mirror. Can I do something like: - I start with disk #0 - add mirror on disk #1 (resilver) - replace first disk (#0) with disk #2 (resilver) Casper
2018 Sep 17
7
add keys and certificate to forwarded agent on remote host
Apologies if this post is inappropriate to this list; please redirect me if so. Our team uses ssh extensively for server access and maintenance (Debian). An issue is acting as root when operating, for example, over ansible and keeping a record of who performed the actions, something ssh certificates solves well. The problem is then to automate certificate issuance since it would be pretty
2005 Nov 17
2
ZFS portable between architectures, yes or no?
On the intro page (http://www.opensolaris.org/os/community/zfs/intro/) it states that "ZFS has ''adaptive endianness'' to cope with different byte order on different platforms". However in the PDF (ZFS: The Last Word in File Systems) on page 3 it states "Not portable between platforms (e.g. x86 to/from SPARC)" So is it or isn''t it? Thanks, John. This
2006 Jan 04
8
Using same ZFS under different kernel versions
I build two zfs filesystems using b29 (from brandz). I then re-installed solaris express b28, preserving the zfs filesystems. When I tried to "zpool import" my zfs filesystems I got a kernel panic: > debugging crash dump vmcore.0 (32-bit) from blackbird > operating system: 5.11 snv_28 (i86pc) > panic message: > ZFS: bad checksum (read on /dev/dsk/c1d0p0 off 24d5e000: zio