similar to: Other zvols for swap and dump?

Displaying 20 results from an estimated 2000 matches similar to: "Other zvols for swap and dump?"

2009 Mar 31
3
Bad SWAP performance from zvol
I''ve upgraded my system from ufs to zfs (root pool). By default, it creates a zvol for dump and swap. It''s a 4GB Ultra-45 and every late night/morning I run a job which takes around 2GB of memory. With a zvol swap, the system becomes unusable and the Sun Ray client often goes into "26B". So I removed the zvol swap and now I have a standard swap partition. The
2009 Jun 08
4
[caiman-discuss] Can not delete swap on AI sparc
Hi Richard, Richard Robinson wrote: > I should add that I also used truss and saw the same ENOMEM error. I am on a 4Gb system with swap -l reporting > > swapfile dev swaplo blocks free > /dev/zvol/dsk/rpool/swap 181,1 8 4194296 4194296 > > and I was trying to follow the directions for increasing swap here: >
2007 Jan 26
10
UFS on zvol: volblocksize and maxcontig
Hi all! First off, if this has been discussed, please point me in that direction. I have searched high and low and really can''t find much info on the subject. We have a large-ish (200gb) UFS file system on a Sun Enterprise 250 that is being shared with samba (lots of files, mostly random IO). OS is Solaris 10u3. Disk set is 7x36gb 10k scsi, 4 internal 3 external. For several
2006 Jul 25
1
Some initial experiences.
Hi folks, I''ve been asked to post my experiences here, so... well, here goes :) (in no particular order). We can view bugs in kernel/xen via bugs.opensolaris.org, but not file them against kernel/xen. Could that be added to the list of categories for submission? There''s kind of a composite set of problems with bootadm(1M). While the modified bootadm will update
2006 Oct 31
3
zfs: zvols minor #''s changing and causing probs w/ volumes
Team, **Please respond to me and my coworker listed in the Cc, since neither one of us are on this alias** QUICK PROBLEM DESCRIPTION: Cu created a dataset which contains all the zvols for a particular zone. The zone is then given access to all the zvols in the dataset using a match statement in the zoneconfig (see long problem description for details). After the initial boot of the zone
2009 Jun 16
3
Adding zvols to a DomU
I''m trying to add extra zvols to a Solaris10 DomU, sv_113 Dom0 I can use virsh attach-disk <name> <zvol> hdb --device phy to attach the zvol as c0d1. Replacing hdb by hdd gives me c1d1 but then that is it. Being able to attach several more zvols would be nice but even being able to get at c1d0 would be useful Am I missing something or can I only attach to hda/hdb/hdd?
2007 Sep 11
4
ext3 on zvols journal performance pathologies?
I''ve been seeing read and write performance pathologies with Linux ext3 over iSCSI to zvols, especially with small writes. Does running a journalled filesystem to a zvol turn the block storage into swiss cheese? I am considering serving ext3 journals (and possibly swap too) off a raw, hardware-mirrored device. Before I do (and I''ll write up any results) I''d like to know
2007 Jan 29
3
dumpadm and using dumpfile on zfs?
Hi All, I''d like to set up dumping to a file. This file is on a mirrored pool using zfs. It seems that the dump setup doesn''t work with zfs. This worked for both a standard UFS slice and a SVM mirror using zfs. Is there something that I''m doing wrong, or is this not yet supported on ZFS? Note this is Solaris 10 Update 3, but I don''t think that should
2009 Dec 31
6
zvol (slow) vs file (fast) performance snv_130
Hello, I was doing performance testing, validating zvol performance in particularly, and found that zvol write performance to be slow ~35-44MB/s at 1MB blocksize writes. I then tested the underlying zfs file system with the same test and got 121MB/s. Is there any way to fix this? I really would like to have compatible performance between the zfs filesystem and the zfs zvols. # first test is a
2011 Jul 10
3
How create a FAT filesystem on a zvol?
The `lofiadm'' man page describes how to export a file as a block device and then use `mkfs -F pcfs'' to create a FAT filesystem on it. Can''t I do the same thing by first creating a zvol and then creating a FAT filesystem on it? Nothing I''ve tried seems to work. Isn''t the zvol just another block device? -- -Gary Mills- -Unix Group-
2008 Apr 12
5
ZVOL access permissions?
How can I set up a ZVOL that''s accessible by non-root users, too? The intent is to use sparse ZVOLs as raw disks in virtualization (reducing overhead compared to file-based virtual volumes). Thanks, -mg This message posted from opensolaris.org
2006 Sep 06
2
creating zvols in a non-global zone (or ''Doctor, it hurts when I do this'')
A colleague just asked if zfs delegation worked with zvols too. Thought I''d give it a go and got myself in a mess (tank/linkfixer is the delegated dataset): root at non-global / # zfs create -V 500M tank/linkfixer/foo cannot create device links for ''tank/linkfixer/foo'': permission denied cannot create ''tank/linkfixer/foo'': permission denied Ok, so
2009 Nov 20
1
Using local disk for cache on an iSCSI zvol...
I''m just wondering if anyone has tried this, and what the performance has been like. Scenario: I''ve got a bunch of v20z machines, with 2 disks. One has the OS on it, and the other is free. As these are disposable client machines, I''m not going to mirror the OS disk. I have a disk server with a striped mirror zpool, carved into a bunch of zvols, each exported via
2013 Nov 22
1
FreeBSD 10-BETA3 - zfs clone of zvol snapshot is not created
Hi, am I doing something wrong, ZFS does not support that or there is a bug that zvol clone does not show up under /dev/zvol after creating it from other zvol snapshot? # zfs list -t all | grep local local 136G 76.8G 144K none local/home 117G 76.8G 117G /home local/vm 18.4G 76.8G 144K
2007 Dec 06
17
HVM Windows: disk image vs. zvol - which is better?
Hi everyone! I''m planning to migrate a little firm''s Windows server to a HVM machine. The test setup woks fine (it''s basically an MSSQL server tied pack of programs for an accountancy office), all programs can reach the mssql from the outside on the hvm windows. If fact it turned out to be so good, that I''m even planning to simply copy the win image(s) to the
2009 Sep 10
3
zfs send of a cloned zvol
Hi, I have a question, let''s say I have a zvol named vol1 which is a clone of a snapshot of another zvol (its origin property is tank/myvol at mysnap). If I send this zvol to a different zpool through a zfs send does it send the origin too that is, does an automatic promotion happen or do I end up whith a broken zvol? Best regards. Maurilio. -- This message posted from
2009 Oct 17
3
zvol used apparently greater than volsize for sparse volume
What does it mean for the reported value of a zvol volsize to be less than the product of used and compressratio? For example, # zfs get -p all home1/home1mm01 NAME PROPERTY VALUE SOURCE home1/home1mm01 type volume - home1/home1mm01 creation 1254440045 - home1/home1mm01 used 14902492672
2007 Oct 18
2
characterizing I/O on a per zvol basis.
Hey all - Time for my silly question of the day, and before I bust out vi and dtrace... If there a simple, existing way I can observe the read / write / IOPS on a per-zvol basis? If not, is there interest in having one? Cheers! Nathan.
2008 Apr 08
6
lucreate error: Cannot determine the physical boot device ...
# lucreate -n B85 Analyzing system configuration. Hi, after typing # lucreate -n B85 I get the following error: No name for current boot environment. INFORMATION: The current boot environment is not named - assigning name <BE1>. Current boot environment is named <BE1>. Creating initial configuration for primary boot environment <BE1>. ERROR: Unable to determine major and
2010 Jul 16
1
Making a zvol unavailable to iSCSI trips up ZFS
I''ve been experimenting with a two system setup in snv_134 where each system exports a zvol via COMSTAR iSCSI. One system imports both its own zvol and the one from the other system and puts them together in a ZFS mirror. I manually faulted the zvol on one system by physically removing some drives. What I expect to happen is that ZFS will fault the zvol pool and the iSCSI stack will