similar to: capacity query

Displaying 20 results from an estimated 200 matches similar to: "capacity query"

2008 Jun 26
3
[Bug 2334] New: zpool destroy panics after zfs_force_umount_stress
http://defect.opensolaris.org/bz/show_bug.cgi?id=2334 Summary: zpool destroy panics after zfs_force_umount_stress Classification: Development Product: zfs-crypto Version: unspecified Platform: Other OS/Version: Solaris Status: NEW Severity: major Priority: P2 Component: other AssignedTo:
2009 Jun 29
7
ZFS - SWAP and lucreate..
Good morning everybody I was migrating my ufs ? rootfilesystem to a zfs ? one, but was a little upset finding out that it became bigger (what was clear because of the swap and dump size). Now I am questioning myself if it is possible to set the swap and dump size by using the lucreate ? command (I wanna try it again but on less space). Unfortunately I did not find any advice in manpages.
2009 Jun 08
4
[caiman-discuss] Can not delete swap on AI sparc
Hi Richard, Richard Robinson wrote: > I should add that I also used truss and saw the same ENOMEM error. I am on a 4Gb system with swap -l reporting > > swapfile dev swaplo blocks free > /dev/zvol/dsk/rpool/swap 181,1 8 4194296 4194296 > > and I was trying to follow the directions for increasing swap here: >
2008 Apr 03
3
[Bug 971] New: zfs key -l fails after unloading (keyscope=dataset)
http://defect.opensolaris.org/bz/show_bug.cgi?id=971 Summary: zfs key -l fails after unloading (keyscope=dataset) Classification: Development Product: zfs-crypto Version: unspecified Platform: Other OS/Version: Solaris Status: NEW Severity: major Priority: P2 Component: other AssignedTo:
2007 Jul 18
0
zfs and zones newbie question
Hello I have a fresh Solaris 10/6 install. My disk have 20Gb but I used only 10GB for the base solaris installation and home directory. I have 10GB free without any partition. I would like to use my free space to store my zones using zfs filesystem. In example: /zones /zone/zona1 /zone/zone2 How I can create "/zones" filesystem ? zpool create zones ????? ?????: What I must
2007 Mar 14
4
solaris 10 + dovecot (1)
hi... i try to make dovecot (v1.0.r26) in solaris 10 and i have the following problem # ./configure --with-ldap ... ... Install prefix ...................... : /usr/local File offsets ........................ : 64bit I/O loop method ..................... : poll File change notification method ..... : none Building with SSL support ........... : yes (OpenSSL) Building with IPv6 support ..........
2008 Aug 26
5
Problem w/ b95 + ZFS (version 11) - seeing fair number of errors on multiple machines
Hi, After upgrading to b95 of OSOL/Indiana, and doing a ZFS upgrade to the newer revision, all arrays I have using ZFS mirroring are displaying errors. This started happening immediately after ZFS upgrades. Here is an example: ormandj at neutron.corenode.com:~$ zpool status pool: rpool state: DEGRADED status: One or more devices has experienced an unrecoverable error. An attempt was
2005 Dec 16
1
a problem in building dovecot @ opensolaris
Hi, I'm trying to build dovecot from snapshot dovecot-20051215.tar.gz on OpenSolaris Nevada build 28. I have added to the PATH directories for make, gcc and automake/autoconf (in BASH): $ export PATH=/opt/sfw/bin:/usr/ccs/bin:/usr/sfw/bin:$PATH then executed --- $ aclocal $ libtoolize --force $ automake --add-missing $ autoheader $ autoconf --- And then make finished with --- Install prefix
2006 Sep 01
4
create ZFS pool(s)/volume(s) during jumpstart instalation
Anyone have idea how to create ZFS pool and volumes during jumstart instalation ? (I''m using JumpStart Enterprise Toolkit) kloczek
2008 Jan 24
1
zfs showing more filesystem using ls than df actually has
Platform T2000 SunOS ccluatdwunix1 5.10 Generic_125100-10 sun4v sparc SUNW,Sun-Fire-T200 I have a user that stated zfs is allocating more file system space than actually available via ls command versus what df -k shows. He stated he used the mkfile to verify if ZFS quota was working. He executes "ls -s" to report usage which reports more allocated than available from "df
2009 Oct 15
8
sub-optimal ZFS performance
Hello, ZFS is behaving strange on a OSOL laptop, your thoughts are welcome. I am running OSOL on my laptop, currently b124 and i found that the performance of ZFS is not optimal in all situations. If i check the how much space the package cache for pkg(1) uses, it takes a bit longer on this host than on comparable machine to which i transferred all the data. user at host:/var/pkg$ time
2007 Feb 10
16
How to backup a slice ? - newbie
... though I tried, read and typed the last 4 hours; still no clue. Please, can anyone give a clear idea on how this works: Get the content of c0d1s1 to c0d0s7 ? c0d1s1 is pool home and active; c0d0s7 is not active. I have followed the suggestion on http://www.opensolaris.org/os/community/zfs/demos/zfs_demo.pdf % sudo zfs snapshot home at backup % zfs list NAME USED AVAIL REFER
2008 Nov 12
0
ZFS sens state
Does zfs send maintain an internal state or lock? I attempted to restart a send that was interrupted when the sending system rebooted and the send hangs. Here''s the last few lines of the truss output: # zfs send -i tue live/fs at wed | ssh staging zfs receive -F -v tank/backup/fs at wed 6074: open("/dev/zfs", O_RDWR) = 3 6074: fstat64(3,
2007 Jul 17
3
Small problem with src/lib/mountpoint.c [now with patch attached!]
Greetings - [SIGH: I hit the "Send" button instead of "Attach". Here's Take 2...] Whilst playing with getting quotas from NFS-mounted filestores I've just discovered a slight problem with src/lib/mountpoint.c I had things working on a Solaris 10 test machine which had the mailstore mounted 'normally' using an entry in /etc/vfstab. However when I changed
2009 Oct 17
3
zvol used apparently greater than volsize for sparse volume
What does it mean for the reported value of a zvol volsize to be less than the product of used and compressratio? For example, # zfs get -p all home1/home1mm01 NAME PROPERTY VALUE SOURCE home1/home1mm01 type volume - home1/home1mm01 creation 1254440045 - home1/home1mm01 used 14902492672
2005 Dec 15
2
Patch: More of kqueue() support.
Hi, the attached patch contains these changes: 2005-12-15 22:18 Vaclav Haisman <v.haisman@sh.cvut.cz> * src/lib/ioloop-kqueue.c: Fix IO_ERROR behaviour. * src/lib/ioloop-notify-kqueue.c: New file. * configure.in: Improve kqueue detection and handling. Cleanup. The configure.in changes that are not related to kqueue are there to make autoconf 2.59, automake 1.9.6 and libtool 1.5.20
2007 Jul 17
0
Small problem with src/lib/mountpoint.c
Greetings - Whilst playing with getting quotas from NFS-mounted filestores I've just discovered a slight problem with src/lib/mountpoint.c I had things working on a Solaris 10 test machine which had the mailstore mounted 'normally' using an entry in /etc/vfstab. However when I changed to use the automounter for the mailstore obtaining the quota broke. The problem lies with
2018 May 08
1
mount failing client to gluster cluster.
Hi, On a debian 9 client, ======== root at kvm01:/var/lib/libvirt# dpkg -l glusterfs-client 8><--- ii glusterfs-client 3.8.8-1 amd64 clustered file-system (client package) root at kvm01:/var/lib/libvirt# ======= I am trying to to do a mount to a Centos 7 gluster setup, ======= [root at glustep1 libvirt]# rpm -q glusterfs glusterfs-4.0.2-1.el7.x86_64
2011 Oct 04
6
zvol space consumption vs ashift, metadata packing
I sent a zvol from host a, to host b, twice. Host b has two pools, one ashift=9, one ashift=12. I sent the zvol to each of the pools on b. The original source pool is ashift=9, and an old revision (2009_06 because it''s still running xen). I sent it twice, because something strange happened on the first send, to the ashift=12 pool. "zfs list -o space" showed figures at
2009 Mar 31
3
Bad SWAP performance from zvol
I''ve upgraded my system from ufs to zfs (root pool). By default, it creates a zvol for dump and swap. It''s a 4GB Ultra-45 and every late night/morning I run a job which takes around 2GB of memory. With a zvol swap, the system becomes unusable and the Sun Ray client often goes into "26B". So I removed the zvol swap and now I have a standard swap partition. The