Displaying 20 results from an estimated 20000 matches similar to: "ZFS Upgrade"
2009 Oct 13
14
How to resize ZFS partion or add a new one?
Hi,
I have the following partions on my laptop, Inspiron 6000, from fdisk:
1 Other OS 0 11 12 0
2 EXT LBA 12 2561 2550 26
3 Active Solaris2 2562 9728 7167 74
First one is for Dell utilities. Second one is NTFS and the third is ZFS.
I am currently using OpenSolaris 2009.06
2010 May 28
21
expand zfs for OpenSolaris running inside vm
hello, all
I am have constraint disk space (only 8GB) while running os inside vm. Now i
want to add more. It is easy to add for vm but how can i update fs in os?
I cannot use autoexpand because it doesn''t implemented in my system:
$ uname -a
SunOS sopen 5.11 snv_111b i86pc i386 i86pc
If it was 171 it would be grate, right?
Doing following:
o added new virtual HDD (it becomes
2011 Aug 08
2
rpool recover not using zfs send/receive
Is it possible to recover the rpool with only a tar/star archive of the root filesystem? I have used the zfs send/receive methods and that work without a problem.
What I am trying to do is recreate the rpool and underlying zfs filesystems (rpool/ROOT, rpool/s10_uXXXXXX, rpool/dump, rpool/swap, rpool/export, and rpool/export/home). I then mount the pool at a alternate root and restore the tar
2008 Aug 04
16
zpool upgrade wrecked GRUB
Machine is running x86 snv_94 after recent upgrade from opensolaris 2008.05. ZFS and zpool reported no troubles except suggesting upgrade for from ver.10 to ver.11. seemed like a good idea at the time. system up for several days after that point then took down for some unrelated maintenance.
now will not boot the opensol, drops to grub prompt, no menus.
zfs was mirrored on two disks c6d0s0 and
2009 Dec 04
2
USB sticks show on one set of devices in zpool, different devices in format
Hello,
I had snv_111b running for a while on a HP DL160G5. With two 16GB USB sticks comprising the mirrored rpool for boot. And four 1TB drives comprising another pool, pool1, for data.
So that''s been working just fine for a few months. Yesterday I get it into my mind to upgrade the OS to latest, then was snv_127. That worked, and all was well. Also did an upgrade to the
2008 Dec 26
19
separate home "partition"?
(i use the term loosely because i know that zfs likes whole volumes better)
when installing ubuntu, i got in the habit of using a separate partition for my home directory so that my data and gnome settings would all remain intact when i reinstalled or upgraded.
i''m running osol 2008.11 on an ultra 20, which has only two drives. i''ve got all my data located in my home directory,
2010 Mar 29
19
sharing a ssd between rpool and l2arc
Hi,
as Richard Elling wrote earlier:
"For more background, low-cost SSDs intended for the boot market are
perfect candidates. Take a X-25V @ 40GB and use 15-20 GB for root
and the rest for an L2ARC. For small form factor machines or machines
with max capacity of 8GB of RAM (a typical home system) this can make a
pleasant improvement over a HDD-only implementation."
For the upcoming
2009 Dec 27
7
How to destroy your system in funny way with ZFS
Hi all,
I installed another OpenSolaris (snv_129) in VirtualBox 3.1.0 on Windows because snv_130 doesn''t boot anymore after installation of VirtualBox guest additions. Older builds before snv_129 were running fine too. I like some features of this OS, but now I end with something funny.
I installed default snv_129, installed guest additions -> reboot, set
2008 Aug 26
5
Problem w/ b95 + ZFS (version 11) - seeing fair number of errors on multiple machines
Hi,
After upgrading to b95 of OSOL/Indiana, and doing a ZFS upgrade to the newer
revision, all arrays I have using ZFS mirroring are displaying errors. This
started happening immediately after ZFS upgrades. Here is an example:
ormandj at neutron.corenode.com:~$ zpool status
pool: rpool
state: DEGRADED
status: One or more devices has experienced an unrecoverable error. An
attempt was
2011 Nov 22
3
SUMMARY: mounting datasets from a read-only pool with aid of tmpfs
Hello all,
I''d like to report a tricky situation and a workaround
I''ve found useful - hope this helps someone in similar
situations.
To cut the long story short, I could not properly mount
some datasets from a readonly pool, which had a non-"legacy"
mountpoint attribute value set, but the mountpoint was not
available (directory absent or not empty). In this case
2010 Apr 29
39
Best practice for full stystem backup - equivelent of ufsdump/ufsrestore
I''m looking for a way to backup my entire system, the rpool zfs pool to an external HDD so that it can be recovered in full if the internal HDD fails. Previously with Solaris 10 using UFS I would use ufsdump and ufsrestore, which worked so well, I was very confident with it. Now ZFS doesn''t have an exact replacement of this so I need to find a best practice to replace it.
2009 Jan 05
3
ZFS import on pool with same name?
I have an OpenSolaris snv_101 box with ZFS on it. (Sun Ultra 20 M2)
zpool name is rpool. The
I have a 2nd hard drive in the box that I am trying to recover the ZFS
data from (long story but that HD became unbootable after installing IPS
on the machine)
Both drives have a pool named "rpool", so I can''t import the rpool from
the 2nd drive.
root at hyperion:~# zpool status
2010 Feb 18
2
Killing an EFI label
Since this seems to be a ubiquitous problem for people running ZFS, even
though it''s really a general Solaris admin issue, I''m guessing the
expertise is actually here, so I''m asking here.
I found lots of online pages telling how to do it.
None of them were correct or complete. I think. I seem to have
accomplished it in a somewhat hackish fashion, possibly not
2011 Jan 28
2
ZFS root clone problem
(for some reason I cannot find my original thread..so I''m reposting it)
I am trying to move my data off of a 40gb 3.5" drive to a 40gb 2.5" drive. This is in a Netra running Solaris 10.
Originally what I did was:
zpool attach -f rpool c0t0d0 c0t2d0.
Then I did an installboot on c0t2d0s0.
Didnt work. I was not able to boot from my second drive (c0t2d0).
I cannot remember
2008 Sep 13
3
Restore a ZFS Root Mirror
Hi all,
after installing OpenSolaris 2008.05 in VirtualBox I''ve created a ZFS Root Mirror by:
"zfs attach rpool Disk B"
and it works like a charm. Now I tried to restore the rpool from the worst Case
Scenario: The Disk the System was installed to (Disk A) fails.
I replaced the Disk A with another virtual Disk C and tried to restore the rpool, but
my Problem is that I
2010 Aug 28
4
ufs root to zfs root liveupgrade?
hi all
Try to learn how UFS root to ZFS root liveUG work.
I download the vbox image of s10u8, it come up as UFS root.
add a new disks (16GB)
create zpool rpool
run lucreate -n zfsroot -p rpool
run luactivate zfsroot
run lustatus it do show zfsroot will be active in next boot
init 6
but it come up with UFS root,
lustatus show ufsroot active
zpool rpool is mounted but not used by boot
Is this a
2009 Aug 23
23
incremental backup with zfs to file
FULL backup to a file
zfs snapshot -r rpool at 0908
zfs send -Rv rpool at 0908 > /net/remote/rpool/snaps/rpool.0908
INCREMENTAL backup to a file
zfs snapshot -i rpool at 0908 rpool at 090822
zfs send -Rv rpool at 090822 > /net/remote/rpool/snaps/rpool.090822
As I understand the latter gives a file with changes between 0908 and
090822. Is this correct?
How do I restore those files? I know
2008 Jul 26
1
Expanding a root pool
I''m attempting to expand a root pool for a VMware VM that is on an 8GB virtual disk. I mirrored it to a 20GB disk and detached the 8GB disk. I did "installgrub" to install grub onto the second virtual disk, but I get a kernel panic when booting. Is there an extra step I need to perform to get this working?
Basically I did this:
1. Created a new 20GB virtual disk.
2. Booted
2010 Sep 10
3
zpool upgrade and zfs upgrade behavior on b145
Not sure what the best list to send this to is right now, so I have selected
a few, apologies in advance.
A couple questions. First I have a physical host (call him bob) that was
just installed with b134 a few days ago. I upgraded to b145 using the
instructions on the Illumos wiki yesterday. The pool has been upgraded (27)
and the zfs file systems have been upgraded (5).
chris at bob:~# zpool
2010 Sep 10
3
zpool upgrade and zfs upgrade behavior on b145
Not sure what the best list to send this to is right now, so I have selected
a few, apologies in advance.
A couple questions. First I have a physical host (call him bob) that was
just installed with b134 a few days ago. I upgraded to b145 using the
instructions on the Illumos wiki yesterday. The pool has been upgraded (27)
and the zfs file systems have been upgraded (5).
chris at bob:~# zpool