Displaying 20 results from an estimated 300 matches similar to: "Using same ZFS under different kernel versions"
2007 Mar 10
3
long character string problem
Hi All
I am having 2 very long character strings (550chars) and I want to put them as
expressions together with c(). The problem is that I also get these
double-quotes, as seen below in 'fct'. How can I remove these double-quotes? I
tried as.name() but it did not work (because of size?). These are creating
trouble with subsequent programs, which I tested with strings that for some
2006 Jul 20
1
tracking an error back to a file
Hi. I''m in the process of writing an introductory paper on ZFS.
The paper is meant to be something that could be given to a systems
admin at a site to introduce ZFS and document common procedures for
using ZFS.
In the paper, I want to document the method for identifying which
file has a checksum error. In previous discussions on this alias,
I''ve used the following
2010 Sep 17
3
ZFS Dataset lost structure
After a crash, in my zpool tree, some dataset report this we i do a ls -la:
brwxrwxrwx 2 777 root 0, 0 Oct 18 2009 mail-cts
also if i set
zfs set mountpoint=legacy dataset
and then i mount the dataset to other location
before the directory tree was only :
dataset
- vdisk.raw
The file was a backing device of a Xen VM, but i cannot access the directory structure of this dataset.
However i
2013 May 07
4
Rake db:migrate
Hello,
I''m trying to use command rake db:migrate, but in MySQL I continuous get
an empty set and I do know why.
#/db/seeds.rb
User.create(username: "lorenz", password: "mypass", admin: true)
When I run:
rake db:seed RAILS_ENV=development --trace
I received:
** Invoke db:seed (first_time)
** Execute db:seed
** Invoke db:abort_if_pending_migrations (first_time)
**
2008 Dec 15
15
Need Help Invalidating Uberblock
I have a ZFS pool that has been corrupted. The pool contains a single device which was actually a file on UFS. The machine was accidentally halted and now the pool is corrupt. There are (of course) no backups and I''ve been asked to recover the pool. The system panics when trying to do anything with the pool.
root@:/$ zpool status
panic[cpu1]/thread=fffffe8000758c80: assertion failed:
2007 Feb 11
0
unable to mount legacy vol - panic in zfs:space_map_remove - zdb crashes
I have a 100gb SAN lun in a pool, been running ok for about 6 months. panicked the system this morning. system was running S10U2. In the course of troubleshooting I''ve installed the latest recommended bundle including kjp 118833-36 and zfs patch 124204-03
created as:
zpool create zfspool01 /dev/dsk/emcpower0c
zfs create zfspool01/nb60openv
zfs set mountpoint=legacy zfspool01/nb60openv
2010 Feb 04
1
Bug in as.character? (PR#14206)
A long formula which is converted using as.character, looses its last
part: ``diagonal = 1e-12)''
Shorter formula is ok though.
Best,
H??vard
************
Browse[2]> formula.str
y ~ -1 + b1 + b2 + b3 + b4 + b5 + b6 + b7 + b8 + b9 + b10 + b11 +
b12 + b13 + b14 + b15 + b16 + b17 + b18 + b19 + b20 + b21 +
b22 + b23 + b24 + b25 + b26 + b27 + b28 + b29 + b30 + b31 +
b32 +
2011 Oct 25
4
[LLVMdev] is anyone using the sparc backend?
Chris,
> FYI, I completely agree that we shouldn't remove the sparc target. However, it is currently unmaintained, and has some pretty glaring bugs. Is anyone interested in stepping up to maintain the target?
>
> -Chris
We can not let it die and I volunteer to keep the Sparc alive!
My SPARCstation pizzaboxes can be ready for action in a day or two.
Pawel Wodnicki
> On Oct
2011 Oct 26
0
[LLVMdev] is anyone using the sparc backend?
Hi Pawel,
> We can not let it die and I volunteer to keep the Sparc alive!
>
> My SPARCstation pizzaboxes can be ready for action in a day or two.
there are three sparc buildbots which are hooked up to lab.llvm.org. They are
currently idle because no builders have been scheduled to use them:
# gcc54 36G 0.5 GHz TI UltraSparc IIe (Hummingbird) / 1.5 GB RAM / Sun
Netra T1 200 /
2012 Dec 20
3
Pool performance when nearly full
Hi
I know some of this has been discussed in the past but I can''t quite find the exact information I''m seeking
(and I''d check the ZFS wikis but the websites are down at the moment).
Firstly, which is correct, free space shown by "zfs list" or by "zpool iostat" ?
zfs list:
used 50.3 TB, free 13.7 TB, total = 64 TB, free = 21.4%
zpool iostat:
used
2011 Nov 05
4
ZFS Recovery: What do I try next?
I would like to pick the brains of the ZFS experts on this list: What
would you do next to try and recover this zfs pool?
I have a ZFS RAIDZ1 pool named bank0 that I cannot import. It was
composed of 4 1.5 TiB disks. One disk is totally dead. Another had
SMART errors, but using GNU ddrescue I was able to copy all the data
off successfully.
I have copied all 3 remaining disks as images using
2011 Nov 25
1
Recovering from kernel panic / reboot cycle importing pool.
Yesterday morning I awoke to alerts from my SAN that one of my OS disks was faulty, FMA said it was in hardware failure. By the time I got to work (1.5 hours after the email) ALL of my pools were in a degraded state, and "tank" my primary pool had kicked in two hot spares because it was so discombobulated.
------------------- EMAIL -------------------
List of faulty resources:
2006 May 19
11
tracking error to file
In my testing, I''ve found the following error:
zpool status -v
pool: local
state: ONLINE
status: One or more devices has experienced an error resulting in data
corruption. Applications may be affected.
action: Restore the file in question if possible. Otherwise restore the
entire pool from backup.
see: http://www.sun.com/msg/ZFS-8000-8A
scrub: none requested
2011 Nov 08
1
Single-disk rpool with inconsistent checksums, import fails
Hello all,
I have an oi_148a PC with a single root disk, and since
recently it fails to boot - hangs after the copyright
message whenever I use any of my GRUB menu options.
Booting with an oi_148a LiveUSB I had around since
installation, I ran some zdb traversals over the rpool
and zpool import attempts. The imports fail by running
the kernel out of RAM (as recently discussed in the
list with
2006 Oct 31
0
6351451 snv_28 on v20z stopped tracking time
Author: sethg
Repository: /hg/zfs-crypto/gate
Revision: c41c7b006581b93b863e2d58e1d492a5449a0ba8
Log message:
6351451 snv_28 on v20z stopped tracking time
Files:
update: usr/src/uts/common/os/clock.c
2013 Apr 22
7
Multiple lon lat points in the map with ggplot2
Hello R users,
For the last few days I am struggling with the following task:
my data.frame:
A1 A2 A3 B1 B2
B3
58.81 53.292 54.501 13.013 17.39 19.407 56.02 56.251 54.033 20.099 13.15
10.411 55.376 53.099 57.625 13.396 21.031 13.22 58.584 53.194 54.218
13.038 16.854 19.289 55.7 55.921 53.847 19.942 13.153 9.828 55.093 52.934
2008 Nov 08
9
How does zfs COW deal with ''..'' in brother directory?
Hi matt,
I have some problems about understanding zfs COW implemention. Suppose
b and c are both children dir of a, if c changes, there will be new
versions of both a and c, namely c'' and a''.
a a''
b c c''
Because ''..'' in b points to a before this change, shall we modify b
to let ''..'' point to a''? If yes,
2006 Jan 03
4
zfs object sets and datasets
Hi All,
I am looking at trying to understand the conceptual model of how data is
grouped and partitioned within a storge pool. In looking at the on disk
document that Tabriz sent out a few weeks ago I see that object sets are
the grouping which ZFS uses to group objects that are related.
Specifially to aid in format and layout of like objects in to a set.
So, for example 1 potential object
2010 Jan 30
3
Checksum fletcher4 or sha256 ?
Hi,
I''m atmost ready to deploy my new homeserver for final testing.
Before I want to be sure that nothing big is left untouched.
Reading ZFS Admin Guide About the checksum method, there''s no advice about it.
The default is fletcher4. there''s also SHA256
Now the sha256 is pretty ''heavy'' to calculate, so I think that it''s left out because can
2005 Nov 19
11
ZFS related panic!
> My current zfs setup lookst like this:
> > homepool 3.63G 34.1G 8K /homepool
> > homepool/db 61.6M 34.1G 8.50K /var/db
> > homepool/db/pgsql 61.5M 34.1G 61.5M
> > /var/db/pgsql
> > homepool/home 3.57G 34.1G 10.0K /users
> > homepool/home/carrie 8K 34.1G 8K
> > /users/carrie
> >