Displaying 20 results from an estimated 200 matches similar to: "getting inodeno for zfs from vnode in vfs kernel layer"
2008 Jun 18
0
getting inode for zfs from vnode/vfs layer in kernel
i need to get inodeno on ZFS and i am not able to find how to find it in
kernel at vfs layer.
i have vnode pointer and i am doing VTOZ to get znode but printing z_id
from znode pointer
gives me deadbeef(unitialized) , can somebody point me how to get that?
i looked at zfs_getattr code and it does similar thing which i am doing
but its able to get me
inode no in getattribute structure(node
2008 Nov 26
2
zfs znode changes getting lost
In place of padding in zfs znode i added a new field. stored an integer value and am able to see saved information.
but after reboot it is not there. If i was able to access before reboot so it must be in memory. I think i need to save it to disk.
how does one force zfs znode to disk.
right now i dont do anything special for it. Just made an ioctl, accessed znode and made changes.
example in
2007 May 09
5
Refactor zfs_zget()
Hi,
Since almost all operations in the FUSE low-level API identify files by inode
number, I''ve been using zfs_zget() to get the corresponding znode/vnode in
order to call the corresponding VFS function in zfs_vnops.c.
However, there are some cases when zfs_zget() behaves slightly different than
I need:
1) If zp->z_unlinked != 0 then zfs_zget() returns ENOENT. I need it to return
2006 Nov 02
11
ZFS and memory usage.
ZFS works really stable on FreeBSD, but I''m biggest problem is how to
control ZFS memory usage. I''ve no idea how to leash that beast.
FreeBSD has a backpresure mechanism. I can register my function so it
will be called when there are memory problems, which I do. I using it
for ARC layer.
Even with this in place under heavy load the kernel panics, because
memory with KM_SLEEP
2010 Sep 17
3
ZFS Dataset lost structure
After a crash, in my zpool tree, some dataset report this we i do a ls -la:
brwxrwxrwx 2 777 root 0, 0 Oct 18 2009 mail-cts
also if i set
zfs set mountpoint=legacy dataset
and then i mount the dataset to other location
before the directory tree was only :
dataset
- vdisk.raw
The file was a backing device of a Xen VM, but i cannot access the directory structure of this dataset.
However i
2006 Jul 20
1
tracking an error back to a file
Hi. I''m in the process of writing an introductory paper on ZFS.
The paper is meant to be something that could be given to a systems
admin at a site to introduce ZFS and document common procedures for
using ZFS.
In the paper, I want to document the method for identifying which
file has a checksum error. In previous discussions on this alias,
I''ve used the following
2006 Mar 17
1
acquiring duplicate lock of same type: "vnode interlock"
I think I've read somewhere about panic during early root mount, fsck
etc.. Perhaps this might be related:
Full dmesg: http://people.freebsd.org/~ariff/misc/dmesg.boot.amd64
[....]
acquiring duplicate lock of same type: "vnode interlock"
1st vnode interlock @ kern/vfs_vnops.c:791
2nd vnode interlock @ kern/vfs_subr.c:2018
KDB: stack backtrace:
witness_checkorder() at
2013 Jul 24
1
NFS deadlock on 9.2-Beta1
Two machines (NFS Server: running ZFS / Client: disk-less), both are running FreeBSD r253506. The NFS client starts to deadlock processes within a few hours. It usually gets worse from there on. The processes stay in "D" state. I haven't been able to reproduce it when I want it to happen. I only have to wait a few hours until the deadlocks occur when traffic to the client machine
2006 Jun 13
1
printing vnode page list
I''m trying to debug a problem that requires I print the
dirty pages (v_pages) list. Any suggestions?
This message posted from opensolaris.org
2006 May 19
11
tracking error to file
In my testing, I''ve found the following error:
zpool status -v
pool: local
state: ONLINE
status: One or more devices has experienced an error resulting in data
corruption. Applications may be affected.
action: Restore the file in question if possible. Otherwise restore the
entire pool from backup.
see: http://www.sun.com/msg/ZFS-8000-8A
scrub: none requested
2007 Aug 09
5
Unremovable file in ZFS filesystem.
I managed to create a link in a ZFS directory that I can''t remove. Session as follows:
# ls
bayes.lock.router.3981 bayes_journal user_prefs
# ls -li bayes.lock.router.3981
bayes.lock.router.3981: No such file or directory
# ls
bayes.lock.router.3981 bayes_journal user_prefs
# /usr/sbin/unlink bayes.lock.router.3981
unlink: No such file or directory
# find . -print
2008 Nov 08
9
How does zfs COW deal with ''..'' in brother directory?
Hi matt,
I have some problems about understanding zfs COW implemention. Suppose
b and c are both children dir of a, if c changes, there will be new
versions of both a and c, namely c'' and a''.
a a''
b c c''
Because ''..'' in b points to a before this change, shall we modify b
to let ''..'' point to a''? If yes,
2003 May 21
8
system slowdown - vnode related
I woke up to a frozen box this morning - it froze up a few more times
before I got a handle on it.
Basically, the box runs idle but refuses to do disk IO, or does it
-very- slowly.
Top shows processes stuck in 'ffsvget', 'inode', and 'vlruwk' state.
I can get the box responsive again by setting sysctl
kern.maxvnods=100000. It starts up with kern.maxnodes=36079. I
2007 Feb 11
0
unable to mount legacy vol - panic in zfs:space_map_remove - zdb crashes
I have a 100gb SAN lun in a pool, been running ok for about 6 months. panicked the system this morning. system was running S10U2. In the course of troubleshooting I''ve installed the latest recommended bundle including kjp 118833-36 and zfs patch 124204-03
created as:
zpool create zfspool01 /dev/dsk/emcpower0c
zfs create zfspool01/nb60openv
zfs set mountpoint=legacy zfspool01/nb60openv
2006 Oct 13
24
Self-tuning recordsize
Would it be worthwhile to implement heuristics to auto-tune
''recordsize'', or would that not be worth the effort?
--
Regards,
Jeremy
2006 Aug 25
4
Looking for confirmation.
Hi.
I''ve almost all file system functions working.
I started to run some heavy file system regression tests. They work. fsx
wasn''t able to break my port, but the test you can find here:
http://people.freebsd.org/~kan/fsstress.tar.gz
broke it. My kernel panics on this assertion (zfs_dir.c):
749: mutex_exit(&dzp->z_lock);
750:
751: error =
2006 Mar 23
17
Poor performance on NFS-exported ZFS volumes
I''m seeing some pretty pitiful performance using ZFS on a NFS server, with a ZFS volume exported (only with rw=host.foo.com,root=host.foo.com opts) and mounted on a Linux host running kernel 2.4.31. This linux kernel I''m working with is limited in that I can only do NFSv2 mounts... irregardless of that aspect, I''m sure something''s amiss.
I mounted the zfs-based
2007 Mar 28
20
Gzip compression for ZFS
Adam,
With the blog entry[1] you''ve made about gzip for ZFS, it raises
a couple of questions...
1) It would appear that a ZFS filesystem can support files of
varying compression algorithm. If a file is compressed using
method A but method B is now active, if I truncate the file
and rewrite it, is A or B used?
2) The question of whether or not to use bzip2 was raised in
the
2007 Sep 14
5
ZFS Space Map optimalization
I have a huge problem with space maps on thumper. Space maps takes over 3GB
and write operations generates massive read operations.
Before every spa sync phase zfs reads space maps from disk.
I decided to turn on compression for pool ( only for pool, not filesystems ) and it helps.
Now space maps, intent log, spa history are compressed.
Not I''m thinking about disabling checksums. All
2011 Nov 08
1
Single-disk rpool with inconsistent checksums, import fails
Hello all,
I have an oi_148a PC with a single root disk, and since
recently it fails to boot - hangs after the copyright
message whenever I use any of my GRUB menu options.
Booting with an oi_148a LiveUSB I had around since
installation, I ran some zdb traversals over the rpool
and zpool import attempts. The imports fail by running
the kernel out of RAM (as recently discussed in the
list with