Displaying 20 results from an estimated 120 matches similar to: "Stale NFS file handle"
2012 Aug 24
1
ext4 issues
Hi, I have seen that there are issues with gluster on ext4. Just to be
clear, is this something which is only related to clients using nfs,
i.e., can I happily use gluster (without downgrading kernel) if all
clients are using gluster native client?
Thanks,
/jon
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
2013 Jun 11
1
cluster.min-free-disk working?
Hi,
have a system consisting of four bricks, using 3.3.2qa3. I used the
command
gluster volume set glusterKumiko cluster.min-free-disk 20%
Two of the bricks where empty, and two were full to just under 80% when
building the volume.
Now, when syncing data (from a primary system), and using min-free-disk
20% I thought new data would go to the two empty bricks, but gluster
does not seem
2013 Jul 15
4
GlusterFS 3.4.0 and 3.3.2 released!
Hi All,
3.4.0 and 3.3.2 releases of GlusterFS are now available. GlusterFS 3.4.0
can be downloaded from [1]
and release notes are available at [2]. Upgrade instructions can be
found at [3].
If you would like to propose bug fix candidates or minor features for
inclusion in 3.4.1, please add them at [4].
3.3.2 packages can be downloaded from [5].
A big note of thanks to everyone who helped in
2013 Feb 27
1
Slow read performance
Help please-
I am running 3.3.1 on Centos using a 10GB network. I get reasonable write speeds, although I think they could be faster. But my read speeds are REALLY slow.
Executive summary:
On gluster client-
Writes average about 700-800MB/s
Reads average about 70-80MB/s
On server-
Writes average about 1-1.5GB/s
Reads average about 2-3GB/s
Any thoughts?
Here are some additional details:
2009 May 19
2
ESTALE error while synching
Hi, I am wondering how rsync-3.0.6 react if it encounters ESTALE error while
synching? If I remember correctly then the rsync-2.6.0 skipping that
file/dir in case of ESTALE error.
Jignesh.
-------------- next part --------------
HTML attachment scrubbed and removed
2011 Sep 21
3
[PATCH] Btrfs: fix orphan cleanup regression
In fixing how we deal with bad inodes, we had a regression in the orphan cleanup
code, since it expects to get a bad inode back. So fix it to deal with getting
-ESTALE back by deleting the orphan item manually and moving on. Thanks,
Reported-by: Simon Kirby <sim@hostway.ca>
Signed-off-by: Josef Bacik <josef@redhat.com>
---
fs/btrfs/inode.c | 36
2006 Oct 31
3
ocfs2 error messages
I received the following error messages in the system logs. Is this
anything to be concerned with?
kernel: (4074,0):ocfs2_populate_inode:234 ERROR: Invalid dinode:
i_ino=1293597, i_blkno=1293597, signature = INODE01, flags = 0x0
kernel: (4074,0):ocfs2_read_locked_inode:389 ERROR: populate inode
failed! i_blkno=1293597, i_ino=1293597
kernel: (4074,0):ocfs2_iget:131 ERROR: status = -116
2004 Nov 23
1
NFS, test52, maildir, freebsd 4.x and OS X (Panther) Server
I've been working on setting up a Dovecot server with a maildir
mailboxes stored over NFS. Initially, everything was working great, and
then I started seeing this error as previously discussed on the list in
my maillog:
(mail_index_expunge): assertion failed: (!t->view->external)
Timo, in another thread, suggested to remove line 247 in
mail-index-transaction.c in the test52
2009 Mar 03
3
[PATCH 1/1] OCFS2: anti stale inode for nfs (V6)
For nfs exporting, ocfs2_get_dentry() returns the dentry for fh.
ocfs2_get_dentry() may read from disk(when inode not in memory) without
any cross cluster lock. this leads to load a stale inode.
this patch fixes above problem.
solution is that in case of inode is not in memory, we get the cluster lock(PR)
of alloc inode where the inode in question is allocated from(this causes node
on which
2005 Apr 27
4
Setting location for indexes
We have NFS-mounted user directories which cause significant problems with
stale file handles:
dovecot: IMAP(uid): fstat() failed with file /home/uid/mail/.imap/INBOX/dovecot.index.log: Stale NFS file handle
(a) why can't it just close the file and re-open ?
(b) is there a way to specify with the 'default_mail_env' (or in some other way)
the location of **only** the indexes (and let
2008 Oct 23
2
[PATCH 1/1] OCFS2: fix for nfs getting stale inode.
Ocfs2 supports exporting.
PROBLEM:
There are 2 problems
(1) Current version of ocfs2_get_dentry() may read from disk
the inode WITHOUT any cross cluster lock. This may lead to load a stale inode.
(2) for deleting an inode, ocfs2_remove_inode() doesn't sync/checkpoint to disk.
This also may lead ocfs2_get_dentry() from other node read out stale inode.
PROBLEM DETAIL:
for problem (1),
For
2011 Mar 24
3
fstat no such file or directory error
Periodically, we will have users IMAP sessions get hung. Their client
sees a message like:
The server responded: 'SERVERBUG] Internal error occurred. Refer to
server log for more information
On the server, we see the following in the logs:
IMAP(<username>): Transaction log dovecot.index.log: duplicate
transaction log sequence (37)
IMAP(<username>): fscking index file
2006 Dec 29
3
[git patches] ocfs2 fixes
Hi Linus,
Here are some 2.6.20 fixes for ocfs2. The patch by Zhen Wei isn't
really a fix, but a very small amount of support for a feature which is
mostly implemented in ocfs2-tools. Considering it's just a single attribute
export via configfs, I'd say it's pretty safe to merge.
Please pull from 'upstream-linus' branch of
2010 Dec 09
3
ZFS send/receive while write is enabled on receive side?
Hi all, from much of the documentation I''ve seen, the advice is to set
readonly=on on volumes on the receiving side during send/receive
operations. Is this still a requirement?
I''ve been trying the send/receive while NOT setting the receiver to
readonly and haven''t seen any problems even though we''re traversing
and ls''ing the dirs within the receiving
2009 Sep 08
1
Using Dovecot code in a non-LGPL package
Hi Timo (and all!),
Venaktesh is a friend of mine working on a patch to the venerable Alpine
mail reader so it can read Maildir format mailboxes on disk.
Venkatesh wanted to know if he could use some code from Dovecot's
lib-storage (safely opening/reading Maildirs, including accounting for NFS
and ESTALE) and the UID->message mapping code.
This way, Dovecot and Alpine would agree on
2018 Jan 03
0
stale file handle on gluster NFS client when trying to remove a directory
Hi all,
I haven't found any root cause or workaround for this yet. Can any one help
me in underatanding the issue?
Regards,
Jeevan.
On Dec 21, 2017 8:20 PM, "Jeevan Patnaik" <g1patnaik at gmail.com> wrote:
> Hi,
>
>
> After running rm -rf on a directory, the files under it got deleted, but
> the directory was not deleted and was showing stale file handle
2018 Jan 03
1
stale file handle on gluster NFS client when trying to remove a directory
An ESTALE error usually means the gfid could not be found. Does repeating
the "rm -rf" delete the directory?
Regards,
Nithya
On 3 January 2018 at 12:16, Jeevan Patnaik <g1patnaik at gmail.com> wrote:
> Hi all,
>
> I haven't found any root cause or workaround for this yet. Can any one
> help me in underatanding the issue?
>
> Regards,
> Jeevan.
>
>
2013 May 20
2
[PATCH] Btrfs: fix estale with btrfs send
This fixes bugzilla 57491. If we take a snapshot of a fs with a unlink ongoing
and then try to send that root we will run into problems. When comparing with a
parent root we will search the parents and the send roots commit_root, which if
we''ve just created the snapshot will include the file that needs to be evicted
by the orphan cleanup. So when we find a changed extent we will try
2017 Dec 21
2
stale file handle on gluster NFS client when trying to remove a directory
Hi,
After running rm -rf on a directory, the files under it got deleted, but
the directory was not deleted and was showing stale file handle error.
After 18 minutes, I'm able to delete the directory. So could anyone help me
in knowing what could have happened or when in general I get such errors.
The following is NFS log:
[2017-12-21 13:56:01.592256] I [MSGID: 108019]
2006 Apr 26
2
being courier_imap compatible
OK, I am running (for many years now) an existing courier_imap maildir
++ setup with exim as the MTA. Has been fine. I have my own custom
ldap schema for user account definition and authorization.
I installed dovecot 1.0 beta 7 as a test and have it listening on a
different port. I would like, if possible, to set things up so that
a user could use either dovecot or courier. Having to