similar to: Setting location for indexes

Displaying 20 results from an estimated 10000 matches similar to: "Setting location for indexes"

2012 Aug 23
1
Stale NFS file handle
Hi, I'm a bit curious of error messages of the type "remote operation failed: Stale NFS file handle". All clients using the file system use Gluster Native Client, so why should stale nfs file handle be reported? Regards, /jon -------------- next part -------------- An HTML attachment was scrubbed... URL:
2011 Mar 24
3
fstat no such file or directory error
Periodically, we will have users IMAP sessions get hung. Their client sees a message like: The server responded: 'SERVERBUG] Internal error occurred. Refer to server log for more information On the server, we see the following in the logs: IMAP(<username>): Transaction log dovecot.index.log: duplicate transaction log sequence (37) IMAP(<username>): fscking index file
2004 Nov 23
1
NFS, test52, maildir, freebsd 4.x and OS X (Panther) Server
I've been working on setting up a Dovecot server with a maildir mailboxes stored over NFS. Initially, everything was working great, and then I started seeing this error as previously discussed on the list in my maillog: (mail_index_expunge): assertion failed: (!t->view->external) Timo, in another thread, suggested to remove line 247 in mail-index-transaction.c in the test52
2009 Mar 03
3
[PATCH 1/1] OCFS2: anti stale inode for nfs (V6)
For nfs exporting, ocfs2_get_dentry() returns the dentry for fh. ocfs2_get_dentry() may read from disk(when inode not in memory) without any cross cluster lock. this leads to load a stale inode. this patch fixes above problem. solution is that in case of inode is not in memory, we get the cluster lock(PR) of alloc inode where the inode in question is allocated from(this causes node on which
2008 Oct 23
2
[PATCH 1/1] OCFS2: fix for nfs getting stale inode.
Ocfs2 supports exporting. PROBLEM: There are 2 problems (1) Current version of ocfs2_get_dentry() may read from disk the inode WITHOUT any cross cluster lock. This may lead to load a stale inode. (2) for deleting an inode, ocfs2_remove_inode() doesn't sync/checkpoint to disk. This also may lead ocfs2_get_dentry() from other node read out stale inode. PROBLEM DETAIL: for problem (1), For
2018 Jan 03
1
stale file handle on gluster NFS client when trying to remove a directory
An ESTALE error usually means the gfid could not be found. Does repeating the "rm -rf" delete the directory? Regards, Nithya On 3 January 2018 at 12:16, Jeevan Patnaik <g1patnaik at gmail.com> wrote: > Hi all, > > I haven't found any root cause or workaround for this yet. Can any one > help me in underatanding the issue? > > Regards, > Jeevan. > >
2018 Jan 03
0
stale file handle on gluster NFS client when trying to remove a directory
Hi all, I haven't found any root cause or workaround for this yet. Can any one help me in underatanding the issue? Regards, Jeevan. On Dec 21, 2017 8:20 PM, "Jeevan Patnaik" <g1patnaik at gmail.com> wrote: > Hi, > > > After running rm -rf on a directory, the files under it got deleted, but > the directory was not deleted and was showing stale file handle
2017 Dec 21
2
stale file handle on gluster NFS client when trying to remove a directory
Hi, After running rm -rf on a directory, the files under it got deleted, but the directory was not deleted and was showing stale file handle error. After 18 minutes, I'm able to delete the directory. So could anyone help me in knowing what could have happened or when in general I get such errors. The following is NFS log: [2017-12-21 13:56:01.592256] I [MSGID: 108019]
2006 Aug 20
2
maildir autodetect failing. what's wrong with this setup?
i'm setting up for virtual_users w/ maildir. i have in my dovecot.conf: auth default { passdb passwd-file { args = /etc/dovecot/imapd.passwd } userdb static { args = static uid=300 gid=300 home=/var/dovecot/mail mail=maildir:/var/dovecot/mail/%d/%n/Maildir nice=10 } } default_mail_env = maildir:/var/dovecot/mail/%d/%n/Maildir:INDEX=/var/dovecot/indexes/%d/%n in
2010 Mar 16
3
I stream read - stale NFS file handle (reboot of server)
In the old days NFS Shared Path had a static handle (ie a number), normal based on some number pulled out of the file system/inode. To fix (well work around) a security issue, for about 10+ years now, when a NFS server reboots, it generates a new random handle for the NFS Share. (sever may generate a new random handle per mount request) The NFS Stale Handle happens when the client is still
2006 Apr 26
2
being courier_imap compatible
OK, I am running (for many years now) an existing courier_imap maildir ++ setup with exim as the MTA. Has been fine. I have my own custom ldap schema for user account definition and authorization. I installed dovecot 1.0 beta 7 as a test and have it listening on a different port. I would like, if possible, to set things up so that a user could use either dovecot or courier. Having to
2009 Feb 17
1
[PATCH 1/1] OCFS2: anti stale inode for nfs (V3)
For nfs exporting, ocfs2_get_dentry() returns the dentry for fh. ocfs2_get_dentry() may read from disk(when inode not in memory) without any cross cluster lock. this leads to load a stale inode. this patch fixes above problem. solution is that in case of inode is not in memory, we get the cluster lock(PR) of alloc inode where the inode in question is allocated from(this causes node on which
2006 Sep 04
2
Dovecot problem in RHEL4
Hi all, I have some problem with dovecot in RHEL4. IMAP server is dropping connection after woorking for some time. It seems that the login process count keeps increasing?(stale login process are not removed) Any points to sort the problem? -- Adarsh
2013 Sep 26
8
[Bug 69830] New: Kernel 3.11 and newer hang at X11 start with NVIDIA GeForce GT 430
https://bugs.freedesktop.org/show_bug.cgi?id=69830 Priority: medium Bug ID: 69830 Assignee: nouveau at lists.freedesktop.org Summary: Kernel 3.11 and newer hang at X11 start with NVIDIA GeForce GT 430 Severity: major Classification: Unclassified OS: Linux (All) Reporter: alan at signal11.us
2009 May 19
2
ESTALE error while synching
Hi, I am wondering how rsync-3.0.6 react if it encounters ESTALE error while synching? If I remember correctly then the rsync-2.6.0 skipping that file/dir in case of ESTALE error. Jignesh. -------------- next part -------------- HTML attachment scrubbed and removed
2017 Nov 29
0
ls performance on directories with small number of items
The -l flag is causing a metadata lookup for every file in the directory. The way the ls command does that it's with individual fstat calls to each directory entry. That's a lot of tiny network round trips with fops that don't even fill a standard frame thus each frame has a high percentage of overhead for tcp. Add to that the replica check to ensure you're not getting stale data
2005 Dec 02
1
Dovecot strange behaviors on Solaris
Folks - I've been using Dovecot for some time now, but have only recently upgraded to 1.0alpha4. I'm struggling to get my configuration right and I'm seeing some truly bizarre issues. My Setup: Solaris 2.8 / Thunderbird / Procmail / Dovecot 1.0alpha4 My mail setup is only for me, so Dovecot runs at an odd port and under my userID. All my mail is in my home directory, which
2009 Mar 06
2
[PATCH 1/1] OCFS2: anti stale inode for nfs (for 1.4git)
Back porting from mainline. For nfs exporting, ocfs2_get_dentry() returns the dentry for fh. ocfs2_get_dentry() may read from disk(when inode not in memory) without any cross cluster lock. this leads to load a stale inode. this patch fixes above problem. solution is that in case of inode is not in memory, we get the cluster lock(PR) of alloc inode where the inode in question is allocated
2017 Nov 27
4
ls performance on directories with small number of items
Hi, I have a situation where an apache web server is trying to locate the IndexDocument for a directory on a gluster volume. This URL is being hit roughly 20 times per second. There is only 1 file in this directory. However, the parent directory does have a large number of items (+123,000 files and dirs) and we are performing operations to move these files into 2 levels of
2011 Sep 21
3
[PATCH] Btrfs: fix orphan cleanup regression
In fixing how we deal with bad inodes, we had a regression in the orphan cleanup code, since it expects to get a bad inode back. So fix it to deal with getting -ESTALE back by deleting the orphan item manually and moving on. Thanks, Reported-by: Simon Kirby <sim@hostway.ca> Signed-off-by: Josef Bacik <josef@redhat.com> --- fs/btrfs/inode.c | 36