Displaying 20 results from an estimated 400 matches similar to: "nfs-alpha feedback"
2011 Jun 09
1
NFS problem
Hi,
I got the same problem as Juergen,
My volume is a simple replicated volume with 2 host and GlusterFS 3.2.0
Volume Name: poolsave
Type: Replicate
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: ylal2950:/soft/gluster-data
Brick2: ylal2960:/soft/gluster-data
Options Reconfigured:
diagnostics.brick-log-level: DEBUG
network.ping-timeout: 20
performance.cache-size: 512MB
2011 Aug 24
1
Input/output error
Hi, everyone.
Its nice meeting you.
I am poor at English....
I am writing this because I'd like to update GlusterFS to 3.2.2-1,and I want
to change from gluster mount to nfs mount.
I have installed GlusterFS 3.2.1 one week ago,and replication 2 server.
OS:CentOS5.5 64bit
RPM:glusterfs-core-3.2.1-1
glusterfs-fuse-3.2.1-1
command
gluster volume create syncdata replica 2 transport tcp
2017 Jun 28
2
setting gfid on .trashcan/... failed - total outage
Hello,
recently we had two times a partial gluster outage followed by a total
outage of all four nodes. Looking into the gluster mailing list i found
a very similar case in
http://lists.gluster.org/pipermail/gluster-users/2016-June/027124.html
but i'm not sure if this issue is fixed...
even this outage happened on glusterfs 3.7.18 which gets no more updates
since ~.20 i would kindly ask
2017 Jun 29
0
setting gfid on .trashcan/... failed - total outage
On Wed, 2017-06-28 at 14:42 +0200, Dietmar Putz wrote:
> Hello,
>
> recently we had two times a partial gluster outage followed by a total?
> outage of all four nodes. Looking into the gluster mailing list i found?
> a very similar case in?
> http://lists.gluster.org/pipermail/gluster-users/2016-June/027124.html
If you are talking about a crash happening on bricks, were you
2017 Jun 29
1
setting gfid on .trashcan/... failed - total outage
Hello Anoop,
thank you for your reply....
answers inside...
best regards
Dietmar
On 29.06.2017 10:48, Anoop C S wrote:
> On Wed, 2017-06-28 at 14:42 +0200, Dietmar Putz wrote:
>> Hello,
>>
>> recently we had two times a partial gluster outage followed by a total
>> outage of all four nodes. Looking into the gluster mailing list i found
>> a very similar case
2006 Jun 04
1
text bubble (rectangle)?
Dear R wizards: sorry to bug everyone twice in one day.
I would like to annotate my graph by putting text strings into
rectangle boxes with a little cartoon-like bubble with a lid pointing
to a specific location. I can draw some sort of bubble-with-lid using
the R primitives. (has anyone done something like this already?)
the problem where I am stuck is that the width of the rectangle must
2000 Jun 20
2
Multiple Services on one Server
Newbie question!
We currently are running a product call
TAS from Syntax Corporation and would like to move to Samba. I have review
the documentation and cannot find how to set up muliple services on one
server. I tried using the Netbios name = and the include statement to
bring in another smb.conf file but I don't think I'm on the right track.
2007 Aug 21
4
Samba on Debian with shares on nfs-filesystem
Hello,
Actually I have some troubels concerning samba (3.0.24) on debian (4.0):
I would like to have some samba-shares on a nfs-filesystem.
NFS works fine for the Linux clients and the Linux server. The nfs
filesystem is mounted from an other server on /home.
The next step is to set up samba to have some shares on the nfs Filesystem
(/home/$LOGIN, anonymized configuration attached).
2017 Dec 21
2
stale file handle on gluster NFS client when trying to remove a directory
Hi,
After running rm -rf on a directory, the files under it got deleted, but
the directory was not deleted and was showing stale file handle error.
After 18 minutes, I'm able to delete the directory. So could anyone help me
in knowing what could have happened or when in general I get such errors.
The following is NFS log:
[2017-12-21 13:56:01.592256] I [MSGID: 108019]
2018 Oct 08
2
missing group affiliation on ad dc
Hi Rowland,
>> Hi,
>>
>> I've a strange problem. I migrated my NT4 PDC to a ad on my debian
>> stretch (samba version is 4.5.12).
>>
>> The Domain Controller has some shares for my users.
>>
>> One user just told me he can't access the share...before the
>> migration he was able to access the share btw! So I checked the ACL's
2010 Jul 13
5
Re-exporting an NFS mount.. Possible?
I have an issue that is not all that unique, so I'm hoping someone has
done it before.
On the client end I have some very old RedHat based systems. On the
server end is a Windows 2008 system running NFS server software. The
clients mount the server resource as an NFS2 mount but some compliance
issues were discovered with the setup. For various reasons, updating
the client is not an option at
2017 Jul 30
1
Lose gnfs connection during test
Hi all
I use Distributed-Replicate(12 x 2 = 24) hot tier plus
Distributed-Replicate(36 x (6 + 2) = 288) cold tier with gluster3.8.4
for performance test. When i set client/server.event-threads as small
values etc 2, it works ok. But if set client/server.event-threads as big
values etc 32, the netconnects will always become un-available during
the test, with following error messages in stree
2008 Apr 07
1
NFS, acls, proto, and "kernel: svc: unknown version"
Hi all,
1) My NFS3 clients don't display or obey existing non-POSIX ACLs on
files of NFS3-mounted exports.
2) setfacl on the client throws error and fails :
# setfacl -m u:stowler:rw testfile.text
setfacl: testfile.text: Operation not supported
3) at time of client mount the server's /var/log/messages shows
"kernel: svc: unknown version (3)".
Any thoughts greatly
2010 Oct 04
1
samba 3.3 - poor performance (compared to NFS)
I have a system that I'm vetting as a NAS server. It has a 2.0TB XFS filesystem mounted on /storage and I'm doing benchmarks using nfs3, nfs4, and samba. I'm testing via iozone by mounting the filesystem from my "nas client" box and then running iozone on the mounted filesystem. NFS seems pretty fast - ie, several orders of magnitude faster than samba, and I'm
2010 Apr 08
1
ZFS monitoring - best practices?
We''re starting to grow our ZFS environment and really need to start
standardizing our monitoring procedures.
OS tools are great for spot troubleshooting and sar can be used for
some trending, but we''d really like to tie this into an SNMP based
system that can generate graphs for us (via RRD or other).
Whether or not we do this via our standard enterprise monitoring tool
or
2018 Jan 03
0
stale file handle on gluster NFS client when trying to remove a directory
Hi all,
I haven't found any root cause or workaround for this yet. Can any one help
me in underatanding the issue?
Regards,
Jeevan.
On Dec 21, 2017 8:20 PM, "Jeevan Patnaik" <g1patnaik at gmail.com> wrote:
> Hi,
>
>
> After running rm -rf on a directory, the files under it got deleted, but
> the directory was not deleted and was showing stale file handle
2020 Jul 13
2
Dovecot permission denied errors on NFS after upgrade to 2.2.17
>>>>> "Mark" == Mark Moseley <moseleymark at gmail.com> writes:
Mark> This is just me throwing things out to look at, but did the
Mark> client mount on the old server use NFS3 and the new upgraded
Mark> client uses NFS4? Sometimes that can cause weirdness with id
Mark> mapping.?
Another thing to check is selinux, is it enabled? It's one of those
2018 Jan 03
1
stale file handle on gluster NFS client when trying to remove a directory
An ESTALE error usually means the gfid could not be found. Does repeating
the "rm -rf" delete the directory?
Regards,
Nithya
On 3 January 2018 at 12:16, Jeevan Patnaik <g1patnaik at gmail.com> wrote:
> Hi all,
>
> I haven't found any root cause or workaround for this yet. Can any one
> help me in underatanding the issue?
>
> Regards,
> Jeevan.
>
>
2012 Apr 27
3
OT Open Cobol
Hi,
Has anyone on this list had any success with installing and using Open
Cobol on RH or CentOS?
Any pointers would be very welcome.
I have Googled and am working with Vince on the Open Cobol mailing list
to try to resolve installation problems.
ChrisG
2015 Jun 16
2
Best practices for mailbox network file storage?
Is there a current best practice for a shared filesystem backend for mailboxes (in my case, all virtual accounts)?? NFS 3/4, GFS2, GlusterFS, other?? In the case of the latter ones, is the use of Director necessary?? Any specific anti-recommendations?? Also, if it matters, I *think* I want to use BackBlaze racks for storage.? Also, I've seen recommendations for the filesystem underneath like