search for: posixs

Displaying 20 results from an estimated 6644 matches for "posixs".

Did you mean: posix
2018 Jan 16
1
[Possibile SPAM] Re: Problem with Gluster 3.12.4, VM and sharding
I've just done all the steps to reproduce the problem. Tha VM volume has been created via "qemu-img create -f qcow2 Test-vda2.qcow2 20G" on the gluster volume mounted via FUSE. I've tried also to create the volume with preallocated metadata, which moves the problem a bit far away (in time). The volume is a replice 3 arbiter 1 volume hosted on XFS bricks. Here are the
2019 Oct 05
3
How to turn on SMB3 POSIX extensions in Samba server?
Hi, I want to use the SMB3 POSIX extensions in the latest Samba (with SMB3.1.1, vers=3.1.1). By following the user manual, I have added the "posix" mount option when mounting, but it shows the following error messages. Error messages: [xxxxx] CIFS VFS: Server does not support mounting with posix SMB3.11 extensions. [xxxxx] CIFS VFS: cifs_mount failed w/return code = -95 Mount option:
2018 Jan 16
1
[Possibile SPAM] Re: Problem with Gluster 3.12.4, VM and sharding
An update: I've tried, for my tests, to create the vm volume as qemu-img create -f qcow2 -o preallocation=full gluster://gluster1/Test/Test-vda.img 20G et voila ! No errors at all, neither in bricks' log file (the "link failed" message disappeared), neither in VM (no corruption and installed succesfully). I'll do another test with a fully preallocated raw image. Il
2017 Jun 28
2
setting gfid on .trashcan/... failed - total outage
Hello, recently we had two times a partial gluster outage followed by a total outage of all four nodes. Looking into the gluster mailing list i found a very similar case in http://lists.gluster.org/pipermail/gluster-users/2016-June/027124.html but i'm not sure if this issue is fixed... even this outage happened on glusterfs 3.7.18 which gets no more updates since ~.20 i would kindly ask
2018 Jan 16
1
[Possibile SPAM] Re: Problem with Gluster 3.12.4, VM and sharding
I've made the test with raw image format (preallocated too) and the corruption problem is still there (but without errors in bricks' log file). What does the "link" error in bricks log files means ? I've seen the source code looking for the lines where it happens and it seems a warning (it doesn't imply a failure). Il 16/01/2018 17:39, Ing. Luca Lazzeroni - Trend
2006 Mar 07
4
POSIX time zone codes
The manual entry for as.POSIX says this about time zone codes... Usage as.POSIXct(x, tz = "") tz A timezone specification to be used for the conversion... but it fails to mention what these "specifications" are. So far, I have tried... as.POSIX(x, tz="UTC") ... works, gives UTC times as.POSIX(x, tz="UTC") ... works, gives EST times as.POSIX(x,
2018 Jan 17
1
[Possibile SPAM] Re: Problem with Gluster 3.12.4, VM and sharding
On Tue, Jan 16, 2018 at 10:47 PM, Ing. Luca Lazzeroni - Trend Servizi Srl < luca at gvnet.it> wrote: > I've made the test with raw image format (preallocated too) and the > corruption problem is still there (but without errors in bricks' log file). > > What does the "link" error in bricks log files means ? > > I've seen the source code looking for the
2018 Jan 17
1
[Possibile SPAM] Re: Problem with Gluster 3.12.4, VM and sharding
Of course. Here's the full log. Please, note that in FUSE mode everything works apparently without problems. I've installed 4 vm and updated them without problems. Il 17/01/2018 11:00, Krutika Dhananjay ha scritto: > > > On Tue, Jan 16, 2018 at 10:47 PM, Ing. Luca Lazzeroni - Trend Servizi > Srl <luca at gvnet.it <mailto:luca at gvnet.it>> wrote: > >
2011 Jan 04
2
[LLVMdev] LLVM for ARM target
At the last step of building llvm for arm target, I am unable to build llvm-gcc. I am trying the follwing options:   $ ../llvm-gcc/configure --target=arm-linux --enable-pic --program-prefix=llvm- --prefix=/llvm/install --disable-multilib --disable-optimized --enable-bindings =none --enable-llvm=$PWD/../llvm-2.8 --enable-languages=c,c++ --disable-bootstr ap   Then I get the following error when I
2018 Jan 17
1
[Possibile SPAM] Re: Problem with Gluster 3.12.4, VM and sharding
Really? Then which protocol exactly do you see this issue with? libgfapi? NFS? -Krutika On Wed, Jan 17, 2018 at 3:59 PM, Ing. Luca Lazzeroni - Trend Servizi Srl < luca at gvnet.it> wrote: > Of course. Here's the full log. Please, note that in FUSE mode everything > works apparently without problems. I've installed 4 vm and updated them > without problems. > > >
2017 Jun 29
0
setting gfid on .trashcan/... failed - total outage
On Wed, 2017-06-28 at 14:42 +0200, Dietmar Putz wrote: > Hello, > > recently we had two times a partial gluster outage followed by a total? > outage of all four nodes. Looking into the gluster mailing list i found? > a very similar case in? > http://lists.gluster.org/pipermail/gluster-users/2016-June/027124.html If you are talking about a crash happening on bricks, were you
2018 Jun 04
5
6.0.1-rc2 has been tagged
Hi, The 6.0.1-rc2 release has been tagged. Testers may begin testing and reporting results. -Tom
2018 Jan 17
1
[Possibile SPAM] Re: Problem with Gluster 3.12.4, VM and sharding
I actually use FUSE and it works. If i try to use "libgfapi" direct interface to gluster in qemu-kvm, the problem appears. Il 17/01/2018 11:35, Krutika Dhananjay ha scritto: > Really? Then which protocol exactly do you see this issue with? > libgfapi? NFS? > > -Krutika > > On Wed, Jan 17, 2018 at 3:59 PM, Ing. Luca Lazzeroni - Trend Servizi > Srl <luca at
2017 Jun 29
1
setting gfid on .trashcan/... failed - total outage
Hello Anoop, thank you for your reply.... answers inside... best regards Dietmar On 29.06.2017 10:48, Anoop C S wrote: > On Wed, 2017-06-28 at 14:42 +0200, Dietmar Putz wrote: >> Hello, >> >> recently we had two times a partial gluster outage followed by a total >> outage of all four nodes. Looking into the gluster mailing list i found >> a very similar case
2018 Apr 26
7
6.0.1-rc1 has been tagged
Hi, I've just tagged the 6.0.1-rc1 release. Testers may begin testing and uploading binaries. Also, any tester who has not tested 5.0.2-rc1 and would like to do so please try to do this before Friday, because I would like to tag 5.0.2-final then. As a reminder to users and developers, May 18 is the deadline for submitting merge requests for 6.0.1, so there is still time to get bug fixes
2018 Jan 17
1
[Possibile SPAM] Re: Problem with Gluster 3.12.4, VM and sharding
Hi, after our IRC chat I've rebuilt a virtual machine with FUSE based virtual disk. Everything worked flawlessly. Now I'm sending you the output of the requested getfattr command on the disk image: # file: TestFUSE-vda.qcow2 trusted.afr.dirty=0x000000000000000000000000 trusted.gfid=0x40ffafbbe987445692bb31295fa40105
2018 Jan 16
1
Problem with Gluster 3.12.4, VM and sharding
Also to help isolate the component, could you answer these: 1. on a different volume with shard not enabled, do you see this issue? 2. on a plain 3-way replicated volume (no arbiter), do you see this issue? On Tue, Jan 16, 2018 at 4:03 PM, Krutika Dhananjay <kdhananj at redhat.com> wrote: > Please share the volume-info output and the logs under /var/log/glusterfs/ > from all your
2023 Jan 25
2
[PATCH 00/12] acl: remove remaining posix acl handlers
Hey everyone, after we finished the introduction of the new posix acl api last cycle we still left the generic POSIX ACL xattr handler around for two reasons. First, because a few filesystems relied on the ->list() method of the generic POSIX ACL xattr handlers in their ->listxattr() inode operation. Second, during inode initalization in inode_init_always() the registered xattr handlers in
2018 Jan 18
1
[Possibile SPAM] Re: Problem with Gluster 3.12.4, VM and sharding
Another update. I've setup a replica 3 volume without sharding and tried to install a VM on a qcow2 volume on that device; however the result is the same and the vm image has been corrupted, exactly at the same point. Here's the volume info of the create volume: Volume Name: gvtest Type: Replicate Volume ID: e2ddf694-ba46-4bc7-bc9c-e30803374e9d Status: Started Snapshot Count: 0 Number
2011 Jan 02
0
[LLVMdev] LLVM for ARM target
On Jan 1, 2011, at 8:29 PM, akramul azim wrote: > Hi, > I am planning to follow the steps to install LLVM for the ARM-target (Processor: Xscale, Architecture: armv5te): > > 1. Install binutils-2.21 (downloaded from gnu.org) > > Steps: > > $ ./configure --target=arm-unknown-linux-gnueabi --program-prefix=arm- --prefix > =/llvm/arm