similar to: Zpool with raidz+mirror = wrong size displayed?

Displaying 20 results from an estimated 1000 matches similar to: "Zpool with raidz+mirror = wrong size displayed?"

2017 Jul 10
0
[Gluster-devel] gfid and volume-id extended attributes lost
@ pranith , yes . we can get the pid on all removexattr call and also print the backtrace of the glusterfsd process when trigerring removing xattr. I will write the script and reply back. On Sat, Jul 8, 2017 at 7:06 AM, Pranith Kumar Karampuri <pkarampu at redhat.com > wrote: > Ram, > As per the code, self-heal was the only candidate which *can* do > it. Could you check
2017 Jul 07
0
[Gluster-devel] gfid and volume-id extended attributes lost
We lost the attributes on all the bricks on servers glusterfs2 and glusterfs3 again. [root at glusterfs2 Log_Files]# gluster volume info Volume Name: StoragePool Type: Distributed-Disperse Volume ID: 149e976f-4e21-451c-bf0f-f5691208531f Status: Started Number of Bricks: 20 x (2 + 1) = 60 Transport-type: tcp Bricks: Brick1: glusterfs1sds:/ws/disk1/ws_brick Brick2: glusterfs2sds:/ws/disk1/ws_brick
2017 Jul 10
0
[Gluster-devel] gfid and volume-id extended attributes lost
Ram, If you see it again, you can use this. I am going to send out a patch for the code path which can lead to removal of gfid/volume-id tomorrow. On Mon, Jul 10, 2017 at 5:19 PM, Sanoj Unnikrishnan <sunnikri at redhat.com> wrote: > Please use the systemtap script(https://paste.fedoraproject.org/paste/ > EGDa0ErwX0LV3y-gBYpfNA) to check which process is invoking remove xattr
2017 Jul 08
2
[Gluster-devel] gfid and volume-id extended attributes lost
Ram, As per the code, self-heal was the only candidate which *can* do it. Could you check logs of self-heal daemon and the mount to check if there are any metadata heals on root? +Sanoj Sanoj, Is there any systemtap script we can use to detect which process is removing these xattrs? On Sat, Jul 8, 2017 at 2:58 AM, Ankireddypalle Reddy <areddy at commvault.com> wrote: >
2017 Jul 13
0
[Gluster-devel] gfid and volume-id extended attributes lost
Ram, I sent https://review.gluster.org/17765 to fix the possibility in bulk removexattr. But I am not sure if this is indeed the reason for this issue. On Mon, Jul 10, 2017 at 6:30 PM, Ankireddypalle Reddy <areddy at commvault.com> wrote: > Thanks for the swift turn around. Will try this out and let you know. > > > > Thanks and Regards, > > Ram > > *From:*
2017 Jul 10
2
[Gluster-devel] gfid and volume-id extended attributes lost
Please use the systemtap script( https://paste.fedoraproject.org/paste/EGDa0ErwX0LV3y-gBYpfNA) to check which process is invoking remove xattr calls. It prints the pid, tid and arguments of all removexattr calls. I have checked for these fops at the protocol/client and posix translators. To run the script .. 1) install systemtap and dependencies. 2) install glusterfs-debuginfo 3) change the path
2017 Jul 10
2
[Gluster-devel] gfid and volume-id extended attributes lost
Thanks for the swift turn around. Will try this out and let you know. Thanks and Regards, Ram From: Pranith Kumar Karampuri [mailto:pkarampu at redhat.com] Sent: Monday, July 10, 2017 8:31 AM To: Sanoj Unnikrishnan Cc: Ankireddypalle Reddy; Gluster Devel (gluster-devel at gluster.org); gluster-users at gluster.org Subject: Re: [Gluster-devel] gfid and volume-id extended attributes lost Ram,
2019 Aug 01
1
guestmount mounts gets corrupted somehow? [iscsi lvm guestmount windows filesystem rsync]
Hello everybody, I been trying to debug a problem for a month now and can use some insights and advice. This is the setup, I got two linux ha storage node providing iscsi disk, the disks is mounted on two linux kvm host and one backup server. The iscsi disk has lvm on it, the logical volume groups are visible on all servers. On the backup server I have the following running: # guestmount
2012 Jun 05
2
best practises for mail systems
hello! Can someone point me to some best practices in building high-available scalable mail system or! share your own success stories. I've read article in LJ "Building a Scalable High-Availability E-Mail System with Active Directory and More" but it seemed to be outdated and there's a single point of failure (Master node). What I want to achieve: high-available,
2011 Feb 01
1
zpool-poolname has 99 threads
After an upgrade of a busy server to Oracle Solaris 10 9/10, I notice a process called zpool-poolname that has 99 threads. This seems to be a limit, as it never goes above that. It is lower on workstations. The `zpool'' man page says only: Processes Each imported pool has an associated process, named zpool- poolname. The threads in this process are the pool''s
2015 Aug 12
1
[PATCH 1/2] disk-create: Allow preallocation off/metadata/full.
For raw, this allows "off" as a synonym for "sparse" (to make it consistent with qcow2). For qcow2, this allows "sparse" as a synonym for "off". It also adds qcow2 "full" preallocation, which is actually mapped to the qemu option "falloc" (see arguments about this on the qemu-devel mailing list, which we lost). This also updates the
2008 May 04
3
Some bugs/inconsistencies.
Hi. I''m working on getting the most recent ZFS to the FreeBSD''s CVS. Because of the huge amount of changes, I decided to work on ZFS regression tests, so I''m more or less sure nothing broke in the meantime. (Yes, I know about ZFS testsuite, but unfortunately I wasn''t able to port it to FreeBSD, it was just too much work. I''m afraid it is too
2006 May 09
3
Possible corruption after disk hiccups...
I''m not sure exactly what happened with my box here, but something caused a hiccup on multiple sata disks... May 9 16:40:33 sol scsi: [ID 107833 kern.warning] WARNING: /pci at 0,0/pci10de,5c at 9/pci-ide at a/ide at 0 (ata6): May 9 16:47:43 sol scsi: [ID 107833 kern.warning] WARNING: /pci at 0,0/pci-ide at 7/ide at 1 (ata3): May 9 16:47:43 sol timeout: abort request, target=0
2013 Dec 03
0
Problem booting guest with more than 8 disks
Hello All, On my host machine, I'm using kvm, libvirt, ceph and ubuntu versions as follows: >> QEMU emulator version 1.5.0 (Debian 1.5.0+dfsg-3ubuntu5), Copyright (c) 2003-2008 Fabrice Bellard >> root at kitt:~# virsh --version: 1.1.1 >> ceph version 0.67.4 (ad85b8bfafea6232d64cb7ba76a8b6e8252fa0c7) >> VERSION="13.10, Saucy Salamander" >> Linux kitt
2012 Jun 05
1
[ Re: best practises for mail systems]
Hello, If disk space and bandwidth are affordable (and from your setup it seems that they are affordable as you have everything locally) I would split the mail storage completely and use replication in between n-master servers (n=2 for your case). The replication is not yet fully tested, but Timo is actively working on this feature. The fear of lossing the imap session does not make sense (at
2006 Oct 24
3
determining raidz pool configuration
Hi all, Sorry for the newbie question, but I''ve looked at the docs and haven''t been able to find an answer for this. I''m working with a system where the pool has already been configured and want to determine what the configuration is. I had thought that''d be with zpool status -v <poolname>, but it doesn''t seem to agree with the
2007 Sep 18
0
Short how to for RHEL5 as dom0 and RHEL5 as domU
Hey, I was looking for step by step installation for RHEL5 as dom0 and RHEL5 as domU, but could not find it.I put some instructions, hope it will save some time for somebody. If you will find some errors please fix them :) I assume that you already have rhel 5 up and running (webserver as well).I will install guest system in paravirtualized mode (mine processor does not support full
2002 Aug 13
2
Samba 2.2.3a on HP-UX 10.20
Hi, when I am connecting to samba, openingor saving a MS Word document, nearly every time there is a core-file. In the log-file there I found following message: [2002/08/12 15:13:44, 0] lib/util_sec.c:assert_gid(111) Failed to set gid privileges to (-1,-24) now set to (0,0) uid=(0,0) [2002/08/12 15:13:44, 0] lib/util.c:smb_panic(1064) PANIC: failed to set gid What does it mean? The
2017 Jul 07
3
[Gluster-devel] gfid and volume-id extended attributes lost
On Fri, Jul 7, 2017 at 9:25 PM, Ankireddypalle Reddy <areddy at commvault.com> wrote: > 3.7.19 > These are the only callers for removexattr and only _posix_remove_xattr has the potential to do removexattr as posix_removexattr already makes sure that it is not gfid/volume-id. And surprise surprise _posix_remove_xattr happens only from healing code of afr/ec. And this can only happen
2010 Mar 31
1
[PATCH] Add LocalCopy transfer method to transfer local files to a target
Also changes command line parsing to require a pool to be specified when using libvirt output, meaning storage will always be copied. --- MANIFEST | 1 + lib/Sys/VirtV2V/Connection.pm | 74 +++++++++--------- lib/Sys/VirtV2V/Connection/LibVirt.pm | 18 +++-- lib/Sys/VirtV2V/Connection/LibVirtXML.pm | 11 ++- lib/Sys/VirtV2V/Target/LibVirt.pm