similar to: rhel nfs bug with 5.5 - nfsd: blocked for more then 120 sec

Displaying 20 results from an estimated 3000 matches similar to: "rhel nfs bug with 5.5 - nfsd: blocked for more then 120 sec"

2010 May 19
3
mail location filesystem noatime, nodiratime?
Will Dovecot be negatively impacted if I change my XFS mount options to noatime,nodiratime? Thanks. -- Stan
2014 May 16
2
Re-mount a drive using its label name
Hello everybody, I would like your opinion on the following question, why this happens in centos and how to fix this (or a possible work around). I have a drive with no partitions and formatted with xfs filesystem. I give the drive a custom label "mydrive" and I mount it under /dev/mountpnts/mydrive. Then, I add a corresponding entry to fstab. These, are the steps I followed,
2013 Jun 21
1
LVM + XFS + external log + snapshots
Hi all, So I have an XFS file system within LVM which has an external log. My mount option in FSTAB is; /dev/vg_spock_data/lv_data /data xfs logdev=/dev/sdc1,nobarrier,logbufs=8,noatime,nodiratime 1 1 All is well no issues and very fast. Now I'd like to snapshot this bad boy and then run rsnapshot to create a few days backup. A snapshot volume is created w/o issue; lvcreate -L250G -s
2023 May 02
1
'error=No space left on device' but, there is plenty of space all nodes
Hi Gluster users, We are seeing 'error=No space left on device' issue and hoping someone might could advise? We are using a 12 node glusterfs v10.4 distributed vsftpd backup cluster for years (not new) and recently 2 weeks ago upgraded to v9 > v10.4. I do not know if the upgrade is related to this new issue. We are seeing a new issue 'error=No space left on device' error
2023 May 04
1
'error=No space left on device' but, there is plenty of space all nodes
Hi,Have you checked inode usage (df -i /lvbackups/brick ) ? Best Regards,Strahil Nikolov On Tuesday, May 2, 2023, 3:05 AM, brandon at thinkhuge.net wrote: Hi Gluster users, We are seeing 'error=No space left on device' issue and hoping someone might could advise? We are using a 12 node glusterfs v10.4 distributed vsftpd backup cluster for years (not new) and recently 2 weeks ago
2011 Jan 18
3
disk quotas + centos 5,5 +xfs
Hi all, is any one aware quotas not working in 5,5? I'm using XFS as a file system. My fstab has the appropriate usrquota,grpquota but when I try to run; quotacheck -cug /foo I get; quotacheck: Can't find filesystem to check or filesystem not mounted with quota option. I already have a large amount of data in the mount. Do quotas get implemented on only empty filesystems and
2023 Jun 07
1
Using glusterfs for virtual machines with qcow2 images
Hi Chris, here is a link to the settings needed for VM storage: https://github.com/gluster/glusterfs/blob/03592930239c3b43cbbdce17607c099ae075fd6d/extras/group-virt.example#L4 You can also ask in ovirt-users for real-world settings.Test well before changing production!!! IMPORTANT: ONCE SHARDING IS ENABLED, IT CANNOT BE DISABLED !!! Best Regards,Strahil Nikolov? On Mon, Jun 5, 2023 at 13:55,
2023 Jun 01
1
Using glusterfs for virtual machines with qcow2 images
Chris: Whilst I don't know what is the issue nor the root cause of your issue with using GlusterFS with Proxmox, but I am going to guess that you might already know that Proxmox "natively" supports Ceph, which the Wikipedia article for it says that it is a distributed object storage system. Maybe that might work better with Proxmox? Hope this helps. Sorry that I wasn't able
2023 Jun 01
1
[EXT] [Glusterusers] Using glusterfs for virtual machines with qco
We use qcow2 with libvirt based kvm on many small clusters and have found it to be exremely reliable though maybe not the fastest, though some of that is most of our storage is SATA SSDs in a software RAID1 config for each brick. What problems are you running into? You just mention 'problems' -wk On 6/1/23 8:42 AM, Christian Schoepplein wrote: > Hi, > > we'd like to use
2023 Jun 01
3
Using glusterfs for virtual machines with qcow2 images
Hi, we'd like to use glusterfs for Proxmox and virtual machines with qcow2 disk images. We have a three node glusterfs setup with one volume and Proxmox is attached and VMs are created, but after some time, and I think after much i/o is going on for a VM, the data inside the virtual machine gets corrupted. When I copy files from or to our glusterfs directly everything is OK, I've
2023 Jun 07
1
Using glusterfs for virtual machines with qcow2 images
Hi everybody Regarding the issue with mount, usually I am using this systemd service to bring up the mount points: /etc/systemd/system/glusterfsmounts.service [Unit] Description=Glustermounting Requires=glusterd.service Wants=glusterd.service After=network.target network-online.target glusterd.service [Service] Type=simple RemainAfterExit=true ExecStartPre=/usr/sbin/gluster volume list
2023 Jun 02
1
[EXT] [Glusterusers] Using glusterfs for virtual machines with qco
On oVirt / Redhat Virtualization, the following Gluster volumes settings are recommended to be applied (preferably at the creation of the volume) These settings are important for data reliability, ( Note that Replica 3 or Replica 2+1 is expected ) performance.quick-read=off performance.read-ahead=off performance.io-cache=off performance.low-prio-threads=32 network.remote-dio=enable
2023 Jun 02
1
[EXT] [Glusterusers] Using glusterfs for virtual machines with qco
Try turn off this options: performance.write-behind performance.flush-behind --- Gilberto Nunes Ferreira (47) 99676-7530 - Whatsapp / Telegram Em sex., 2 de jun. de 2023 ?s 07:55, Guillaume Pavese < guillaume.pavese at interactiv-group.com> escreveu: > On oVirt / Redhat Virtualization, > the following Gluster volumes settings are recommended to be applied > (preferably at
2023 Jun 05
1
[EXT] [Glusterusers] Using glusterfs for virtual machines with qco
Hi Gilberto, hi all, thanks a lot for all your answers. At first I changed both settings mentioned below and first test look good. Before changing the settings I was able to crash a new installed VM every time after a fresh installation by producing much i/o, e.g. when installing Libre Office. This always resulted in corrupt files inside the VM, but researching the qcow2 file with the
2012 Sep 17
1
Odd issue with XFS quotas reporting 8192.0EB in use
Hi All, I have this very odd issue where after a migration we have some users who have quotas showing 8192.0EB in use, when in reality on disk there is only a couple hundred megabytes. The host in question is a: * CentOS 6.3 host with mostly latest patches (except latest glibc patches) * running in SELinux Enforcing mode (Permissive doesn't seem to help) * all file systems are on Logical
2020 Nov 06
1
Centos 8 and xfs_quota
Folks I'm trying to use xfs_quota to keep track of disk space usage for my users. The documentation states that I should specify "uquota" as an option on the mount data in /etc/fstab. Yet, I cannot find the entry in fstab that corresponds to the Logical Volume that ends up being mounted on /home1. The system in question was installed on a single disk system. Later, after
2012 Aug 10
3
CentOS 6 kvm disk write performance
I have 2 similar servers. Since upgrading one from CentOS 5.5 to 6, disk write performance in kvm guest VMs is much worse. There are many, many posts about optimising kvm, many mentioning disk performance in CentOS 5 vs 6. I've tried various changes to speed up write performance, but northing's made a significant difference so far: - Install virtio disk drivers in guest - update the
2018 Apr 30
1
Gluster rebalance taking many years
I cannot calculate the number of files normally Through df -i I got the approximate number of files is 63694442 [root at CentOS-73-64-minimal ~]# df -i Filesystem Inodes IUsed IFree IUse% Mounted on /dev/md2 131981312 30901030 101080282 24% / devtmpfs 8192893 435 8192458 1% /dev tmpfs
2023 May 04
1
'error=No space left on device' but, there is plenty of space all nodes
Hi Strahil and Gluster users, Yes I had checked but, checked again and only 1% inode usage. 99% free. Same every node. Example: [root at nybaknode1 ]# df -i /lvbackups/brick Filesystem Inodes IUsed IFree IUse% Mounted on /dev/mapper/vgbackups-lvbackups 3108921344 93602 3108827742 1% /lvbackups [root at nybaknode1 ]# I neglected to clarify in
2014 Feb 27
3
Centos 6.5 on USB stick performance / stalls
Hi, I am running C6.5 64bit on an USB stick connected to a HP DL360G7; It is usually an ESXi host but eg for firmware updates (not available on SPP) I use this local installation. Problem are the lags and unresponsiveness we are seeing for example when running yum update in the installation phase. The whole system stalls but there is no io in vmstat. From what google tells us, this is a