Displaying 20 results from an estimated 1000 matches similar to: "problem expanding a volume"
2010 Mar 02
2
crash when using the cp command to copy files off a striped gluster dir but not when using rsync
Hi,
I've got this strange problem where a striped endpoint will crash when
I try to use cp to copy files off of it but not when I use rsync to
copy files off:
[user at gluster5 user]$ cp -r Python-2.6.4/ ~/tmp/
cp: reading `Python-2.6.4/Lib/lib2to3/tests/data/fixers/myfixes/__init__.py':
Software caused connection abort
cp: closing
2024 Jun 11
1
[EXT] Replace broken host, keeping the existing bricks
Hi,
The method depends a bit if you use a distributed-only system (like me) or a replicated setting.
I'm using a distributed-only setting (many bricks on different servers, but no replication). All my servers boot via network, i.e., on a start, it's like a new host.
To rescue the old bricks, just set up a new server this the same OS, the same IP and and the same hostname (!very
2013 Jul 02
1
files do not show up on gluster volume
I am trying to touch files on a mounted gluster mount point.
gluster1:/gv0 24G 786M 22G 4% /mnt
[root at centos63 ~]# cd /mnt
[root at centos63 mnt]# ll
total 0
[root at centos63 mnt]# touch hi
[root at centos63 mnt]# ll
total 0
The files don't show up after I ls them, but if I try to do a mv operation
something very strange happens:
[root at centos63 mnt]# mv /tmp/hi .
mv:
2011 Feb 24
0
No subject
which is a stripe of the gluster storage servers, this is the
performance I get (note use a file size > amount of RAM on client and
server systems, 13GB in this case) :
4k block size :
111 pir4:/pirstripe% /sb/admin/scripts/nfsSpeedTest -s 13g -y
pir4: Write test (dd): 142.281 MB/s 1138.247 mbps 93.561 seconds
pir4: Read test (dd): 274.321 MB/s 2194.570 mbps 48.527 seconds
testing from 8k -
2014 Jul 16
1
anaconda, kickstart, lvm over raid, logvol --grow, centos7 mystery
I am testing some kickstarts on ESXi virtual machine with pair of 16GB disks.
Partitioning is lvm over raid.
If i am using "logvol --grow i get "ValueError: not enough free space in volume group"
Only workaround i can find is to add --maxsize=XXX where XXX is at least 640MB less than available.
(10 extents or 320Mb per created logical volume)
Following snippet is failing with
2008 Jan 03
4
physical device
Hi!
I'' ve a xen virtual machine, and i'' d like to assign cdrw device to it. My
physical device on domain0 is /dev/hda, and in the configuration file i''
ve this:
disk = [ ''phy:vg0/vx-disk,sda1,w'', ''phy:vg0/vx-swap,sda2,w'',
''phy:vg0/vx-data,sda5,w'' ]
I'' ve tried to assign the cdrom to it in many ways i
2009 Jan 06
3
Booting issue on CentOS 5.2+Xen 3.3.0
Hello All,
I finally compiled xen 3.3.0 using make World, installed it but I cannot get
it to Boot, I always get kernel panic, kernel not sync kind of errors, I
suppose it''s some hard disk or partition issue, my setup is as follows.
CentOS 5.2 x86_64 on a HP ML-115
Grub, menu.list
Entries
CentOS xen''s line
title CentOS (2.6.18-92.1.22.el5xen)
root
2017 Dec 21
3
Wrong volume size with df
Sure!
> 1 - output of gluster volume heal <volname> info
Brick pod-sjc1-gluster1:/data/brick1/gv0
Status: Connected
Number of entries: 0
Brick pod-sjc1-gluster2:/data/brick1/gv0
Status: Connected
Number of entries: 0
Brick pod-sjc1-gluster1:/data/brick2/gv0
Status: Connected
Number of entries: 0
Brick pod-sjc1-gluster2:/data/brick2/gv0
Status: Connected
Number of entries: 0
Brick
2018 Jan 10
0
Blocking IO when hot tier promotion daemon runs
Hi,
Can you send the volume info, and volume status output and the tier logs.
And I need to know the size of the files that are being stored.
On Tue, Jan 9, 2018 at 9:51 PM, Tom Fite <tomfite at gmail.com> wrote:
> I've recently enabled an SSD backed 2 TB hot tier on my 150 TB 2 server / 3
> bricks per server distributed replicated volume.
>
> I'm seeing IO get blocked
2024 Nov 05
1
Add an arbiter when have multiple bricks at same server.
Yes but I want to add.
Is it the same logic?
---
Gilberto Nunes Ferreira
+55 (47) 99676-7530
Proxmox VE
VinChin Backup & Restore
Em ter., 5 de nov. de 2024, 14:09, Aravinda <aravinda at kadalu.tech> escreveu:
> Hello Gilberto,
>
> You can create a Arbiter volume using three bricks. Two of them will be
> data bricks and one will be Arbiter brick.
>
> gluster volume
2018 Jan 18
2
Blocking IO when hot tier promotion daemon runs
Hi Tom,
The volume info doesn't show the hot bricks. I think you have took the
volume info output before attaching the hot tier.
Can you send the volume info of the current setup where you see this issue.
The logs you sent are from a later point in time. The issue is hit
earlier than the logs what is available in the log. I need the logs
from an earlier time.
And along with the entire tier
2018 Jan 10
2
Blocking IO when hot tier promotion daemon runs
The sizes of the files are extremely varied, there are millions of small
(<1 MB) files and thousands of files larger than 1 GB.
Attached is the tier log for gluster1 and gluster2. These are full of
"demotion failed" messages, which is also shown in the status:
[root at pod-sjc1-gluster1 gv0]# gluster volume tier gv0 status
Node Promoted files Demoted files
2009 Oct 27
2
oldschool xen kernel on sid
Hello!
In an attempt to finally get a recent Xen dom0 kernel booted on this
pesky amd64 system (see
<http://lists.xensource.com/archives/html/xen-users/2009-10/msg00485.html>
for the (short) story), I wanted to give this one one a try:
Marco Nenciarini <mnencia at prato.linux.it> wrote:
> I've applied latest forward ported patches from
>
>
2022 Apr 24
3
Installing mdadm and C7 on new computer
On 04/23/2022 09:19 PM, H wrote:
> On 04/19/2022 09:57 AM, Roberto Ragusa wrote:
>> On 4/18/22 1:27 PM, H wrote:
>>> I have a new computer with 2 x 2TB SSDs where I wanted to install C7 and use mdadm for RAID1 configuration and encrypting the /home partition. On the net I found https://tuxfixer.com/centos-7-installation-with-lvm-raid-1-mirroring/ which I adopted slightly with
2023 Jun 01
3
Using glusterfs for virtual machines with qcow2 images
Hi,
we'd like to use glusterfs for Proxmox and virtual machines with qcow2
disk images. We have a three node glusterfs setup with one volume and
Proxmox is attached and VMs are created, but after some time, and I think
after much i/o is going on for a VM, the data inside the virtual machine
gets corrupted. When I copy files from or to our glusterfs
directly everything is OK, I've
2018 Jan 02
0
Wrong volume size with df
For what it's worth here, after I added a hot tier to the pool, the brick
sizes are now reporting the correct size of all bricks combined instead of
just one brick.
Not sure if that gives you any clues for this... maybe adding another brick
to the pool would have a similar effect?
On Thu, Dec 21, 2017 at 11:44 AM, Tom Fite <tomfite at gmail.com> wrote:
> Sure!
>
> > 1 -
2024 Nov 05
1
Add an arbiter when have multiple bricks at same server.
If you create a volume with replica 2 arbiter 1
you create 2 data bricks that are mirrored (makes 2 file copies)
+
you create 1 arbiter that holds metadata of all files on these bricks.
You "can" create all on the same server, but this makes no sense,
because when the server goes down, no files on these disks are
accessible anymore,
hence why bestpractice is to spread out over 3
2012 Jun 28
2
Strange du/df behaviour.
Hi all.
I have currently a server:
cat /etc/redhat-release
CentOS release 5.7 (Final)
uname -a
Linux host.domain.com 2.6.18-274.18.1.el5 #1 SMP Thu Feb 9 12:45:44
EST 2012 x86_64 x86_64 x86_64 GNU/Linux
I have there a filesystem mounted:
/dev/vg0/paczki /home/paczki-workdir ext4
defaults,noatime 0 0
on which df gives strange output:
LANG=C df -h
2018 Jan 18
0
Blocking IO when hot tier promotion daemon runs
Thanks for the info, Hari. Sorry about the bad gluster volume info, I
grabbed that from a file not realizing it was out of date. Here's a current
configuration showing the active hot tier:
[root at pod-sjc1-gluster1 ~]# gluster volume info
Volume Name: gv0
Type: Tier
Volume ID: d490a9ec-f9c8-4f10-a7f3-e1b6d3ced196
Status: Started
Snapshot Count: 13
Number of Bricks: 8
Transport-type: tcp
Hot
2017 Dec 21
0
Wrong volume size with df
Could youplease provide following -
1 - output of gluster volume heal <volname> info
2 - /var/log/glusterfs - provide log file with mountpoint-volumename.log
3 - output of gluster volume <volname> info
4 - output of gluster volume <volname> status
5 - Also, could you try unmount the volume and mount it again and check the size?
----- Original Message -----
From: