Displaying 15 results from an estimated 15 matches for "virt2".
Did you mean:
virt
2017 Nov 15
1
unable to remove brick, pleas help
Hi,
I am trying to remove a brick, from a server which is no longer part of the
gluster pool, but I keep running into errors for which I cannot find
answers on google.
[root at virt2 ~]# gluster peer status
Number of Peers: 3
Hostname: srv1
Uuid: 2bed7e51-430f-49f5-afbc-06f8cec9baeb
State: Peer in Cluster (Disconnected)
Hostname: srv3
Uuid: 0e78793c-deca-4e3b-a36f-2333c8f91825
State: Peer in Cluster (Connected)
Hostname: srv4
Uuid: 1a6eedc6-59eb-4329-b091-2b9bc6f0834f
State:...
2010 Dec 13
2
Deploying libvirt with live migration
I have two physical servers: Virt1 and Virt2. I'm setting up live
migration with CentOS 5.5 between the two. I've done this by NFS
mounting /etc/libvirt and /var/lib/libvirt/images on both servers. This
is working well for me except for one thing.
I see the same list of VMs on each server (as expected), but each server
(Virt1 and...
2017 Dec 21
0
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...ume info output for this volume?
# gluster volume info virt_images
Volume Name: virt_images
Type: Replicate
Volume ID: 9f3c8273-4d9d-4af2-a4e7-4cb4a51e3594
Status: Started
Snapshot Count: 2
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: virt3:/data/virt_images/brick
Brick2: virt2:/data/virt_images/brick
Brick3: printserver:/data/virt_images/brick (arbiter)
Options Reconfigured:
features.quota-deem-statfs: on
features.inode-quota: on
features.quota: on
features.barrier: disable
features.scrub: Active
features.bitrot: on
nfs.rpc-auth-allow: on
server.allow-insecure: on
user.c...
2017 Dec 21
2
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...ges
> > Type: Replicate
> > Volume ID: 9f3c8273-4d9d-4af2-a4e7-4cb4a51e3594
> > Status: Started
> > Snapshot Count: 2
> > Number of Bricks: 1 x (2 + 1) = 3
> > Transport-type: tcp
> > Bricks:
> > Brick1: virt3:/data/virt_images/brick
> > Brick2: virt2:/data/virt_images/brick
> > Brick3: printserver:/data/virt_images/brick (arbiter)
> > Options Reconfigured:
> > features.quota-deem-statfs: on
> > features.inode-quota: on
> > features.quota: on
> > features.barrier: disable
> > features.scrub: Active
>...
2017 Dec 22
2
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...o virt_images
>
> Volume Name: virt_images
> Type: Replicate
> Volume ID: 9f3c8273-4d9d-4af2-a4e7-4cb4a51e3594
> Status: Started
> Snapshot Count: 2
> Number of Bricks: 1 x (2 + 1) = 3
> Transport-type: tcp
> Bricks:
> Brick1: virt3:/data/virt_images/brick
> Brick2: virt2:/data/virt_images/brick
> Brick3: printserver:/data/virt_images/brick (arbiter)
> Options Reconfigured:
> features.quota-deem-statfs: on
> features.inode-quota: on
> features.quota: on
> features.barrier: disable
> features.scrub: Active
> features.bitrot: on
> nfs.rpc-au...
2018 Feb 05
0
[ovirt-users] VM paused due unknown storage error
...s
> Status of volume: data
> Gluster process TCP Port RDMA Port Online
> Pid
> ------------------------------------------------------------
> ------------------
> Brick virt1:/gluster/brick2/data 49152 0 Y
> 3536
> Brick virt2:/gluster/brick2/data 49152 0 Y
> 3557
> Brick virt3:/gluster/brick2/data 49152 0 Y
> 3523
> Self-heal Daemon on localhost N/A N/A Y
> 32056
> Self-heal Daemon on virt2 N/A N/A...
2017 Dec 22
0
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...ain: 0
No. of heal failed entries: 1
I was immediately able to touch the file, so gluster was okay about
it, however heal info still showed the file for a while:
# gluster volume heal virt_images info
Brick virt3:/data/virt_images/brick
/fedora27.qcow2
Status: Connected
Number of entries: 1
Brick virt2:/data/virt_images/brick
/fedora27.qcow2
Status: Connected
Number of entries: 1
Brick printserver:/data/virt_images/brick
/fedora27.qcow2
Status: Connected
Number of entries: 1
Now heal info shows 0 entries, and the two data bricks have the same
md5sum, so it's back in sync.
I have a few...
2017 Dec 22
0
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...was immediately able to touch the file, so gluster was okay about
> it, however heal info still showed the file for a while:
> # gluster volume heal virt_images info
> Brick virt3:/data/virt_images/brick
> /fedora27.qcow2
> Status: Connected
> Number of entries: 1
>
> Brick virt2:/data/virt_images/brick
> /fedora27.qcow2
> Status: Connected
> Number of entries: 1
>
> Brick printserver:/data/virt_images/brick
> /fedora27.qcow2
> Status: Connected
> Number of entries: 1
>
>
>
> Now heal info shows 0 entries, and the two data bricks have the...
2017 Dec 20
2
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
Hi,
I have the following volume:
Volume Name: virt_images
Type: Replicate
Volume ID: 9f3c8273-4d9d-4af2-a4e7-4cb4a51e3594
Status: Started
Snapshot Count: 2
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: virt3:/data/virt_images/brick
Brick2: virt2:/data/virt_images/brick
Brick3: printserver:/data/virt_images/brick (arbiter)
Options Reconfigured:
features.quota-deem-statfs: on
features.inode-quota: on
features.quota: on
features.barrier: disable
features.scrub: Active
features.bitrot: on
nfs.rpc-auth-allow: on
server.allow-insecure: on
user.c...
2017 Dec 21
0
Gluster replicate 3 arbiter 1 in split brain. gluster cli seems unaware
...wing volume:
>
> Volume Name: virt_images
> Type: Replicate
> Volume ID: 9f3c8273-4d9d-4af2-a4e7-4cb4a51e3594
> Status: Started
> Snapshot Count: 2
> Number of Bricks: 1 x (2 + 1) = 3
> Transport-type: tcp
> Bricks:
> Brick1: virt3:/data/virt_images/brick
> Brick2: virt2:/data/virt_images/brick
> Brick3: printserver:/data/virt_images/brick (arbiter)
> Options Reconfigured:
> features.quota-deem-statfs: on
> features.inode-quota: on
> features.quota: on
> features.barrier: disable
> features.scrub: Active
> features.bitrot: on
> nfs.rpc-au...
2020 Sep 28
1
Centos8: Glusterd do not start correctly when I startup or reboot all server together
...systemctl enable --now glusterd
gluster peer probe virt1
gluster peer status
sh creavolume.sh gfsvol1 301G /gfsvol1 xfs
# NOTE: this is a my shell script to create fs on lvm
mkdir /gfsvol1/brick1
gluster pool list
gluster volume create gfsvol1 replica 2 virt1:/gfsvol1/brick1
virt2:/gfsvol1/brick1 force
gluster volume start gfsvol1
gluster volume info gfsvol1
gluster volume status gfsvol1
gluster volume heal gfsvol1
# add to /etc/fstab
vi /etc/fstab
virt1:/gfsvol2 /virt-gfs glusterfs defaults,noatime,_netdev 0
0
mkdir /virt-gfs
mount -a
All work...
2005 Aug 15
16
swig_up
Tracing down some things to add in validators and I''ve run across
something that kinda bothers me...
In order to implement validators you have to override the clone method.
The directors seems to be set up to specifically handle this situation.
However, whenever C++ calls back to the object''s methods the swig_get_up
function is returning false. It seems like swig_up
2020 Oct 02
0
Centos8: Glusterd do not start correctly when I startup or reboot all server together
The systemd glusterd.service unit modify do not resolve my problem
The solution is mount the glusterfs volume with this line into
/etc/fstab:
virt2:/gfsvol2 /virt-gfs glusterfs defaults,_netdev,noauto,x-systemd.automount,x-systemd.device-timeout=20,x-systemd.requires=glusterd.service 0 0
run
systemctl daemon-reload
and run this for mount the volume
systemctl restart virt\\x2dgfs.mount
Now all work fine, also when I reboot or when I p...
2012 Oct 16
1
Migrating a LV backed guest
I have a KVM VM that's backed by a logical volume on local disk.
I'd like to copy / move it to an identically configured host.
[root at virt2 cquinn]# virsh migrate --copy-storage-all --verbose --persistent node1.www qemu+ssh://10.102.1.11/system
error: Unable to read from monitor: Connection reset by peer
How should I effectively troubleshoot this? Am I misunderstanding how virsh migrate is supposed to work?
Regards,
Corey Quinn
2010 Sep 24
2
grep contents of file on remote server
...Y-DATA-/FAKE-KEY-DATA-/FAKE-KEY-DATA-==
bluethundr at lbsd8-2.summitnjhome.com==: No such file or directory
grep: root at bt-laptop: No such file or directory
[/code]
Ultimately, what I would like to do is script this in order to
automate this process:
[code]
#!/bin/sh
HOSTS="sum1 sum2 virt1 virt2 virt3 virt4 virt5 virt6 virt7"
SSHDIR=~/.ssh
RSYNC=/usr/local/bin/rsync
KEYFILE=/home/bluethundr/.ssh/id_rsa.pub
CAT='/bin/cat'
GREP='/bin/grep'
for h in $HOSTS ; do
scp $KEYFILE root@$h:~/
if [ $? = 0 ]; then
echo ; echo ; echo
echo "KEY TRANSFERRED TO $h"...