Displaying 20 results from an estimated 1000 matches similar to: "Kickstart and lvm"
2017 Apr 23
0
Proper way to remove a qemu-nbd-mounted volume usnig lvm
I either haven't searched for the right thing or the web doesn't contain
the answer.
I have used the following to mount an image and now I need to know the
proper way to reverse the process.
qemu-nbd -c /dev/nbd0 <qcow2 image using lvm>
vgscan --cache (had to use --cache to get the qemu-nbd volume to
be recognized, lvmetad is running)
vgchange -ay
2010 Oct 25
2
interesting kvm lvm collision issue
I've been running into a reproducible problem when using default LVM volume
group names to present block devices for virtual machines in KVM, and I'm
wondering why it is happening.
On dom0 I make a default VolGroup00 for the operating system. I make a
second VolGroup01 for logical volumes that will be block devices for virtual
systems.
In VolGroup01, I make two lv's for one system:
2011 Jul 22
0
Strange problem with LVM, device-mapper, and software RAID...
Running on a up-to-date CentOS 5.6 x86_64 machine:
[heller at ravel ~]$ uname -a
Linux ravel.60villagedrive 2.6.18-238.19.1.el5 #1 SMP Fri Jul 15 07:31:24 EDT 2011 x86_64 x86_64 x86_64 GNU/Linux
with a TYAN Computer Corp S4881 motherboard, which has a nVidia 4
channel SATA controller. It also has a Marvell Technology Group Ltd.
88SX7042 PCI-e 4-port SATA-II (rev 02).
This machine has a 120G
2016 Sep 03
4
mounting lvm2
CentOS 7 for everything
I am probably never using lvm again. I have never not once actually
benefited from its capabilities but this is probably the third or fourth
time it has gotten in my way.
Issue is that I use the tools so rarely that I can't recall what needs
to be done to properly mount the drives on another system and finding
that info on the web is becoming more difficult as the
2013 Mar 23
2
"Can't find root device" with lvm root after moving drive on CentOS 6.3
I have an 8-core SuperMicro Xeon server with CentOS 6.3. The OS is
installed on a 120 GB SSD connected by SATA, the machine also contains
an Areca SAS controller with 24 drives connected. The motherboard is a
SuperMicro X9DA7.
When I installed the OS, I used the default options, which creates an
LVM volume group to contain / and /home, and keeps /boot and /boot/efi
outside the volume group.
2013 Sep 25
3
Best Practice to remove an ISCSI LVM from a system
Hi,
I'd like to know what would be the best way to remove an iscsi lvm
storage from a server. (removing all reference to that storage etc.)
The storage in question will be reset and reformated and used on a
different server; so no LVM export is needed.
Do I have to do lvremove ..., vgremove ..., pvremove ... and do an
iscsiadm -m node -T ... -p ... -u and
iscsiadm -m node -o delete -T ...
2012 Aug 30
2
[PATCH v2] daemon: collect list of called external commands
guestfsd calls many different tools. Keeping track of all of them is
error prone. This patch introduces a new helper macro to put the command
string into its own ELF section:
GUESTFSD_EXT_CMD(C_variable, command_name);
This syntax makes it still possible to grep for used command names.
The actual usage of the collected list could be like this:
objcopy -j .guestfsd_ext_cmds -O binary
2017 Oct 01
3
LVM not activating on reboot
On 30/09/2017 17:49, Gordon Messmer wrote:
> On 09/30/2017 08:30 AM, Duncan Brown wrote:
>> However on a reboot, boot fails if I add that entry to fstab:
>> 'Timed out waiting for device dev-mapper-vg03\x2dstorage.device'
>>
>> I then have to activate it again with vgchange. I'm guessing I'm
>> going to need a grub option, or do something with
2007 Jun 13
1
lvm
Hello,
I have a physical volume with no volume group.
# /sbin/pvscan -n
WARNING: only considering physical volumes in no volume group
PV /dev/sdg2 lvm2 [148.95 GB]
Total: 1 [148.95 GB] / in use: 0 [0 ] / in no VG: 1 [148.95 GB]
Can I just create a volume group -- for example:
# vgcreate my_volume_group /dev/sdg2
and then activate it:
# vgchange -a y my_volume_group
I
2012 Aug 30
1
[PATCH] collect list of called external commands
guestfsd calls many different tools. Keeping track of all of them is
error prone. This patch introduces a new helper macro to put the command
string into its own ELF section:
GUESTFS_EXT_CMD(C_variable, command_name);
This syntax makes it still possible to grep for used command names.
The actual usage of the collected list could be like this:
objcopy -j .guestfs_ext_cmds -O binary
2015 Apr 01
1
can't mount an LVM volume inCentos 5.10
I have a degraded raid array (originally raid-10, now only two drives)
that contains an LVM volume. I can see in the appended text that the
Xen domains are there but I don't see how to mount them. No doubt this
is just ignorance on my part but I wonder if anyone would care to
direct me? I want to be able to retrieve dom-0 and one of the dom-Us
to do data recovery, the others are of
2013 Sep 20
1
Creating 38TB ext4 FS
mkfs.ext4 fails to create 38TB file system on CentOS 6.4 64bit with this
error:
mkfs.ext4: Size of device /dev/vg02/vtapes too big to be expressed in 32
bits
using a blocksize of 4096.
More details follow:
# uname -a
Linux tzbackup 2.6.32-358.18.1.el6.x86_64 #1 SMP Wed Aug 28 17:19:38 UTC
2013 x86_64 x86_64 x86_64 GNU/Linux
# fdisk -l /dev/sdc
Disk /dev/sdc: 41996.7 GB, 41996727091200
2011 Jun 02
2
increase harddisk diskspace Failed to suspend LogVol00
Hi,
I want to increase my harddisk space and receive the following error
# lvextend -l +323 /dev/VolGroup00/LogVol00
Extending logical volume LogVol00 to 48.97 GB
device-mapper: reload ioctl failed: Invalid argument
Failed to suspend LogVol00
Can you help me please?
# fdisk -l
Disk /dev/hda: 54.7 GB, 54759997440 bytes
255 heads, 63 sectors/track, 6657 cylinders
Units =
2012 Apr 12
1
CentOS 6.2 anaconda bug?
I have a kickstart file with the following partitioning directives:
part /boot --fstype ext3 --onpart=sda1
part pv.100000 --onpart=sda2 --noformat
volgroup vol0 pv.100000 --noformat
logvol / --vgname=vol0 --name=lvol1 --useexisting --fstype=ext4
logvol /tmp --vgname=vol0 --name=lvol2 --useexisting --fstype=ext4
logvol swap --vgname=vol0 --name=lvol3 --useexisting
logvol /data --vgname=vol0
2017 Dec 18
2
Upgrading from Gluster 3.8 to 3.12
Hi,
I have a cluster of 10 servers all running Fedora 24 along with
Gluster 3.8. I'm planning on doing rolling upgrades to Fedora 27 with
Gluster 3.12. I saw the documentation and did some testing but I
would like to run my plan through some (more?) educated minds.
The current setup is:
Volume Name: vol0
Distributed-Replicate
Number of Bricks: 2 x (2 + 1) = 6
Bricks:
Brick1:
2016 Jul 26
0
[PATCH 2/5] daemon: lvm-filter: start lvmetad better
Currently lvmetad is started in init, and thus using the system
(= appliance) configuration of lvm. Later on, in the daemon, a local
copy of the lvm configuration is setup, and set it for use using the
LVM_SYSTEM_DIR environment variable: this means only the programmes
executed by the daemon will use the local lvm configuration, and not
lvmetad.
Thus manually start lvmetad from the daemon, right
2017 Dec 19
0
Upgrading from Gluster 3.8 to 3.12
On Tue, Dec 19, 2017 at 1:10 AM, Ziemowit Pierzycki <ziemowit at pierzycki.com>
wrote:
> Hi,
>
> I have a cluster of 10 servers all running Fedora 24 along with
> Gluster 3.8. I'm planning on doing rolling upgrades to Fedora 27 with
> Gluster 3.12. I saw the documentation and did some testing but I
> would like to run my plan through some (more?) educated minds.
>
2013 Jul 09
2
Re: Libvirt and Glusterfs
> Hi,
>
> I'm trying to use qemu native glusterfs integration with libvirt. It's
> all working well from the qemu side, but libvirt fails to start a domain
> with a gluster drive or attach a drive.
> I have exactly the same error as this person:
> https://www.redhat.com/archives/libvirt-users/2013-April/msg00204.html
>
> I use qemu 1.5.1 with glusterfs 3.4 beta 4
2013 Jul 10
1
Re: Libvirt and Glusterfs
On 07/09/2013 08:18 PM, Olivier Mauras wrote:
> On 2013-07-09 09:40, Vijay Bellur wrote:
>
>>> Hi, I'm trying to use qemu native glusterfs integration with libvirt.
>>> It's all working well from the qemu side, but libvirt fails to start
>>> a domain with a gluster drive or attach a drive. I have exactly the
>>> same error as this person:
2017 Jul 27
0
[PATCH v2] daemon: Remove GUESTFSD_EXT_CMD.
GUESTFSD_EXT_CMD was used by OpenSUSE to track which external commands
are run by the daemon and package those commands into the appliance.
It is no longer used by recent SUSE builds, so remove it.
Thanks: Pino Toscano, Olaf Hering.
---
daemon/9p.c | 3 +-
daemon/available.c | 7 +--
daemon/base64.c | 6 +--
daemon/blkid.c | 10 ++---