similar to: [parted-devel] "Error: Can't have a partition outside the disk!"

Displaying 20 results from an estimated 70 matches similar to: "[parted-devel] "Error: Can't have a partition outside the disk!""

2012 Feb 06
3
[PATCH 1/3] NEW API: add a new api zero_fs
Add the new api zero_fs to erase the filesystem signatures on a device but now erase any data. Signed-off-by: Wanlong Gao <gaowanlong at cn.fujitsu.com> --- daemon/zero.c | 17 +++++++++++++++++ generator/generator_actions.ml | 14 ++++++++++++++ src/MAX_PROC_NR | 2 +- 3 files changed, 32 insertions(+), 1 deletions(-) diff --git a/daemon/zero.c
2009 Dec 07
2
eve online error after splash screen
My wine version is 1.0.1 on Ubuntu 9.04 64 bit. When I start up eve on the command prompt the splash screen loads then it aborts with this error message: fixme:ntdll:NtQuerySystemInformation info_class SYSTEM_PROCESSOR_PERFORMANCE_INFORMATION fixme:heap:HeapSetInformation 0x9f0000 0 0x33fcac 4
2016 Aug 21
0
Kickstart issue with UEFi
On 08/19/2016 11:35 PM, Phil Manuel wrote: > The install fails under UEFi due to the fact the partitions are not cleared, and it doesn?t have any space to continue. Is there an extra step I need to do to remove the original partitions before the new layout will work ? Yes. If you have pre-existing RAID partitions, they'll automatically assemble before Anaconda runs, and the installer
2016 Aug 22
2
Kickstart issue with UEFi
On Sun, 21 Aug 2016 at 12:51 Gordon Messmer <gordon.messmer at gmail.com> wrote: > On 08/19/2016 11:35 PM, Phil Manuel wrote: > > The install fails under UEFi due to the fact the partitions are not > cleared, and it doesn?t have any space to continue. Is there an extra step > I need to do to remove the original partitions before the new layout will > work ? > > >
2013 Jan 05
2
BUG btrfs fi show displays stale btrfs volume
I''ve filed this bug under util-linux, because I think wipefs isn''t deleting all btrfs metadata it could. But ultimately it appears to be a btrfs bug because nothing else sees the stale volume. https://bugzilla.redhat.com/show_bug.cgi?id=889888#c15 btrfs-progs-0.20.rc1.20121017git92d9eec-1.fc18.x86_64 e2fs-progs-1.42.5-1.fc18.x86_64 kernel 3.7.1-2 Brand new 80GB virtual disk,
2013 Feb 04
1
Should we always do wipefs before mkfs? Discuss ...
https://bugzilla.redhat.com/show_bug.cgi?id=907554 In particular comments 2 & 3. We could change libguestfs's guestfs_mkfs (internally) so it always does an implicit wipefs on the filesystem. wipefs is not too onerous -- in particular I believe it only writes to a few chosen areas of the disk. Especially considering that we're about to run mkfs anyway which for some filesystems
2014 Jan 19
0
Re: mke2fs fails due to wipefs --force
On Sun, Jan 19, 2014 at 07:42:58PM +0100, Olaf Hering wrote: > My wipefs has no --force option, as a result mke2fs command in guestfish fails. > > Looks like 72dd398679cd0bb803daf306d12558369615ba70 needs an adjustment > to make use of do_wipefs? > Also str_wipefs is not used in that file. Also daemon/zero.c has a wipefs_has_force_option function. It seems like it would be better
2018 Jul 31
1
Issues booting centos7 [dracut is failing to enable centos/root, centos/swap LVs]
Finally, I was able to find a solution for this. I'm posting the result here, just in case that it's useful for someone else in the future ... The problem was not in initramfs (dracut), but on the fact that /dev/sda2 partition(PV), after I resized it, somehow got 2 partition table signatures. One was set as a "dos" partition with an offset "0x1fe" and second set as
2012 Feb 08
2
[PATCH v3 0/2] Add a new api wipefs
Hi Rich: This is a alternative patch, if you like the api name called wipefs ;-P Thanks -Wanlong Gao Wanlong Gao (2): NEW API: add a new api wipefs virt-format: erase the filesystem signatures before erase partitions TODO | 2 -- daemon/zero.c | 17 +++++++++++++++++ format/format.c | 19 ++++---------------
2014 Jan 29
1
[PATCH] Proposal for a pacifier option with mkdiskimage
Hallo, Mattias, Du meintest am 28.01.14: >> Yes, we should probably fix that. The tool was originally designed >> for images, not for hardware devices, and it doesn't make much sense >> to zero the whole hardware device like that. > Probably not. But a left over backup GPT at the end of a "disk" (hard > disk, thumb drive, SSD, eMMC) now confuses some UEFI
2014 Jan 19
2
mke2fs fails due to wipefs --force
My wipefs has no --force option, as a result mke2fs command in guestfish fails. Looks like 72dd398679cd0bb803daf306d12558369615ba70 needs an adjustment to make use of do_wipefs? Also str_wipefs is not used in that file. Olaf
2015 Mar 23
0
xfs fsck error metadata corruption
On Mon, Mar 23, 2015 at 2:24 PM, Stephen Drotar <stephen at artifex360.com> wrote: > Hi, > > Everytime I restart Centos 7 I receive a error saying? > > metadata is corrupt > > and then I need to go through the process of mount and unmount the disk uuid then run > > xfs_repair {some uuid} > or > xfs_repair -L {some uuid} which ultimately corrupts
2019 Feb 26
0
Problem with mdadm, raid1 and automatically adds any disk to raid
On 2/24/19 9:01 PM, Jobst Schmalenbach wrote: > I tried to delete the MDX, I removed the disks by failing them, then removing each array md0, md1 and md2. > I also did > > dd if=/dev/zero of=/dev/sdX bs=512 seek=$(($(blockdev --getsz /dev/sdX)-1024)) count=1024 Clearing the initial sectors doesn't do anything to clear the data in the partitions.? They don't become blank
2020 Sep 14
0
erasing a disk
what if you just dd the first 1GB of the disk and the last GB of the disk (the last because of RAID signatures of some controllers that write to the end of the disk) Look at this article and modify accordingly https://zedt.eu/tech/linux/using-dd-to-repeatedly-erase-a-specific-range-of-sectors-on-the-hard-disk/ Also, use wipefs -a (Gordon Messmer answered faster than me) On Mon, Sep 14, 2020 at
2014 Sep 28
0
Re: Why libguestfs guest exist exceptionally?
On Sun, Sep 28, 2014 at 11:04:11PM +0800, Zhi Yong Wu wrote: > On Sun, Sep 28, 2014 at 6:48 PM, Richard W.M. Jones <rjones@redhat.com> wrote: > > On Sun, Sep 28, 2014 at 06:33:53PM +0800, Zhi Yong Wu wrote: > >> On Sun, Sep 28, 2014 at 6:26 PM, Richard W.M. Jones <rjones@redhat.com> wrote: > >> > Which version of libguestfs is this? I think you'll
2014 Sep 10
1
Questions about gluster reblance
Hello, Recently I spent a bit time understanding rebalance since I want to know its performance given that there could be more and more bricks to be added into my glusterfs volume and there will be more and more files and directories in the existing glusterfs volume. During the test I saw something which I'm really confused about. Steps: SW versions: glusterfs 3.4.4 + centos 6.5 Inital
2014 Sep 28
2
Re: Why libguestfs guest exist exceptionally?
On Sun, Sep 28, 2014 at 11:31 PM, Richard W.M. Jones <rjones@redhat.com> wrote: > On Sun, Sep 28, 2014 at 11:04:11PM +0800, Zhi Yong Wu wrote: >> On Sun, Sep 28, 2014 at 6:48 PM, Richard W.M. Jones <rjones@redhat.com> wrote: >> > On Sun, Sep 28, 2014 at 06:33:53PM +0800, Zhi Yong Wu wrote: >> >> On Sun, Sep 28, 2014 at 6:26 PM, Richard W.M. Jones
2012 Feb 10
3
[PATCH-v4 0/2] Add a new api wipefs
Hi Rich: This is a latest patch. ;-P Comments? Thanks -Wanlong Gao Wanlong Gao (2): NEW API: add a new api wipefs virt-format: erase the filesystem signatures before erase partitions TODO | 2 -- daemon/zero.c | 17 +++++++++++++++++ format/format.c | 19 ++++--------------- generator/generator_actions.ml | 14
2017 Apr 28
2
[PATCH] common/options: Change drv struct to store drive index instead of device name.
The device name is only used by guestfish (when using the -N option to prepare drives). We constructed the device name very naively, basically ‘sprintf ("/dev/sd%c", next_drive)’. This stores the device index instead, and only constructs the device name in guestfish. Also the device name is constructed properly using guestfs_int_drive_name so it can cope with #drives > 26. ---
2019 Feb 26
2
Problem with mdadm, raid1 and automatically adds any disk to raid
> On 2/24/19 9:01 PM, Jobst Schmalenbach wrote: >> I tried to delete the MDX, I removed the disks by failing them, then >> removing each array md0, md1 and md2. >> I also did >> >> dd if=/dev/zero of=/dev/sdX bs=512 seek=$(($(blockdev --getsz >> /dev/sdX)-1024)) count=1024 > > > Clearing the initial sectors doesn't do anything to clear the