Displaying 20 results from an estimated 7000 matches similar to: "Ext3 and undeletion - A way how it could work."
2002 Jul 24
0
ext3 orphaned inode list error
Hello,
last fsck says my /tmp and my /home files system has errors, here is the
log from fsck using the read-only option:
How to recover from this savely?
How to avoid this in future?
i am using kernel 2.4.18 with the acl-patches from acl.bestbits.at and the
patches for the fileutils and e2fsprogs.
If you need more informations please mail.
2002 Jul 24
2
ext3 orphaned inode list
Hello,
last fsck says my /tmp and my /home files system has errors, here is the
log from fsck using the read-only option:
How to recover from this savely?
How to avoid this in future?
i am using kernel 2.4.18 with the acl-patches from acl.bestbits.at and the
patches for the fileutils and e2fsprogs.
If you need more informations please mail.
2002 May 07
3
inodes 100% full, how do I know?
How can you know beforehand, without running fsck, that all inodes are used
of a particular ext3 filesystem? Default systemtools use output from df,
which shows only a 50% usage of the filesystem, and pretend nothing is
wrong, while you really cant't move or copy a file to it. So I only found
out when running fsck.
This is my output from fsck (RH7.2, stock kernel, stock? ext3):
root# fsck
2005 Jul 08
2
Accidentally issued "mkswap" on ext3 fs -- recovery possible?
Hi,
I accidentally issued "mkswap" on a used ext3 fs partition (~30G) :-/
I have analyzed the behaviour of mkswap using two test files and it
appears to only change "some" bytes:
--8<--
--- swap2.xxd 2005-07-04 21:00:10.157261360 +0200
+++ swap1.xxd 2005-07-04 21:00:01.894517488 +0200
@@ -62,7 +62,7 @@
00003d0: 0000 0000 0000 0000 0000 0000 0000 0000 ................
2002 Dec 01
1
another idiot and ext3 - The inode is from a bad block in the inode table
i too am an idiot.. and request help
here's the sccop
i originally formatted the drive with 3 partitions and installed rh8
/boot
/
swap
everything worked.
my plan was to have this drive only be a data drive so i inatlled rh8 on
another drive and mounted this one on /share
so /share looked like
/share
/bin
/sbin
/mp3
/usr
/etc
/dev
and so on
so i deleted all the
2003 Jan 23
3
e2fsck too old / ext3 HTREE errors
hi,
i have a tricky situation here and and don't know where to start with
the solution.
i happened to set up a linux system, but i had to install the base
system on another machine. i used e2fsprogs 1.30-WIP (30-Sep-2002) for
this.
now the machine is up and running, but i have to use e2fsprogs 1.27
(8-Mar-2002). it's a debian system and i wanted to stay sane & stable,
hence using
2007 Apr 04
1
fsck.ext3 reporting large file I cannot find
I am checking a file system (ext3) as shown below. It is actually a fresh file
system, as I had deleted all partitions and created an ext3 file system. But
when I run the check with the verbose option, it says I have one large file.
Am I missing something here, or is it odd that I cannot find this large file it
is reporting? I reviewed some documentation on ext3 file systems, and
experimented
2009 Dec 29
2
ext3 partition size
Hi all,
I am running fedora 11 with kernel 2.6.30.9-102.fc11.x86_64 #1 SMP
Fri Dec 4 00:18:53 EST 2009 x86_64 x86_64 x86_64 GNU/Linux. I am
noticing a partition on my drive is reporting incorrect size with "df",
the partition is ext3 size 204GB with about 79GB actual usage, the "df"
result show the partition size to be 111GB, 93GB is missing. Please
advice on what
2006 Sep 24
1
Retaining undelete data on ext3
Having just spent a day trying to recover a deleted ext3 file
for a friend, I'm wondering about this way of maintining
undelete information in ext3, like is done for ext2:
The last step in the deletion process would be to put back
the (previously zeroed) block pointers. Since it gets logged
to the journal, I _think_ that this should be safe. The worst
that would happen is that, if the plug
2001 Nov 28
1
Reading output from "debugfs -R stat <8>"
Hi,
I'm trying to understand if my journal is fragmented. Here's some output
from `debugfs -R "stat <8>" /dev/sda3`:
Inode: 8 Type: regular Mode: 0600 Flags: 0x0 Generation: 0
User: 0 Group: 0 Size: 104857600
File ACL: 0 Directory ACL: 0
Links: 1 Blockcount: 205016
Fragment: Address: 0 Number: 0 Size: 0
ctime: 0x3c0442fd -- Tue Nov 27
2007 Oct 24
1
Problem with file system
While I untar a large archive on xfs , ext3 (ver 1.3 and ver 1.4) file
systems , on ppc processor and kernel ver 2.6.21 , I get an error. Also
sometimes, on ext3 (1.3 and 1.4) the file system goes read-only while
untarring.
The same tar file when untarred on a i386 machine works properly.
ERROR:
--------------
tar: Skipping to next header
gzip: stdin: invalid compressed data--crc error
tar:
2008 Jul 09
1
netCDF to TIFF
Greetings R users!
I am working with the ENSEMBLE climate data (10 min resolution daily
temperatures images for all of Europe 1950-2006). The data comes
packaged in a single netCDF file. I would like to read the data in and
export a subset (2002-2006) as geotiffs (one image per day). So far, I
can successfully read in the data and view the images within an R
display window. However, I have yet to
2011 May 08
1
questions about the output of "gee" and its summary
Dear R-helpers,
I am using the package "gee" to run a marginal model.
Here is the output.
In my simulated data, both x and z are time-varying, so I include their
interaction terms with time indicator (i.e. tind=0, if time 1, and 1 if time
2)
The data is simulated, so the true parameter of z both at time 1 and time 2
is 5, which is very close from the model output
for time 1, z =
2003 Sep 18
2
BIG Bug on Linux EXT3 file system, or "/bin/ls" problem?? Please Help!!!
Hi, all,
I "found" a big bug either related with Linux Ext3 file system or "bin/ls"
command. The problem is:
File system is no more than 220GB, but "/bin/ls" reports holding >370GB
data! The system is Redhat 8.0, with a Generic Kernel 2.4.20. Any
suggestions are greately appreciated. more information can be available upon
requests.
The following is the
2009 Apr 17
2
E2fsck and large file
How big is a file that e2fsck considers it to be a large file?
814611 blocks used (42.79%)
0 bad blocks
1 large file <----- that
Thanks
John Nelson
2005 Jun 20
0
[patch 2/3] fs/ext3/resize.c: fix sparse warnings
An embedded and charset-unspecified text was scrubbed...
Name: sparse-fs_ext3_resize.patch
URL: <http://listman.redhat.com/archives/ext3-users/attachments/20050620/8e33ffa8/attachment.ksh>
2005 Feb 15
1
[PATCH] ext3: Fix sparse -Wbitwise warnings.
Signed-off-by: Alexey Dobriyan <adobriyan at mail.ru>
---
fs/ext3/resize.c | 20 ++++++++++----------
fs/ext3/super.c | 8 ++++----
include/linux/ext3_fs.h | 2 +-
3 files changed, 15 insertions(+), 15 deletions(-)
Index: linux-warnings/include/linux/ext3_fs.h
===================================================================
---
2005 Jul 08
1
filesystem fragmentation stats?
Let me preface this by saying "Yes, I know *nix filesystems don't need
to worry about fragmentation".
That said, is there a way to check the overall level of fragmentation of
a live ext3 filesystem? I know about filefrag, but that's for specific
files. And I think e2fsck tells you, but only if you take the
filesystem offline for the scan. Is there anything that will give
2005 Mar 29
0
setAs between a new S4 class and "POSIXlt"
Dear R core team
Please apologize for posting the same question twice on R-help and
R-devel. Since I was not sure which list is appropriate I tried R-help (Tue
Mar 22), but got no answer.
Now I do not know if the formulation of my question was unclear or the
question is not so easy to answer or to easy (what I do not hope).
My problem: I create a new S4 class, containing one slot, data (of
2005 Mar 22
0
setAs between a new S4 class and "POSIXlt"
Dear R gurus
I've a question concerning the transformation of a new S4 class
(that I defined) and the existing class "POSIXlt". I did the
following:
## Definition of the new class:
setClass("dtime",
representation(data = "POSIXlt"),
prototype(data = as.POSIXlt("2004/06/01")))
## Transformation between the new class "dtime"