similar to: Bitrot and panics

Displaying 20 results from an estimated 1000 matches similar to: "Bitrot and panics"

2007 Jun 09
2
zfs bug
dd if=/dev/zero of=sl1 bs=512 count=256000 dd if=/dev/zero of=sl2 bs=512 count=256000 dd if=/dev/zero of=sl3 bs=512 count=256000 dd if=/dev/zero of=sl4 bs=512 count=256000 zpool create -m /export/test1 test1 raidz /export/sl1 /export/sl2 /export/sl3 zpool add -f test1 /export/sl4 dd if=/dev/zero of=sl4 bs=512 count=256000 zpool scrub test1 panic. and message like on image. This message posted
2018 Apr 17
2
Bitrot - Restoring bad file
Hi, I have a question regarding bitrot detection. Following the RedHat manual (https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/bitrot-restore_corrupt_file) I am trying out bad-file-restoration after bitrot. "gluster volume bitrot VOLNAME status" gets me the GFIDs that are corrupt and on which Host this happens. As far as I can tell
2018 Apr 18
0
Bitrot - Restoring bad file
On 04/17/2018 06:25 PM, Omar Kohl wrote: > Hi, > > I have a question regarding bitrot detection. > > Following the RedHat manual (https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/bitrot-restore_corrupt_file) I am trying out bad-file-restoration after bitrot. > > "gluster volume bitrot VOLNAME status" gets me the
2017 Sep 22
0
how to verify bitrot signed file manually?
ok, from bitrot code I figured out gluster using sha256 hashing algo. Now coming to the problem, during scrub run in my cluster some of my files were marked as bad in few set of nodes. I just wanted to confirm bad file. so, I have used "sha256sum" tool in Linux to manually get file hash. here is the result. file-1, file-2 marked as bad by scrub and file-3 is healthy. file-1 sha256
2017 Sep 29
1
how to verify bitrot signed file manually?
Hi Amudhan, Sorry for the late response as I was busy with other things. You are right bitrot uses sha256 for checksum. If file-1, file-2 are marked bad, the I/O should be errored out with EIO. If that is not happening, we need to look further into it. But what's the file contents of file-1 and file-2 on the replica bricks ? Are they matching ? Thanks and Regards, Kotresh HR On Mon, Sep 25,
2018 Apr 18
1
Bitrot strange behavior
Hi Sweta, Thanks, this drive me some more questions: 1. What is the reason of delaying signature creation ? 2. As a same file (replicated or dispersed) having different signature thought bricks is by definition an error, it would be good to triggered it during a scrub, or with a different tool. Is something like this planned ? Cheers ? C?dric Lemarchand > On 18 Apr 2018, at 07:53, Sweta
2018 Apr 16
2
Bitrot strange behavior
Hello, I am playing around with the bitrot feature and have some questions: 1. when a file is created, the "trusted.bit-rot.signature? attribute seems only created approximatively 120 seconds after its creations (the cluster is idle and there is only one file living on it). Why ? Is there a way to make this attribute generated at the same time of the file creation ? 2. corrupting a file
2017 Sep 25
2
how to verify bitrot signed file manually?
resending mail. On Fri, Sep 22, 2017 at 5:30 PM, Amudhan P <amudhan83 at gmail.com> wrote: > ok, from bitrot code I figured out gluster using sha256 hashing algo. > > > Now coming to the problem, during scrub run in my cluster some of my files > were marked as bad in few set of nodes. > I just wanted to confirm bad file. so, I have used "sha256sum" tool in >
2018 Apr 18
0
Bitrot strange behavior
Hi Cedric, Any file is picked up for signing by the bitd process after the predetermined wait of 120 seconds. This default value is captured in the volume option 'features.expiry-time' and is configurable - in your case, it can be set to 0 or 1. Point 2 is correct. A file corrupted before the bitrot signature is generated will not be successfully detected by the scrubber. That would
2017 Oct 03
1
how to verify bitrot signed file manually?
my volume is distributed disperse volume 8+2 EC. file1 and file2 are different files lying in same brick. I am able to read the file from mount point without any issue because of EC it reads rest of the available blocks in other nodes. my question is "file1" sha256 value matches bitrot signature value but still, it is also marked as bad by scrubber daemon. why is that? On Fri, Sep
2017 Sep 21
2
how to verify bitrot signed file manually?
Hi, I have a file in my brick which was signed by bitrot and latter when running scrub it was marked as bad. Now, I want to verify file again manually. just to clarify my doubt how can I do this? regards Amudhan -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170921/f69ff7be/attachment.html>
2005 Jan 27
2
Samba not recognizing secondary groups
Hi all. I'm having a strange issue with a server running Samba 3.0.2a. The server has a single share, and under that are subfolders that contain documents. Each subfolder is owned by a different folder-specific group, and users belong to the groups that own the folders to which they need access. But suddenly one user can't access any folders. Windows (the client) suggests that she
2017 Nov 06
0
how to verify bitrot signed file manually?
Any update? On Fri, Oct 13, 2017 at 1:14 PM, Amudhan P <amudhan83 at gmail.com> wrote: > any update?. > > why is it marked bad? > > Any way to find out what happened to the file? > > > On Tue, Oct 3, 2017 at 12:44 PM, Amudhan P <amudhan83 at gmail.com> wrote: > >> >> my volume is distributed disperse volume 8+2 EC. >> file1 and file2 are
2012 Jan 15
22
Does raidzN actually protect against bitrot? If yes - how?
"Does raidzN actually protect against bitrot?" That''s a kind of radical, possibly offensive, question formula that I have lately. Reading up on theory of RAID5, I grasped the idea of the write hole (where one of the sectors of the stripe, such as the parity data, doesn''t get written - leading to invalid data upon read). In general, I think the same applies to bitrot of
2012 Feb 19
3
Non-parametric test for repeated measures and post-hoc single comparisons in R?
Some attribute x from 17 individuals was recorded repeatedly on 6 time points using a Likert scale with 7 distractors. Which statistical test(s) can I apply to check whether the changes along the 6 time points were significant? set.seed( 123 ) x <- matrix( sample( 1:7, 17*6, repl=T ), nrow = 17, byrow = TRUE, dimnames = list(1:17, paste( 'T', 1:6, sep='' )) ) I found
2017 Jun 19
2
total outage - almost
Hi, we use a bunch of replicated gluster volumes as a backend for our backup. Yesterday I noticed that some synthetic backups failed because of I/O errors. Today I ran "find /gluster_vol -type f | xargs md5sum" and got loads of I/O errors. The brick log file shows the below errors [2017-06-19 13:42:33.554875] E [MSGID: 116020] [bit-rot-stub.c:566:br_stub_check_bad_object]
2017 Jun 19
0
total outage - almost
Hi, I checked the attributes of one of the files with I/O errors root at chastcvtprd04:~# getfattr -d -e hex -m - /data/glusterfs/Server_Standard/1I-1-14/brick/Server_Standard/CV_MAGNETIC/V_1050932/CHUNK_11126559/SFILE_CONTAINER_014 getfattr: Removing leading '/' from absolute path names # file:
2006 Mar 30
39
Proposal: ZFS Hot Spare support
As mentioned last night, we''ve been reviewing a proposal for hot spare support in ZFS. Below you can find a current draft of the proposed interfaces. This has not yet been submitted for ARC review, but comments are welcome. Note that this does not include any enhanced FMA diagnosis to determine when a device is "faulted". This will come in a follow-on project, of which some
2009 Dec 12
2
Deleting contents of /tmp on shutdown
I use to have a line of code in /etc/init.d/syslog (I think this was the file) to delete the contents of my /tmp directory on shutdown. It originally came from fedora. I have searched for it but can't find it again. Anyone happen to have it? TIA
2023 Jan 19
1
really large number of skipped files after a scrub
Hi, Just to follow up my first observation from this email from december: automatic scheduled scrubs that not happen. We have now upgraded glusterfs from 7.4 to 10.1, and now see that the automated scrubs ARE running now. Not sure why they didn't in 7.4, but issue solved. :-) MJ On Mon, 12 Dec 2022 at 13:38, cYuSeDfZfb cYuSeDfZfb <cyusedfzfb at gmail.com> wrote: > Hi, > > I