Dmitry Antipov
2020-Sep-29 09:18 UTC
[Gluster-users] Weird full heal on Distributed-Disperse volume with sharding
For the testing purposes, I've set up a localhost-only setup with 6x16M
ramdisks (formatted as ext4) mounted (with '-o user_xattr') at
/tmp/ram/{0,1,2,3,4,5} and SHARD_MIN_BLOCK_SIZE lowered to 4K. Finally
the volume is:
Volume Name: test
Type: Distributed-Replicate
Volume ID: 241d6679-7cd7-48b4-bdc5-8bc1c9940ac3
Status: Started
Snapshot Count: 0
Number of Bricks: 2 x 3 = 6
Transport-type: tcp
Bricks:
Brick1: [local-ip]:/tmp/ram/0
Brick2: [local-ip]:/tmp/ram/1
Brick3: [local-ip]:/tmp/ram/2
Brick4: [local-ip]:/tmp/ram/3
Brick5: [local-ip]:/tmp/ram/4
Brick6: [local-ip]:/tmp/ram/5
Options Reconfigured:
features.shard-block-size: 64KB
features.shard: on
storage.fips-mode-rchecksum: on
transport.address-family: inet
nfs.disable: on
performance.client-io-threads: off
Then I mount it under /mnt/test:
# mount -t glusterfs [local-ip]:/test /mnt/test
and create 4M file on it:
# dd if=/dev/random of=/mnt/test/file0 bs=1M count=4
This creates 189 shards of 64K each, in /tmp/ram/?/.shard:
/tmp/ram/0/.shard: 24
/tmp/ram/1/.shard: 24
/tmp/ram/2/.shard: 24
/tmp/ram/3/.shard: 39
/tmp/ram/4/.shard: 39
/tmp/ram/5/.shard: 39
To simulate data loss I just remove 2 arbitrary .shard directories,
for example:
# rm -rfv /tmp/ram/0/.shard /tmp/ram/5/.shard
Finally, I do full heal:
# gluster volume heal test full
and successfully got all shards under /tmp/ram/{0,5}.shard back.
But the things seems going weird for the following volume:
Volume Name: test
Type: Distributed-Disperse
Volume ID: aa621c7e-1693-427a-9fd5-d7b38c27035e
Status: Started
Snapshot Count: 0
Number of Bricks: 2 x (2 + 1) = 6
Transport-type: tcp
Bricks:
Brick1: [local-ip]:/tmp/ram/0
Brick2: [local-ip]:/tmp/ram/1
Brick3: [local-ip]:/tmp/ram/2
Brick4: [local-ip]:/tmp/ram/3
Brick5: [local-ip]:/tmp/ram/4
Brick6: [local-ip]:/tmp/ram/5
Options Reconfigured:
features.shard: on
features.shard-block-size: 64KB
storage.fips-mode-rchecksum: on
transport.address-family: inet
nfs.disable: on
After creating 4M file as before, I've got the same 189 shards
but 32K each. After deleting /tmp/ram/{0,5}/.shard and full heal,
I was able to get all shards back. But, after deleting
/tmp/ram/{3,4}/.shard and full heal, I've ended up with the following:
/tmp/ram/0/.shard:
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.10
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.11
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.12
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.13
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.14
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.15
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.16
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.17
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.2
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.22
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.23
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.27
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.28
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.3
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.31
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.34
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.35
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.37
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.39
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.4
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.40
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.44
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.45
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.46
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.47
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.53
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.54
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.55
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.57
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.58
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.6
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.63
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.7
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.9
/tmp/ram/1/.shard:
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.10
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.11
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.12
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.13
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.14
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.15
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.16
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.17
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.2
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.22
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.23
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.27
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.28
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.3
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.31
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.34
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.35
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.37
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.39
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.4
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.40
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.44
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.45
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.46
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.47
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.53
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.54
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.55
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.57
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.58
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.6
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.63
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.7
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.9
/tmp/ram/2/.shard:
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.10
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.11
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.12
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.13
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.14
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.15
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.16
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.17
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.2
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.22
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.23
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.27
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.28
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.3
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.31
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.34
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.35
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.37
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.39
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.4
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.40
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.44
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.45
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.46
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.47
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.53
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.54
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.55
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.57
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.58
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.6
-rw-r--r-- 2 root root 32768 Sep 29 12:01
951d7c52-7230-420b-b8bb-da887fffd41e.63
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.7
-rw-r--r-- 2 root root 32768 Sep 29 12:01 951d7c52-7230-420b-b8bb-da887fffd41e.9
So, /tmp/ram/{3,4}/.shard was not recovered. Even worse, /tmp/ram/5/.shard
has disappeared completely. And of course this breaks all I/O on
/mnt/test/file0,
for example:
# dd if=/dev/random of=/mnt/test/file0 bs=1M count=4
dd: error writing '/mnt/test/file0': No such file or directory
dd: closing output file '/mnt/test/file0': No such file or directory
Any ideas on what's going on here?
Dmitry
Xavi Hernandez
2020-Sep-30 05:58 UTC
[Gluster-users] [Gluster-devel] Weird full heal on Distributed-Disperse volume with sharding
Hi Dmitry, my comments below... On Tue, Sep 29, 2020 at 11:19 AM Dmitry Antipov <dmantipov at yandex.ru> wrote:> For the testing purposes, I've set up a localhost-only setup with 6x16M > ramdisks (formatted as ext4) mounted (with '-o user_xattr') at > /tmp/ram/{0,1,2,3,4,5} and SHARD_MIN_BLOCK_SIZE lowered to 4K. Finally > the volume is: > > Volume Name: test > Type: Distributed-Replicate > Volume ID: 241d6679-7cd7-48b4-bdc5-8bc1c9940ac3 > Status: Started > Snapshot Count: 0 > Number of Bricks: 2 x 3 = 6 > Transport-type: tcp > Bricks: > Brick1: [local-ip]:/tmp/ram/0 > Brick2: [local-ip]:/tmp/ram/1 > Brick3: [local-ip]:/tmp/ram/2 > Brick4: [local-ip]:/tmp/ram/3 > Brick5: [local-ip]:/tmp/ram/4 > Brick6: [local-ip]:/tmp/ram/5 > Options Reconfigured: > features.shard-block-size: 64KB > features.shard: on > storage.fips-mode-rchecksum: on > transport.address-family: inet > nfs.disable: on > performance.client-io-threads: off > > Then I mount it under /mnt/test: > > # mount -t glusterfs [local-ip]:/test /mnt/test > > and create 4M file on it: > > # dd if=/dev/random of=/mnt/test/file0 bs=1M count=4 > > This creates 189 shards of 64K each, in /tmp/ram/?/.shard: > > /tmp/ram/0/.shard: 24 > /tmp/ram/1/.shard: 24 > /tmp/ram/2/.shard: 24 > /tmp/ram/3/.shard: 39 > /tmp/ram/4/.shard: 39 > /tmp/ram/5/.shard: 39 > > To simulate data loss I just remove 2 arbitrary .shard directories, > for example: > > # rm -rfv /tmp/ram/0/.shard /tmp/ram/5/.shard > > Finally, I do full heal: > > # gluster volume heal test full > > and successfully got all shards under /tmp/ram/{0,5}.shard back. > > But the things seems going weird for the following volume: > > Volume Name: test > Type: Distributed-Disperse > Volume ID: aa621c7e-1693-427a-9fd5-d7b38c27035e > Status: Started > Snapshot Count: 0 > Number of Bricks: 2 x (2 + 1) = 6 > Transport-type: tcp > Bricks: > Brick1: [local-ip]:/tmp/ram/0 > Brick2: [local-ip]:/tmp/ram/1 > Brick3: [local-ip]:/tmp/ram/2 > Brick4: [local-ip]:/tmp/ram/3 > Brick5: [local-ip]:/tmp/ram/4 > Brick6: [local-ip]:/tmp/ram/5 > Options Reconfigured: > features.shard: on > features.shard-block-size: 64KB > storage.fips-mode-rchecksum: on > transport.address-family: inet > nfs.disable: on > > After creating 4M file as before, I've got the same 189 shards > but 32K each.This is normal. A dispersed volume writes encoded fragments of each block in each brick. In this case it's a 2+1 configuration, so each block is divided into 2 fragments. A third fragment is generated for redundancy and stored on the third brick.> After deleting /tmp/ram/{0,5}/.shard and full heal, > I was able to get all shards back. But, after deleting > /tmp/ram/{3,4}/.shard and full heal, I've ended up with the following: >This is not right. A disperse 2+1 configuration only supports a single failure. Wiping 2 fragments from the same file makes the file unrecoverable. Disperse works using the Reed-Solomon erasure code, which requires at least 2 healthy fragments to recover the data (in a 2+1 configuration). If you want to be able to recover from 2 disk failures, you need to create a 4+2 configuration. To make it more clear: a 2+1 configuration is like a traditional RAID5 with 3 disks. If you lose 2 disks, data is lost. A 4+2 is similar to a RAID6. Regards, Xavi> /tmp/ram/0/.shard: > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.10 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.11 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.12 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.13 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.14 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.15 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.16 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.17 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.2 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.22 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.23 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.27 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.28 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.3 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.31 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.34 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.35 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.37 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.39 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.4 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.40 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.44 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.45 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.46 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.47 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.53 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.54 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.55 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.57 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.58 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.6 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.63 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.7 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.9 > > /tmp/ram/1/.shard: > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.10 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.11 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.12 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.13 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.14 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.15 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.16 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.17 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.2 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.22 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.23 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.27 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.28 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.3 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.31 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.34 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.35 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.37 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.39 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.4 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.40 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.44 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.45 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.46 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.47 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.53 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.54 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.55 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.57 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.58 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.6 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.63 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.7 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.9 > > /tmp/ram/2/.shard: > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.10 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.11 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.12 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.13 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.14 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.15 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.16 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.17 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.2 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.22 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.23 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.27 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.28 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.3 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.31 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.34 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.35 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.37 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.39 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.4 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.40 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.44 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.45 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.46 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.47 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.53 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.54 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.55 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.57 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.58 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.6 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.63 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.7 > -rw-r--r-- 2 root root 32768 Sep 29 12:01 > 951d7c52-7230-420b-b8bb-da887fffd41e.9 > > So, /tmp/ram/{3,4}/.shard was not recovered. Even worse, /tmp/ram/5/.shard > has disappeared completely. And of course this breaks all I/O on > /mnt/test/file0, > for example: > > # dd if=/dev/random of=/mnt/test/file0 bs=1M count=4 > dd: error writing '/mnt/test/file0': No such file or directory > dd: closing output file '/mnt/test/file0': No such file or directory > > Any ideas on what's going on here?> Dmitry > _______________________________________________ > > Community Meeting Calendar: > > Schedule - > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > Bridge: https://bluejeans.com/441850968 > > > > > Gluster-devel mailing list > Gluster-devel at gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-devel > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200930/87a11eb8/attachment.html>