On 13/09/17 06:21, Gaurav Yadav wrote:> Please provide the output of gluster volume info, gluster 
> volume status and gluster peer status.
>
> Apart? from above info, please provide glusterd logs, 
> cmd_history.log.
>
> Thanks
> Gaurav
>
> On Tue, Sep 12, 2017 at 2:22 PM, lejeczek 
> <peljasz at yahoo.co.uk <mailto:peljasz at yahoo.co.uk>> wrote:
>
>     hi everyone
>
>     I have 3-peer cluster with all vols in replica mode, 9
>     vols.
>     What I see, unfortunately, is one brick fails in one
>     vol, when it happens it's always the same vol on the
>     same brick.
>     Command: gluster vol status $vol - would show brick
>     not online.
>     Restarting glusterd with systemclt does not help, only
>     system reboot seem to help, until it happens, next time.
>
>     How to troubleshoot this weird misbehaviour?
>     many thanks, L.
>
>     .
>     _______________________________________________
>     Gluster-users mailing list
>     Gluster-users at gluster.org
>     <mailto:Gluster-users at gluster.org>
>     http://lists.gluster.org/mailman/listinfo/gluster-users
>     <http://lists.gluster.org/mailman/listinfo/gluster-users>
>
>
hi, here:
$ gluster vol info C-DATA
Volume Name: C-DATA
Type: Replicate
Volume ID: 18ffba73-532e-4a4d-84da-fceea52f8c2e
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: 
10.5.6.49:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-C-DATA
Brick2: 
10.5.6.100:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-C-DATA
Brick3: 
10.5.6.32:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-C-DATA
Options Reconfigured:
performance.md-cache-timeout: 600
performance.cache-invalidation: on
performance.stat-prefetch: on
features.cache-invalidation-timeout: 600
features.cache-invalidation: on
performance.io-thread-count: 64
performance.cache-size: 128MB
cluster.self-heal-daemon: enable
features.quota-deem-statfs: on
changelog.changelog: on
geo-replication.ignore-pid-check: on
geo-replication.indexing: on
features.inode-quota: on
features.quota: on
performance.readdir-ahead: on
nfs.disable: on
transport.address-family: inet
performance.cache-samba-metadata: on
$ gluster vol status C-DATA
Status of volume: C-DATA
Gluster process???????????????????????????? TCP Port? RDMA 
Port Online? Pid
------------------------------------------------------------------------------
Brick 10.5.6.49:/__.aLocalStorages/0/0-GLUS
TERs/0GLUSTER-C-DATA???????????????????? N/A?????? N/A 
N?????? N/A
Brick 10.5.6.100:/__.aLocalStorages/0/0-GLU
STERs/0GLUSTER-C-DATA??????????????????? 49152???? 0 Y?????? 
9376
Brick 10.5.6.32:/__.aLocalStorages/0/0-GLUS
TERs/0GLUSTER-C-DATA???????????????????? 49152???? 0 Y?????? 
8638
Self-heal Daemon on localhost?????????????? N/A?????? N/A 
Y?????? 387879
Quota Daemon on localhost?????????????????? N/A?????? N/A 
Y?????? 387891
Self-heal Daemon on rider.private.ccnr.ceb.
private.cam.ac.uk?????????????????????????? N/A?????? N/A 
Y?????? 16439
Quota Daemon on rider.private.ccnr.ceb.priv
ate.cam.ac.uk?????????????????????????????? N/A?????? N/A 
Y?????? 16451
Self-heal Daemon on 10.5.6.32?????????????? N/A?????? N/A 
Y?????? 7708
Quota Daemon on 10.5.6.32?????????????????? N/A?????? N/A 
Y?????? 8623
Self-heal Daemon on 10.5.6.17?????????????? N/A?????? N/A 
Y?????? 20549
Quota Daemon on 10.5.6.17?????????????????? N/A?????? N/A 
Y?????? 9337
Task Status of Volume C-DATA
------------------------------------------------------------------------------
There are no active volume tasks
.