Varadharajan S
2015-May-19 11:40 UTC
[Gluster-users] Distributed volume going to Read only mode if any of the Brick is not available
Hi,
We are using Ubuntu 14.04 server and for storage purpose we configured
gluster 3.5 as distributed volume and find the below details,
1).4 Servers - 14.04 Ubuntu Server and each server disks free spaces are
configured as ZFS raiddz2 volume
2). Each server has /pool/gluster zfs volume and capacity as - 5 TB,8 TB,6
TB and 10 TB
3). Bricks are - rep1,rep2,rep3 and st1 and all the bricks are connected as
Distributed Volume and mounted on each system as,
For E.x in rep1 -> mount -t glusterfs rep1:/glustervol /data.
rep2 -> mount -t glusterfs rep2:/glustervol /data
rep3 -> mount -t glusterfs rep3:/glustervol /data
st1 -> mount -t glusterfs st1:/glustervol /data
So we get /data is having around 29 TB and all our applications data's are
stored in /data mount point.
*Details about volume:*
volume glustervol-client-0
type protocol/client
option send-gids true
option password b217da9d1d8b-bb55
option username 9d76-4553-8c75
option transport-type tcp
option remote-subvolume /pool/gluster
option remote-host rep1
option ping-timeout 42
end-volume
volume glustervol-client-1
type protocol/client
option send-gids true
option password b217da9d1d8b-bb55
option username jkd76-4553-5347
option transport-type tcp
option remote-subvolume /pool/gluster
option remote-host rep2
option ping-timeout 42
end-volume
volume glustervol-client-2
type protocol/client
option send-gids true
option password b217da9d1d8b-bb55
option username 19d7-5a190c2
option transport-type tcp
option remote-subvolume /pool/gluster
option remote-host rep3
option ping-timeout 42
end-volume
volume glustervol-client-3
type protocol/client
option send-gids true
option password b217da9d1d8b-bb55
option username c75-5436b5a168347
option transport-type tcp
option remote-subvolume /pool/gluster
option remote-host st1
option ping-timeout 42
end-volume
volume glustervol-dht
type cluster/distribute
subvolumes glustervol-client-0 glustervol-client-1 glustervol-client-2
glustervol-client-3
end-volume
volume glustervol-write-behind
type performance/write-behind
subvolumes glustervol-dht
end-volume
volume glustervol-read-ahead
type performance/read-ahead
subvolumes glustervol-write-behind
end-volume
volume glustervol-io-cache
type performance/io-cache
subvolumes glustervol-read-ahead
end-volume
volume glustervol-quick-read
type performance/quick-read
subvolumes glustervol-io-cache
end-volume
volume glustervol-open-behind
type performance/open-behind
subvolumes glustervol-quick-read
end-volume
volume glustervol-md-cache
type performance/md-cache
subvolumes glustervol-open-behind
end-volume
volume glustervol
type debug/io-stats
option count-fop-hits off
option latency-measurement off
subvolumes glustervol-md-cache
end-volume
ap at rep3:~$ sudo gluster volume info
Volume Name: glustervol
Type: Distribute
Volume ID: 165b-XXXXX
Status: Started
Number of Bricks: 4
Transport-type: tcp
Bricks:
Brick1: rep1:/pool/gluster
Brick2: rep2:/pool/gluster
Brick3: rep3:/pool/gluster
Brick4: st1:/pool/gluster
*Problem:*
If we shutdown any of the bricks , the volume size is reduced (this is ok)
but from the other servers , i can see my mount point /data but it's only
listing contents and i can't write or edit any single files/folders.
*Solution Required:*
If anyone brick is not available, From other servers should allow for Write
and edit functions
Please let us know, what can i try further ?
Regards,
Varad
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://www.gluster.org/pipermail/gluster-users/attachments/20150519/a3596cb8/attachment.html>
Atin Mukherjee
2015-May-19 14:32 UTC
[Gluster-users] Distributed volume going to Read only mode if any of the Brick is not available
On 19 May 2015 17:10, "Varadharajan S" <rajanvaradhu at gmail.com> wrote:> > Hi, > > We are using Ubuntu 14.04 server and for storage purpose we configuredgluster 3.5 as distributed volume and find the below details,> > 1).4 Servers - 14.04 Ubuntu Server and each server disks free spaces areconfigured as ZFS raiddz2 volume> > 2). Each server has /pool/gluster zfs volume and capacity as - 5 TB,8TB,6 TB and 10 TB> > 3). Bricks are - rep1,rep2,rep3 and st1 and all the bricks are connectedas Distributed Volume and mounted on each system as,> > For E.x in rep1 -> mount -t glusterfs rep1:/glustervol /data. > rep2 -> mount -t glusterfs rep2:/glustervol /data > rep3 -> mount -t glusterfs rep3:/glustervol /data > st1 -> mount -t glusterfs st1:/glustervol /data > > So we get /data is having around 29 TB and all our applications data'sare stored in /data mount point.> > Details about volume: > > volume glustervol-client-0 > type protocol/client > option send-gids true > option password b217da9d1d8b-bb55 > option username 9d76-4553-8c75 > option transport-type tcp > option remote-subvolume /pool/gluster > option remote-host rep1 > option ping-timeout 42 > end-volume > > volume glustervol-client-1 > type protocol/client > option send-gids true > option password b217da9d1d8b-bb55 > option username jkd76-4553-5347 > option transport-type tcp > option remote-subvolume /pool/gluster > option remote-host rep2 > option ping-timeout 42 > end-volume > > volume glustervol-client-2 > type protocol/client > option send-gids true > option password b217da9d1d8b-bb55 > option username 19d7-5a190c2 > option transport-type tcp > option remote-subvolume /pool/gluster > option remote-host rep3 > option ping-timeout 42 > end-volume > > volume glustervol-client-3 > type protocol/client > option send-gids true > option password b217da9d1d8b-bb55 > option username c75-5436b5a168347 > option transport-type tcp > option remote-subvolume /pool/gluster > option remote-host st1 > > option ping-timeout 42 > end-volume > > volume glustervol-dht > type cluster/distribute > subvolumes glustervol-client-0 glustervol-client-1glustervol-client-2 glustervol-client-3> end-volume > > volume glustervol-write-behind > type performance/write-behind > subvolumes glustervol-dht > end-volume > > volume glustervol-read-ahead > type performance/read-ahead > subvolumes glustervol-write-behind > end-volume > > volume glustervol-io-cache > type performance/io-cache > subvolumes glustervol-read-ahead > end-volume > > volume glustervol-quick-read > type performance/quick-read > subvolumes glustervol-io-cache > end-volume > > volume glustervol-open-behind > type performance/open-behind > subvolumes glustervol-quick-read > end-volume > > volume glustervol-md-cache > type performance/md-cache > subvolumes glustervol-open-behind > end-volume > > volume glustervol > type debug/io-stats > option count-fop-hits off > option latency-measurement off > subvolumes glustervol-md-cache > end-volume > > > ap at rep3:~$ sudo gluster volume info > > Volume Name: glustervol > Type: Distribute > Volume ID: 165b-XXXXX > Status: Started > Number of Bricks: 4 > Transport-type: tcp > Bricks: > Brick1: rep1:/pool/gluster > Brick2: rep2:/pool/gluster > Brick3: rep3:/pool/gluster > Brick4: st1:/pool/gluster > > Problem: > > If we shutdown any of the bricks , the volume size is reduced (this isok) but from the other servers , i can see my mount point /data but it's only listing contents and i can't write or edit any single files/folders.> > Solution Required: > > If anyone brick is not available, From other servers should allow forWrite and edit functions This is expected since you are using distributed volume. You wouldn't be able to write/edit files belonging to the brick which is down. Solution would be to migrate to distributed replicate volume.> > Please let us know, what can i try further ? > > Regards, > Varad > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://www.gluster.org/mailman/listinfo/gluster-users-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150519/864616f6/attachment.html>