Ravishankar N
2015-Jun-18 11:18 UTC
[Gluster-users] mount with one alive node in replica set
On 06/18/2015 04:25 PM, ????? ??????? wrote:> Thank you for you answer! > > I check recomendation: > 1. On first node closed all connection from second node by iptables. > Check that on both nodes "gluster peer status" return "Disconnected". > Check that on both nodes share was mounted and work well like local > file system. > 2. Rebooted second node (remind first node closed by iptables). Second > node booted without problem and proccesses of glusterfs started: > # ps aux | grep [g]luster > root 4145 0.0 0.0 375692 16076 ? Ssl 13:35 0:00 > /usr/sbin/glusterd -p /var/run/glusterd.pid > > "gluster peer status" return "Disconnected"and volume started on > localhost: > # gluster volume info > Volume Name: files > Type: Replicate > Volume ID: 41067184-d57a-4132-a997-dbd47c974b40 > Status: Started > Number of Bricks: 1 x 2 = 2 > Transport-type: tcp > Bricks: > Brick1: xxx1:/storage/gluster_brick_repofiles > Brick2: xxx2:/storage/gluster_brick_repofiles > > ?ut I cann't mount this volume: > # cat /etc/fstab |grep gluster > 127.0.0.1:/files /repo glusterfs rw,_netdev 0 0 > # mount /repo > Mount failed. Please check the log file for more details. > > Part of log I have sent in first message. > > If I will open first node by iptables I could mount without problem, > but what must I do, when I lost one node and I have probability reboot > another node? > >`gluster volume start <volname> force` doesn't work?> > 2015-06-17 18:46 GMT+03:00 Ravishankar N <ravishankar at redhat.com > <mailto:ravishankar at redhat.com>>: > > > > On 06/17/2015 07:04 PM, ????? ??????? wrote: > > If we turn off one server, another will be work and mounted > volume will be use without problem. > But if we rebooted our another server, when first was turned > off (or gluster was stopped on this server), our volume cann't > mount (glusterd started). > > If both nodes are down and you bring up only one node, glusterd > will not start the volume (i.e. the brick, nfs and glustershd > processes) automatically. It waits for the other node's glusterd > also to be up so that they are in sync. You can override this > behavior by doing a `gluster volume start <volname> force` to > bring up the gluster process only on this node and then mount the > volume. > > -Ravi > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150618/0e6e8295/attachment.html>
Игорь Бирюлин
2015-Jun-18 15:21 UTC
[Gluster-users] mount with one alive node in replica set
Sorry, I didn't check this, because after reboot my second node I had looked "gluster volume info" only and I have found "Status: Started". Now I've checked your recomendation and you are right! "gluster volume start <volname> force" didn't changed output of "gluster volume info" but I have mounted my share! Thank you very much for your advice! But why does "gluster volume info" show that my volname "started" before "gluster volume start <volname> force"? 2015-06-18 14:18 GMT+03:00 Ravishankar N <ravishankar at redhat.com>:> > > On 06/18/2015 04:25 PM, ????? ??????? wrote: > > Thank you for you answer! > > I check recomendation: > 1. On first node closed all connection from second node by iptables. > Check that on both nodes "gluster peer status" return "Disconnected". > Check that on both nodes share was mounted and work well like local file > system. > 2. Rebooted second node (remind first node closed by iptables). Second > node booted without problem and proccesses of glusterfs started: > # ps aux | grep [g]luster > root 4145 0.0 0.0 375692 16076 ? Ssl 13:35 0:00 > /usr/sbin/glusterd -p /var/run/glusterd.pid > > "gluster peer status" return "Disconnected" and volume started on > localhost: > # gluster volume info > Volume Name: files > Type: Replicate > Volume ID: 41067184-d57a-4132-a997-dbd47c974b40 > Status: Started > Number of Bricks: 1 x 2 = 2 > Transport-type: tcp > Bricks: > Brick1: xxx1:/storage/gluster_brick_repofiles > Brick2: xxx2:/storage/gluster_brick_repofiles > > ?ut I cann't mount this volume: > # cat /etc/fstab |grep gluster > 127.0.0.1:/files /repo > glusterfs rw,_netdev 0 0 > # mount /repo > Mount failed. Please check the log file for more details. > > Part of log I have sent in first message. > > If I will open first node by iptables I could mount without problem, but > what must I do, when I lost one node and I have probability reboot > another node? > > > `gluster volume start <volname> force` doesn't work? > > > 2015-06-17 18:46 GMT+03:00 Ravishankar N <ravishankar at redhat.com>: > >> >> >> On 06/17/2015 07:04 PM, ????? ??????? wrote: >> >>> If we turn off one server, another will be work and mounted volume will >>> be use without problem. >>> But if we rebooted our another server, when first was turned off (or >>> gluster was stopped on this server), our volume cann't mount (glusterd >>> started). >>> >> If both nodes are down and you bring up only one node, glusterd will not >> start the volume (i.e. the brick, nfs and glustershd processes) >> automatically. It waits for the other node's glusterd also to be up so that >> they are in sync. You can override this behavior by doing a `gluster volume >> start <volname> force` to bring up the gluster process only on this node >> and then mount the volume. >> >> -Ravi >> > > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150618/bc5cedd4/attachment.html>