Hi Atin, Please find the below details :- [image: Inline image 1] [image: Inline image 2] Now when i set the optimize for the virt storage under ovirt and restart glusterd service on any node...it start failing the quorum.. [image: Inline image 3] [image: Inline image 4] Thanks, Punit On Mon, Jul 20, 2015 at 10:44 AM, Punit Dambiwal <hypunit at gmail.com> wrote:> HI Atin, > > Apologies for the delay response... > > 1. When you added the brick was the command successful? > >> Yes..it was successful.. > 2. If volume status is failing what's output its throwing in the console > and how about the glusterd log? > >> I will reproduce the issue again and update you.. > > On Mon, Jul 13, 2015 at 11:46 AM, Atin Mukherjee <amukherj at redhat.com> > wrote: > >> >> >> On 07/13/2015 05:19 AM, Punit Dambiwal wrote: >> > Hi Sathees, >> > >> > With 3 bricks i can get the gluster volume status....but after added >> more >> > bricks....can not get gluster volume status.... >> The information is still incomplete in respect to analyze the problem. >> Further questions: >> >> 1. When you added the brick was the command successful? >> 2. If volume status is failing what's output its throwing in the console >> and how about the glusterd log? >> >> ~Atin >> > >> > On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN <sasundar at redhat.com> >> wrote: >> > >> >> On 07/11/2015 02:46 PM, Atin Mukherjee wrote: >> >> >> >>> >> >>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote: >> >>> >> >>>> Hi, >> >>>> >> >>>> I have deployed one replica 3 storage...but i am facing some issue >> with >> >>>> quorum... >> >>>> >> >>>> Let me elaborate more :- >> >>>> >> >>>> 1. I have 3 node machines and every machine has 5 HDD(Bricks)...No >> >>>> RAID...Just JBOD... >> >>>> 2. Gluster working fine when just add 3 HDD as below :- >> >>>> >> >>>> B HDD from server 1 >> >>>> B HDD from server 2 >> >>>> B HDD from server 3 >> >>>> >> >>>> But when i add more bricks as below :- >> >>>> >> >>>> ----------------------- >> >>>> [root at stor1 ~]# gluster volume info >> >>>> >> >>>> Volume Name: 3TB >> >>>> Type: Distributed-Replicate >> >>>> Volume ID: 5be9165c-3402-4083-b3db-b782da2fb8d8 >> >>>> Status: Stopped >> >>>> Number of Bricks: 5 x 3 = 15 >> >>>> Transport-type: tcp >> >>>> Bricks: >> >>>> Brick1: stor1:/bricks/b/vol1 >> >>>> Brick2: stor2:/bricks/b/vol1 >> >>>> Brick3: stor3:/bricks/b/vol1 >> >>>> Brick4: stor1:/bricks/c/vol1 >> >>>> Brick5: stor2:/bricks/c/vol1 >> >>>> Brick6: stor3:/bricks/c/vol1 >> >>>> Brick7: stor1:/bricks/d/vol1 >> >>>> Brick8: stor2:/bricks/d/vol1 >> >>>> Brick9: stor3:/bricks/d/vol1 >> >>>> Brick10: stor1:/bricks/e/vol1 >> >>>> Brick11: stor2:/bricks/e/vol1 >> >>>> Brick12: stor3:/bricks/e/vol1 >> >>>> Brick13: stor1:/bricks/f/vol1 >> >>>> Brick14: stor2:/bricks/f/vol1 >> >>>> Brick15: stor3:/bricks/f/vol1 >> >>>> Options Reconfigured: >> >>>> nfs.disable: off >> >>>> user.cifs: enable >> >>>> auth.allow: * >> >>>> performance.quick-read: off >> >>>> performance.read-ahead: off >> >>>> performance.io-cache: off >> >>>> performance.stat-prefetch: off >> >>>> cluster.eager-lock: enable >> >>>> network.remote-dio: enable >> >>>> cluster.quorum-type: auto >> >>>> cluster.server-quorum-type: server >> >>>> storage.owner-uid: 36 >> >>>> storage.owner-gid: 36 >> >>>> -------------------------------- >> >>>> >> >>>> Brick added successfully without any error but after 1 min quorum >> failed >> >>>> and gluster stop working... >> >>>> >> >>> Punit, >> >> >> >> And what do you mean by quorum failed ? >> >> What is effect that you are seeing ? >> >> Could you provide output of 'gluster volume status' as well ? >> >> >> >> -- Sathees >> >> >> >> >> >> What do log files say? >> >>> >> >>>> Thanks, >> >>>> Punit >> >>>> >> >>>> >> >>>> >> >>>> _______________________________________________ >> >>>> Gluster-users mailing list >> >>>> Gluster-users at gluster.org >> >>>> http://www.gluster.org/mailman/listinfo/gluster-users >> >>>> >> >>>> >> >> >> > >> >> -- >> ~Atin >> > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150720/a08bc09e/attachment-0001.html> -------------- next part -------------- A non-text attachment was scrubbed... Name: image.png Type: image/png Size: 81873 bytes Desc: not available URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150720/a08bc09e/attachment-0004.png> -------------- next part -------------- A non-text attachment was scrubbed... Name: image.png Type: image/png Size: 78693 bytes Desc: not available URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150720/a08bc09e/attachment-0005.png> -------------- next part -------------- A non-text attachment was scrubbed... Name: image.png Type: image/png Size: 70052 bytes Desc: not available URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150720/a08bc09e/attachment-0006.png> -------------- next part -------------- A non-text attachment was scrubbed... Name: image.png Type: image/png Size: 12638 bytes Desc: not available URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150720/a08bc09e/attachment-0007.png> -------------- next part -------------- A non-text attachment was scrubbed... Name: 273 20-Jul-15.jpg Type: image/jpeg Size: 41536 bytes Desc: not available URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150720/a08bc09e/attachment-0004.jpg> -------------- next part -------------- A non-text attachment was scrubbed... Name: 274 20-Jul-15.jpg Type: image/jpeg Size: 149076 bytes Desc: not available URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150720/a08bc09e/attachment-0005.jpg> -------------- next part -------------- A non-text attachment was scrubbed... Name: 275 20-Jul-15.jpg Type: image/jpeg Size: 191154 bytes Desc: not available URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150720/a08bc09e/attachment-0006.jpg> -------------- next part -------------- A non-text attachment was scrubbed... Name: 277 20-Jul-15.jpg Type: image/jpeg Size: 145245 bytes Desc: not available URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150720/a08bc09e/attachment-0007.jpg>
In one of your earlier email you mentioned that after adding a brick volume status stopped working. Can you point me to the glusterd log for that transaction? ~Atin On 07/20/2015 12:11 PM, Punit Dambiwal wrote:> Hi Atin, > > Please find the below details :- > > [image: Inline image 1] > > [image: Inline image 2] > > Now when i set the optimize for the virt storage under ovirt and restart > glusterd service on any node...it start failing the quorum.. > > [image: Inline image 3] > > [image: Inline image 4] > > Thanks, > Punit > > On Mon, Jul 20, 2015 at 10:44 AM, Punit Dambiwal <hypunit at gmail.com> wrote: > >> HI Atin, >> >> Apologies for the delay response... >> >> 1. When you added the brick was the command successful? >>>> Yes..it was successful.. >> 2. If volume status is failing what's output its throwing in the console >> and how about the glusterd log? >>>> I will reproduce the issue again and update you.. >> >> On Mon, Jul 13, 2015 at 11:46 AM, Atin Mukherjee <amukherj at redhat.com> >> wrote: >> >>> >>> >>> On 07/13/2015 05:19 AM, Punit Dambiwal wrote: >>>> Hi Sathees, >>>> >>>> With 3 bricks i can get the gluster volume status....but after added >>> more >>>> bricks....can not get gluster volume status.... >>> The information is still incomplete in respect to analyze the problem. >>> Further questions: >>> >>> 1. When you added the brick was the command successful? >>> 2. If volume status is failing what's output its throwing in the console >>> and how about the glusterd log? >>> >>> ~Atin >>>> >>>> On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN <sasundar at redhat.com> >>> wrote: >>>> >>>>> On 07/11/2015 02:46 PM, Atin Mukherjee wrote: >>>>> >>>>>> >>>>>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote: >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> I have deployed one replica 3 storage...but i am facing some issue >>> with >>>>>>> quorum... >>>>>>> >>>>>>> Let me elaborate more :- >>>>>>> >>>>>>> 1. I have 3 node machines and every machine has 5 HDD(Bricks)...No >>>>>>> RAID...Just JBOD... >>>>>>> 2. Gluster working fine when just add 3 HDD as below :- >>>>>>> >>>>>>> B HDD from server 1 >>>>>>> B HDD from server 2 >>>>>>> B HDD from server 3 >>>>>>> >>>>>>> But when i add more bricks as below :- >>>>>>> >>>>>>> ----------------------- >>>>>>> [root at stor1 ~]# gluster volume info >>>>>>> >>>>>>> Volume Name: 3TB >>>>>>> Type: Distributed-Replicate >>>>>>> Volume ID: 5be9165c-3402-4083-b3db-b782da2fb8d8 >>>>>>> Status: Stopped >>>>>>> Number of Bricks: 5 x 3 = 15 >>>>>>> Transport-type: tcp >>>>>>> Bricks: >>>>>>> Brick1: stor1:/bricks/b/vol1 >>>>>>> Brick2: stor2:/bricks/b/vol1 >>>>>>> Brick3: stor3:/bricks/b/vol1 >>>>>>> Brick4: stor1:/bricks/c/vol1 >>>>>>> Brick5: stor2:/bricks/c/vol1 >>>>>>> Brick6: stor3:/bricks/c/vol1 >>>>>>> Brick7: stor1:/bricks/d/vol1 >>>>>>> Brick8: stor2:/bricks/d/vol1 >>>>>>> Brick9: stor3:/bricks/d/vol1 >>>>>>> Brick10: stor1:/bricks/e/vol1 >>>>>>> Brick11: stor2:/bricks/e/vol1 >>>>>>> Brick12: stor3:/bricks/e/vol1 >>>>>>> Brick13: stor1:/bricks/f/vol1 >>>>>>> Brick14: stor2:/bricks/f/vol1 >>>>>>> Brick15: stor3:/bricks/f/vol1 >>>>>>> Options Reconfigured: >>>>>>> nfs.disable: off >>>>>>> user.cifs: enable >>>>>>> auth.allow: * >>>>>>> performance.quick-read: off >>>>>>> performance.read-ahead: off >>>>>>> performance.io-cache: off >>>>>>> performance.stat-prefetch: off >>>>>>> cluster.eager-lock: enable >>>>>>> network.remote-dio: enable >>>>>>> cluster.quorum-type: auto >>>>>>> cluster.server-quorum-type: server >>>>>>> storage.owner-uid: 36 >>>>>>> storage.owner-gid: 36 >>>>>>> -------------------------------- >>>>>>> >>>>>>> Brick added successfully without any error but after 1 min quorum >>> failed >>>>>>> and gluster stop working... >>>>>>> >>>>>> Punit, >>>>> >>>>> And what do you mean by quorum failed ? >>>>> What is effect that you are seeing ? >>>>> Could you provide output of 'gluster volume status' as well ? >>>>> >>>>> -- Sathees >>>>> >>>>> >>>>> What do log files say? >>>>>> >>>>>>> Thanks, >>>>>>> Punit >>>>>>> >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Gluster-users mailing list >>>>>>> Gluster-users at gluster.org >>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users >>>>>>> >>>>>>> >>>>> >>>> >>> >>> -- >>> ~Atin >>> >> >> >-- ~Atin