On 07/20/2015 12:44 PM, Punit Dambiwal wrote:> Hi Atin,
>
> Earlier i was adding the brick and create the volume through ovirt
> itself....this time i did it through commandline...i added all 15
> bricks 5*3 through command line....
>
> It seems before optimize for virt storage it all ok...but after that
> one quorum start and it failed the whole...
>
"Optimize for Virt" action in oVirt sets the following volume options.
group = > virt
storage.owner-uid => 36
storage.owner-gid => 36
server.allow-insecure => on
network.ping-timeout => 10
Regards,
Ramesh
> On Mon, Jul 20, 2015 at 3:07 PM, Atin Mukherjee <amukherj at redhat.com
> <mailto:amukherj at redhat.com>> wrote:
>
> In one of your earlier email you mentioned that after adding a brick
> volume status stopped working. Can you point me to the glusterd
> log for
> that transaction?
>
> ~Atin
>
> On 07/20/2015 12:11 PM, Punit Dambiwal wrote:
> > Hi Atin,
> >
> > Please find the below details :-
> >
> > [image: Inline image 1]
> >
> > [image: Inline image 2]
> >
> > Now when i set the optimize for the virt storage under ovirt and
> restart
> > glusterd service on any node...it start failing the quorum..
> >
> > [image: Inline image 3]
> >
> > [image: Inline image 4]
> >
> > Thanks,
> > Punit
> >
> > On Mon, Jul 20, 2015 at 10:44 AM, Punit Dambiwal <hypunit at
gmail.com <mailto:hypunit at gmail.com>> wrote:
> >
> >> HI Atin,
> >>
> >> Apologies for the delay response...
> >>
> >> 1. When you added the brick was the command successful?
> >>>> Yes..it was successful..
> >> 2. If volume status is failing what's output its throwing
in
> the console
> >> and how about the glusterd log?
> >>>> I will reproduce the issue again and update you..
> >>
> >> On Mon, Jul 13, 2015 at 11:46 AM, Atin Mukherjee
> <amukherj at redhat.com <mailto:amukherj at redhat.com>>
> >> wrote:
> >>
> >>>
> >>>
> >>> On 07/13/2015 05:19 AM, Punit Dambiwal wrote:
> >>>> Hi Sathees,
> >>>>
> >>>> With 3 bricks i can get the gluster volume
status....but
> after added
> >>> more
> >>>> bricks....can not get gluster volume status....
> >>> The information is still incomplete in respect to analyze
the
> problem.
> >>> Further questions:
> >>>
> >>> 1. When you added the brick was the command successful?
> >>> 2. If volume status is failing what's output its
throwing in
> the console
> >>> and how about the glusterd log?
> >>>
> >>> ~Atin
> >>>>
> >>>> On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN
> <sasundar at redhat.com <mailto:sasundar at redhat.com>>
> >>> wrote:
> >>>>
> >>>>> On 07/11/2015 02:46 PM, Atin Mukherjee wrote:
> >>>>>
> >>>>>>
> >>>>>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote:
> >>>>>>
> >>>>>>> Hi,
> >>>>>>>
> >>>>>>> I have deployed one replica 3
storage...but i am facing
> some issue
> >>> with
> >>>>>>> quorum...
> >>>>>>>
> >>>>>>> Let me elaborate more :-
> >>>>>>>
> >>>>>>> 1. I have 3 node machines and every
machine has 5
> HDD(Bricks)...No
> >>>>>>> RAID...Just JBOD...
> >>>>>>> 2. Gluster working fine when just add 3
HDD as below :-
> >>>>>>>
> >>>>>>> B HDD from server 1
> >>>>>>> B HDD from server 2
> >>>>>>> B HDD from server 3
> >>>>>>>
> >>>>>>> But when i add more bricks as below :-
> >>>>>>>
> >>>>>>> -----------------------
> >>>>>>> [root at stor1 ~]# gluster volume info
> >>>>>>>
> >>>>>>> Volume Name: 3TB
> >>>>>>> Type: Distributed-Replicate
> >>>>>>> Volume ID:
5be9165c-3402-4083-b3db-b782da2fb8d8
> >>>>>>> Status: Stopped
> >>>>>>> Number of Bricks: 5 x 3 = 15
> >>>>>>> Transport-type: tcp
> >>>>>>> Bricks:
> >>>>>>> Brick1: stor1:/bricks/b/vol1
> >>>>>>> Brick2: stor2:/bricks/b/vol1
> >>>>>>> Brick3: stor3:/bricks/b/vol1
> >>>>>>> Brick4: stor1:/bricks/c/vol1
> >>>>>>> Brick5: stor2:/bricks/c/vol1
> >>>>>>> Brick6: stor3:/bricks/c/vol1
> >>>>>>> Brick7: stor1:/bricks/d/vol1
> >>>>>>> Brick8: stor2:/bricks/d/vol1
> >>>>>>> Brick9: stor3:/bricks/d/vol1
> >>>>>>> Brick10: stor1:/bricks/e/vol1
> >>>>>>> Brick11: stor2:/bricks/e/vol1
> >>>>>>> Brick12: stor3:/bricks/e/vol1
> >>>>>>> Brick13: stor1:/bricks/f/vol1
> >>>>>>> Brick14: stor2:/bricks/f/vol1
> >>>>>>> Brick15: stor3:/bricks/f/vol1
> >>>>>>> Options Reconfigured:
> >>>>>>> nfs.disable: off
> >>>>>>> user.cifs: enable
> >>>>>>> auth.allow: *
> >>>>>>> performance.quick-read: off
> >>>>>>> performance.read-ahead: off
> >>>>>>> performance.io-cache: off
> >>>>>>> performance.stat-prefetch: off
> >>>>>>> cluster.eager-lock: enable
> >>>>>>> network.remote-dio: enable
> >>>>>>> cluster.quorum-type: auto
> >>>>>>> cluster.server-quorum-type: server
> >>>>>>> storage.owner-uid: 36
> >>>>>>> storage.owner-gid: 36
> >>>>>>> --------------------------------
> >>>>>>>
> >>>>>>> Brick added successfully without any error
but after 1 min
> quorum
> >>> failed
> >>>>>>> and gluster stop working...
> >>>>>>>
> >>>>>> Punit,
> >>>>>
> >>>>> And what do you mean by quorum failed ?
> >>>>> What is effect that you are seeing ?
> >>>>> Could you provide output of 'gluster volume
status' as well ?
> >>>>>
> >>>>> -- Sathees
> >>>>>
> >>>>>
> >>>>> What do log files say?
> >>>>>>
> >>>>>>> Thanks,
> >>>>>>> Punit
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>>
_______________________________________________
> >>>>>>> Gluster-users mailing list
> >>>>>>> Gluster-users at gluster.org
<mailto:Gluster-users at gluster.org>
> >>>>>>>
http://www.gluster.org/mailman/listinfo/gluster-users
> >>>>>>>
> >>>>>>>
> >>>>>
> >>>>
> >>>
> >>> --
> >>> ~Atin
> >>>
> >>
> >>
> >
>
> --
> ~Atin
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://www.gluster.org/pipermail/gluster-users/attachments/20150720/897b3856/attachment.html>