Khoi Mai
2014-Jun-09 18:11 UTC
[Gluster-users] Gluster-users Digest, Vol 74, Issue 3, Message 7:
Andy does the disconnect appear mostly from a CLI change? I'm trying to investigate the same issue with "disconnects". Several of my clients report rpc_ping_timer_check of not responding in 10secs. which is the limit i've set, but I am not fully understanding what trips that limit? Prasanth is the gluster update going to be made available for gluster 3.4.4 ? Khoi Mai From: gluster-users-request at gluster.org To: gluster-users at gluster.org Date: 06/03/2014 06:58 AM Subject: Gluster-users Digest, Vol 74, Issue 3 Sent by: gluster-users-bounces at gluster.org Send Gluster-users mailing list submissions to gluster-users at gluster.org To subscribe or unsubscribe via the World Wide Web, visit http://supercolony.gluster.org/mailman/listinfo/gluster-users or, via email, send a message with subject or body 'help' to gluster-users-request at gluster.org You can reach the person managing the list at gluster-users-owner at gluster.org When replying, please edit your Subject line so it is more specific than "Re: Contents of Gluster-users digest..." Today's Topics: 1. Distributed volumes (yalla.gnan.kumar at accenture.com) 2. Re: Distributed volumes (Michael DePaulo) 3. Re: Distributed volumes (Franco Broi) 4. Re: recommended upgrade procedure from gluster-3.2.7 to gluster-3.5.0 (Todd Pfaff) 5. Re: Unavailability during self-heal for large volumes (Laurent Chouinard) 6. Re: [Gluster-devel] autodelete in snapshots (M S Vishwanath Bhat) 7. Brick on just one host constantly going offline (Andrew Lau) 8. Re: Unavailability during self-heal for large volumes (Pranith Kumar Karampuri) 9. Re: Brick on just one host constantly going offline (Pranith Kumar Karampuri) 10. Re: Brick on just one host constantly going offline (Andrew Lau) 11. Re: Brick on just one host constantly going offline (Pranith Kumar Karampuri) 12. Re: Distributed volumes (yalla.gnan.kumar at accenture.com) 13. Re: Distributed volumes (Franco Broi) 14. Re: Distributed volumes (yalla.gnan.kumar at accenture.com) 15. Re: [Gluster-devel] autodelete in snapshots (Kaushal M) 16. Re: Distributed volumes (Franco Broi) 17. Re: Distributed volumes (yalla.gnan.kumar at accenture.com) 18. Re: Distributed volumes (Kaushal M) 19. Re: Distributed volumes (yalla.gnan.kumar at accenture.com) 20. Re: Distributed volumes (Franco Broi) 21. Re: Distributed volumes (Vijay Bellur) 22. NFS ACL Support in Gluster 3.4 (Indivar Nair) 23. Re: NFS ACL Support in Gluster 3.4 (Santosh Pradhan) ---------------------------------------------------------------------- Message: 1 Date: Mon, 2 Jun 2014 12:26:09 +0000 From: <yalla.gnan.kumar at accenture.com> To: <gluster-users at gluster.org> Subject: [Gluster-users] Distributed volumes Message-ID: <67765C71374B974FBFD2AD05AF438EFF0BD564B6 at 048-CH1MPN3-362.048d.mgd.msft.net> Content-Type: text/plain; charset="us-ascii" Hi All, I have created a distributed volume of 1 GB , using two bricks from two different servers. I have written 7 files whose sizes are a total of 1 GB. How can I check that files are distributed on both the bricks ? Thanks Kumar ________________________________ This message is for the designated recipient only and may contain privileged, proprietary, or otherwise confidential information. If you have received it in error, please notify the sender immediately and delete the original. Any other use of the e-mail by you is prohibited. Where allowed by local law, electronic communications with Accenture and its affiliates, including e-mail and instant messaging (including content), may be scanned by our systems for the purposes of information security and assessment of internal compliance with Accenture policy. ______________________________________________________________________________________ www.accenture.com -------------- next part -------------- An HTML attachment was scrubbed... URL: < http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140602/fef11c67/attachment-0001.html>------------------------------ Message: 2 Date: Mon, 2 Jun 2014 09:04:44 -0400 From: Michael DePaulo <mikedep333 at gmail.com> To: yalla.gnan.kumar at accenture.com Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes Message-ID: <CAMKht8gtijHHZhza3T5zv5hRij9PzwrVvkxtSHDXYRkmGDYpEA at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 On Mon, Jun 2, 2014 at 8:26 AM, <yalla.gnan.kumar at accenture.com> wrote:> Hi All, > > > > I have created a distributed volume of 1 GB , using two bricks from two > different servers. > > I have written 7 files whose sizes are a total of 1 GB. > > How can I check that files are distributed on both the bricks ? > > > > > > Thanks > > KumarHi Kumar, You can use standard file browsing commands like "cd" and "ls" on both of the bricks. The volume's files will show up as regular files on the underlying filesystem. You can manually verify that files that exist on brick 1 do not exist on brick 2, and vica-versa. For example, here's me running file browsing commands on my replicated volume's brick: mike at nostromo:/data1/brick1/gv1 :( [7] $ ls -latr total 24 drwxr-xr-x. 3 root root 4096 Dec 19 22:21 homes drwxr-xr-x. 3 root root 4096 May 3 17:55 .. drw-------. 261 root root 4096 May 3 18:38 .glusterfs drwxr-xr-x. 4 root root 4096 May 3 21:02 . mike at nostromo:/data1/brick1/gv1 :) [8] $ sudo du -s -h homes/ .glusterfs/ [sudo] password for mike: 34G homes/ 252M .glusterfs/ -Mike ------------------------------ Message: 3 Date: Mon, 02 Jun 2014 21:05:07 +0800 From: Franco Broi <franco.broi at iongeo.com> To: yalla.gnan.kumar at accenture.com Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes Message-ID: <1401714307.17051.23.camel at tc1> Content-Type: text/plain; charset="UTF-8" Just do an ls on the bricks, the paths are the same as the mounted filesystem. On Mon, 2014-06-02 at 12:26 +0000, yalla.gnan.kumar at accenture.com wrote:> Hi All, > > > > I have created a distributed volume of 1 GB , using two bricks from > two different servers. > > I have written 7 files whose sizes are a total of 1 GB. > > How can I check that files are distributed on both the bricks ? > > > > > > Thanks > > Kumar > > > > > ______________________________________________________________________ > > > This message is for the designated recipient only and may contain > privileged, proprietary, or otherwise confidential information. If you > have received it in error, please notify the sender immediately and > delete the original. Any other use of the e-mail by you is prohibited. > Where allowed by local law, electronic communications with Accenture > and its affiliates, including e-mail and instant messaging (including > content), may be scanned by our systems for the purposes of > information security and assessment of internal compliance with > Accenture policy. >______________________________________________________________________________________> > www.accenture.com > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://supercolony.gluster.org/mailman/listinfo/gluster-users------------------------------ Message: 4 Date: Mon, 2 Jun 2014 11:56:11 -0400 (EDT) From: Todd Pfaff <pfaff at rhpcs.mcmaster.ca> To: Pranith Kumar Karampuri <pkarampu at redhat.com> Cc: Susant Palai <spalai at redhat.com>, Venkatesh Somyajulu <vsomyaju at redhat.com>, gluster-users at gluster.org Subject: Re: [Gluster-users] recommended upgrade procedure from gluster-3.2.7 to gluster-3.5.0 Message-ID: <alpine.LMD.2.02.1406021149500.1729 at rhpcserv.rhpcs.mcmaster.ca> Content-Type: TEXT/PLAIN; charset=US-ASCII; format=flowed On Sun, 1 Jun 2014, Pranith Kumar Karampuri wrote:> > > ----- Original Message ----- >> From: "Todd Pfaff" <pfaff at rhpcs.mcmaster.ca> >> To: "Pranith Kumar Karampuri" <pkarampu at redhat.com> >> Cc: gluster-users at gluster.org >> Sent: Saturday, May 31, 2014 7:18:23 PM >> Subject: Re: [Gluster-users] recommended upgrade procedure fromgluster-3.2.7 to gluster-3.5.0>> >> thanks, pranith, that was very helpful! >> >> i followed your advice, it ran and completed, and now i'm left withthese>> results on the removed brick (before commit): >> >> find /2/scratch/ | wc -l >> 83083 >> >> find /2/scratch/ -type f | wc -l >> 16 >> >> find /2/scratch/ -type d | wc -l >> 70243 >> >> find /2/scratch/ -type l | wc -l >> 12824 >> >> find /2/scratch/ ! -type d -a ! -type f | wc -l >> 12824 >> >> find /2/scratch/.glusterfs -type l | wc -l >> 12824 >> >> find /2/scratch/* | wc -l >> 12873 >> >> find /2/scratch/* -type d | wc -l >> 12857 >> >> so it looks like i have 16 files and 12857 directories left in/2/scratch,>> and 12824 links under /2/scratch/.glusterfs/. >> >> my first instinct is to ignore (and remove) the many remainingdirectories>> that are empty and only look closer at those that contain the 16remaining>> files. >> >> can i ignore the links under /2/scratch/.glusterfs? >> >> as for the 16 files that remain, i can migrate them manually ifnecessary>> but i'll first look at all the brick filesystems to see if they already >> exist elsewhere in some form. >> >> do you recommend i do anything else? > > Your solutions are good :-). Could you please send us the > configuration, logs of the setup so that we can debug why those files > didn't move? It would be good if we can find the reason for it and fix > it in the next releases so that this issue is prevented.sure, i'd be happy to help. what exactly should i send you in terms of configuration? just my /etc/glusterfs/glusterd.vol? output of some gluster commands? other? in terms of logs, what do you want to see? do you want this file in its entirety? -rw------- 1 root root 145978018 May 31 08:10 /var/log/glusterfs/scratch-rebalance.log anything else?> > CC developers who work on this feature to look into the issue. > > Just curious, did the remove-brick status output say if any failures > happened?i don't recall seeing anything in the remove-brick status command output that indicated any failures. tp> > Pranith > >> >> thanks, >> tp >> >> >> On Fri, 30 May 2014, Pranith Kumar Karampuri wrote: >> >>> >>> >>> ----- Original Message ----- >>>> From: "Todd Pfaff" <pfaff at rhpcs.mcmaster.ca> >>>> To: gluster-users at gluster.org >>>> Sent: Saturday, May 31, 2014 1:58:33 AM >>>> Subject: Re: [Gluster-users] recommended upgrade procedure from >>>> gluster-3.2.7 to gluster-3.5.0 >>>> >>>> On Sat, 24 May 2014, Todd Pfaff wrote: >>>> >>>>> I have a gluster distributed volume that has been running nicelywith>>>>> gluster-3.2.7 for the past two years and I now want to upgrade thisto>>>>> gluster-3.5.0. >>>>> >>>>> What is the recommended procedure for such an upgrade? Is itnecessary>>>>> to >>>>> upgrade from 3.2.7 to 3.3 to 3.4 to 3.5, or can I safely transitionfrom>>>>> 3.2.7 directly to 3.5.0? >>>> >>>> nobody responded so i decided to wing it and hope for the best. >>>> >>>> i also decided to go directly from 3.2.7 to 3.4.3 and not bother with >>>> 3.5 yet. >>>> >>>> the volume is distributed across 13 bricks. formerly these were in13>>>> nodes, 1 brick per node, but i recently lost one of these nodes. >>>> i've moved the brick from the dead node to be a second brick in oneof>>>> the remaining 12 nodes. i currently have this state: >>>> >>>> gluster volume status >>>> Status of volume: scratch >>>> Gluster process Port Online Pid >>>>------------------------------------------------------------------------------>>>> Brick 172.16.1.1:/1/scratch 49152 Y 6452 >>>> Brick 172.16.1.2:/1/scratch 49152 Y 10783 >>>> Brick 172.16.1.3:/1/scratch 49152 Y 10164 >>>> Brick 172.16.1.4:/1/scratch 49152 Y 10465 >>>> Brick 172.16.1.5:/1/scratch 49152 Y 10186 >>>> Brick 172.16.1.6:/1/scratch 49152 Y 10388 >>>> Brick 172.16.1.7:/1/scratch 49152 Y 10386 >>>> Brick 172.16.1.8:/1/scratch 49152 Y 10215 >>>> Brick 172.16.1.9:/1/scratch 49152 Y 11059 >>>> Brick 172.16.1.10:/1/scratch 49152 Y 9238 >>>> Brick 172.16.1.11:/1/scratch 49152 Y 9466 >>>> Brick 172.16.1.12:/1/scratch 49152 Y 10777 >>>> Brick 172.16.1.1:/2/scratch 49153 Y 6461 >>>> >>>> >>>> what i want to do next is remove Brick 172.16.1.1:/2/scratch and have >>>> all files it contains redistributed across the other 12 bricks. >>>> >>>> what's the correct procedure for this? is it as simple as: >>>> >>>> gluster volume remove-brick scratch 172.16.1.1:/2/scratch start >>>> >>>> and then wait for all files to be moved off that brick? or do i also >>>> have to do: >>>> >>>> gluster volume remove-brick scratch 172.16.1.1:/2/scratch commit >>>> >>>> and then wait for all files to be moved off that brick? or do i also >>>> have to do something else, such as a rebalance, to cause the files to >>>> be moved? >>> >>> 'gluster volume remove-brick scratch 172.16.1.1:/2/scratch start'does>>> start the process of migrating all the files to the other bricks. Youneed>>> to observe the progress of the process using 'gluster volumeremove-brick>>> scratch 172.16.1.1:/2/scratch status' Once this command says'completed'>>> You should execute 'gluster volume remove-brick scratch >>> 172.16.1.1:/2/scratch commit' to completely remove this brick from the >>> volume. I am a bit paranoid so I would check that no files are leftbehind>>> by doing a find on the brick 172.16.1.1:/2/scratch just before issuingthe>>> 'commit' :-). >>> >>> Pranith. >>> >>>> >>>> how do i know when everything has been moved safely to other bricksand>>>> the then-empty brick is no longer involved in the cluster? >>>> >>>> thanks, >>>> tp >>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users >>>> >>> >>> >> > >------------------------------ Message: 5 Date: Mon, 2 Jun 2014 19:26:40 +0000 From: Laurent Chouinard <laurent.chouinard at ubisoft.com> To: Pranith Kumar Karampuri <pkarampu at redhat.com> Cc: "gluster-users at gluster.org" <gluster-users at gluster.org> Subject: Re: [Gluster-users] Unavailability during self-heal for large volumes Message-ID: <95ea1865fac2484980d020c6a3b7f0cd at MSR-MAIL-EXCH02.ubisoft.org> Content-Type: text/plain; charset="utf-8"> Laurent, > This has been improved significantly in afr-v2 (enhanced version ofreplication> translator in gluster) which will be released with 3.6 I believe. Theissue happens> because of the directory self-heal in the older versions. In the newversion per file> healing in a directory is performed instead of Full directory healat-once which was> creating a lot of traffic. Unfortunately This is too big a change tobackport to older> releases :-(. > > PranithHi Pranith, Thank you for this information. Do you think there is a way to limit/throttle the current directory self-heal then? I don't mind if it takes a long time. Alternatively, is there a way to completely disable the complete healing system? I would consider running a manual healing operation by STAT'ing every file, which would allow me to throttle the speed to a more manageable level. Thanks, Laurent Chouinard ------------------------------ Message: 6 Date: Tue, 3 Jun 2014 01:23:48 +0530 From: M S Vishwanath Bhat <msvbhat at gmail.com> To: Vijay Bellur <vbellur at redhat.com> Cc: Seema Naik <senaik at redhat.com>, Gluster Devel <gluster-devel at gluster.org>, gluster-users at gluster.org Subject: Re: [Gluster-users] [Gluster-devel] autodelete in snapshots Message-ID: <CA+H6b3MyWZruhrrykB6Vmv6B1++cgPwb=f+iSqdHnECJQvvmEQ at mail.gmail.com> Content-Type: text/plain; charset="utf-8" On 3 June 2014 01:02, M S Vishwanath Bhat <msvbhat at gmail.com> wrote:> > > > On 2 June 2014 20:22, Vijay Bellur <vbellur at redhat.com> wrote: > >> On 04/23/2014 05:50 AM, Vijay Bellur wrote: >> >>> On 04/20/2014 11:42 PM, Lalatendu Mohanty wrote: >>> >>>> On 04/16/2014 11:39 AM, Avra Sengupta wrote: >>>> >>>>> The whole purpose of introducing the soft-limit is, that at anypoint>>>>> of time the number of >>>>> snaps should not exceed the hard limit. If we trigger auto-delete on >>>>> hitting hard-limit, then >>>>> the purpose itself is lost, because at that point we would be takinga>>>>> snap, making the limit >>>>> hard-limit + 1, and then triggering auto-delete, which violates the >>>>> sanctity of the hard-limit. >>>>> Also what happens when we are at hard-limit + 1, and another snap is >>>>> issued, while auto-delete >>>>> is yet to process the first delete. At that point we end up at >>>>> hard-limit + 1. Also what happens >>>>> if for a particular snap the auto-delete fails. >>>>> >>>>> We should see the hard-limit, as something set by the admin keepingin>>>>> mind the resource consumption >>>>> and at no-point should we cross this limit, come what may. If we hit >>>>> this limit, the create command >>>>> should fail asking the user to delete snaps using the "snapshot >>>>> delete" command. >>>>> >>>>> The two options Raghavendra mentioned are applicable for the >>>>> soft-limit only, in which cases on >>>>> hitting the soft-limit >>>>> >>>>> 1. Trigger auto-delete >>>>> >>>>> or >>>>> >>>>> 2. Log a warning-message, for the user saying the number of snaps is >>>>> exceeding the snap-limit and >>>>> display the number of available snaps >>>>> >>>>> Now which of these should happen also depends on the user, becausethe>>>>> auto-delete option >>>>> is configurable. >>>>> >>>>> So if the auto-delete option is set as true, auto-delete should be >>>>> triggered and the above message >>>>> should also be logged. >>>>> >>>>> But if the option is set as false, only the message should belogged.>>>>> >>>>> This is the behaviour as designed. Adding Rahul, and Seema in the >>>>> mail, to reflect upon the >>>>> behaviour as well. >>>>> >>>>> Regards, >>>>> Avra >>>>> >>>> >>>> This sounds correct. However we need to make sure that the usage or >>>> documentation around this should be good enough , so that users >>>> understand the each of the limits correctly. >>>> >>>> >>> It might be better to avoid the usage of the term "soft-limit". >>> soft-limit as used in quota and other places generally has an alerting >>> connotation. Something like "auto-deletion-limit" might be better. >>> >>> >> I still see references to "soft-limit" and auto deletion seems to get >> triggered upon reaching soft-limit. >> >> Why is the ability to auto delete not configurable? It does seem pretty >> nasty to go about deleting snapshots without obtaining explicit consent >> from the user. >> > > I agree with Vijay here. It's not good to delete a snap (even though itis> oldest) without the explicit consent from user. > > FYI It took me more than 2 weeks to figure out that my snaps weregetting> autodeleted after reaching "soft-limit". For all I know I had not done > anything and my snap restore were failing. > > I propose to remove the terms "soft" and "hard" limit. I believe there > should be a limit (just "limit") after which all snapshot creates should > fail with proper error messages. And there can be a water-mark afterwhich> user should get warning messages. So below is my proposal. > > *auto-delete + snap-limit: *If the snap-limit is set to *n*, next snap > create (n+1th) will succeed *only if* *if auto-delete is set toon/true/1*> and oldest snap will get deleted automatically. If autodelete is set to > off/false/0 , (n+1)th snap create will fail with proper error messagefrom> gluster CLI command. But again by default autodelete should be off. > > *snap-water-mark*: This should come in picture only if autodelete is > turned off. It should not have any meaning if auto-delete is turned ON. > Basically it's usage is to give the user warning that limit almost being > reached and it is time for admin to decide which snaps should be deleted > (or which should be kept) > > *my two cents* >Adding gluster-users as well. -MS> > -MS > > >> >> Cheers, >> >> Vijay >> >> _______________________________________________ >> Gluster-devel mailing list >> Gluster-devel at gluster.org >> http://supercolony.gluster.org/mailman/listinfo/gluster-devel >> > >-------------- next part -------------- An HTML attachment was scrubbed... URL: < http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140603/58732995/attachment-0001.html>------------------------------ Message: 7 Date: Tue, 3 Jun 2014 08:40:25 +1000 From: Andrew Lau <andrew at andrewklau.com> To: "gluster-users at gluster.org List" <gluster-users at gluster.org> Subject: [Gluster-users] Brick on just one host constantly going offline Message-ID: <CAD7dF9dCb-f_pNkiu51P7BsPqbeonE+OOuXH84ni4fD_poR0kA at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 Hi, Just a short post as I've since nuked the test environment. I've had this case where in a 2 node gluster replica, the brick of the first host is constantly going offline. gluster volume status would report host 1's brick is offline. The quorum would kick in, putting the whole cluster into a read only state. This has only recently been happening w/ gluster 3.5 and it normally happens after about 3-4 days of 500GB or so data transfer. Has anyone noticed this before? The only way to bring it back was to: killall glusterfsd ; killall -9 glusterfsd ; killall glusterd ; glusterd Thanks, Andrew ------------------------------ Message: 8 Date: Mon, 2 Jun 2014 20:42:36 -0400 (EDT) From: Pranith Kumar Karampuri <pkarampu at redhat.com> To: Laurent Chouinard <laurent.chouinard at ubisoft.com> Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Unavailability during self-heal for large volumes Message-ID: <1921256894.15836933.1401756156535.JavaMail.zimbra at redhat.com> Content-Type: text/plain; charset=utf-8 ----- Original Message -----> From: "Laurent Chouinard" <laurent.chouinard at ubisoft.com> > To: "Pranith Kumar Karampuri" <pkarampu at redhat.com> > Cc: gluster-users at gluster.org > Sent: Tuesday, June 3, 2014 12:56:40 AM > Subject: RE: [Gluster-users] Unavailability during self-heal for largevolumes> > > Laurent, > > This has been improved significantly in afr-v2 (enhanced version of > > replication > > translator in gluster) which will be released with 3.6 I believe. Theissue> > happens > > because of the directory self-heal in the older versions. In the new > > version per file > > healing in a directory is performed instead of Full directory healat-once> > which was > > creating a lot of traffic. Unfortunately This is too big a change to > > backport to older > > releases :-(. > > > > Pranith > > > Hi Pranith, > > Thank you for this information. > > Do you think there is a way to limit/throttle the current directoryself-heal> then? I don't mind if it takes a long time. > > Alternatively, is there a way to completely disable the complete healing > system? I would consider running a manual healing operation by STAT'ing > every file, which would allow me to throttle the speed to a moremanageable> level.gluster volume set <volume-name> cluster.self-heal-daemon off would disable glustershd performing automatic healing. Pranith> > Thanks, > > Laurent Chouinard >------------------------------ Message: 9 Date: Mon, 2 Jun 2014 20:56:14 -0400 (EDT) From: Pranith Kumar Karampuri <pkarampu at redhat.com> To: Andrew Lau <andrew at andrewklau.com> Cc: "gluster-users at gluster.org List" <gluster-users at gluster.org> Subject: Re: [Gluster-users] Brick on just one host constantly going offline Message-ID: <1543691713.15838872.1401756974277.JavaMail.zimbra at redhat.com> Content-Type: text/plain; charset=utf-8 ----- Original Message -----> From: "Andrew Lau" <andrew at andrewklau.com> > To: "gluster-users at gluster.org List" <gluster-users at gluster.org> > Sent: Tuesday, June 3, 2014 4:10:25 AM > Subject: [Gluster-users] Brick on just one host constantly going offline > > Hi, > > Just a short post as I've since nuked the test environment. > > I've had this case where in a 2 node gluster replica, the brick of the > first host is constantly going offline. > > gluster volume status > > would report host 1's brick is offline. The quorum would kick in, > putting the whole cluster into a read only state. This has only > recently been happening w/ gluster 3.5 and it normally happens after > about 3-4 days of 500GB or so data transfer.Could you check mount logs to see if there are ping timer expiry messages for disconnects? If you see them, then it is very likely that you are hitting throttling problem fixed by http://review.gluster.org/7531 Pranith> > Has anyone noticed this before? The only way to bring it back was to: > > killall glusterfsd ; killall -9 glusterfsd ; killall glusterd ; glusterd > > > Thanks, > Andrew > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://supercolony.gluster.org/mailman/listinfo/gluster-users >------------------------------ Message: 10 Date: Tue, 3 Jun 2014 11:12:44 +1000 From: Andrew Lau <andrew at andrewklau.com> To: Pranith Kumar Karampuri <pkarampu at redhat.com> Cc: "gluster-users at gluster.org List" <gluster-users at gluster.org> Subject: Re: [Gluster-users] Brick on just one host constantly going offline Message-ID: <CAD7dF9c0005xCo_RKZW9L-cqP_TqGPpNd1B7-2FWRzEXZ_Rvvw at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 Hi Pranith, On Tue, Jun 3, 2014 at 10:56 AM, Pranith Kumar Karampuri <pkarampu at redhat.com> wrote:> > > ----- Original Message ----- >> From: "Andrew Lau" <andrew at andrewklau.com> >> To: "gluster-users at gluster.org List" <gluster-users at gluster.org> >> Sent: Tuesday, June 3, 2014 4:10:25 AM >> Subject: [Gluster-users] Brick on just one host constantly goingoffline>> >> Hi, >> >> Just a short post as I've since nuked the test environment. >> >> I've had this case where in a 2 node gluster replica, the brick of the >> first host is constantly going offline. >> >> gluster volume status >> >> would report host 1's brick is offline. The quorum would kick in, >> putting the whole cluster into a read only state. This has only >> recently been happening w/ gluster 3.5 and it normally happens after >> about 3-4 days of 500GB or so data transfer. > > Could you check mount logs to see if there are ping timer expirymessages for disconnects?> If you see them, then it is very likely that you are hitting throttlingproblem fixed by http://review.gluster.org/7531>Ah, that makes sense as it was the only volume which had that ping timeout setting. I also did see the timeout messages in the logs when I was checking. So is this merged in 3.5.1 ?> Pranith > >> >> Has anyone noticed this before? The only way to bring it back was to: >> >> killall glusterfsd ; killall -9 glusterfsd ; killall glusterd ;glusterd>> >> >> Thanks, >> Andrew >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://supercolony.gluster.org/mailman/listinfo/gluster-users >>------------------------------ Message: 11 Date: Mon, 2 Jun 2014 22:14:53 -0400 (EDT) From: Pranith Kumar Karampuri <pkarampu at redhat.com> To: Andrew Lau <andrew at andrewklau.com> Cc: "gluster-users at gluster.org List" <gluster-users at gluster.org> Subject: Re: [Gluster-users] Brick on just one host constantly going offline Message-ID: <1518143519.15897096.1401761693764.JavaMail.zimbra at redhat.com> Content-Type: text/plain; charset=utf-8 ----- Original Message -----> From: "Andrew Lau" <andrew at andrewklau.com> > To: "Pranith Kumar Karampuri" <pkarampu at redhat.com> > Cc: "gluster-users at gluster.org List" <gluster-users at gluster.org> > Sent: Tuesday, June 3, 2014 6:42:44 AM > Subject: Re: [Gluster-users] Brick on just one host constantly goingoffline> > Hi Pranith, > > On Tue, Jun 3, 2014 at 10:56 AM, Pranith Kumar Karampuri > <pkarampu at redhat.com> wrote: > > > > > > ----- Original Message ----- > >> From: "Andrew Lau" <andrew at andrewklau.com> > >> To: "gluster-users at gluster.org List" <gluster-users at gluster.org> > >> Sent: Tuesday, June 3, 2014 4:10:25 AM > >> Subject: [Gluster-users] Brick on just one host constantly goingoffline> >> > >> Hi, > >> > >> Just a short post as I've since nuked the test environment. > >> > >> I've had this case where in a 2 node gluster replica, the brick ofthe> >> first host is constantly going offline. > >> > >> gluster volume status > >> > >> would report host 1's brick is offline. The quorum would kick in, > >> putting the whole cluster into a read only state. This has only > >> recently been happening w/ gluster 3.5 and it normally happens after > >> about 3-4 days of 500GB or so data transfer. > > > > Could you check mount logs to see if there are ping timer expirymessages> > for disconnects? > > If you see them, then it is very likely that you are hittingthrottling> > problem fixed by http://review.gluster.org/7531 > > > > Ah, that makes sense as it was the only volume which had that ping > timeout setting. I also did see the timeout messages in the logs when > I was checking. So is this merged in 3.5.1 ?Yes! http://review.gluster.org/7570 Pranith> > > Pranith > > > >> > >> Has anyone noticed this before? The only way to bring it back was to: > >> > >> killall glusterfsd ; killall -9 glusterfsd ; killall glusterd ;glusterd> >> > >> > >> Thanks, > >> Andrew > >> _______________________________________________ > >> Gluster-users mailing list > >> Gluster-users at gluster.org > >> http://supercolony.gluster.org/mailman/listinfo/gluster-users > >> >------------------------------ Message: 12 Date: Tue, 3 Jun 2014 07:21:21 +0000 From: <yalla.gnan.kumar at accenture.com> To: <franco.broi at iongeo.com> Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes Message-ID: <67765C71374B974FBFD2AD05AF438EFF0BD586A4 at 048-CH1MPN3-362.048d.mgd.msft.net> Content-Type: text/plain; charset="utf-8" Hi, I have created a distributed volume on my gluster node. I have attached this volume to a VM on openstack. The size is 1 GB. I have written files close to 1 GB onto the Volume. But when I do a ls inside the brick directory , the volume is present only on one gluster server brick. But it is empty on another server brick. Files are meant to be spread across both the bricks according to distributed volume definition. On the VM: -------------- # ls -al total 1013417 drwxr-xr-x 3 root root 4096 Jun 1 22:03 . drwxrwxr-x 3 root root 1024 Jun 1 21:24 .. -rw------- 1 root root 31478251520 Jun 1 21:52 file -rw------- 1 root root 157391257600 Jun 1 21:54 file1 -rw------- 1 root root 629565030400 Jun 1 21:55 file2 -rw------- 1 root root 708260659200 Jun 1 21:59 file3 -rw------- 1 root root 6295650304 Jun 1 22:01 file4 -rw------- 1 root root 39333801984 Jun 1 22:01 file5 -rw------- 1 root root 78643200000 Jun 1 22:04 file6 drwx------ 2 root root 16384 Jun 1 21:24 lost+found ---------- # du -sch * 20.0M file 100.0M file1 400.0M file2 454.0M file3 4.0M file4 11.6M file5 0 file6 16.0K lost+found 989.7M total ------------------------ On the gluster server nodes: ----------------------- root at primary:/export/sdd1/brick# ll total 12 drwxr-xr-x 2 root root 4096 Jun 2 04:08 ./ drwxr-xr-x 4 root root 4096 May 27 08:42 ../ root at primary:/export/sdd1/brick# -------------------------- root at secondary:/export/sdd1/brick# ll total 1046536 drwxr-xr-x 2 root root 4096 Jun 2 08:51 ./ drwxr-xr-x 4 root root 4096 May 27 08:43 ../ -rw-rw-rw- 1 108 115 1073741824 Jun 2 09:35 volume-0ec560be-997f-46da-9ec8-e9d6627f2de1 root at secondary:/export/sdd1/brick# --------------------------------- Thanks Kumar -----Original Message----- From: Franco Broi [mailto:franco.broi at iongeo.com] Sent: Monday, June 02, 2014 6:35 PM To: Gnan Kumar, Yalla Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes Just do an ls on the bricks, the paths are the same as the mounted filesystem. On Mon, 2014-06-02 at 12:26 +0000, yalla.gnan.kumar at accenture.com wrote:> Hi All, > > > > I have created a distributed volume of 1 GB , using two bricks from > two different servers. > > I have written 7 files whose sizes are a total of 1 GB. > > How can I check that files are distributed on both the bricks ? > > > > > > Thanks > > Kumar > > > > > ______________________________________________________________________ > > > This message is for the designated recipient only and may contain > privileged, proprietary, or otherwise confidential information. If you > have received it in error, please notify the sender immediately and > delete the original. Any other use of the e-mail by you is prohibited. > Where allowed by local law, electronic communications with Accenture > and its affiliates, including e-mail and instant messaging (including > content), may be scanned by our systems for the purposes of > information security and assessment of internal compliance with > Accenture policy. > ______________________________________________________________________ > ________________ > > www.accenture.com > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://supercolony.gluster.org/mailman/listinfo/gluster-users________________________________ This message is for the designated recipient only and may contain privileged, proprietary, or otherwise confidential information. If you have received it in error, please notify the sender immediately and delete the original. Any other use of the e-mail by you is prohibited. Where allowed by local law, electronic communications with Accenture and its affiliates, including e-mail and instant messaging (including content), may be scanned by our systems for the purposes of information security and assessment of internal compliance with Accenture policy. ______________________________________________________________________________________ www.accenture.com ------------------------------ Message: 13 Date: Tue, 03 Jun 2014 15:25:46 +0800 From: Franco Broi <franco.broi at iongeo.com> To: yalla.gnan.kumar at accenture.com Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes Message-ID: <1401780346.2236.299.camel at tc1> Content-Type: text/plain; charset="UTF-8" What do gluster vol info and gluster vol status give you? On Tue, 2014-06-03 at 07:21 +0000, yalla.gnan.kumar at accenture.com wrote:> Hi, > > I have created a distributed volume on my gluster node. I haveattached this volume to a VM on openstack. The size is 1 GB. I have written files close to 1 GB onto the> Volume. But when I do a ls inside the brick directory , the volume ispresent only on one gluster server brick. But it is empty on another server brick. Files are meant to be> spread across both the bricks according to distributed volumedefinition.> > On the VM: > -------------- > > # ls -al > total 1013417 > drwxr-xr-x 3 root root 4096 Jun 1 22:03 . > drwxrwxr-x 3 root root 1024 Jun 1 21:24 .. > -rw------- 1 root root 31478251520 Jun 1 21:52 file > -rw------- 1 root root 157391257600 Jun 1 21:54 file1 > -rw------- 1 root root 629565030400 Jun 1 21:55 file2 > -rw------- 1 root root 708260659200 Jun 1 21:59 file3 > -rw------- 1 root root 6295650304 Jun 1 22:01 file4 > -rw------- 1 root root 39333801984 Jun 1 22:01 file5 > -rw------- 1 root root 78643200000 Jun 1 22:04 file6 > drwx------ 2 root root 16384 Jun 1 21:24 lost+found > ---------- > # du -sch * > 20.0M file > 100.0M file1 > 400.0M file2 > 454.0M file3 > 4.0M file4 > 11.6M file5 > 0 file6 > 16.0K lost+found > 989.7M total > ------------------------ > > > On the gluster server nodes: > ----------------------- > root at primary:/export/sdd1/brick# ll > total 12 > drwxr-xr-x 2 root root 4096 Jun 2 04:08 ./ > drwxr-xr-x 4 root root 4096 May 27 08:42 ../ > root at primary:/export/sdd1/brick# > -------------------------- > > root at secondary:/export/sdd1/brick# ll > total 1046536 > drwxr-xr-x 2 root root 4096 Jun 2 08:51 ./ > drwxr-xr-x 4 root root 4096 May 27 08:43 ../ > -rw-rw-rw- 1 108 115 1073741824 Jun 2 09:35volume-0ec560be-997f-46da-9ec8-e9d6627f2de1> root at secondary:/export/sdd1/brick# > --------------------------------- > > > Thanks > Kumar > > > > > > > > > -----Original Message----- > From: Franco Broi [mailto:franco.broi at iongeo.com] > Sent: Monday, June 02, 2014 6:35 PM > To: Gnan Kumar, Yalla > Cc: gluster-users at gluster.org > Subject: Re: [Gluster-users] Distributed volumes > > Just do an ls on the bricks, the paths are the same as the mountedfilesystem.> > On Mon, 2014-06-02 at 12:26 +0000, yalla.gnan.kumar at accenture.com > wrote: > > Hi All, > > > > > > > > I have created a distributed volume of 1 GB , using two bricks from > > two different servers. > > > > I have written 7 files whose sizes are a total of 1 GB. > > > > How can I check that files are distributed on both the bricks ? > > > > > > > > > > > > Thanks > > > > Kumar > > > > > > > > > > ______________________________________________________________________ > > > > > > This message is for the designated recipient only and may contain > > privileged, proprietary, or otherwise confidential information. If you > > have received it in error, please notify the sender immediately and > > delete the original. Any other use of the e-mail by you is prohibited. > > Where allowed by local law, electronic communications with Accenture > > and its affiliates, including e-mail and instant messaging (including > > content), may be scanned by our systems for the purposes of > > information security and assessment of internal compliance with > > Accenture policy. > > ______________________________________________________________________ > > ________________ > > > > www.accenture.com > > _______________________________________________ > > Gluster-users mailing list > > Gluster-users at gluster.org > > http://supercolony.gluster.org/mailman/listinfo/gluster-users > > > > > ________________________________ > > This message is for the designated recipient only and may containprivileged, proprietary, or otherwise confidential information. If you have received it in error, please notify the sender immediately and delete the original. Any other use of the e-mail by you is prohibited. Where allowed by local law, electronic communications with Accenture and its affiliates, including e-mail and instant messaging (including content), may be scanned by our systems for the purposes of information security and assessment of internal compliance with Accenture policy.>______________________________________________________________________________________> > www.accenture.com------------------------------ Message: 14 Date: Tue, 3 Jun 2014 07:29:24 +0000 From: <yalla.gnan.kumar at accenture.com> To: <franco.broi at iongeo.com> Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes Message-ID: <67765C71374B974FBFD2AD05AF438EFF0BD586B9 at 048-CH1MPN3-362.048d.mgd.msft.net> Content-Type: text/plain; charset="utf-8" root at secondary:/export/sdd1/brick# gluster volume info Volume Name: dst Type: Distribute Status: Started Number of Bricks: 2 Transport-type: tcp Bricks: Brick1: primary:/export/sdd1/brick Brick2: secondary:/export/sdd1/brick -----Original Message----- From: Franco Broi [mailto:franco.broi at iongeo.com] Sent: Tuesday, June 03, 2014 12:56 PM To: Gnan Kumar, Yalla Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes What do gluster vol info and gluster vol status give you? On Tue, 2014-06-03 at 07:21 +0000, yalla.gnan.kumar at accenture.com wrote:> Hi, > > I have created a distributed volume on my gluster node. I haveattached this volume to a VM on openstack. The size is 1 GB. I have written files close to 1 GB onto the> Volume. But when I do a ls inside the brick directory , the volume ispresent only on one gluster server brick. But it is empty on another server brick. Files are meant to be> spread across both the bricks according to distributed volumedefinition.> > On the VM: > -------------- > > # ls -al > total 1013417 > drwxr-xr-x 3 root root 4096 Jun 1 22:03 . > drwxrwxr-x 3 root root 1024 Jun 1 21:24 .. > -rw------- 1 root root 31478251520 Jun 1 21:52 file > -rw------- 1 root root 157391257600 Jun 1 21:54 file1 > -rw------- 1 root root 629565030400 Jun 1 21:55 file2 > -rw------- 1 root root 708260659200 Jun 1 21:59 file3 > -rw------- 1 root root 6295650304 Jun 1 22:01 file4 > -rw------- 1 root root 39333801984 Jun 1 22:01 file5 > -rw------- 1 root root 78643200000 Jun 1 22:04 file6 > drwx------ 2 root root 16384 Jun 1 21:24 lost+found > ---------- > # du -sch * > 20.0M file > 100.0M file1 > 400.0M file2 > 454.0M file3 > 4.0M file4 > 11.6M file5 > 0 file6 > 16.0K lost+found > 989.7M total > ------------------------ > > > On the gluster server nodes: > ----------------------- > root at primary:/export/sdd1/brick# ll > total 12 > drwxr-xr-x 2 root root 4096 Jun 2 04:08 ./ drwxr-xr-x 4 root root > 4096 May 27 08:42 ../ root at primary:/export/sdd1/brick# > -------------------------- > > root at secondary:/export/sdd1/brick# ll > total 1046536 > drwxr-xr-x 2 root root 4096 Jun 2 08:51 ./ > drwxr-xr-x 4 root root 4096 May 27 08:43 ../ > -rw-rw-rw- 1 108 115 1073741824 Jun 2 09:35 > volume-0ec560be-997f-46da-9ec8-e9d6627f2de1 > root at secondary:/export/sdd1/brick# > --------------------------------- > > > Thanks > Kumar > > > > > > > > > -----Original Message----- > From: Franco Broi [mailto:franco.broi at iongeo.com] > Sent: Monday, June 02, 2014 6:35 PM > To: Gnan Kumar, Yalla > Cc: gluster-users at gluster.org > Subject: Re: [Gluster-users] Distributed volumes > > Just do an ls on the bricks, the paths are the same as the mountedfilesystem.> > On Mon, 2014-06-02 at 12:26 +0000, yalla.gnan.kumar at accenture.com > wrote: > > Hi All, > > > > > > > > I have created a distributed volume of 1 GB , using two bricks from > > two different servers. > > > > I have written 7 files whose sizes are a total of 1 GB. > > > > How can I check that files are distributed on both the bricks ? > > > > > > > > > > > > Thanks > > > > Kumar > > > > > > > > > > ____________________________________________________________________ > > __ > > > > > > This message is for the designated recipient only and may contain > > privileged, proprietary, or otherwise confidential information. If > > you have received it in error, please notify the sender immediately > > and delete the original. Any other use of the e-mail by you isprohibited.> > Where allowed by local law, electronic communications with Accenture > > and its affiliates, including e-mail and instant messaging > > (including content), may be scanned by our systems for the purposes > > of information security and assessment of internal compliance with > > Accenture policy. > > ____________________________________________________________________ > > __ > > ________________ > > > > www.accenture.com > > _______________________________________________ > > Gluster-users mailing list > > Gluster-users at gluster.org > > http://supercolony.gluster.org/mailman/listinfo/gluster-users > > > > > ________________________________ > > This message is for the designated recipient only and may containprivileged, proprietary, or otherwise confidential information. If you have received it in error, please notify the sender immediately and delete the original. Any other use of the e-mail by you is prohibited. Where allowed by local law, electronic communications with Accenture and its affiliates, including e-mail and instant messaging (including content), may be scanned by our systems for the purposes of information security and assessment of internal compliance with Accenture policy.> ______________________________________________________________________ > ________________ > > www.accenture.com------------------------------ Message: 15 Date: Tue, 3 Jun 2014 13:02:54 +0530 From: Kaushal M <kshlmster at gmail.com> To: Gluster Devel <gluster-devel at gluster.org> Cc: "gluster-users at gluster.org" <gluster-users at gluster.org> Subject: Re: [Gluster-users] [Gluster-devel] autodelete in snapshots Message-ID: <CAOujamXNzzsNZW0jX9gNPzP7JzUnM1gGDCnNiu7R=ygWTc4oFQ at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 I agree as well. We shouldn't be deleting any data without the explicit consent of the user. The approach proposed by MS is better than the earlier approach. ~kaushal On Tue, Jun 3, 2014 at 1:02 AM, M S Vishwanath Bhat <msvbhat at gmail.com> wrote:> > > > On 2 June 2014 20:22, Vijay Bellur <vbellur at redhat.com> wrote: >> >> On 04/23/2014 05:50 AM, Vijay Bellur wrote: >>> >>> On 04/20/2014 11:42 PM, Lalatendu Mohanty wrote: >>>> >>>> On 04/16/2014 11:39 AM, Avra Sengupta wrote: >>>>> >>>>> The whole purpose of introducing the soft-limit is, that at anypoint>>>>> of time the number of >>>>> snaps should not exceed the hard limit. If we trigger auto-delete on >>>>> hitting hard-limit, then >>>>> the purpose itself is lost, because at that point we would be takinga>>>>> snap, making the limit >>>>> hard-limit + 1, and then triggering auto-delete, which violates the >>>>> sanctity of the hard-limit. >>>>> Also what happens when we are at hard-limit + 1, and another snap is >>>>> issued, while auto-delete >>>>> is yet to process the first delete. At that point we end up at >>>>> hard-limit + 1. Also what happens >>>>> if for a particular snap the auto-delete fails. >>>>> >>>>> We should see the hard-limit, as something set by the admin keepingin>>>>> mind the resource consumption >>>>> and at no-point should we cross this limit, come what may. If we hit >>>>> this limit, the create command >>>>> should fail asking the user to delete snaps using the "snapshot >>>>> delete" command. >>>>> >>>>> The two options Raghavendra mentioned are applicable for the >>>>> soft-limit only, in which cases on >>>>> hitting the soft-limit >>>>> >>>>> 1. Trigger auto-delete >>>>> >>>>> or >>>>> >>>>> 2. Log a warning-message, for the user saying the number of snaps is >>>>> exceeding the snap-limit and >>>>> display the number of available snaps >>>>> >>>>> Now which of these should happen also depends on the user, becausethe>>>>> auto-delete option >>>>> is configurable. >>>>> >>>>> So if the auto-delete option is set as true, auto-delete should be >>>>> triggered and the above message >>>>> should also be logged. >>>>> >>>>> But if the option is set as false, only the message should belogged.>>>>> >>>>> This is the behaviour as designed. Adding Rahul, and Seema in the >>>>> mail, to reflect upon the >>>>> behaviour as well. >>>>> >>>>> Regards, >>>>> Avra >>>> >>>> >>>> This sounds correct. However we need to make sure that the usage or >>>> documentation around this should be good enough , so that users >>>> understand the each of the limits correctly. >>>> >>> >>> It might be better to avoid the usage of the term "soft-limit". >>> soft-limit as used in quota and other places generally has an alerting >>> connotation. Something like "auto-deletion-limit" might be better. >>> >> >> I still see references to "soft-limit" and auto deletion seems to get >> triggered upon reaching soft-limit. >> >> Why is the ability to auto delete not configurable? It does seem pretty >> nasty to go about deleting snapshots without obtaining explicit consentfrom>> the user. > > > I agree with Vijay here. It's not good to delete a snap (even though itis> oldest) without the explicit consent from user. > > FYI It took me more than 2 weeks to figure out that my snaps weregetting> autodeleted after reaching "soft-limit". For all I know I had not done > anything and my snap restore were failing. > > I propose to remove the terms "soft" and "hard" limit. I believe there > should be a limit (just "limit") after which all snapshot creates should > fail with proper error messages. And there can be a water-mark afterwhich> user should get warning messages. So below is my proposal. > > auto-delete + snap-limit: If the snap-limit is set to n, next snapcreate> (n+1th) will succeed only if if auto-delete is set to on/true/1 andoldest> snap will get deleted automatically. If autodelete is set to off/false/0,> (n+1)th snap create will fail with proper error message from gluster CLI > command. But again by default autodelete should be off. > > snap-water-mark: This should come in picture only if autodelete isturned> off. It should not have any meaning if auto-delete is turned ON.Basically> it's usage is to give the user warning that limit almost being reachedand> it is time for admin to decide which snaps should be deleted (or which > should be kept) > > *my two cents* > > -MS > >> >> >> Cheers, >> >> Vijay >> >> _______________________________________________ >> Gluster-devel mailing list >> Gluster-devel at gluster.org >> http://supercolony.gluster.org/mailman/listinfo/gluster-devel > > > > _______________________________________________ > Gluster-devel mailing list > Gluster-devel at gluster.org > http://supercolony.gluster.org/mailman/listinfo/gluster-devel >------------------------------ Message: 16 Date: Tue, 03 Jun 2014 15:34:01 +0800 From: Franco Broi <franco.broi at iongeo.com> To: yalla.gnan.kumar at accenture.com Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes Message-ID: <1401780841.2236.304.camel at tc1> Content-Type: text/plain; charset="UTF-8" Ok, what you have is a single large file (must be filesystem image??). Gluster will not stripe files, it writes different whole files to different bricks. On Tue, 2014-06-03 at 07:29 +0000, yalla.gnan.kumar at accenture.com wrote:> root at secondary:/export/sdd1/brick# gluster volume info > > Volume Name: dst > Type: Distribute > Status: Started > Number of Bricks: 2 > Transport-type: tcp > Bricks: > Brick1: primary:/export/sdd1/brick > Brick2: secondary:/export/sdd1/brick > > > > -----Original Message----- > From: Franco Broi [mailto:franco.broi at iongeo.com] > Sent: Tuesday, June 03, 2014 12:56 PM > To: Gnan Kumar, Yalla > Cc: gluster-users at gluster.org > Subject: Re: [Gluster-users] Distributed volumes > > > What do gluster vol info and gluster vol status give you? > > On Tue, 2014-06-03 at 07:21 +0000, yalla.gnan.kumar at accenture.com > wrote: > > Hi, > > > > I have created a distributed volume on my gluster node. I haveattached this volume to a VM on openstack. The size is 1 GB. I have written files close to 1 GB onto the> > Volume. But when I do a ls inside the brick directory , the volumeis present only on one gluster server brick. But it is empty on another server brick. Files are meant to be> > spread across both the bricks according to distributed volumedefinition.> > > > On the VM: > > -------------- > > > > # ls -al > > total 1013417 > > drwxr-xr-x 3 root root 4096 Jun 1 22:03 . > > drwxrwxr-x 3 root root 1024 Jun 1 21:24 .. > > -rw------- 1 root root 31478251520 Jun 1 21:52 file > > -rw------- 1 root root 157391257600 Jun 1 21:54 file1 > > -rw------- 1 root root 629565030400 Jun 1 21:55 file2 > > -rw------- 1 root root 708260659200 Jun 1 21:59 file3 > > -rw------- 1 root root 6295650304 Jun 1 22:01 file4 > > -rw------- 1 root root 39333801984 Jun 1 22:01 file5 > > -rw------- 1 root root 78643200000 Jun 1 22:04 file6 > > drwx------ 2 root root 16384 Jun 1 21:24 lost+found > > ---------- > > # du -sch * > > 20.0M file > > 100.0M file1 > > 400.0M file2 > > 454.0M file3 > > 4.0M file4 > > 11.6M file5 > > 0 file6 > > 16.0K lost+found > > 989.7M total > > ------------------------ > > > > > > On the gluster server nodes: > > ----------------------- > > root at primary:/export/sdd1/brick# ll > > total 12 > > drwxr-xr-x 2 root root 4096 Jun 2 04:08 ./ drwxr-xr-x 4 root root > > 4096 May 27 08:42 ../ root at primary:/export/sdd1/brick# > > -------------------------- > > > > root at secondary:/export/sdd1/brick# ll > > total 1046536 > > drwxr-xr-x 2 root root 4096 Jun 2 08:51 ./ > > drwxr-xr-x 4 root root 4096 May 27 08:43 ../ > > -rw-rw-rw- 1 108 115 1073741824 Jun 2 09:35 > > volume-0ec560be-997f-46da-9ec8-e9d6627f2de1 > > root at secondary:/export/sdd1/brick# > > --------------------------------- > > > > > > Thanks > > Kumar > > > > > > > > > > > > > > > > > > -----Original Message----- > > From: Franco Broi [mailto:franco.broi at iongeo.com] > > Sent: Monday, June 02, 2014 6:35 PM > > To: Gnan Kumar, Yalla > > Cc: gluster-users at gluster.org > > Subject: Re: [Gluster-users] Distributed volumes > > > > Just do an ls on the bricks, the paths are the same as the mountedfilesystem.> > > > On Mon, 2014-06-02 at 12:26 +0000, yalla.gnan.kumar at accenture.com > > wrote: > > > Hi All, > > > > > > > > > > > > I have created a distributed volume of 1 GB , using two bricks from> > > two different servers. > > > > > > I have written 7 files whose sizes are a total of 1 GB. > > > > > > How can I check that files are distributed on both the bricks ? > > > > > > > > > > > > > > > > > > Thanks > > > > > > Kumar > > > > > > > > > > > > > > > ____________________________________________________________________ > > > __ > > > > > > > > > This message is for the designated recipient only and may contain > > > privileged, proprietary, or otherwise confidential information. If > > > you have received it in error, please notify the sender immediately > > > and delete the original. Any other use of the e-mail by you isprohibited.> > > Where allowed by local law, electronic communications with Accenture> > > and its affiliates, including e-mail and instant messaging > > > (including content), may be scanned by our systems for the purposes > > > of information security and assessment of internal compliance with > > > Accenture policy. > > > ____________________________________________________________________ > > > __ > > > ________________ > > > > > > www.accenture.com > > > _______________________________________________ > > > Gluster-users mailing list > > > Gluster-users at gluster.org > > > http://supercolony.gluster.org/mailman/listinfo/gluster-users > > > > > > > > > > ________________________________ > > > > This message is for the designated recipient only and may containprivileged, proprietary, or otherwise confidential information. If you have received it in error, please notify the sender immediately and delete the original. Any other use of the e-mail by you is prohibited. Where allowed by local law, electronic communications with Accenture and its affiliates, including e-mail and instant messaging (including content), may be scanned by our systems for the purposes of information security and assessment of internal compliance with Accenture policy.> > ______________________________________________________________________ > > ________________ > > > > www.accenture.com > > >------------------------------ Message: 17 Date: Tue, 3 Jun 2014 07:39:58 +0000 From: <yalla.gnan.kumar at accenture.com> To: <franco.broi at iongeo.com> Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes Message-ID: <67765C71374B974FBFD2AD05AF438EFF0BD586DF at 048-CH1MPN3-362.048d.mgd.msft.net> Content-Type: text/plain; charset="utf-8" I have created distributed volume, created a 1 GB volume on it, and attached it to the VM and created a filesystem on it. How to verify that the files in the vm are distributed across both the bricks on two servers ? -----Original Message----- From: Franco Broi [mailto:franco.broi at iongeo.com] Sent: Tuesday, June 03, 2014 1:04 PM To: Gnan Kumar, Yalla Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes Ok, what you have is a single large file (must be filesystem image??). Gluster will not stripe files, it writes different whole files to different bricks. On Tue, 2014-06-03 at 07:29 +0000, yalla.gnan.kumar at accenture.com wrote:> root at secondary:/export/sdd1/brick# gluster volume info > > Volume Name: dst > Type: Distribute > Status: Started > Number of Bricks: 2 > Transport-type: tcp > Bricks: > Brick1: primary:/export/sdd1/brick > Brick2: secondary:/export/sdd1/brick > > > > -----Original Message----- > From: Franco Broi [mailto:franco.broi at iongeo.com] > Sent: Tuesday, June 03, 2014 12:56 PM > To: Gnan Kumar, Yalla > Cc: gluster-users at gluster.org > Subject: Re: [Gluster-users] Distributed volumes > > > What do gluster vol info and gluster vol status give you? > > On Tue, 2014-06-03 at 07:21 +0000, yalla.gnan.kumar at accenture.com > wrote: > > Hi, > > > > I have created a distributed volume on my gluster node. I haveattached this volume to a VM on openstack. The size is 1 GB. I have written files close to 1 GB onto the> > Volume. But when I do a ls inside the brick directory , the volumeis present only on one gluster server brick. But it is empty on another server brick. Files are meant to be> > spread across both the bricks according to distributed volumedefinition.> > > > On the VM: > > -------------- > > > > # ls -al > > total 1013417 > > drwxr-xr-x 3 root root 4096 Jun 1 22:03 . > > drwxrwxr-x 3 root root 1024 Jun 1 21:24 .. > > -rw------- 1 root root 31478251520 Jun 1 21:52 file > > -rw------- 1 root root 157391257600 Jun 1 21:54 file1 > > -rw------- 1 root root 629565030400 Jun 1 21:55 file2 > > -rw------- 1 root root 708260659200 Jun 1 21:59 file3 > > -rw------- 1 root root 6295650304 Jun 1 22:01 file4 > > -rw------- 1 root root 39333801984 Jun 1 22:01 file5 > > -rw------- 1 root root 78643200000 Jun 1 22:04 file6 > > drwx------ 2 root root 16384 Jun 1 21:24 lost+found > > ---------- > > # du -sch * > > 20.0M file > > 100.0M file1 > > 400.0M file2 > > 454.0M file3 > > 4.0M file4 > > 11.6M file5 > > 0 file6 > > 16.0K lost+found > > 989.7M total > > ------------------------ > > > > > > On the gluster server nodes: > > ----------------------- > > root at primary:/export/sdd1/brick# ll > > total 12 > > drwxr-xr-x 2 root root 4096 Jun 2 04:08 ./ drwxr-xr-x 4 root root > > 4096 May 27 08:42 ../ root at primary:/export/sdd1/brick# > > -------------------------- > > > > root at secondary:/export/sdd1/brick# ll total 1046536 > > drwxr-xr-x 2 root root 4096 Jun 2 08:51 ./ > > drwxr-xr-x 4 root root 4096 May 27 08:43 ../ > > -rw-rw-rw- 1 108 115 1073741824 Jun 2 09:35 > > volume-0ec560be-997f-46da-9ec8-e9d6627f2de1 > > root at secondary:/export/sdd1/brick# > > --------------------------------- > > > > > > Thanks > > Kumar > > > > > > > > > > > > > > > > > > -----Original Message----- > > From: Franco Broi [mailto:franco.broi at iongeo.com] > > Sent: Monday, June 02, 2014 6:35 PM > > To: Gnan Kumar, Yalla > > Cc: gluster-users at gluster.org > > Subject: Re: [Gluster-users] Distributed volumes > > > > Just do an ls on the bricks, the paths are the same as the mountedfilesystem.> > > > On Mon, 2014-06-02 at 12:26 +0000, yalla.gnan.kumar at accenture.com > > wrote: > > > Hi All, > > > > > > > > > > > > I have created a distributed volume of 1 GB , using two bricks from> > > two different servers. > > > > > > I have written 7 files whose sizes are a total of 1 GB. > > > > > > How can I check that files are distributed on both the bricks ? > > > > > > > > > > > > > > > > > > Thanks > > > > > > Kumar > > > > > > > > > > > > > > > ____________________________________________________________________ > > > __ > > > > > > > > > This message is for the designated recipient only and may contain > > > privileged, proprietary, or otherwise confidential information. If > > > you have received it in error, please notify the sender immediately > > > and delete the original. Any other use of the e-mail by you isprohibited.> > > Where allowed by local law, electronic communications with Accenture> > > and its affiliates, including e-mail and instant messaging > > > (including content), may be scanned by our systems for the purposes > > > of information security and assessment of internal compliance with > > > Accenture policy. > > > ____________________________________________________________________ > > > __ > > > ________________ > > > > > > www.accenture.com > > > _______________________________________________ > > > Gluster-users mailing list > > > Gluster-users at gluster.org > > > http://supercolony.gluster.org/mailman/listinfo/gluster-users > > > > > > > > > > ________________________________ > > > > This message is for the designated recipient only and may containprivileged, proprietary, or otherwise confidential information. If you have received it in error, please notify the sender immediately and delete the original. Any other use of the e-mail by you is prohibited. Where allowed by local law, electronic communications with Accenture and its affiliates, including e-mail and instant messaging (including content), may be scanned by our systems for the purposes of information security and assessment of internal compliance with Accenture policy.> > ______________________________________________________________________ > > ________________ > > > > www.accenture.com > > >------------------------------ Message: 18 Date: Tue, 3 Jun 2014 13:19:08 +0530 From: Kaushal M <kshlmster at gmail.com> To: yalla.gnan.kumar at accenture.com Cc: "gluster-users at gluster.org" <gluster-users at gluster.org> Subject: Re: [Gluster-users] Distributed volumes Message-ID: <CAOujamXxBoF3QAbYgyAOvGY7fimNgkGBv8Z-gotpchh0xgEePA at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 You have only 1 file on the gluster volume, the 1GB disk image/volume that you created. This disk image is attached to the VM as a file system, not the gluster volume. So whatever you do in the VM's file system, affects just the 1 disk image. The files, directories etc. you created, are inside the disk image. So you still have just one file on the gluster volume, not many as you are assuming. On Tue, Jun 3, 2014 at 1:09 PM, <yalla.gnan.kumar at accenture.com> wrote:> I have created distributed volume, created a 1 GB volume on it, andattached it to the VM and created a filesystem on it. How to verify that the files in the vm are> distributed across both the bricks on two servers ? > > > > -----Original Message----- > From: Franco Broi [mailto:franco.broi at iongeo.com] > Sent: Tuesday, June 03, 2014 1:04 PM > To: Gnan Kumar, Yalla > Cc: gluster-users at gluster.org > Subject: Re: [Gluster-users] Distributed volumes > > > Ok, what you have is a single large file (must be filesystem image??). > Gluster will not stripe files, it writes different whole files todifferent bricks.> > On Tue, 2014-06-03 at 07:29 +0000, yalla.gnan.kumar at accenture.com > wrote: >> root at secondary:/export/sdd1/brick# gluster volume info >> >> Volume Name: dst >> Type: Distribute >> Status: Started >> Number of Bricks: 2 >> Transport-type: tcp >> Bricks: >> Brick1: primary:/export/sdd1/brick >> Brick2: secondary:/export/sdd1/brick >> >> >> >> -----Original Message----- >> From: Franco Broi [mailto:franco.broi at iongeo.com] >> Sent: Tuesday, June 03, 2014 12:56 PM >> To: Gnan Kumar, Yalla >> Cc: gluster-users at gluster.org >> Subject: Re: [Gluster-users] Distributed volumes >> >> >> What do gluster vol info and gluster vol status give you? >> >> On Tue, 2014-06-03 at 07:21 +0000, yalla.gnan.kumar at accenture.com >> wrote: >> > Hi, >> > >> > I have created a distributed volume on my gluster node. I haveattached this volume to a VM on openstack. The size is 1 GB. I have written files close to 1 GB onto the>> > Volume. But when I do a ls inside the brick directory , the volumeis present only on one gluster server brick. But it is empty on another server brick. Files are meant to be>> > spread across both the bricks according to distributed volumedefinition.>> > >> > On the VM: >> > -------------- >> > >> > # ls -al >> > total 1013417 >> > drwxr-xr-x 3 root root 4096 Jun 1 22:03 . >> > drwxrwxr-x 3 root root 1024 Jun 1 21:24 .. >> > -rw------- 1 root root 31478251520 Jun 1 21:52 file >> > -rw------- 1 root root 157391257600 Jun 1 21:54 file1 >> > -rw------- 1 root root 629565030400 Jun 1 21:55 file2 >> > -rw------- 1 root root 708260659200 Jun 1 21:59 file3 >> > -rw------- 1 root root 6295650304 Jun 1 22:01 file4 >> > -rw------- 1 root root 39333801984 Jun 1 22:01 file5 >> > -rw------- 1 root root 78643200000 Jun 1 22:04 file6 >> > drwx------ 2 root root 16384 Jun 1 21:24 lost+found >> > ---------- >> > # du -sch * >> > 20.0M file >> > 100.0M file1 >> > 400.0M file2 >> > 454.0M file3 >> > 4.0M file4 >> > 11.6M file5 >> > 0 file6 >> > 16.0K lost+found >> > 989.7M total >> > ------------------------ >> > >> > >> > On the gluster server nodes: >> > ----------------------- >> > root at primary:/export/sdd1/brick# ll >> > total 12 >> > drwxr-xr-x 2 root root 4096 Jun 2 04:08 ./ drwxr-xr-x 4 root root >> > 4096 May 27 08:42 ../ root at primary:/export/sdd1/brick# >> > -------------------------- >> > >> > root at secondary:/export/sdd1/brick# ll total 1046536 >> > drwxr-xr-x 2 root root 4096 Jun 2 08:51 ./ >> > drwxr-xr-x 4 root root 4096 May 27 08:43 ../ >> > -rw-rw-rw- 1 108 115 1073741824 Jun 2 09:35 >> > volume-0ec560be-997f-46da-9ec8-e9d6627f2de1 >> > root at secondary:/export/sdd1/brick# >> > --------------------------------- >> > >> > >> > Thanks >> > Kumar >> > >> > >> > >> > >> > >> > >> > >> > >> > -----Original Message----- >> > From: Franco Broi [mailto:franco.broi at iongeo.com] >> > Sent: Monday, June 02, 2014 6:35 PM >> > To: Gnan Kumar, Yalla >> > Cc: gluster-users at gluster.org >> > Subject: Re: [Gluster-users] Distributed volumes >> > >> > Just do an ls on the bricks, the paths are the same as the mountedfilesystem.>> > >> > On Mon, 2014-06-02 at 12:26 +0000, yalla.gnan.kumar at accenture.com >> > wrote: >> > > Hi All, >> > > >> > > >> > > >> > > I have created a distributed volume of 1 GB , using two bricksfrom>> > > two different servers. >> > > >> > > I have written 7 files whose sizes are a total of 1 GB. >> > > >> > > How can I check that files are distributed on both the bricks ? >> > > >> > > >> > > >> > > >> > > >> > > Thanks >> > > >> > > Kumar >> > > >> > > >> > > >> > > >> > >____________________________________________________________________>> > > __ >> > > >> > > >> > > This message is for the designated recipient only and may contain >> > > privileged, proprietary, or otherwise confidential information. If >> > > you have received it in error, please notify the sender immediately >> > > and delete the original. Any other use of the e-mail by you isprohibited.>> > > Where allowed by local law, electronic communications withAccenture>> > > and its affiliates, including e-mail and instant messaging >> > > (including content), may be scanned by our systems for the purposes >> > > of information security and assessment of internal compliance with >> > > Accenture policy. >> > >____________________________________________________________________>> > > __ >> > > ________________ >> > > >> > > www.accenture.com >> > > _______________________________________________ >> > > Gluster-users mailing list >> > > Gluster-users at gluster.org >> > > http://supercolony.gluster.org/mailman/listinfo/gluster-users >> > >> > >> > >> > >> > ________________________________ >> > >> > This message is for the designated recipient only and may containprivileged, proprietary, or otherwise confidential information. If you have received it in error, please notify the sender immediately and delete the original. Any other use of the e-mail by you is prohibited. Where allowed by local law, electronic communications with Accenture and its affiliates, including e-mail and instant messaging (including content), may be scanned by our systems for the purposes of information security and assessment of internal compliance with Accenture policy.>> >______________________________________________________________________>> > ________________ >> > >> > www.accenture.com >> >> >> > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://supercolony.gluster.org/mailman/listinfo/gluster-users------------------------------ Message: 19 Date: Tue, 3 Jun 2014 08:20:46 +0000 From: <yalla.gnan.kumar at accenture.com> To: <kshlmster at gmail.com> Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes Message-ID: <67765C71374B974FBFD2AD05AF438EFF0BD58729 at 048-CH1MPN3-362.048d.mgd.msft.net> Content-Type: text/plain; charset="utf-8" Hi, So , in which scenario, does the distributed volumes have files on both the bricks ? -----Original Message----- From: Kaushal M [mailto:kshlmster at gmail.com] Sent: Tuesday, June 03, 2014 1:19 PM To: Gnan Kumar, Yalla Cc: Franco Broi; gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes You have only 1 file on the gluster volume, the 1GB disk image/volume that you created. This disk image is attached to the VM as a file system, not the gluster volume. So whatever you do in the VM's file system, affects just the 1 disk image. The files, directories etc. you created, are inside the disk image. So you still have just one file on the gluster volume, not many as you are assuming. On Tue, Jun 3, 2014 at 1:09 PM, <yalla.gnan.kumar at accenture.com> wrote:> I have created distributed volume, created a 1 GB volume on it, and > attached it to the VM and created a filesystem on it. How to verifythat the files in the vm are distributed across both the bricks on two servers ?> > > > -----Original Message----- > From: Franco Broi [mailto:franco.broi at iongeo.com] > Sent: Tuesday, June 03, 2014 1:04 PM > To: Gnan Kumar, Yalla > Cc: gluster-users at gluster.org > Subject: Re: [Gluster-users] Distributed volumes > > > Ok, what you have is a single large file (must be filesystem image??). > Gluster will not stripe files, it writes different whole files todifferent bricks.> > On Tue, 2014-06-03 at 07:29 +0000, yalla.gnan.kumar at accenture.com > wrote: >> root at secondary:/export/sdd1/brick# gluster volume info >> >> Volume Name: dst >> Type: Distribute >> Status: Started >> Number of Bricks: 2 >> Transport-type: tcp >> Bricks: >> Brick1: primary:/export/sdd1/brick >> Brick2: secondary:/export/sdd1/brick >> >> >> >> -----Original Message----- >> From: Franco Broi [mailto:franco.broi at iongeo.com] >> Sent: Tuesday, June 03, 2014 12:56 PM >> To: Gnan Kumar, Yalla >> Cc: gluster-users at gluster.org >> Subject: Re: [Gluster-users] Distributed volumes >> >> >> What do gluster vol info and gluster vol status give you? >> >> On Tue, 2014-06-03 at 07:21 +0000, yalla.gnan.kumar at accenture.com >> wrote: >> > Hi, >> > >> > I have created a distributed volume on my gluster node. I haveattached this volume to a VM on openstack. The size is 1 GB. I have written files close to 1 GB onto the>> > Volume. But when I do a ls inside the brick directory , the volumeis present only on one gluster server brick. But it is empty on another server brick. Files are meant to be>> > spread across both the bricks according to distributed volumedefinition.>> > >> > On the VM: >> > -------------- >> > >> > # ls -al >> > total 1013417 >> > drwxr-xr-x 3 root root 4096 Jun 1 22:03 . >> > drwxrwxr-x 3 root root 1024 Jun 1 21:24 .. >> > -rw------- 1 root root 31478251520 Jun 1 21:52 file >> > -rw------- 1 root root 157391257600 Jun 1 21:54 file1 >> > -rw------- 1 root root 629565030400 Jun 1 21:55 file2 >> > -rw------- 1 root root 708260659200 Jun 1 21:59 file3 >> > -rw------- 1 root root 6295650304 Jun 1 22:01 file4 >> > -rw------- 1 root root 39333801984 Jun 1 22:01 file5 >> > -rw------- 1 root root 78643200000 Jun 1 22:04 file6 >> > drwx------ 2 root root 16384 Jun 1 21:24 lost+found >> > ---------- >> > # du -sch * >> > 20.0M file >> > 100.0M file1 >> > 400.0M file2 >> > 454.0M file3 >> > 4.0M file4 >> > 11.6M file5 >> > 0 file6 >> > 16.0K lost+found >> > 989.7M total >> > ------------------------ >> > >> > >> > On the gluster server nodes: >> > ----------------------- >> > root at primary:/export/sdd1/brick# ll total 12 drwxr-xr-x 2 root root >> > 4096 Jun 2 04:08 ./ drwxr-xr-x 4 root root >> > 4096 May 27 08:42 ../ root at primary:/export/sdd1/brick# >> > -------------------------- >> > >> > root at secondary:/export/sdd1/brick# ll total 1046536 >> > drwxr-xr-x 2 root root 4096 Jun 2 08:51 ./ >> > drwxr-xr-x 4 root root 4096 May 27 08:43 ../ >> > -rw-rw-rw- 1 108 115 1073741824 Jun 2 09:35 >> > volume-0ec560be-997f-46da-9ec8-e9d6627f2de1 >> > root at secondary:/export/sdd1/brick# >> > --------------------------------- >> > >> > >> > Thanks >> > Kumar >> > >> > >> > >> > >> > >> > >> > >> > >> > -----Original Message----- >> > From: Franco Broi [mailto:franco.broi at iongeo.com] >> > Sent: Monday, June 02, 2014 6:35 PM >> > To: Gnan Kumar, Yalla >> > Cc: gluster-users at gluster.org >> > Subject: Re: [Gluster-users] Distributed volumes >> > >> > Just do an ls on the bricks, the paths are the same as the mountedfilesystem.>> > >> > On Mon, 2014-06-02 at 12:26 +0000, yalla.gnan.kumar at accenture.com >> > wrote: >> > > Hi All, >> > > >> > > >> > > >> > > I have created a distributed volume of 1 GB , using two bricks >> > > from two different servers. >> > > >> > > I have written 7 files whose sizes are a total of 1 GB. >> > > >> > > How can I check that files are distributed on both the bricks ? >> > > >> > > >> > > >> > > >> > > >> > > Thanks >> > > >> > > Kumar >> > > >> > > >> > > >> > > >> > > _________________________________________________________________ >> > > ___ >> > > __ >> > > >> > > >> > > This message is for the designated recipient only and may contain >> > > privileged, proprietary, or otherwise confidential information. >> > > If you have received it in error, please notify the sender >> > > immediately and delete the original. Any other use of the e-mail byyou is prohibited.>> > > Where allowed by local law, electronic communications with >> > > Accenture and its affiliates, including e-mail and instant >> > > messaging (including content), may be scanned by our systems for >> > > the purposes of information security and assessment of internal >> > > compliance with Accenture policy. >> > > _________________________________________________________________ >> > > ___ >> > > __ >> > > ________________ >> > > >> > > www.accenture.com >> > > _______________________________________________ >> > > Gluster-users mailing list >> > > Gluster-users at gluster.org >> > > http://supercolony.gluster.org/mailman/listinfo/gluster-users >> > >> > >> > >> > >> > ________________________________ >> > >> > This message is for the designated recipient only and may containprivileged, proprietary, or otherwise confidential information. If you have received it in error, please notify the sender immediately and delete the original. Any other use of the e-mail by you is prohibited. Where allowed by local law, electronic communications with Accenture and its affiliates, including e-mail and instant messaging (including content), may be scanned by our systems for the purposes of information security and assessment of internal compliance with Accenture policy.>> > ___________________________________________________________________ >> > ___ >> > ________________ >> > >> > www.accenture.com >> >> >> > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://supercolony.gluster.org/mailman/listinfo/gluster-users------------------------------ Message: 20 Date: Tue, 03 Jun 2014 16:22:16 +0800 From: Franco Broi <franco.broi at iongeo.com> To: yalla.gnan.kumar at accenture.com Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes Message-ID: <1401783736.2236.333.camel at tc1> Content-Type: text/plain; charset="UTF-8" On Tue, 2014-06-03 at 08:20 +0000, yalla.gnan.kumar at accenture.com wrote:> Hi, > > So , in which scenario, does the distributed volumes have files on boththe bricks ? If you make more than 1 file.> > > -----Original Message----- > From: Kaushal M [mailto:kshlmster at gmail.com] > Sent: Tuesday, June 03, 2014 1:19 PM > To: Gnan Kumar, Yalla > Cc: Franco Broi; gluster-users at gluster.org > Subject: Re: [Gluster-users] Distributed volumes > > You have only 1 file on the gluster volume, the 1GB disk image/volumethat you created. This disk image is attached to the VM as a file system, not the gluster volume. So whatever you do in the VM's file system, affects just the 1 disk image. The files, directories etc. you created, are inside the disk image. So you still have just one file on the gluster volume, not many as you are assuming.> > > > On Tue, Jun 3, 2014 at 1:09 PM, <yalla.gnan.kumar at accenture.com> wrote: > > I have created distributed volume, created a 1 GB volume on it, and > > attached it to the VM and created a filesystem on it. How to verifythat the files in the vm are distributed across both the bricks on two servers ?> > > > > > > > -----Original Message----- > > From: Franco Broi [mailto:franco.broi at iongeo.com] > > Sent: Tuesday, June 03, 2014 1:04 PM > > To: Gnan Kumar, Yalla > > Cc: gluster-users at gluster.org > > Subject: Re: [Gluster-users] Distributed volumes > > > > > > Ok, what you have is a single large file (must be filesystem image??). > > Gluster will not stripe files, it writes different whole files todifferent bricks.> > > > On Tue, 2014-06-03 at 07:29 +0000, yalla.gnan.kumar at accenture.com > > wrote: > >> root at secondary:/export/sdd1/brick# gluster volume info > >> > >> Volume Name: dst > >> Type: Distribute > >> Status: Started > >> Number of Bricks: 2 > >> Transport-type: tcp > >> Bricks: > >> Brick1: primary:/export/sdd1/brick > >> Brick2: secondary:/export/sdd1/brick > >> > >> > >> > >> -----Original Message----- > >> From: Franco Broi [mailto:franco.broi at iongeo.com] > >> Sent: Tuesday, June 03, 2014 12:56 PM > >> To: Gnan Kumar, Yalla > >> Cc: gluster-users at gluster.org > >> Subject: Re: [Gluster-users] Distributed volumes > >> > >> > >> What do gluster vol info and gluster vol status give you? > >> > >> On Tue, 2014-06-03 at 07:21 +0000, yalla.gnan.kumar at accenture.com > >> wrote: > >> > Hi, > >> > > >> > I have created a distributed volume on my gluster node. I haveattached this volume to a VM on openstack. The size is 1 GB. I have written files close to 1 GB onto the> >> > Volume. But when I do a ls inside the brick directory , thevolume is present only on one gluster server brick. But it is empty on another server brick. Files are meant to be> >> > spread across both the bricks according to distributed volumedefinition.> >> > > >> > On the VM: > >> > -------------- > >> > > >> > # ls -al > >> > total 1013417 > >> > drwxr-xr-x 3 root root 4096 Jun 1 22:03 . > >> > drwxrwxr-x 3 root root 1024 Jun 1 21:24 .. > >> > -rw------- 1 root root 31478251520 Jun 1 21:52 file > >> > -rw------- 1 root root 157391257600 Jun 1 21:54 file1 > >> > -rw------- 1 root root 629565030400 Jun 1 21:55 file2 > >> > -rw------- 1 root root 708260659200 Jun 1 21:59 file3 > >> > -rw------- 1 root root 6295650304 Jun 1 22:01 file4 > >> > -rw------- 1 root root 39333801984 Jun 1 22:01 file5 > >> > -rw------- 1 root root 78643200000 Jun 1 22:04 file6 > >> > drwx------ 2 root root 16384 Jun 1 21:24 lost+found > >> > ---------- > >> > # du -sch * > >> > 20.0M file > >> > 100.0M file1 > >> > 400.0M file2 > >> > 454.0M file3 > >> > 4.0M file4 > >> > 11.6M file5 > >> > 0 file6 > >> > 16.0K lost+found > >> > 989.7M total > >> > ------------------------ > >> > > >> > > >> > On the gluster server nodes: > >> > ----------------------- > >> > root at primary:/export/sdd1/brick# ll total 12 drwxr-xr-x 2 root root> >> > 4096 Jun 2 04:08 ./ drwxr-xr-x 4 root root > >> > 4096 May 27 08:42 ../ root at primary:/export/sdd1/brick# > >> > -------------------------- > >> > > >> > root at secondary:/export/sdd1/brick# ll total 1046536 > >> > drwxr-xr-x 2 root root 4096 Jun 2 08:51 ./ > >> > drwxr-xr-x 4 root root 4096 May 27 08:43 ../ > >> > -rw-rw-rw- 1 108 115 1073741824 Jun 2 09:35 > >> > volume-0ec560be-997f-46da-9ec8-e9d6627f2de1 > >> > root at secondary:/export/sdd1/brick# > >> > --------------------------------- > >> > > >> > > >> > Thanks > >> > Kumar > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > -----Original Message----- > >> > From: Franco Broi [mailto:franco.broi at iongeo.com] > >> > Sent: Monday, June 02, 2014 6:35 PM > >> > To: Gnan Kumar, Yalla > >> > Cc: gluster-users at gluster.org > >> > Subject: Re: [Gluster-users] Distributed volumes > >> > > >> > Just do an ls on the bricks, the paths are the same as the mountedfilesystem.> >> > > >> > On Mon, 2014-06-02 at 12:26 +0000, yalla.gnan.kumar at accenture.com > >> > wrote: > >> > > Hi All, > >> > > > >> > > > >> > > > >> > > I have created a distributed volume of 1 GB , using two bricks > >> > > from two different servers. > >> > > > >> > > I have written 7 files whose sizes are a total of 1 GB. > >> > > > >> > > How can I check that files are distributed on both the bricks ? > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > Thanks > >> > > > >> > > Kumar > >> > > > >> > > > >> > > > >> > > > >> > > _________________________________________________________________ > >> > > ___ > >> > > __ > >> > > > >> > > > >> > > This message is for the designated recipient only and may contain> >> > > privileged, proprietary, or otherwise confidential information. > >> > > If you have received it in error, please notify the sender > >> > > immediately and delete the original. Any other use of the e-mailby you is prohibited.> >> > > Where allowed by local law, electronic communications with > >> > > Accenture and its affiliates, including e-mail and instant > >> > > messaging (including content), may be scanned by our systems for > >> > > the purposes of information security and assessment of internal > >> > > compliance with Accenture policy. > >> > > _________________________________________________________________ > >> > > ___ > >> > > __ > >> > > ________________ > >> > > > >> > > www.accenture.com > >> > > _______________________________________________ > >> > > Gluster-users mailing list > >> > > Gluster-users at gluster.org > >> > > http://supercolony.gluster.org/mailman/listinfo/gluster-users > >> > > >> > > >> > > >> > > >> > ________________________________ > >> > > >> > This message is for the designated recipient only and may containprivileged, proprietary, or otherwise confidential information. If you have received it in error, please notify the sender immediately and delete the original. Any other use of the e-mail by you is prohibited. Where allowed by local law, electronic communications with Accenture and its affiliates, including e-mail and instant messaging (including content), may be scanned by our systems for the purposes of information security and assessment of internal compliance with Accenture policy.> >> > ___________________________________________________________________ > >> > ___ > >> > ________________ > >> > > >> > www.accenture.com > >> > >> > >> > > > > > > > > _______________________________________________ > > Gluster-users mailing list > > Gluster-users at gluster.org > > http://supercolony.gluster.org/mailman/listinfo/gluster-users >------------------------------ Message: 21 Date: Tue, 03 Jun 2014 14:21:39 +0530 From: Vijay Bellur <vbellur at redhat.com> To: yalla.gnan.kumar at accenture.com, kshlmster at gmail.com Cc: gluster-users at gluster.org Subject: Re: [Gluster-users] Distributed volumes Message-ID: <538D8C9B.3010709 at redhat.com> Content-Type: text/plain; charset=ISO-8859-1; format=flowed On 06/03/2014 01:50 PM, yalla.gnan.kumar at accenture.com wrote:> Hi, > > So , in which scenario, does the distributed volumes have files on boththe bricks ?> >Reading the documentation for various volume types [1] can be useful to obtain answers for questions of this nature. -Vijay [1] https://github.com/gluster/glusterfs/blob/master/doc/admin-guide/en-US/markdown/admin_setting_volumes.md ------------------------------ Message: 22 Date: Tue, 3 Jun 2014 17:10:57 +0530 From: Indivar Nair <indivar.nair at techterra.in> To: Gluster Users <gluster-users at gluster.org> Subject: [Gluster-users] NFS ACL Support in Gluster 3.4 Message-ID: <CALuPYL0CykF9Q41SKsWtOSRXnAqor02mfR6W9FAth18XwK=cXQ at mail.gmail.com> Content-Type: text/plain; charset="utf-8" Hi All, I recently upgraded a Gluster 3.3.1 installation to Gluster 3.4. It was a straight forward upgrade using Yum. The OS is CentOS 6.3. The main purpose of the upgrade was to get ACL Support on NFS exports. But it doesn't seem to be working. I mounted the gluster volume using the following options - mount -t nfs -o vers=3,mountproto=tcp,acl <gluster_server>:/volume /mnt The getfacl or setfacl commands does not work on any dir/files on this mount. The plan is to re-export the NFS Mounts using Samba+CTDB. NFS mounts seem to give better performance than Gluster Mounts. Am I missing something? Regards, Indivar Nair -------------- next part -------------- An HTML attachment was scrubbed... URL: < http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140603/e543c5e8/attachment-0001.html>------------------------------ Message: 23 Date: Tue, 03 Jun 2014 17:26:56 +0530 From: Santosh Pradhan <spradhan at redhat.com> To: Indivar Nair <indivar.nair at techterra.in>, Gluster Users <gluster-users at gluster.org> Subject: Re: [Gluster-users] NFS ACL Support in Gluster 3.4 Message-ID: <538DB808.5050802 at redhat.com> Content-Type: text/plain; charset="iso-8859-1"; Format="flowed" I guess Gluster 3.5 has fixed the NFS-ACL issues and getfacl/setfacl works there. Regards, Santosh On 06/03/2014 05:10 PM, Indivar Nair wrote:> Hi All, > > I recently upgraded a Gluster 3.3.1 installation to Gluster 3.4. > It was a straight forward upgrade using Yum. > The OS is CentOS 6.3. > > The main purpose of the upgrade was to get ACL Support on NFS exports. > But it doesn't seem to be working. > > I mounted the gluster volume using the following options - > > mount -t nfs -o vers=3,mountproto=tcp,acl <gluster_server>:/volume /mnt > > The getfacl or setfacl commands does not work on any dir/files on this > mount. > > The plan is to re-export the NFS Mounts using Samba+CTDB. > NFS mounts seem to give better performance than Gluster Mounts. > > Am I missing something? > > Regards, > > > Indivar Nair > > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://supercolony.gluster.org/mailman/listinfo/gluster-users-------------- next part -------------- An HTML attachment was scrubbed... URL: < http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140603/04fb291d/attachment-0001.html>------------------------------ _______________________________________________ Gluster-users mailing list Gluster-users at gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users End of Gluster-users Digest, Vol 74, Issue 3 ******************************************** ** This email and any attachments may contain information that is confidential and/or privileged for the sole use of the intended recipient. Any use, review, disclosure, copying, distribution or reliance by others, and any forwarding of this email or its contents, without the express permission of the sender is strictly prohibited by law. If you are not the intended recipient, please contact the sender immediately, delete the e-mail and destroy all copies. ** -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140609/1e778d84/attachment.html>