search for: 4e35

Displaying 16 results from an estimated 16 matches for "4e35".

Did you mean: 435
2013 Aug 26
2
[JOBS] Ruby on Rails - Freelancing
...n email to rubyonrails-talk+unsubscribe-/JYPxA39Uh5TLH3MbocFF+G/Ez6ZCGd0@public.gmane.org To post to this group, send email to rubyonrails-talk-/JYPxA39Uh5TLH3MbocFF+G/Ez6ZCGd0@public.gmane.org To view this discussion on the web visit https://groups.google.com/d/msgid/rubyonrails-talk/ddab70b1-40c0-4e35-bfed-9247f473605e%40googlegroups.com. For more options, visit https://groups.google.com/groups/opt_out.
2013 Jul 05
0
Sign in and sign out
...n email to rubyonrails-talk+unsubscribe-/JYPxA39Uh5TLH3MbocFF+G/Ez6ZCGd0@public.gmane.org To post to this group, send email to rubyonrails-talk-/JYPxA39Uh5TLH3MbocFF+G/Ez6ZCGd0@public.gmane.org To view this discussion on the web visit https://groups.google.com/d/msgid/rubyonrails-talk/7e2a0b3d-675e-4e35-8d24-ee5a0ea683c3%40googlegroups.com. For more options, visit https://groups.google.com/groups/opt_out.
2018 Jan 02
2
"file changed as we read it" message during tar file creation on GlusterFS
...thank you very much for your support and sorry for the late. > Below you can find the output of ?gluster volume info tier2? command and the gluster software stack version: > > gluster volume info > > Volume Name: tier2 > Type: Distributed-Disperse > Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c > Status: Started > Snapshot Count: 0 > Number of Bricks: 6 x (4 + 2) = 36 > Transport-type: tcp > Bricks: > Brick1: s01-stg:/gluster/mnt1/brick > Brick2: s02-stg:/gluster/mnt1/brick > Brick3: s03-stg:/gluster/mnt1/brick > Brick4: s01-stg:/gluster/mnt2/b...
2017 Dec 29
0
"file changed as we read it" message during tar file creation on GlusterFS
Hi Nithya, thank you very much for your support and sorry for the late. Below you can find the output of ?gluster volume info tier2? command and the gluster software stack version: gluster volume info Volume Name: tier2 Type: Distributed-Disperse Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c Status: Started Snapshot Count: 0 Number of Bricks: 6 x (4 + 2) = 36 Transport-type: tcp Bricks: Brick1: s01-stg:/gluster/mnt1/brick Brick2: s02-stg:/gluster/mnt1/brick Brick3: s03-stg:/gluster/mnt1/brick Brick4: s01-stg:/gluster/mnt2/brick Brick5: s02-stg:/gluster/mnt2/brick Bric...
2018 Jan 02
0
"file changed as we read it" message during tar file creation on GlusterFS
...rt and sorry for the late. >> Below you can find the output of ?gluster volume info tier2? command >> and the gluster software stack version: >> >> gluster volume info >> >> Volume Name: tier2 >> Type: Distributed-Disperse >> Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c >> Status: Started >> Snapshot Count: 0 >> Number of Bricks: 6 x (4 + 2) = 36 >> Transport-type: tcp >> Bricks: >> Brick1: s01-stg:/gluster/mnt1/brick >> Brick2: s02-stg:/gluster/mnt1/brick >> Brick3: s03-stg:/gluster/mnt1/brick >...
2017 Dec 29
2
"file changed as we read it" message during tar file creation on GlusterFS
Hi Mauro, What version of Gluster are you running and what is your volume configuration? IIRC, this was seen because of mismatches in the ctime returned to the client. I don't think there were issues with the files but I will leave it to Ravi and Raghavendra to comment. Regards, Nithya On 29 December 2017 at 04:10, Mauro Tridici <mauro.tridici at cmcc.it> wrote: > > Hi All,
2018 Jan 02
1
"file changed as we read it" message during tar file creation on GlusterFS
...late. >>> Below you can find the output of ?gluster volume info tier2? command and the gluster software stack version: >>> >>> gluster volume info >>> >>> Volume Name: tier2 >>> Type: Distributed-Disperse >>> Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c >>> Status: Started >>> Snapshot Count: 0 >>> Number of Bricks: 6 x (4 + 2) = 36 >>> Transport-type: tcp >>> Bricks: >>> Brick1: s01-stg:/gluster/mnt1/brick >>> Brick2: s02-stg:/gluster/mnt1/brick >>> Brick3:...
2017 Sep 25
2
df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
Dear Gluster Users, I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options: [root at s01 tier2]# gluster volume info Volume Name: tier2 Type: Distributed-Disperse Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c Status: Started Snapshot Count: 0 Number of Bricks: 6 x (4 + 2) = 36 Transport-type: tcp Bricks: Brick1: s01-stg:/gluster/mnt1/brick Brick2: s02-stg:/gluster/mnt1/brick Brick3: s03-stg:/gluster/mnt1/brick Brick4: s01-stg:/gluster/mnt2/brick Brick5: s02-stg:/gluster/mnt2/brick Bric...
2017 Sep 26
0
df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
...ndpoint mount error on gluster client v.3.10.5 + core dump Dear Gluster Users, I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options: [root at s01 tier2]# gluster volume info Volume Name: tier2 Type: Distributed-Disperse Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c Status: Started Snapshot Count: 0 Number of Bricks: 6 x (4 + 2) = 36 Transport-type: tcp Bricks: Brick1: s01-stg:/gluster/mnt1/brick Brick2: s02-stg:/gluster/mnt1/brick Brick3: s03-stg:/gluster/mnt1/brick Brick4: s01-stg:/gluster/mnt2/brick Brick5: s02-stg:/gluster/mnt2/...
2017 Sep 26
2
df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
...5 + core dump > > Dear Gluster Users, > > I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options: > > [root at s01 tier2]# gluster volume info > > Volume Name: tier2 > Type: Distributed-Disperse > Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c > Status: Started > Snapshot Count: 0 > Number of Bricks: 6 x (4 + 2) = 36 > Transport-type: tcp > Bricks: > Brick1: s01-stg:/gluster/mnt1/brick > Brick2: s02-stg:/gluster/mnt1/brick > Brick3: s03-stg:/gluster/mnt1/brick > Brick4: s01-stg:/gluster/mnt2/b...
2017 Sep 26
2
df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
...5 + core dump > > Dear Gluster Users, > > I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options: > > [root at s01 tier2]# gluster volume info > > Volume Name: tier2 > Type: Distributed-Disperse > Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c > Status: Started > Snapshot Count: 0 > Number of Bricks: 6 x (4 + 2) = 36 > Transport-type: tcp > Bricks: > Brick1: s01-stg:/gluster/mnt1/brick > Brick2: s02-stg:/gluster/mnt1/brick > Brick3: s03-stg:/gluster/mnt1/brick > Brick4: s01-stg:/gluster/mnt2/b...
2017 Sep 26
0
df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
...ndpoint mount error on gluster client v.3.10.5 + core dump Dear Gluster Users, I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options: [root at s01 tier2]# gluster volume info Volume Name: tier2 Type: Distributed-Disperse Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c Status: Started Snapshot Count: 0 Number of Bricks: 6 x (4 + 2) = 36 Transport-type: tcp Bricks: Brick1: s01-stg:/gluster/mnt1/brick Brick2: s02-stg:/gluster/mnt1/brick Brick3: s03-stg:/gluster/mnt1/brick Brick4: s01-stg:/gluster/mnt2/brick Brick5: s02-stg:/gluster/mnt2/...
2017 Sep 27
0
df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
...uster Users, >> >> I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options: >> >> [root at s01 tier2]# gluster volume info >> >> Volume Name: tier2 >> Type: Distributed-Disperse >> Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c >> Status: Started >> Snapshot Count: 0 >> Number of Bricks: 6 x (4 + 2) = 36 >> Transport-type: tcp >> Bricks: >> Brick1: s01-stg:/gluster/mnt1/brick >> Brick2: s02-stg:/gluster/mnt1/brick >> Brick3: s03-stg:/gluster/mnt1/brick >...
2017 Oct 26
0
not healing one file
Hey Richard, Could you share the following informations please? 1. gluster volume info <volname> 2. getfattr output of that file from all the bricks getfattr -d -e hex -m . <brickpath/filepath> 3. glustershd & glfsheal logs Regards, Karthik On Thu, Oct 26, 2017 at 10:21 AM, Amar Tumballi <atumball at redhat.com> wrote: > On a side note, try recently released health
2017 Oct 26
3
not healing one file
On a side note, try recently released health report tool, and see if it does diagnose any issues in setup. Currently you may have to run it in all the three machines. On 26-Oct-2017 6:50 AM, "Amar Tumballi" <atumball at redhat.com> wrote: > Thanks for this report. This week many of the developers are at Gluster > Summit in Prague, will be checking this and respond next
2017 Oct 26
2
not healing one file
...E4137F7AD15038E (8fe34aaf-dc9d-4f02-a681-0588c244f7d0) on home-client-2 [2017-10-25 10:14:19.537606] W [MSGID: 108015] [afr-self-heal-entry.c:56:afr_selfheal_entry_delete] 0-home-replicate-0: expunging file a3f5a769-8859-48e3-96ca-60a988eb9358/BF04731FA5E6649A4FFE6450F9BAA160A476BC07 (77217517-17cc-4e35-977b-a58a3b4d215c) on home-client-2 [2017-10-25 10:14:19.556844] W [MSGID: 108015] [afr-self-heal-entry.c:56:afr_selfheal_entry_delete] 0-home-replicate-0: expunging file a3f5a769-8859-48e3-96ca-60a988eb9358/24AF654EF4D076B1E152C1DFDB156E975CB89C2E (7c6e44ab-92ec-4dd4-93a2-04e339b97bf2) on home-cli...