Displaying 16 results from an estimated 16 matches for "4e35".
Did you mean:
435
2013 Aug 26
2
[JOBS] Ruby on Rails - Freelancing
...n email to rubyonrails-talk+unsubscribe-/JYPxA39Uh5TLH3MbocFF+G/Ez6ZCGd0@public.gmane.org
To post to this group, send email to rubyonrails-talk-/JYPxA39Uh5TLH3MbocFF+G/Ez6ZCGd0@public.gmane.org
To view this discussion on the web visit https://groups.google.com/d/msgid/rubyonrails-talk/ddab70b1-40c0-4e35-bfed-9247f473605e%40googlegroups.com.
For more options, visit https://groups.google.com/groups/opt_out.
2013 Jul 05
0
Sign in and sign out
...n email to rubyonrails-talk+unsubscribe-/JYPxA39Uh5TLH3MbocFF+G/Ez6ZCGd0@public.gmane.org
To post to this group, send email to rubyonrails-talk-/JYPxA39Uh5TLH3MbocFF+G/Ez6ZCGd0@public.gmane.org
To view this discussion on the web visit https://groups.google.com/d/msgid/rubyonrails-talk/7e2a0b3d-675e-4e35-8d24-ee5a0ea683c3%40googlegroups.com.
For more options, visit https://groups.google.com/groups/opt_out.
2018 Jan 02
2
"file changed as we read it" message during tar file creation on GlusterFS
...thank you very much for your support and sorry for the late.
> Below you can find the output of ?gluster volume info tier2? command and the gluster software stack version:
>
> gluster volume info
>
> Volume Name: tier2
> Type: Distributed-Disperse
> Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 6 x (4 + 2) = 36
> Transport-type: tcp
> Bricks:
> Brick1: s01-stg:/gluster/mnt1/brick
> Brick2: s02-stg:/gluster/mnt1/brick
> Brick3: s03-stg:/gluster/mnt1/brick
> Brick4: s01-stg:/gluster/mnt2/b...
2017 Dec 29
0
"file changed as we read it" message during tar file creation on GlusterFS
Hi Nithya,
thank you very much for your support and sorry for the late.
Below you can find the output of ?gluster volume info tier2? command and the gluster software stack version:
gluster volume info
Volume Name: tier2
Type: Distributed-Disperse
Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
Status: Started
Snapshot Count: 0
Number of Bricks: 6 x (4 + 2) = 36
Transport-type: tcp
Bricks:
Brick1: s01-stg:/gluster/mnt1/brick
Brick2: s02-stg:/gluster/mnt1/brick
Brick3: s03-stg:/gluster/mnt1/brick
Brick4: s01-stg:/gluster/mnt2/brick
Brick5: s02-stg:/gluster/mnt2/brick
Bric...
2018 Jan 02
0
"file changed as we read it" message during tar file creation on GlusterFS
...rt and sorry for the late.
>> Below you can find the output of ?gluster volume info tier2? command
>> and the gluster software stack version:
>>
>> gluster volume info
>>
>> Volume Name: tier2
>> Type: Distributed-Disperse
>> Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
>> Status: Started
>> Snapshot Count: 0
>> Number of Bricks: 6 x (4 + 2) = 36
>> Transport-type: tcp
>> Bricks:
>> Brick1: s01-stg:/gluster/mnt1/brick
>> Brick2: s02-stg:/gluster/mnt1/brick
>> Brick3: s03-stg:/gluster/mnt1/brick
>...
2017 Dec 29
2
"file changed as we read it" message during tar file creation on GlusterFS
Hi Mauro,
What version of Gluster are you running and what is your volume
configuration?
IIRC, this was seen because of mismatches in the ctime returned to the
client. I don't think there were issues with the files but I will leave it
to Ravi and Raghavendra to comment.
Regards,
Nithya
On 29 December 2017 at 04:10, Mauro Tridici <mauro.tridici at cmcc.it> wrote:
>
> Hi All,
2018 Jan 02
1
"file changed as we read it" message during tar file creation on GlusterFS
...late.
>>> Below you can find the output of ?gluster volume info tier2? command and the gluster software stack version:
>>>
>>> gluster volume info
>>>
>>> Volume Name: tier2
>>> Type: Distributed-Disperse
>>> Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
>>> Status: Started
>>> Snapshot Count: 0
>>> Number of Bricks: 6 x (4 + 2) = 36
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: s01-stg:/gluster/mnt1/brick
>>> Brick2: s02-stg:/gluster/mnt1/brick
>>> Brick3:...
2017 Sep 25
2
df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
Dear Gluster Users,
I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options:
[root at s01 tier2]# gluster volume info
Volume Name: tier2
Type: Distributed-Disperse
Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
Status: Started
Snapshot Count: 0
Number of Bricks: 6 x (4 + 2) = 36
Transport-type: tcp
Bricks:
Brick1: s01-stg:/gluster/mnt1/brick
Brick2: s02-stg:/gluster/mnt1/brick
Brick3: s03-stg:/gluster/mnt1/brick
Brick4: s01-stg:/gluster/mnt2/brick
Brick5: s02-stg:/gluster/mnt2/brick
Bric...
2017 Sep 26
0
df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
...ndpoint mount error on gluster client v.3.10.5 + core dump
Dear Gluster Users,
I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options:
[root at s01 tier2]# gluster volume info
Volume Name: tier2
Type: Distributed-Disperse
Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
Status: Started
Snapshot Count: 0
Number of Bricks: 6 x (4 + 2) = 36
Transport-type: tcp
Bricks:
Brick1: s01-stg:/gluster/mnt1/brick
Brick2: s02-stg:/gluster/mnt1/brick
Brick3: s03-stg:/gluster/mnt1/brick
Brick4: s01-stg:/gluster/mnt2/brick
Brick5: s02-stg:/gluster/mnt2/...
2017 Sep 26
2
df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
...5 + core dump
>
> Dear Gluster Users,
>
> I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options:
>
> [root at s01 tier2]# gluster volume info
>
> Volume Name: tier2
> Type: Distributed-Disperse
> Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 6 x (4 + 2) = 36
> Transport-type: tcp
> Bricks:
> Brick1: s01-stg:/gluster/mnt1/brick
> Brick2: s02-stg:/gluster/mnt1/brick
> Brick3: s03-stg:/gluster/mnt1/brick
> Brick4: s01-stg:/gluster/mnt2/b...
2017 Sep 26
2
df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
...5 + core dump
>
> Dear Gluster Users,
>
> I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options:
>
> [root at s01 tier2]# gluster volume info
>
> Volume Name: tier2
> Type: Distributed-Disperse
> Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 6 x (4 + 2) = 36
> Transport-type: tcp
> Bricks:
> Brick1: s01-stg:/gluster/mnt1/brick
> Brick2: s02-stg:/gluster/mnt1/brick
> Brick3: s03-stg:/gluster/mnt1/brick
> Brick4: s01-stg:/gluster/mnt2/b...
2017 Sep 26
0
df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
...ndpoint mount error on gluster client v.3.10.5 + core dump
Dear Gluster Users,
I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options:
[root at s01 tier2]# gluster volume info
Volume Name: tier2
Type: Distributed-Disperse
Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
Status: Started
Snapshot Count: 0
Number of Bricks: 6 x (4 + 2) = 36
Transport-type: tcp
Bricks:
Brick1: s01-stg:/gluster/mnt1/brick
Brick2: s02-stg:/gluster/mnt1/brick
Brick3: s03-stg:/gluster/mnt1/brick
Brick4: s01-stg:/gluster/mnt2/brick
Brick5: s02-stg:/gluster/mnt2/...
2017 Sep 27
0
df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
...uster Users,
>>
>> I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options:
>>
>> [root at s01 tier2]# gluster volume info
>>
>> Volume Name: tier2
>> Type: Distributed-Disperse
>> Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
>> Status: Started
>> Snapshot Count: 0
>> Number of Bricks: 6 x (4 + 2) = 36
>> Transport-type: tcp
>> Bricks:
>> Brick1: s01-stg:/gluster/mnt1/brick
>> Brick2: s02-stg:/gluster/mnt1/brick
>> Brick3: s03-stg:/gluster/mnt1/brick
>...
2017 Oct 26
0
not healing one file
Hey Richard,
Could you share the following informations please?
1. gluster volume info <volname>
2. getfattr output of that file from all the bricks
getfattr -d -e hex -m . <brickpath/filepath>
3. glustershd & glfsheal logs
Regards,
Karthik
On Thu, Oct 26, 2017 at 10:21 AM, Amar Tumballi <atumball at redhat.com> wrote:
> On a side note, try recently released health
2017 Oct 26
3
not healing one file
On a side note, try recently released health report tool, and see if it
does diagnose any issues in setup. Currently you may have to run it in all
the three machines.
On 26-Oct-2017 6:50 AM, "Amar Tumballi" <atumball at redhat.com> wrote:
> Thanks for this report. This week many of the developers are at Gluster
> Summit in Prague, will be checking this and respond next
2017 Oct 26
2
not healing one file
...E4137F7AD15038E (8fe34aaf-dc9d-4f02-a681-0588c244f7d0) on home-client-2
[2017-10-25 10:14:19.537606] W [MSGID: 108015] [afr-self-heal-entry.c:56:afr_selfheal_entry_delete] 0-home-replicate-0: expunging file a3f5a769-8859-48e3-96ca-60a988eb9358/BF04731FA5E6649A4FFE6450F9BAA160A476BC07 (77217517-17cc-4e35-977b-a58a3b4d215c) on home-client-2
[2017-10-25 10:14:19.556844] W [MSGID: 108015] [afr-self-heal-entry.c:56:afr_selfheal_entry_delete] 0-home-replicate-0: expunging file a3f5a769-8859-48e3-96ca-60a988eb9358/24AF654EF4D076B1E152C1DFDB156E975CB89C2E (7c6e44ab-92ec-4dd4-93a2-04e339b97bf2) on home-cli...