search for: b048

Displaying 5 results from an estimated 5 matches for "b048".

Did you mean: 2048
2014 Sep 15
0
Re: cgroups inside LXC containers losts memory limits after some time
HI all >After unpredictable time passed (1-5 day ?), cgroups inside LXC >magicaly removed. virsh dumpxml config look like this: <domain type='lxc' id='3566'> <name>puppet</name> <uuid>6d49b280-5686-4e3c-b048-1b5d362fb137</uuid> <memory unit='KiB'>8388608</memory> <currentMemory unit='KiB'>8388608</currentMemory> <memtune> <hard_limit unit='KiB'>8388608</hard_limit> <soft_limit unit='KiB'>8388608</soft...
2014 Sep 15
2
cgroups inside LXC containers losts memory limits after some time
Hi all I have CentOS Linux release 7.0.1406, libvirt 1.2.7 installed. Just after create and start inside LXC container present cgroups. Example for memory: [root@ce7-t1 /]# ls -la /sys/fs/cgroup/memory/ total 0 drwxr-xr-x 2 root root 0 Sep 15 17:14 . drwxr-xr-x 12 root root 280 Sep 15 17:14 .. -rw-r--r-- 1 root root 0 Sep 15 17:14 cgroup.clone_children --w--w--w- 1 root root 0 Sep 15
2017 Oct 26
0
not healing one file
Hey Richard, Could you share the following informations please? 1. gluster volume info <volname> 2. getfattr output of that file from all the bricks getfattr -d -e hex -m . <brickpath/filepath> 3. glustershd & glfsheal logs Regards, Karthik On Thu, Oct 26, 2017 at 10:21 AM, Amar Tumballi <atumball at redhat.com> wrote: > On a side note, try recently released health
2017 Oct 26
3
not healing one file
On a side note, try recently released health report tool, and see if it does diagnose any issues in setup. Currently you may have to run it in all the three machines. On 26-Oct-2017 6:50 AM, "Amar Tumballi" <atumball at redhat.com> wrote: > Thanks for this report. This week many of the developers are at Gluster > Summit in Prague, will be checking this and respond next
2017 Oct 26
2
not healing one file
...n.c:1327:afr_log_selfheal] 0-home-replicate-0: Completed metadata selfheal on 2ecba13f-ffbd-4d52-a61b-f91694e178a1. sources=0 [2] sinks=1 [2017-10-25 10:40:23.202620] I [MSGID: 108026] [afr-self-heal-common.c:1327:afr_log_selfheal] 0-home-replicate-0: Completed data selfheal on 524d2581-252e-4460-b048-2a0e4fcf2cca. sources=0 [2] sinks=1 [2017-10-25 10:40:23.204154] I [MSGID: 108026] [afr-self-heal-metadata.c:52:__afr_selfheal_metadata_do] 0-home-replicate-0: performing metadata selfheal on 524d2581-252e-4460-b048-2a0e4fcf2cca [2017-10-25 10:40:23.208652] I [MSGID: 108026] [afr-self-heal-common...