search for: starvate

Displaying 20 results from an estimated 265 matches for "starvate".

Did you mean: stardate
2005 Jun 15
1
Starvation
Hi, I'v included speex-lib into my http://www.visit.se/~pointless IP-Sound streaming app. When using speex the sound is being garbled for short intervals, approx every 20:th second. Sounds like buffer starvation..... but in between it sounds 100% ok. As I am using other "codecs" with success there might be something that I have missed in the implementation of speex. Using
2007 Apr 06
1
The best way to protect against starvation?
Hello, If an ordinary user runs: -- snip -- cat > starv.c <<EOF main(){ char *point; while(1) { point = ( char * ) malloc(10000); }} EOF cc starv.c while true do ./a.out & done -- snip -- This will fast starv the operating system (FreeBSD 6.2). I have tried to limit the number of processes and the amount of memmory consumed (in login.conf). There is also a file /etc/malloc.conf
2008 Jul 08
6
[Bug 16640] New: NV18/DVI bandwidth starvation issue
http://bugs.freedesktop.org/show_bug.cgi?id=16640 Summary: NV18/DVI bandwidth starvation issue Product: xorg Version: git Platform: Other OS/Version: All Status: NEW Severity: normal Priority: high Component: Driver/nouveau AssignedTo: nouveau at lists.freedesktop.org ReportedBy: d.okias at
1998 Feb 22
0
resource starvation against passwd(1)
Standard apology if old... This demonstrates a resource starvation attack on the setuid root passwd(1) program. In the case I tested it was the Red Hat Linux passwd-0.50-7 program without shadowing. #include <stdio.h> #include <sys/time.h> #include <stdlib.h> #include <unistd.h> #include <sys/resource.h> main () { struct rlimit rl, *rlp; rlp=&rl;
2007 Oct 09
0
DRb - queue starvation?
I think I''ve run into some kind of queue starvation bug w/ DRb - it normally takes minutes per item to process items in the queue, and normally the queue has a bit of a backlog. But when it''s empty (nothing coming in), it seems to stop checking for new work. Has anyone else run into this w/ DRb? Any suggestions for debugging this? Thanks, m
2012 Sep 18
0
Rails Thread Starvation
Hi! I was hoping to receive some help on a difficult matter - I have tried a lot already. http://stackoverflow.com/questions/12468781/rails-thread-starvation That''s the link, I hope it is ok to link there. Thanks! Carl -- You received this message because you are subscribed to the Google Groups "Ruby on Rails: Talk" group. To post to this group, send email to
2003 Oct 18
1
MORE ON : [htb] strange problems !?
hello again, I got some just preliminary results .... hope someone can explain them to me... As I already told I have the following config : egress { class (30kbps) {sfq}; class (50kbps) {sfq}; class (80kbps) {sfq}; class (30kbps) {sfq}; class (50kbps) {sfq}; ...hundreds like this... class (10kbps,default) {sfq}; } What I got is traffic starvation very often for a period of ~30 sec.. as
2014 Feb 27
1
[PATCH RFC v5 4/8] pvqspinlock, x86: Allow unfair spinlock in a real PV environment
On 26/02/14 15:14, Waiman Long wrote: > Locking is always an issue in a virtualized environment as the virtual > CPU that is waiting on a lock may get scheduled out and hence block > any progress in lock acquisition even when the lock has been freed. > > One solution to this problem is to allow unfair lock in a > para-virtualized environment. In this case, a new lock acquirer
2014 Feb 27
1
[PATCH RFC v5 4/8] pvqspinlock, x86: Allow unfair spinlock in a real PV environment
On 26/02/14 15:14, Waiman Long wrote: > Locking is always an issue in a virtualized environment as the virtual > CPU that is waiting on a lock may get scheduled out and hence block > any progress in lock acquisition even when the lock has been freed. > > One solution to this problem is to allow unfair lock in a > para-virtualized environment. In this case, a new lock acquirer
2006 Apr 14
3
Shaping per IP in PPPoE borrowing or sharing Uplink or Downlink
helo again. I think this question i am asking is worth: we know that pppoe-server creates a pppX device on each connection done to it. So, when i have to shape, i have to shape each pppX connection device on itself alone. What i know is that the borrowing method on one device by itself, e.g. ppp0, alone using HTB or the like. this means that i have to create for another device, e.g. ppp1,
2014 Jul 01
2
[RFC PATCH v2] Implement Batched (group) ticket lock
On 07/01/2014 01:35 PM, Peter Zijlstra wrote: > On Sat, Jun 28, 2014 at 02:47:04PM +0530, Raghavendra K T wrote: >> In virtualized environment there are mainly three problems >> related to spinlocks that affects performance. >> 1. LHP (lock holder preemption) >> 2. Lock Waiter Preemption (LWP) >> 3. Starvation/fairness >> >> Though Ticketlocks solve
2014 Jul 01
2
[RFC PATCH v2] Implement Batched (group) ticket lock
On 07/01/2014 01:35 PM, Peter Zijlstra wrote: > On Sat, Jun 28, 2014 at 02:47:04PM +0530, Raghavendra K T wrote: >> In virtualized environment there are mainly three problems >> related to spinlocks that affects performance. >> 1. LHP (lock holder preemption) >> 2. Lock Waiter Preemption (LWP) >> 3. Starvation/fairness >> >> Though Ticketlocks solve
2006 Jun 21
1
Linux Qos : PRIO qdisc works
Hi, all I am interested in Linux related QoS, especially PRIO queue discipline (qdisc). I test some experiments about PRIO qdisc. Host A ------ Router A --- Router B --- Host C | 1Gbps | Host B ---------+ +-------- Host D flow 1 : Host A -> Host C (TCP) flow 2 : Host B -> Host D (UDP) priority 1 queue (pfifo, size 1000 packets ) : flow 1 priority 2
2014 Jun 28
2
[RFC PATCH v2] Implement Batched (group) ticket lock
In virtualized environment there are mainly three problems related to spinlocks that affects performance. 1. LHP (lock holder preemption) 2. Lock Waiter Preemption (LWP) 3. Starvation/fairness Though Ticketlocks solve fairness problem it worsens LWP, LHP problems. Though pv-ticketlocks tried to address these problems we can further improve at the cost of relaxed fairness. The following patch
2014 Jun 28
2
[RFC PATCH v2] Implement Batched (group) ticket lock
In virtualized environment there are mainly three problems related to spinlocks that affects performance. 1. LHP (lock holder preemption) 2. Lock Waiter Preemption (LWP) 3. Starvation/fairness Though Ticketlocks solve fairness problem it worsens LWP, LHP problems. Though pv-ticketlocks tried to address these problems we can further improve at the cost of relaxed fairness. The following patch
2003 Sep 25
1
FreeBSD Security Advisory FreeBSD-SA-03:14.arp [REVISED]
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 ============================================================================= FreeBSD-SA-03:14.arp Security Advisory The FreeBSD Project Topic: denial of service due to ARP resource starvation Category: core Module: sys Announced:
2003 Oct 17
3
[htb] strange problems !?
hi, I have strange problems with HTB and several hundred classes flat structure i.e. root |--50kbps |--30kbps |--50kbps |--80kbps |--100kbps .... several hundred classes like this Ceil is the same as rate. The machine get no more than 2-3% average cpu(2.4Ghz pentium). What happens is that from time to time the traffic got "stalled". I tried numerous things to solve
2015 May 27
1
FREAK/Logjam, and SSL protocols to use
Quoting Gedalya <gedalya at gedalya.net>: > On 05/27/2015 09:55 AM, Rick Romero wrote: >> Quoting Gedalya <gedalya at gedalya.net>: >> >>> On 05/26/2015 10:37 AM, Ron Leach wrote: >>>> https://weakdh.org/sysadmin.html >>>> >>>> includes altering DH parameters length to 2048, and re-specifying the >>>> allowable
2019 May 14
3
[PATCH v2 1/8] vsock/virtio: limit the memory used per-socket
On 2019/5/14 ??1:23, Stefano Garzarella wrote: > On Mon, May 13, 2019 at 05:58:53PM +0800, Jason Wang wrote: >> On 2019/5/10 ??8:58, Stefano Garzarella wrote: >>> Since virtio-vsock was introduced, the buffers filled by the host >>> and pushed to the guest using the vring, are directly queued in >>> a per-socket list avoiding to copy it. >>> These
2019 May 14
3
[PATCH v2 1/8] vsock/virtio: limit the memory used per-socket
On 2019/5/14 ??1:23, Stefano Garzarella wrote: > On Mon, May 13, 2019 at 05:58:53PM +0800, Jason Wang wrote: >> On 2019/5/10 ??8:58, Stefano Garzarella wrote: >>> Since virtio-vsock was introduced, the buffers filled by the host >>> and pushed to the guest using the vring, are directly queued in >>> a per-socket list avoiding to copy it. >>> These