search for: netems

Displaying 20 results from an estimated 82 matches for "netems".

Did you mean: nelems
2006 Apr 26
5
how to change classful netem loss probability?
Hi, I am using netem to add loss and then adding another qdisc within netem according to the wiki. Then i want to change the netem drop probability without having to delete the qdisc and recreate it. I try it but I get invalid argument: thorium-ini hedpe # tc qdisc add dev ath0 root handle 1:0 netem drop 1% thorium-ini hedpe # tc qdisc add dev ath0 parent 1:1 handle 10: xcp capacity 54Mbit
2004 Aug 31
1
netem usage example
I''m trying to setup a netem delay with no luck (using iproute2-2.6.8, compilation broke during arpd compile, so I use the tc binary in the tc/ subdir, there''s also a q_netem.so there). kernel is 2.6.8.1, compile with CPU cycle counter as time reference. I was using sch_delay of 2.6.7 happily with something like: tc qdisc add dev eth0 root 1: delay latency 1ms rate 35M now I use:
2005 Jul 12
0
Teql and NetEm can''t work together
Thanks in advance! Summary: when I load netem and teql together, teql doesn''t work correctly. (If I load teql only, everything is fine) I loaded both netem and teql. Netem is associated with eth0, and teql is associated with both eth0 and eth1. But traffic only goes out of eth1. Attached are the commands that I used to configure teql and netem (on machine 1), and commands to
2005 May 24
3
four tc filter and netem questions
The following (occuring on debian/testing with kernel-image-2.6.8-2-386 version 2.6.8-13 and iproute version 20041019-3) confuses me: # tc qdisc add dev eth0 root handle 1: prio # tc filter add dev eth0 parent 1: proto ip pref 1 handle 1 fw classid 1:2 # tc filter ls dev eth0 filter parent 1: protocol ip pref 1 fw filter parent 1: protocol ip pref 1 fw handle 0x1 classid 1:2 # tc filter del dev
2007 Mar 30
1
Please Help: applying multiple different delays with netem
I''m trying to use tc and netem to delay packets from several different machines as they exit via eth0. Assume two source IPs, 10.0.0.122 and 10.0.0.133. I''d like to delay packets from the first one by 200ms, and packets from the second one by 300 ms. Any other traffic should be sent out normally. Here''s what I tried: # make three classes, 1:1, 1:2, and 1:3: tc qdisc add
2006 Jul 17
1
How to add multiple filters and netem rules on a single interface?
Hi! We want to run TCP streams to several port numbers through one interface, each with a different delay set by Netem. E.g. TCP streams to port 80 could have 50ms delay, while TCP streams to port 81 could have 100 ms delay and so on. We have tried to solve this by using a combination of tc filter and netem rules, but we can''t get it quite right. We are considering one class per port,
2005 Jan 27
2
netem bug?
Hi all, I''m running some tests with netem and I noticed some strange behaviour that looks like a bug: I''m pinging another machine and adding delay with netem. When I tell netem to give me a 10ms delay, it works fine. The problem is that when I ask for a 11ms delay, it gives me 20ms! It happens for any value between 11ms an 20ms, and it repeats for values over 20ms, now
2006 Apr 16
9
how to do probabilistic packet loss in kernel?
Hi, I am using iproute2 to setup fowarding, adding routes like "ip route add 192.168.1.3 via 192.168.1.2" I was wondering where in the kernel I can insert probabilistic packet loss only for forwarded packets? So that for instance I can drop 5% of all forwarded packets? I don''t need help with the actual code, just need help finding where to insert this code :) Thanks! George
2005 Mar 30
5
netem with prio hangs on duplicate
hi i tried the example given on the examples page to duplicate selected traffic like tc qdisc add dev eth0 root handle 1: prio tc qdisc add dev eth0 parent 1:3 handle 3: netem duplicate 40% tc filter add dev eth0 protocol ip parent 1:0 prio 3 u32 match ip dst 11.0.2.2 flowid 1:3 when i ping from 11.0.2.2 to this interface my machine hangs. the same thing works for drop or delay. i would
2007 Apr 24
0
OSPF with Netem
Hi all, I am currently trying to emulate a satellite link, via Netem, on a testbed which is OSPF-enabled. I''d like to set up a Netem box between two routers. Since all routing between routers is dynamic, I''m wondering how to set up OSPF on my Netem box? Could someone indicate me if it is feasible and give me some guidelines to possibly do so? Thanks in advance, Vincent.
2007 Apr 02
1
Please Help: Can''t access bands > 10 on prio qdisc
Hi, I''m trying to set up 15 different delay intervals for packets leaving on an interface, using netems hanging off of a 16-band prio. I''m having trouble adding anything to bands higher than 10. Here''s what I tried: tc qdisc add dev eth0 root handle 1: prio bands 16 \ priomap 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 I want all default traffic to go to the highest p...
2005 Jun 22
0
Help: NIST Net or NetEm ?
Hi, I want to study the multi-path TCP behavior (single TCP connection over multi-paths), and I am setting up the test bed. I have two machines, each with two network cards. I have directly connected two eth0, and I am trying to figure out how connect two eth1 to control the properties of this connection, so that I can study the TCP on multiple different paths (different in terms of latency,
2007 Apr 23
1
Multiple bands with equal priority ?
I''m trying to build a wan latency test environment, where packets from different "remote" locations get delayed by different amounts of time, depending on which remote location we''re pretending they are from. Currently, I''m doing this using the ''prio'' qdisc to obtain multiple bands, and hanging a different netem qdisc off each of the branches
2004 Oct 08
2
Delay packets by 50ms
Hi all, I am trying to solve a tiny problem that is trivial to solve using dummynet (FreeBSD). I just want to add a delay of 50ms to each outgoing packet from an interface. This is to simulate a large pool of multiple modem users so I also need to add b/w limits etc (which seems to be easy to do). From the mailing list I could fine 2 qdiscs that can simulate latency : "delay" &
2005 May 13
1
Qdisc requeue should be void?
There is an design problem with the qdisc interface that causes qlen related bugs in netem, tbf, and other qdisc''s that peek at the top of the queue. The problem is that requeue needs to be called from the dequeue function but requeue can fail. If requeue fails, then the calling qdisc can not properly handle the error. If it returns NULL, then the parent''s expectation about qlen
2005 Oct 11
1
How to do network emulation on incoming traffic?
I''m trying to simulate a satellite link to a Linux server to test application performance. I haven''t used any of the tc stuff before, but I blandly assured people it would be "easy" to set up a simulated long thin pipe on a spare network interface. However, now that I''m exploring, it''s proving quite difficult. Let me start with the general question
2006 Sep 11
1
simple question
Hello everybody, I''m trying to set up netem on a SuSE box (kernel 2.6.16), but I''m having some weird problems, considering that on my Gentoo laptop (kernel 2.6.17) it worked fine. If I write: # tc qdisc add dev eth0 root netem delay 20ms loss 20% and then try pinging another machine over eth0, I get a 24ms delay but no packet loss (while on my laptop I get 21ms and a correct
2006 Apr 27
1
Unsubscribe
> Send LARTC mailing list submissions to > lartc@mailman.ds9a.nl > > To subscribe or unsubscribe via the World Wide Web, visit > http://mailman.ds9a.nl/cgi-bin/mailman/listinfo/lartc > or, via email, send a message with subject or body ''help'' to > lartc-request@mailman.ds9a.nl > > You can reach the person managing the list at >
2017 Sep 28
9
[PATCH net-next] vhost_net: do not stall on zerocopy depletion
From: Willem de Bruijn <willemb at google.com> Vhost-net has a hard limit on the number of zerocopy skbs in flight. When reached, transmission stalls. Stalls cause latency, as well as head-of-line blocking of other flows that do not use zerocopy. Instead of stalling, revert to copy-based transmission. Tested by sending two udp flows from guest to host, one with payload of
2017 Sep 28
9
[PATCH net-next] vhost_net: do not stall on zerocopy depletion
From: Willem de Bruijn <willemb at google.com> Vhost-net has a hard limit on the number of zerocopy skbs in flight. When reached, transmission stalls. Stalls cause latency, as well as head-of-line blocking of other flows that do not use zerocopy. Instead of stalling, revert to copy-based transmission. Tested by sending two udp flows from guest to host, one with payload of