similar to: Hmmm... Weird results with HTB in kernel 2.6.11

Displaying 20 results from an estimated 300 matches similar to: "Hmmm... Weird results with HTB in kernel 2.6.11"

2006 Mar 08
1
negative token/ctokens
In this simple htb setup: # tc -s -d class ls dev eth0 class htb 1:1 root rate 300000bit ceil 300000bit burst 1749b/8 mpu 0b overhead 0b cburst 1749b/8 mpu 0b overhead 0b level 7 Sent 13171835 bytes 13169 pkt (dropped 0, overlimits 0 requeues 0) rate 45848bit 10pps backlog 0b 0p requeues 0 lended: 5272 borrowed: 0 giants: 0 tokens: -84429 ctokens: -84429 class htb 1:2 parent 1:1 prio 0
2005 Nov 14
6
Transfer speed exceeding the ceil
What''s going on here? I''m spewing UDP traffic at this thing, and it is exceeding the ceil. Anyone know how to fix this? class htb 1:613 parent 1:5 leaf 613: prio 6 quantum 2560 rate 20480bit ceil 103360bit burst 15Kb/8 mpu 0b overhead 0b cburst 1728b/8 mpu 0b overhead 0b level 0 Sent 16591370 bytes 4159 pkt (dropped 39449, overlimits 0 requeues 0) rate 412384bit 6pps backlog 0b
2006 May 19
25
iptables CLASSIFY and MARK not working?
I have to match my packets based on MAC address, which I cannot do in the POSTROUTING chain, so I do it in PREROUTING using MARK. Then, I match on the MARK in the POSTROUTING chain to do a CLASSIFY. But this does not seem to work: wireless-r1 bwlimit # iptables -L -v -n -t mangle Chain PREROUTING (policy ACCEPT 3353K packets, 941M bytes) pkts bytes target prot opt in out source
2006 Feb 23
1
1k: 1000 or 1024?
The docs[1][2] suggest it''s 1024, but tc says something else: # tc qdisc add dev eth0 root tbf rate 1kbps latency 50ms burst 1500 # tc -s qdisc ls dev eth0 qdisc tbf 8009: rate 8000bit burst 1499b lat 48.8ms ^^^^^^^ Sent 0 bytes 0 pkt (dropped 0, overlimits 0 requeues 0) rate 0bit 0pps backlog 0b 0p requeues 0 If 1k were 1024, then I would have 8192bit above.
2007 Aug 29
11
tc not matching
Dear all, I''m having real problems getting tc to do anything useful at all. I''m also under pressure to get this fixed before the students start arriving later this month (I work in a university). In short, I want each IP address to be hard limited to 128kbit down, 64kbit up, never to be allowed more bandwidth than this. It is also important that the latency remains
2005 Nov 15
1
UDP transfer speed exceeding the ceil by about 4x
A bit more detail. I have the following htb classes set up... class htb 1:356 parent 1:4 leaf 356: prio 4 quantum 1600 rate 12800bit ceil 51680bit burst 15Kb/8 mpu 0b overhead 0b cburst 1663b/8 mpu 0b overhead 0b level 0 class htb 1:357 parent 1:4 leaf 357: prio 4 quantum 1600 rate 12800bit ceil 51680bit burst 15Kb/8 mpu 0b overhead 0b cburst 1663b/8 mpu 0b overhead 0b level 0 class htb 1:2 root
2007 May 19
1
Re: LARTC Digest, Vol 27, Issue 26
Hi folks...!!! I need to generate qdisc statistics to show my 4 class (10, 20, 30, 40), i`ve all working with HTB and so on, but i need to graph this results e.gwith RRDTOOL. I found a script made in perl, that can to graph my 4 class, but i need to know which IP address on my LAN are using the bandwidth too, in other hand i need to classify the traffic by IP to show. This is an out of my
2006 Feb 24
4
why isn''t 1:1 getting the traffic? [filter question]
With the below script, whenever I ping 10.0.16.10 (which matches the only filter I have), traffic still get''s sent to the default 1:2 class instead of 1:1 and I don''t know why... Any hints? (kernel 2.6.12, iproute2-2.6.15) tc qdisc del dev eth0 root > /dev/null 2>&1 tc qdisc add dev eth0 handle 1: root htb default 2 tc class add dev eth0 classid 1:1 parent 1: htb rate
2012 Nov 20
3
Simple Traffic Shaping Problem
Hello, I have a problem with simple traffic shapping in shorewall, my current configuration is: zones vlan10 ipv4 # interfaces vlan10 vlan10 detect tcpflags,routeback shorewall.conf TC_ENABLED=Simple tcinterfaces vlan10 Internal 1mbit:50kb shorewall show tc Device vlan10: qdisc prio 5: root refcnt 2 bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
2007 Dec 15
1
hfsc and bps
Hi! Do you know somthing about hfsc and bps? There''s no output for speed only for packets. Doesn''t hfsc support such a field? tc -s class show dev eth0 class hfsc 1: root Sent 0 bytes 0 pkt (dropped 0, overlimits 0 requeues 0) rate 0bit 0pps backlog 0b 0p requeues 0 period 0 level 2 class hfsc 1:11 parent 1:1 sc m1 0bit d 18.0ms m2 1000Kbit ul m1 0bit d 0us m2
2010 Apr 22
1
PRIO qdisc + iptables TOS target
i am trying to do some traffic classification using the PRIO qdisc and i am having a few problems. I have a root htb class: tc qdisc add dev imq0 root handle 1: htb default 255 r2q 1 tc class add dev imq0 parent 1: classid 1:1 htb rate 768kbit and a child PRIO tc class add dev imq0 parent 1:1 classid 1:99 htb rate 96kbit ceil 600kbit prio 0 tc qdisc add dev imq0 parent 1:99 prio tc filter
2007 Aug 01
4
tc shown rate larger than ceil (was "Weird rate in HTB")
An earlier exchange about someone seeing the rate larger than the ceiling is posted below. Andy explained the reason for the "above ceiling" rate in Daniel''s output . . . but I just saw an example that doesn''t fit. >> tc output >> class htb 1:14 parent 1:1 leaf 14: prio 1 quantum 3072 rate 256000bit ceil 282000bit burst 1820b/8 mpu 0b overhead 0b
2006 Sep 09
2
Re: htb traffic shaping problem (Charlie Meyer)
Eric, I''ve attached the output from the command you requested below. I''m using kernel 2.6.17.12 and tc version ss060323. I''m have a nforce4 onboard giggabit controller using the forcedeth driver, if that makes a difference. Charlie, I''ll take a look at the Apache model for throttling, thanks. I was going for tc because was going to try to add additional
2005 Nov 11
2
marked packets end up in wrong class
Hello, I have recently started looking at tc and iptables. I have an htb-queue with two classes 1:10 and 1:20 where 1:20 is the default. Then I use iptables to mark all packets I send out on eth1. I then filter marked packets into class 1:10. I expected all packets sent on eth1 to end up in class 1:10, but some packets still go to 1:20. Did I do it wrong? Thank you for any help. regards, David
2007 May 16
5
statistics and calc bandwidth traffic using tc -s qdisc show
2013 May 31
1
Re: Fwd: Does Set Block IO tune take effect in the same session.
On 31.05.2013 07:05, vikrant verma wrote: > Hello Michal, > > Throttling is working now after setting inbound.average = 10 (i have > set other values to 0 ) > But i am observing an inconsistent behavior in the observed throttling > bandwidth > > when i set inbound.average = 10 , i observed bandwidth of 24 kbps > when i set inbound.average = 100 , i observed
2005 Nov 24
0
max latency with htb and wondershaper for voip and p2p
Hi I''m actually using a wondershaper derived script. The problem I have is that when I ping (ping is in high priority queue) the ping time is not as low as I expected with p2p running: 760 packets transmitted, 760 received, 0% packet loss, time 759695ms rtt min/avg/max/mdev = 39.667/62.863/598.539/33.287 ms, pipe 2 With voip, it is much better with the qos but it is still a little
2007 Jun 15
2
HTB question, tokens.
Hi, What exactly are the "tokens"? I thought each token allowed the sending of one byte, that tokens are stored in a bucket that can hold a max of "burst" tokens, and that this bucket is filled with tokens at "rate". But theory does not seem to explain the "tc -s .." output in the examples below. And I can''t figure out why or how... #tc qdisc
2006 Apr 09
2
tc counters "problem"
Hi, I''m using tc and HTB to shape my outgoing ADSL traffic. I was trying to make some graphs on the classes by meassuring the "sent bytes" of each class using rrdtool to store the data (as kbps after conversion). I expected that meassuring the root class I would get values similar that the ones I get measuring the interface counters but they differ by a large amount. Is
2005 Jun 10
2
HTB lended and borrowed
Running "tc -s class show dev eth1" against an HTB qdisc results in the output of class summaries similar to this: class htb 1:10 parent 1:1 leaf 10: prio 0 rate 150000bit ceil 1466Kbit burst 2Kb cburst 2Kb Sent 158641651 bytes 771351 pkt (dropped 0, overlimits 0 requeues 0) rate 8064bit 7pps backlog 0b 0p requeues 0 lended: 680985 borrowed: 90366 giants: 0 tokens: 173443 ctokens: