similar to: dropping tokens

Displaying 20 results from an estimated 40000 matches similar to: "dropping tokens"

2005 May 26
2
Leaky bucket in TC
Hi, I am trying to control bandwidths of the connections by using TC with HTB. But I want to implement Leaky Bucket on this structure so as to have a straight flow. Due to being new in this area, could anyone help me how I could do this? Does TC support leaky bucket? Best regards, Fatih Duzova
2004 Nov 26
4
HTB dropping packets while tokens available?
Hi again, I just ran into a seriously strange glitch, and was wondering if anybody had experienced this... I have an HTB class with an iptables classifier. The iptables classifier shows packets hitting the rule for this class. Now, HTB drops these packets (''dropped'' increases), even although the packets are small, and there are over 10000 tokens available. Is there any
2005 Jun 20
0
Error rate control on queues
Hello, I want to implement error rate control on the queues which ve been created as htb. The aim is to see the results by dropping packets randomly with a determined error rate- may be changed dynamically- on the requested queues while sending data. But I do not know where to begin, and what to use. I would appreciate if anyone could help me and suggest sth. Thanks, Fatih
2005 May 28
2
bandwidth control on htb
Hi, I implemented queues on some ports with htb by filtering. and to have a constant bw, i defined rate and ceil the same. But while watching the bw''s, I ve realized they may sometimes exceed the defined value for short periods. How could I avoid this situation? I thought that implementing a leaky bucket''d be fine but i m not sure about this and how to do it on tc. I would be
2005 Jun 23
7
where s filtering done?
hello, I am trying to find where filtering is done in order to drop packets randomly there (eg.10%of them) for a testing purpose in a school project. I want to do this after packet is filtered and before being sent. I d aprreciate if you could help or suggest. Thanks.
2007 Jun 15
2
HTB question, tokens.
Hi, What exactly are the "tokens"? I thought each token allowed the sending of one byte, that tokens are stored in a bucket that can hold a max of "burst" tokens, and that this bucket is filled with tokens at "rate". But theory does not seem to explain the "tc -s .." output in the examples below. And I can''t figure out why or how... #tc qdisc
2003 Nov 15
3
ctokens vs tokens - HTB.
Hi, One more of those ... 1) what is a ctoken with regard to HTB? Does the ''c'' stand for ceil and if so how is it different from tokens? 2) Is the difference, only in the generation of these tokens? 3) Would the ctokens and token be treated equally once borrowed? Kindly help. Thank you. Regards, suraj. __________________________________ Do you Yahoo!? Protect your
2005 Jan 27
3
negative tokens
I have this: class htb 1:29 parent 1:1 prio 0 rate 3000bit ceil 100000bit burst 1599b cburst 1611b Sent 33233 bytes 772 pkts (dropped 0, overlimits 0) lended: 415 borrowed: 357 giants: 0 tokens: -3756376 ctokens: 128779 or this: class htb 1:21 parent 1:1 prio 0 rate 3000bit ceil 100000bit burst 1599b cburst 1611b Sent 57554 bytes 618 pkts (dropped 0, overlimits 0) lended: 193 borrowed: 425
2006 Mar 08
1
negative token/ctokens
In this simple htb setup: # tc -s -d class ls dev eth0 class htb 1:1 root rate 300000bit ceil 300000bit burst 1749b/8 mpu 0b overhead 0b cburst 1749b/8 mpu 0b overhead 0b level 7 Sent 13171835 bytes 13169 pkt (dropped 0, overlimits 0 requeues 0) rate 45848bit 10pps backlog 0b 0p requeues 0 lended: 5272 borrowed: 0 giants: 0 tokens: -84429 ctokens: -84429 class htb 1:2 parent 1:1 prio 0
2004 Jun 02
1
Avoid dropped packets
I´m shaping with htb with a kernel 2.4.25 compiled with PSCHED_CPU. What i´m noticing (in mrtg) is that traffic coming in to eth0 (internet side) is much bigger than eth1(Lan side). I think this is due to packets being dropped right? In this case, is there any way to avoid this? because right now traffic coming from the router is being wasted right? This is the output of tc: tc -s class show
2005 Jun 09
1
Segmentation fault
Hi, I have a small question about tc''s source code. I m trying to compile the code by itself, and for this purpose i ve created a big .c file that includes all the sources it uses. the reason behind this to make a small modification for my school project. But whenever I try to do "tc qdisc add ..." or "tc class add..", it gives the error segmentation fault, and it
2004 Apr 04
2
IMQ & NAT
Hi all, my IMQ device works OK (thanks to Andy Furniss), but now I''ve problems to attach the traffic in the qdisc''s. This is my conf: ----------------------------------------------------------------------- INET | |eth0 300Kbps ROUTER (NAT) |eth1 | LAN ----------------------------------------------------------------------- MAX=300 tc qdisc add dev imq0 root handle 1: htb
2007 Aug 19
0
HTB qdisc within HTB root qdisc
Hello... Im trying to setup HTB to allow me to shape traffic from two upstreams that meets on single lan0 interface. I prefer to use HTB qdisc within HTB root qdisc for cleaner rules design. Seems that it doesnt work at all. tc -s class show doesnt show any traffic on other classes attached to HTB qdisc. Linux 2.6.20.7 iproute-2.6.20-070313 Weird thing is that tc -s class show that 1: and 2:
2007 Jun 21
1
A HTB problem
My hardware is a Linksys AP with MIPS 300MHz and Linux kernel 2.4.20. The traffics are from two LAN switch ports to WAN port. And the traffics are generate at a rate of 80Mbit. So the total traffic to WAN port is 160Mbit. The shaping works well that the traffic to WAN port is about 50/30Mbit according to configuration. But the priority seems strange when the root rate is 50Mbit. When the root
2007 Oct 16
0
Determining burst/cburst HTB values
Hello. I am wondering if the default (auto-computed) burst and cburst parameters are appropriate values for my application. I have a 100 mbit interface which is connected to a 16 mbit uplink. I am sharing this bandwidth into a set of partitions (tc classes) defined as follows: root class [1:1] |
2003 Jul 14
0
Why arent filteres working? (as I wanted)
Here goes the filter statements: #Initializing traffic control... tc qdisc add dev br0 root handle 1:0 htb #Loading queue disciplines for plis230 network... tc class add dev br0 parent 1:0 classid 1:1 htb rate 512kbit ceil 512kbit #Loading queue disciplines for pmad048 network... tc class add dev br0 parent 1:1 classid 1:20 htb rate 30kbit ceil 30kbit #server queue tc class add dev br0 parent
2004 Dec 02
1
FW: differing rate and ceil values
Hello all, when running tc -s -d class show dev eth, the parent class rate and ceil are different to the configuration? his is the first time I have used HTB so I may have made some basic mistakes. Thanks in advance Shaun Output as follows: class htb 1:1 root rate 760Mbit ceil 760Mbit burst 997641b/8 mpu 0b cburst 997641b/8 mpu 0b level 7 Sent 0 bytes 0 pkts (dropped 0, overlimits 0)
2005 Nov 24
0
max latency with htb and wondershaper for voip and p2p
Hi I''m actually using a wondershaper derived script. The problem I have is that when I ping (ping is in high priority queue) the ping time is not as low as I expected with p2p running: 760 packets transmitted, 760 received, 0% packet loss, time 759695ms rtt min/avg/max/mdev = 39.667/62.863/598.539/33.287 ms, pipe 2 With voip, it is much better with the qos but it is still a little
2007 Sep 19
1
major packet loss with htb
Hi, got a 25/25Mbit connection which is quite stuffed. So I applied htb rules. Uplink: class htb 1:1 root rate 24500Kbit ceil 24500Kbit burst 4661b/8 mpu 0b overhead 0b cburst 4661b/8 mpu 0b overhead 0b level 7 Sent 430600689269 bytes 730147320 pkt (dropped 0, overlimits 0 requeues 0) rate 23057Kbit 5520pps backlog 0b 0p requeues 0 lended: 199673949 borrowed: 0 giants: 0 tokens: -964
2007 Sep 23
0
Giants on IMQ0
Why am I getting giants on imq0? How do I stop that? ========== imq0 Link encap:UNSPEC HWaddr 00-00-00-00-00-00-00-00-00-00-00-00-00-00-00-00 UP RUNNING NOARP MTU:1500 Metric:1 RX packets:39880312 errors:0 dropped:0 overruns:0 frame:0 TX packets:39878715 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1024 RX