Displaying 20 results from an estimated 200 matches similar to: "htb : server trafficcontrol"
2007 May 31
5
IFB & 802.1q
Hello
What I''m looking for is how to configure the Linux QoS module to do
global rate limitation for two (or more) 802.1q pseudo network devices.
I naturally suppose there is a possibility with IFB. I don''t want to use
IMQ because it''s not integrated to my kernel v2.6.21.1 and I didn''t find
IMQ patches for it nor for the iptables package I use (v1.3.7).
2006 Apr 10
1
Where is the documentation for IFB ?
Hi all,
Can you tell me where is the documentation for the new IFB (implemented
in kernels > 2.6.16).
Thanks in advance!
Nikolay
2008 Mar 05
2
Shorewall & IFB
Hello Tom!
After i read and analyze some docs about IFB i decide that for implement
this feature in Shorewall not need more efforts (of course i may be wrong).
If we have 'ifb0' device then we must activate ingress discipline on real
device (f.e. eth2) and redirect 'egress' from it to 'ifb0'.
tc qdisc add dev eth2 ingress
tc filter add dev eth2 parent ffff: protocol ip
2007 Mar 28
4
modprobe ifb
I''m trying to set up a traffic control on ingress attaching a egress qdisc to
the ifb device. The idea is to use a RED algorithm instead of policing
the incoming traffic. After trying with tc-red and not obtaining the
expected results, I decided to try with something easier, and use htb
as bottleneck:
ifconfig ifb0 up
tc qdisc del dev $dev ingress
tc qdisc del dev ifb0 root
tc qdisc add
2007 Sep 19
7
ifb and ppp
Hello!
My goal is to setup an ingress traffic shaping on my PPPOE DSL line with ifb.
My old imq stuff used iptables marks (like ''iptables -t mangle -A
PREROUTING -p tcp --sport 22 -m length --length :500 -j MARK --set-mark 31'')
to classify the traffic and since i am lazy, i tried to to reuse them with
ifb. But no luck: iptables marks the packets well, but tc
2007 Jun 24
5
Traffic shaping on multiple interfaces
I''m trying to setup traffic shaping on my linux gateway/router.
The system has 3 interfaces:
eth0 - My LAN - with IP address 192.168.0.254
eth1 - The ethernet connection to which my ADSL modem is connected. This
has a 10.25.x.x IP, more on this later. The ADSL link has an upstream of
~1.2mbit.
ppp0 - The PPP connection which is my WAN connection, with a real world IP.
The system acts
2007 Mar 28
7
(no subject)
Hi all,
I''m having a hell of a time getting my IFB to work. I know I''ve done
this before, so I''m missing something stupid. Can anybody tell me what
it might be?
Configs as follows:
--------
#!/bin/sh
modprobe ifb numifbs=1000
modprobe act_mirred
modprobe 8021q
brctl addbr br0
brctl setfd br0 0
brctl stp br0 off
brctl addif br0 eth1
brctl addif br0 eth2
ifconfig eth1
2006 Jul 02
3
IFB working
How to use IFB as replacement of IMQ
There is not much documentation about it
Please explain with example
--
Failure seldom stops you. What stops you is the fear of failure.
_______________________________________________
LARTC mailing list
LARTC@mailman.ds9a.nl
http://mailman.ds9a.nl/cgi-bin/mailman/listinfo/lartc
2017 Sep 01
2
[PATCH net-next] virtio-net: invoke zerocopy callback on xmit path if no tx napi
On 2017?08?31? 22:30, Willem de Bruijn wrote:
>> Incomplete results at this stage, but I do see this correlation between
>> flows. It occurs even while not running out of zerocopy descriptors,
>> which I cannot yet explain.
>>
>> Running two threads in a guest, each with a udp socket, each
>> sending up to 100 datagrams, or until EAGAIN, every msec.
>>
2017 Sep 01
2
[PATCH net-next] virtio-net: invoke zerocopy callback on xmit path if no tx napi
On 2017?08?31? 22:30, Willem de Bruijn wrote:
>> Incomplete results at this stage, but I do see this correlation between
>> flows. It occurs even while not running out of zerocopy descriptors,
>> which I cannot yet explain.
>>
>> Running two threads in a guest, each with a udp socket, each
>> sending up to 100 datagrams, or until EAGAIN, every msec.
>>
2007 Mar 22
8
Shape own router
On 2/19/07, Salatiel Filho <salatiel.filho@gmail.com> wrote:
> Well , thanks to imq all my client machines are now shaped and
> everything is great ...
> But now i have a doubt , is there a way to shape the traffic that goes
> to the route [doing a wget from the router for example ]?
>
>
> I have a PREROUTING IMQ0 and a POSTROUTING IMQ1 , everything is
> working like
2017 Sep 28
9
[PATCH net-next] vhost_net: do not stall on zerocopy depletion
From: Willem de Bruijn <willemb at google.com>
Vhost-net has a hard limit on the number of zerocopy skbs in flight.
When reached, transmission stalls. Stalls cause latency, as well as
head-of-line blocking of other flows that do not use zerocopy.
Instead of stalling, revert to copy-based transmission.
Tested by sending two udp flows from guest to host, one with payload
of
2017 Sep 28
9
[PATCH net-next] vhost_net: do not stall on zerocopy depletion
From: Willem de Bruijn <willemb at google.com>
Vhost-net has a hard limit on the number of zerocopy skbs in flight.
When reached, transmission stalls. Stalls cause latency, as well as
head-of-line blocking of other flows that do not use zerocopy.
Instead of stalling, revert to copy-based transmission.
Tested by sending two udp flows from guest to host, one with payload
of
2007 Oct 25
7
TC (HTB) doesn''t work well when network is congested?
Hi,
I have a server and ten clients in a Gigabit network. The server has
125mbps network bandwidth.
I want that the server has 40Mbps bandwidth reserved for client 1 (IP
192.168.5.141), and the
rest bandwidth is for all other clients.
My script looks like this (I use IFB for incoming traffic):
#!/bin/bash
export TC="/sbin/tc"
$TC qdisc add dev ifb0 root handle 1: htb default 30
2017 Sep 30
2
[PATCH net-next] vhost_net: do not stall on zerocopy depletion
On Fri, Sep 29, 2017 at 3:38 PM, Michael S. Tsirkin <mst at redhat.com> wrote:
> On Wed, Sep 27, 2017 at 08:25:56PM -0400, Willem de Bruijn wrote:
>> From: Willem de Bruijn <willemb at google.com>
>>
>> Vhost-net has a hard limit on the number of zerocopy skbs in flight.
>> When reached, transmission stalls. Stalls cause latency, as well as
>>
2017 Sep 30
2
[PATCH net-next] vhost_net: do not stall on zerocopy depletion
On Fri, Sep 29, 2017 at 3:38 PM, Michael S. Tsirkin <mst at redhat.com> wrote:
> On Wed, Sep 27, 2017 at 08:25:56PM -0400, Willem de Bruijn wrote:
>> From: Willem de Bruijn <willemb at google.com>
>>
>> Vhost-net has a hard limit on the number of zerocopy skbs in flight.
>> When reached, transmission stalls. Stalls cause latency, as well as
>>
2006 Aug 18
5
Handle limit in filter
-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA1
Hi,
I''ve written a minimal sort of Perl module that dynamically creates
and destroys traffic control rules for specific IPs. I''m currently
using it for a user bandwidth control application at a client site.
The module essentially gets Ethernet device(s), IP address and in/out
speeds as input and dynamically creates classes, queues
2017 Oct 06
1
[PATCH net-next v2] vhost_net: do not stall on zerocopy depletion
From: Willem de Bruijn <willemb at google.com>
Vhost-net has a hard limit on the number of zerocopy skbs in flight.
When reached, transmission stalls. Stalls cause latency, as well as
head-of-line blocking of other flows that do not use zerocopy.
Instead of stalling, revert to copy-based transmission.
Tested by sending two udp flows from guest to host, one with payload
of
2017 Oct 06
1
[PATCH net-next v2] vhost_net: do not stall on zerocopy depletion
From: Willem de Bruijn <willemb at google.com>
Vhost-net has a hard limit on the number of zerocopy skbs in flight.
When reached, transmission stalls. Stalls cause latency, as well as
head-of-line blocking of other flows that do not use zerocopy.
Instead of stalling, revert to copy-based transmission.
Tested by sending two udp flows from guest to host, one with payload
of
2007 Dec 05
3
Adding qdiscs crashes kernel??
Hi all,
I''m having some problems setting up qdiscs on a bridge.The config looks
a little like this:
ifconfig ifb0 up # Bring up the IFB for this bridge.
tc qdisc add dev eth2 ingress
tc qdisc add dev eth3 ingress
tc qdisc add dev ifb0 root handle 1:0 cbq bandwidth 100Mbit avpkt 1000
cell 8
# Raw qdiscs on each bridge port
tc qdisc add dev eth2 root handle 1:0 cbq bandwidth