[Bloat] fq_codel on bridge multiple subnets?
Dev
dev at logicalwebhost.com
Fri Jan 4 13:20:20 EST 2019
I want to pass multiple hundreds of Mbps across this bridge very consistently, and across multiple subnets to different enterprise gateways which then connect to the internet, will plan a little test to see how this does under load. Hopefully I don’t need special NIC’s to handle it?
> On Jan 4, 2019, at 1:19 AM, Pete Heist <pete at heistp.net> wrote:
>
> It’s a little different for me in that I’m rate limiting on one of the physical interfaces, but otherwise, your setup should reduce latency under load when the Ethernet devices are being used at line rate.
>
> If your WAN interface is enp8s0 and goes out to the Internet, you may want to shape there (htb+fq_codel or cake) depending on what upstream device is in use.
>
> If enp7s6 and enp9s2 are only carrying LAN traffic, and not traffic that goes out to the Internet, fq_codel’s target and interval could be reduced.
>
>> On Jan 4, 2019, at 6:22 AM, Dev <dev at logicalwebhost.com> wrote:
>>
>> Okay, so this is what I have for /etc/network/interfaces (replaced eth0-2 with what Debian Buster actually calls them):
>>
>> auto lo br0
>> iface lo inet loopback
>>
>> allow-hotplug enp8s0
>> iface enp8s0 inet static
>> address 192.168.10.200
>> netmask 255.255.255.0
>> gateway 192.168.10.1
>> dns-nameservers 8.8.8.8
>>
>> iface enp7s6 inet manual
>> tc qdisc add dev enp7s6 root fq_codel
>>
>> iface enp9s2 inet manual
>> tc qdisc add dev enp9s2 root fq_codel
>>
>> # Bridge setup
>> iface br0 inet static
>> bridge_ports enp7s6 enp9s2
>> #bridge_stp on
>> address 192.168.3.50
>> broadcast 192.168.3.255
>> netmask 255.255.255.0
>> gateway 192.168.3.1
>> dns-nameservers 8.8.8.8
>>
>> so my bridge interfaces now show:
>>
>>> : tc -s qdisc show dev enp7s6
>> qdisc fq_codel 0: root refcnt 2 limit 10240p flows 1024 quantum 1514 target 5.0ms interval 100.0ms memory_limit 32Mb ecn
>> Sent 0 bytes 0 pkt (dropped 0, overlimits 0 requeues 0)
>> backlog 0b 0p requeues 0
>> maxpacket 0 drop_overlimit 0 new_flow_count 0 ecn_mark 0
>> new_flows_len 0 old_flows_len 0
>>
>> and
>>
>>> : tc -s qdisc show dev enp9s2
>> qdisc fq_codel 0: root refcnt 2 limit 10240p flows 1024 quantum 1514 target 5.0ms interval 100.0ms memory_limit 32Mb ecn
>> Sent 12212 bytes 80 pkt (dropped 0, overlimits 0 requeues 0)
>> backlog 0b 0p requeues 0
>> maxpacket 0 drop_overlimit 0 new_flow_count 0 ecn_mark 0
>> new_flows_len 0 old_flows_len 0
>>
>> with my bridge like:
>>
>> ip a
>>
>> 5: br0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP group default qlen 1000
>> link/ether 00:04:5a:86:a2:84 brd ff:ff:ff:ff:ff:ff
>> inet 192.168.3.50/24 brd 192.168.3.255 scope global br0
>> valid_lft forever preferred_lft forever
>> inet6 fe80::204:5aff:fe86:a284/64 scope link
>> valid_lft forever preferred_lft forever
>>
>> So do I have it configured right or should I change something? I haven’t gotten a chance to stress test it yet, but will try tomorrow.
>>
>> - Dev
>>
>>> On Jan 3, 2019, at 10:54 AM, Pete Heist <pete at heistp.net> wrote:
>>>
>>>
>>>> On Jan 3, 2019, at 7:12 PM, Toke Høiland-Jørgensen <toke at toke.dk> wrote:
>>>>
>>>> Dev <dev at logicalwebhost.com> writes:
>>>>
>>>>> I’m trying to create a bridge on eth1 and eth2, with a management
>>>>> interface on eth0, then enable fq_codel on the bridge. My bridge
>>>>> interface looks like:
>>>>
>>>> You'll probably want to put FQ-CoDel on the underlying physical
>>>> interfaces, as those are the ones actually queueing the traffic...
>>>
>>> I can confirm that. I'm currently using a bridge on my home router. eth3 and eth4 are bridged, eth4 is connected to the CPE device which goes out to the Internet, eth4 is where queue management is applied, and this works. It does not work to add this to br0…
>>>
>>
>
More information about the Bloat
mailing list