[Cerowrt-devel] Ideas on how to simplify and popularize bufferbloat control for consideration.
David Lang
david at lang.hm
Sat Jul 26 18:24:10 EDT 2014
On Sat, 26 Jul 2014, David Lang wrote:
> On Sat, 26 Jul 2014, Sebastian Moeller wrote:
>
>> On Jul 26, 2014, at 22:39 , David Lang <david at lang.hm> wrote:
>>
>>> by how much tuning is required, I wasn't meaning how frequently to tune,
>>> but how close default settings can come to the performance of a expertly
>>> tuned setup.
>>
>> Good question.
>>
>>>
>>> Ideally the tuning takes into account the characteristics of the hardware
>>> of the link layer. If it's IP encapsulated in something else (ATM, PPPoE,
>>> VPN, VLAN tagging, ethernet with jumbo packet support for example), then
>>> you have overhead from the encapsulation that you would ideally take into
>>> account when tuning things.
>>>
>>> the question I'm talking about below is how much do you loose compared to
>>> the idea if you ignore this sort of thing and just assume that the wire is
>>> dumb and puts the bits on them as you send them? By dumb I mean don't even
>>> allow for inter-packet gaps, don't measure the bandwidth, don't try to
>>> pace inbound connections by the timing of your acks, etc. Just run BQL and
>>> fq_codel and start the BQL sizes based on the wire speed of your link
>>> (Gig-E on the 3800) and shrink them based on long-term passive observation
>>> of the sender.
>>
>> As data talks I just did a quick experiment with my ADSL2+ koine at
>> home. The solid lines in the attached plot show the results for proper
>> shaping with SQM (shaping to 95% of del link rates of downstream and
>> upstream while taking the link layer properties, that is ATM encapsulation
>> and per packet overhead into account) the broken lines show the same system
>> with just the link layer adjustments and per packet overhead adjustments
>> disabled, but still shaping to 95% of link rate (this is roughly equivalent
>> to 15% underestimation of the packet size). The actual theist is
>> netperf-wrappers RRUL (4 tcp streams up, 4 tcp steams down while measuring
>> latency with ping and UDP probes). As you can see from the plot just
>> getting the link layer encapsulation wrong destroys latency under load
>> badly. The host is ~52ms RTT away, and with fq_codel the ping time per leg
>> is just increased one codel target of 5ms each resulting in an modest
>> latency increase of ~10ms with proper shaping for a total of ~65ms, with
>> improper shaping RTTs increase to ~95ms (they almost double), so RTT
>> increases by ~43ms. Also note how the extremes for the broken lines are
>> much worse than for the solid lines. In short I would estimate that a
>> slight misjudgment (15%) results in almost 80% increase of latency under
>> load. In other words getting the rates right matters a lot. (I should also
>> note that in my setup there is a secondary router that limits RTT to max
>> 300ms, otherwise the broken lines might look even worse...)
is this with BQL/fq_codel in both directions or only in one direction?
David Lang
> what is the latency like without BQL and codel? the pre-bufferbloat version?
> (without any traffic shaping)
>
> I agree that going from 65ms to 95ms seems significant, but if the stock
> version goes into up above 1000ms, then I think we are talking about things
> that are 'close'
>
> assuming that latency under load without the improvents got >1000ms
>
> fast-slow (in ms)
> ideal=10
> untuned=43
> bloated > 1000
>
> fast/slow
> ideal = 1.25
> untuned = 1.83
> bloated > 19
>
> slow/fast
> ideal = 0.8
> untuned = 0.55
> bloated = 0.05
>
> rather than looking at how much worse it is than the ideal, look at how much
> closer it is to the ideal than to the bloated version.
>
> David Lang
>
> _______________________________________________
> Cerowrt-devel mailing list
> Cerowrt-devel at lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/cerowrt-devel
>
More information about the Cerowrt-devel
mailing list