* [Make-wifi-fast] IEEE 802.11n/ac Wireless Network Efficiency under different TCP Congestion Controls
@ 2019-12-13 19:53 Dave Taht
2019-12-13 21:05 ` Carlo Augusto Grazia
2019-12-13 21:09 ` [Make-wifi-fast] IEEE 802.11n/ac Wireless Network Efficiency under different TCP Congestion Controls Holland, Jake
0 siblings, 2 replies; 21+ messages in thread
From: Dave Taht @ 2019-12-13 19:53 UTC (permalink / raw)
To: Make-Wifi-fast; +Cc: carloaugusto.grazia
https://sci-hub.tw/10.1109/WiMOB.2019.8923418
It predates the aql work, but the bbr result is puzzling.
--
Make Music, Not War
Dave Täht
CTO, TekLibre, LLC
http://www.teklibre.com
Tel: 1-831-435-0729
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: IEEE 802.11n/ac Wireless Network Efficiency under different TCP Congestion Controls
2019-12-13 19:53 [Make-wifi-fast] IEEE 802.11n/ac Wireless Network Efficiency under different TCP Congestion Controls Dave Taht
@ 2019-12-13 21:05 ` Carlo Augusto Grazia
2019-12-13 21:25 ` [Make-wifi-fast] the future belongs to pacing Dave Taht
2019-12-13 21:09 ` [Make-wifi-fast] IEEE 802.11n/ac Wireless Network Efficiency under different TCP Congestion Controls Holland, Jake
1 sibling, 1 reply; 21+ messages in thread
From: Carlo Augusto Grazia @ 2019-12-13 21:05 UTC (permalink / raw)
To: Dave Taht; +Cc: Make-Wifi-fast
[-- Attachment #1: Type: text/plain, Size: 1237 bytes --]
Hi Dave,
thank you for your email!
Toke told me about AQL a couple of weeks ago, I definitely want to test it
ASAP.
BBR struggles a lot on Wi-Fi interfaces (ones with aggregation) with kernel
4.14 & 4.19.
Anyway, it seems that with BBRv2 on new kernels this problem does not exist
anymore.
Best regards
Carlo
Il giorno ven 13 dic 2019 alle 20:54 Dave Taht <dave.taht@gmail.com> ha
scritto:
> https://sci-hub.tw/10.1109/WiMOB.2019.8923418
>
> It predates the aql work, but the bbr result is puzzling.
>
>
> --
> Make Music, Not War
>
> Dave Täht
> CTO, TekLibre, LLC
> http://www.teklibre.com
> Tel: 1-831-435-0729
>
--
--------------------------------------------------------------------
Carlo Augusto Grazia, Ph. D.
Assistant Professor
--------------------------------------------------------------------
Dept. of Engineering "Enzo Ferrari"
University of Modena and Reggio Emilia
Via Pietro Vivarelli, 10/1 - 41125 - Modena - Italy
Building 26, floor 2, room 28
Tel.: +39-059-2056323
email: carloaugusto.grazia@unimore.it
Link to my personal home page here
<https://web.ing.unimo.it/wiki/index.php/Carlo_Augusto_Grazia>
--------------------------------------------------------------------
[-- Attachment #2: Type: text/html, Size: 2487 bytes --]
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] IEEE 802.11n/ac Wireless Network Efficiency under different TCP Congestion Controls
2019-12-13 19:53 [Make-wifi-fast] IEEE 802.11n/ac Wireless Network Efficiency under different TCP Congestion Controls Dave Taht
2019-12-13 21:05 ` Carlo Augusto Grazia
@ 2019-12-13 21:09 ` Holland, Jake
2019-12-13 21:18 ` Carlo Augusto Grazia
1 sibling, 1 reply; 21+ messages in thread
From: Holland, Jake @ 2019-12-13 21:09 UTC (permalink / raw)
To: Dave Taht, Make-Wifi-fast; +Cc: carloaugusto.grazia
They seem to be saying it's basically because of pacing?
(0 aggregation from TSO...)
On 2019-12-13, 11:54, "Dave Taht" <dave.taht@gmail.com> wrote:
https://sci-hub.tw/10.1109/WiMOB.2019.8923418
It predates the aql work, but the bbr result is puzzling.
--
Make Music, Not War
Dave Täht
CTO, TekLibre, LLC
http://www.teklibre.com
Tel: 1-831-435-0729
_______________________________________________
Make-wifi-fast mailing list
Make-wifi-fast@lists.bufferbloat.net
https://lists.bufferbloat.net/listinfo/make-wifi-fast
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] IEEE 802.11n/ac Wireless Network Efficiency under different TCP Congestion Controls
2019-12-13 21:09 ` [Make-wifi-fast] IEEE 802.11n/ac Wireless Network Efficiency under different TCP Congestion Controls Holland, Jake
@ 2019-12-13 21:18 ` Carlo Augusto Grazia
2019-12-13 23:06 ` Dave Taht
0 siblings, 1 reply; 21+ messages in thread
From: Carlo Augusto Grazia @ 2019-12-13 21:18 UTC (permalink / raw)
To: Holland, Jake; +Cc: Dave Taht, Make-Wifi-fast
[-- Attachment #1: Type: text/plain, Size: 1374 bytes --]
Yes, we have submitted a couple of works in which, changing the BBR Pacing
Gain, the problem is mitigated
Il giorno ven 13 dic 2019 alle 22:09 Holland, Jake <jholland@akamai.com> ha
scritto:
> They seem to be saying it's basically because of pacing?
> (0 aggregation from TSO...)
>
> On 2019-12-13, 11:54, "Dave Taht" <dave.taht@gmail.com> wrote:
>
> https://sci-hub.tw/10.1109/WiMOB.2019.8923418
>
> It predates the aql work, but the bbr result is puzzling.
>
>
> --
> Make Music, Not War
>
> Dave Täht
> CTO, TekLibre, LLC
> http://www.teklibre.com
> Tel: 1-831-435-0729
> _______________________________________________
> Make-wifi-fast mailing list
> Make-wifi-fast@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/make-wifi-fast
>
> --
--------------------------------------------------------------------
Carlo Augusto Grazia, Ph. D.
Assistant Professor
--------------------------------------------------------------------
Dept. of Engineering "Enzo Ferrari"
University of Modena and Reggio Emilia
Via Pietro Vivarelli, 10/1 - 41125 - Modena - Italy
Building 26, floor 2, room 28
Tel.: +39-059-2056323
email: carloaugusto.grazia@unimore.it
Link to my personal home page here
<https://web.ing.unimo.it/wiki/index.php/Carlo_Augusto_Grazia>
--------------------------------------------------------------------
[-- Attachment #2: Type: text/html, Size: 2819 bytes --]
^ permalink raw reply [flat|nested] 21+ messages in thread
* [Make-wifi-fast] the future belongs to pacing
2019-12-13 21:05 ` Carlo Augusto Grazia
@ 2019-12-13 21:25 ` Dave Taht
2020-07-04 17:29 ` [Bloat] " Matt Mathis
` (2 more replies)
0 siblings, 3 replies; 21+ messages in thread
From: Dave Taht @ 2019-12-13 21:25 UTC (permalink / raw)
To: Carlo Augusto Grazia; +Cc: Make-Wifi-fast, bloat, jamshid
and everything we know about the tcp macroscopic model, is obsolete,
according to a provocative paper by matt mathis and Jamshid Mahdavi
in sigcomm.
https://ccronline.sigcomm.org/wp-content/uploads/2019/10/acmdl19-323.pdf
On Fri, Dec 13, 2019 at 1:05 PM Carlo Augusto Grazia
<carloaugusto.grazia@unimore.it> wrote:
>
> Hi Dave,
> thank you for your email!
> Toke told me about AQL a couple of weeks ago, I definitely want to test it ASAP.
> BBR struggles a lot on Wi-Fi interfaces (ones with aggregation) with kernel 4.14 & 4.19.
> Anyway, it seems that with BBRv2 on new kernels this problem does not exist anymore.
>
> Best regards
> Carlo
>
> Il giorno ven 13 dic 2019 alle 20:54 Dave Taht <dave.taht@gmail.com> ha scritto:
>>
>> https://sci-hub.tw/10.1109/WiMOB.2019.8923418
>>
>> It predates the aql work, but the bbr result is puzzling.
>>
>>
>> --
>> Make Music, Not War
>>
>> Dave Täht
>> CTO, TekLibre, LLC
>> http://www.teklibre.com
>> Tel: 1-831-435-0729
>
> --
> --------------------------------------------------------------------
> Carlo Augusto Grazia, Ph. D.
> Assistant Professor
> --------------------------------------------------------------------
> Dept. of Engineering "Enzo Ferrari"
> University of Modena and Reggio Emilia
> Via Pietro Vivarelli, 10/1 - 41125 - Modena - Italy
> Building 26, floor 2, room 28
> Tel.: +39-059-2056323
> email: carloaugusto.grazia@unimore.it
> Link to my personal home page here
> --------------------------------------------------------------------
--
Make Music, Not War
Dave Täht
CTO, TekLibre, LLC
http://www.teklibre.com
Tel: 1-831-435-0729
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] IEEE 802.11n/ac Wireless Network Efficiency under different TCP Congestion Controls
2019-12-13 21:18 ` Carlo Augusto Grazia
@ 2019-12-13 23:06 ` Dave Taht
0 siblings, 0 replies; 21+ messages in thread
From: Dave Taht @ 2019-12-13 23:06 UTC (permalink / raw)
To: Carlo Augusto Grazia; +Cc: Holland, Jake, Make-Wifi-fast
On Fri, Dec 13, 2019 at 1:18 PM Carlo Augusto Grazia
<carloaugusto.grazia@unimore.it> wrote:
>
> Yes, we have submitted a couple of works in which, changing the BBR Pacing Gain, the problem is mitigated
I am always looking for repeatable experiments (scripts and so forth),
to carry the work forward, even if the paper isn't published yet. got
anything in github?
There's some exciting work going on with SCE - among iother things the
possibity of addig ecn support to bbr1 and 2... and I'm also trying to
get a grip on the behaviors of the new
ax chips.
> Il giorno ven 13 dic 2019 alle 22:09 Holland, Jake <jholland@akamai.com> ha scritto:
>>
>> They seem to be saying it's basically because of pacing?
>> (0 aggregation from TSO...)
>>
>> On 2019-12-13, 11:54, "Dave Taht" <dave.taht@gmail.com> wrote:
>>
>> https://sci-hub.tw/10.1109/WiMOB.2019.8923418
>>
>> It predates the aql work, but the bbr result is puzzling.
>>
>>
>> --
>> Make Music, Not War
>>
>> Dave Täht
>> CTO, TekLibre, LLC
>> http://www.teklibre.com
>> Tel: 1-831-435-0729
>> _______________________________________________
>> Make-wifi-fast mailing list
>> Make-wifi-fast@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/make-wifi-fast
>>
> --
> --------------------------------------------------------------------
> Carlo Augusto Grazia, Ph. D.
> Assistant Professor
> --------------------------------------------------------------------
> Dept. of Engineering "Enzo Ferrari"
> University of Modena and Reggio Emilia
> Via Pietro Vivarelli, 10/1 - 41125 - Modena - Italy
> Building 26, floor 2, room 28
> Tel.: +39-059-2056323
> email: carloaugusto.grazia@unimore.it
> Link to my personal home page here
> --------------------------------------------------------------------
--
Make Music, Not War
Dave Täht
CTO, TekLibre, LLC
http://www.teklibre.com
Tel: 1-831-435-0729
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Bloat] the future belongs to pacing
2019-12-13 21:25 ` [Make-wifi-fast] the future belongs to pacing Dave Taht
@ 2020-07-04 17:29 ` Matt Mathis
[not found] ` <mailman.763.1593883755.24343.bloat@lists.bufferbloat.net>
2020-07-06 14:08 ` [Make-wifi-fast] " Luca Muscariello
2 siblings, 0 replies; 21+ messages in thread
From: Matt Mathis @ 2020-07-04 17:29 UTC (permalink / raw)
To: Dave Taht; +Cc: Carlo Augusto Grazia, Make-Wifi-fast, jamshid, bloat
[-- Attachment #1: Type: text/plain, Size: 2975 bytes --]
Be aware that BBR is a moving target. There was an important WiFi fix
that went into BBRv1 in Jan 2019 that didn't make it into lots of
distros... BBRv2 is in the wings and fixes (nearly?) all sharing issues,
but isn't done yet.
Key takeaway: pacing is inevitable, because it saves large content
providers money (more efficient use of the most expensive silicon in the
data center, the switch buffer memory), however to use pacing we walk away
from 30 years of experience with TCP self clock, which is the foundation of
all of our CC research....
Thanks,
--MM--
The best way to predict the future is to create it. - Alan Kay
We must not tolerate intolerance;
however our response must be carefully measured:
too strong would be hypocritical and risks spiraling out of
control;
too weak risks being mistaken for tacit approval.
On Fri, Dec 13, 2019 at 1:25 PM Dave Taht <dave.taht@gmail.com> wrote:
> and everything we know about the tcp macroscopic model, is obsolete,
> according to a provocative paper by matt mathis and Jamshid Mahdavi
> in sigcomm.
>
> https://ccronline.sigcomm.org/wp-content/uploads/2019/10/acmdl19-323.pdf
>
>
>
>
> On Fri, Dec 13, 2019 at 1:05 PM Carlo Augusto Grazia
> <carloaugusto.grazia@unimore.it> wrote:
> >
> > Hi Dave,
> > thank you for your email!
> > Toke told me about AQL a couple of weeks ago, I definitely want to test
> it ASAP.
> > BBR struggles a lot on Wi-Fi interfaces (ones with aggregation) with
> kernel 4.14 & 4.19.
> > Anyway, it seems that with BBRv2 on new kernels this problem does not
> exist anymore.
> >
> > Best regards
> > Carlo
> >
> > Il giorno ven 13 dic 2019 alle 20:54 Dave Taht <dave.taht@gmail.com> ha
> scritto:
> >>
> >> https://sci-hub.tw/10.1109/WiMOB.2019.8923418
> >>
> >> It predates the aql work, but the bbr result is puzzling.
> >>
> >>
> >> --
> >> Make Music, Not War
> >>
> >> Dave Täht
> >> CTO, TekLibre, LLC
> >> http://www.teklibre.com
> >> Tel: 1-831-435-0729 <(831)%20435-0729>
> >
> > --
> > --------------------------------------------------------------------
> > Carlo Augusto Grazia, Ph. D.
> > Assistant Professor
> > --------------------------------------------------------------------
> > Dept. of Engineering "Enzo Ferrari"
> > University of Modena and Reggio Emilia
> > Via Pietro Vivarelli, 10/1 - 41125 - Modena - Italy
> > Building 26, floor 2, room 28
> > Tel.: +39-059-2056323 <+39%20059%20205%206323>
> > email: carloaugusto.grazia@unimore.it
> > Link to my personal home page here
> > --------------------------------------------------------------------
>
>
>
> --
> Make Music, Not War
>
> Dave Täht
> CTO, TekLibre, LLC
> http://www.teklibre.com
> Tel: 1-831-435-0729 <(831)%20435-0729>
> _______________________________________________
> Bloat mailing list
> Bloat@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/bloat
>
[-- Attachment #2: Type: text/html, Size: 4712 bytes --]
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] [Bloat] the future belongs to pacing
[not found] ` <mailman.763.1593883755.24343.bloat@lists.bufferbloat.net>
@ 2020-07-04 17:52 ` Daniel Sterling
2020-07-04 18:02 ` Jonathan Morton
2020-07-04 18:29 ` Sebastian Moeller
0 siblings, 2 replies; 21+ messages in thread
From: Daniel Sterling @ 2020-07-04 17:52 UTC (permalink / raw)
To: Matt Mathis
Cc: Dave Taht, Make-Wifi-fast, Carlo Augusto Grazia, jamshid, bloat
On Sat, Jul 4, 2020 at 1:29 PM Matt Mathis via Bloat
<bloat@lists.bufferbloat.net> wrote:
"pacing is inevitable, because it saves large content providers money
(more efficient use of the most expensive silicon in the data center,
the switch buffer memory), however to use pacing we walk away from 30
years of experience with TCP self clock"
at the risk of asking w/o doing any research,
could someone explain this to a lay person or point to a doc talking
about this more?
What does BBR do that's different from other algorithms? Why does it
break the clock? Before BBR, was the clock the only way TCP did CC?
Also,
I have UBNT "Amplifi" HD wifi units in my house. (HD units only; none
of the "mesh" units. Just HD units connected either via wifi or
wired.) Empirically, I've found that in order to reduce latency, I
need to set cake to about 1/4 of the total possible wifi speed;
otherwise if a large download comes down from my internet link, that
flow causes latency.
That is, if I'm using 5ghz at 20mhz channel width, I need to set
cake's bandwidth argument to 40mbits to prevent video streams /
downloads from impacting latency for any other stream. This is w/o any
categorization at all; no packet marking based on port or anything
else; cake set to "best effort".
Anything higher and when a large amount of data comes thru, something
(assumedly the buffer in the Amplifi HD units) causes 100s of
milliseconds of latency.
Can anyone speak to how BBR would react to this? My ISP is full
gigabit; but cake is going to drop a lot of packets as it throttles
that down to 40mbit before it sends the packets to the wifi AP.
Thanks,
Dan
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] [Bloat] the future belongs to pacing
2020-07-04 17:52 ` [Make-wifi-fast] " Daniel Sterling
@ 2020-07-04 18:02 ` Jonathan Morton
2020-07-04 18:29 ` Sebastian Moeller
1 sibling, 0 replies; 21+ messages in thread
From: Jonathan Morton @ 2020-07-04 18:02 UTC (permalink / raw)
To: Daniel Sterling
Cc: Matt Mathis, Make-Wifi-fast, Carlo Augusto Grazia, bloat, jamshid
> On 4 Jul, 2020, at 8:52 pm, Daniel Sterling <sterling.daniel@gmail.com> wrote:
>
> could someone explain this to a lay person or point to a doc talking
> about this more?
>
> What does BBR do that's different from other algorithms? Why does it
> break the clock? Before BBR, was the clock the only way TCP did CC?
Put simply, BBR directly probes for the capacity and baseline latency of the path, and picks a send rate (implemented using pacing) and a failsafe cwnd to match that. The bandwidth probe looks at the rate of returning acks, so in fact it's still using the ack-clock mechanism, it's just connected much less directly to the send rate than before.
Other TCPs can use pacing as well. In that case the cwnd and RTT estimate are calculated in the normal way, and the send rate (for pacing) is calculated from those. It prevents a sudden opening of the receive or congestion windows from causing a huge burst which would tend to swamp buffers.
- Jonathan Morton
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] [Bloat] the future belongs to pacing
2020-07-04 17:52 ` [Make-wifi-fast] " Daniel Sterling
2020-07-04 18:02 ` Jonathan Morton
@ 2020-07-04 18:29 ` Sebastian Moeller
2020-07-05 6:10 ` Matt Mathis
1 sibling, 1 reply; 21+ messages in thread
From: Sebastian Moeller @ 2020-07-04 18:29 UTC (permalink / raw)
To: Daniel Sterling
Cc: Matt Mathis, Make-Wifi-fast, Carlo Augusto Grazia, bloat, jamshid
> On Jul 4, 2020, at 19:52, Daniel Sterling <sterling.daniel@gmail.com> wrote:
>
> On Sat, Jul 4, 2020 at 1:29 PM Matt Mathis via Bloat
> <bloat@lists.bufferbloat.net> wrote:
> "pacing is inevitable, because it saves large content providers money
> (more efficient use of the most expensive silicon in the data center,
> the switch buffer memory), however to use pacing we walk away from 30
> years of experience with TCP self clock"
>
> at the risk of asking w/o doing any research,
>
> could someone explain this to a lay person or point to a doc talking
> about this more?
>
> What does BBR do that's different from other algorithms?
Well, it does not believe the network (blindly), that is currently it ignores both ECN marks and (sparse) drops as signs of congestion, instead it uses its own rate estimates to set its send rate and cyclically will re-assess its rate estimate. Sufficiently severe drops will be honored. IMHO a somewhat risky approach, that works reasonably well, as often sparse drops are not real signs of congestion but just random drops of say a wifi link (that said, these drops on wifi typically also cause painful latency spikes as wifi often takes heroic measures in attempting retransmitting for several 100s of milliseconds).
> Why does it
> break the clock?
One can argue that there is no real clock to break. TCP gates the release on new packets on the reception of ACK signals from the receiver, this is only a clock, if one does not really care for the equi-temporal period property of a real clock. But for better or worse that is the term that is used. IMHO (and I really am calling this from way out in the left-field) gating would be a better term, but changing the nomenclature probably is not an option at this point.
> Before BBR, was the clock the only way TCP did CC?
No, TCP also interpreted a drop (or rather 3 duplicated ACKs) as signal of congestion and hit the brakes, by halving the congestion window (the amount of data that could be in flight unacknowledged, which roughly correlates with the send rate, if averaged over long enough time windows). BBR explicitly does not do this unless it really is convinced that someone dropped multiple packets purposefully to signal congestion.
In practice it works rather well, in theory it could do with at least an rfc3168 compliant response to ECN marks (which an AQM uses to explicitly signal congestion, unlike a drop an ECN mark is really unambiguous, some hop on the way "told" the flow slow down).
>
> Also,
>
> I have UBNT "Amplifi" HD wifi units in my house. (HD units only; none
> of the "mesh" units. Just HD units connected either via wifi or
> wired.) Empirically, I've found that in order to reduce latency, I
> need to set cake to about 1/4 of the total possible wifi speed;
> otherwise if a large download comes down from my internet link, that
> flow causes latency.
>
> That is, if I'm using 5ghz at 20mhz channel width, I need to set
> cake's bandwidth argument to 40mbits to prevent video streams /
> downloads from impacting latency for any other stream. This is w/o any
> categorization at all; no packet marking based on port or anything
> else; cake set to "best effort".
>
> Anything higher and when a large amount of data comes thru, something
> (assumedly the buffer in the Amplifi HD units) causes 100s of
> milliseconds of latency.
>
> Can anyone speak to how BBR would react to this? My ISP is full
> gigabit; but cake is going to drop a lot of packets as it throttles
> that down to 40mbit before it sends the packets to the wifi AP.
>
> Thanks,
> Dan
> _______________________________________________
> Bloat mailing list
> Bloat@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/bloat
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Bloat] the future belongs to pacing
2020-07-04 18:29 ` Sebastian Moeller
@ 2020-07-05 6:10 ` Matt Mathis
2020-07-05 12:01 ` [Make-wifi-fast] " Sebastian Moeller
0 siblings, 1 reply; 21+ messages in thread
From: Matt Mathis @ 2020-07-05 6:10 UTC (permalink / raw)
To: Sebastian Moeller
Cc: Daniel Sterling, Make-Wifi-fast, Carlo Augusto Grazia, bloat,
Jamshid Mahdavi
[-- Attachment #1: Type: text/plain, Size: 5920 bytes --]
I strongly suggest that people (re)read VJ88 - I do every couple of years,
and still discover things that I overlooked on previous readings.
All of the negative comments about BBR and loss, ECN marks, or unfairness
to cubic were correct for BBRv1 but have been addressed in BBRv2.
My paper has a synopsis of BBR, which is intended to get people started.
See the references in the paper for more info:
[12] Neal Cardwell, Yuchung Cheng, C. Stephen Gunn, Soheil Hassas Yeganeh,
and Van Jacobson. 2016. BBR: Congestion-Based Congestion Control. Queue 14,
5, Pages 50 (October 2016). DOI: https://doi.org/10.1145/3012426.3022184
[13] Neal Cardwell, Yuchung Cheng, C. Stephen Gunn, Soheil Hassas Yeganeh,
and Van Jacobson. 2017. BBR: Congestion-Based Congestion Control. Commun.
ACM 60, 2 (January 2017), 58-66. DOI: https://doi.org/10.1145/3009824
[22] google/bbr. 2019. GitHub repository, retrieved
https://github.com/google/bbr
Key definitions: self clocked: data is triggered by ACKs. All screwy
packet and ACK scheduling in the network is reflected back into the network
on the next RTT.
Paced: data is transmitted on a timer, independent of ACK arrivals (as long
as the ACKs take less than twice the measured minRTT). Thus in bulk
transport there is little or no correlation between data transmissions and
events elsewhere in the network.
Clarification about my earlier WiFi comment: The BBRv1 WiFi fix missed
4.19 LTS, so bad results are "expected" for many distros. If you want to
do useful experiments, you must read https://groups.google.com/g/bbr-dev/ and
start from BBRv2 in [22].
Thanks,
--MM--
The best way to predict the future is to create it. - Alan Kay
We must not tolerate intolerance;
however our response must be carefully measured:
too strong would be hypocritical and risks spiraling out of
control;
too weak risks being mistaken for tacit approval.
On Sat, Jul 4, 2020 at 11:29 AM Sebastian Moeller <moeller0@gmx.de> wrote:
>
>
> > On Jul 4, 2020, at 19:52, Daniel Sterling <sterling.daniel@gmail.com>
> wrote:
> >
> > On Sat, Jul 4, 2020 at 1:29 PM Matt Mathis via Bloat
> > <bloat@lists.bufferbloat.net> wrote:
> > "pacing is inevitable, because it saves large content providers money
> > (more efficient use of the most expensive silicon in the data center,
> > the switch buffer memory), however to use pacing we walk away from 30
> > years of experience with TCP self clock"
> >
> > at the risk of asking w/o doing any research,
> >
> > could someone explain this to a lay person or point to a doc talking
> > about this more?
> >
> > What does BBR do that's different from other algorithms?
>
> Well, it does not believe the network (blindly), that is currently
> it ignores both ECN marks and (sparse) drops as signs of congestion,
> instead it uses its own rate estimates to set its send rate and cyclically
> will re-assess its rate estimate. Sufficiently severe drops will be
> honored. IMHO a somewhat risky approach, that works reasonably well, as
> often sparse drops are not real signs of congestion but just random drops
> of say a wifi link (that said, these drops on wifi typically also cause
> painful latency spikes as wifi often takes heroic measures in attempting
> retransmitting for several 100s of milliseconds).
>
>
> > Why does it
> > break the clock?
>
> One can argue that there is no real clock to break. TCP gates the
> release on new packets on the reception of ACK signals from the receiver,
> this is only a clock, if one does not really care for the equi-temporal
> period property of a real clock. But for better or worse that is the term
> that is used. IMHO (and I really am calling this from way out in the
> left-field) gating would be a better term, but changing the nomenclature
> probably is not an option at this point.
>
> > Before BBR, was the clock the only way TCP did CC?
>
> No, TCP also interpreted a drop (or rather 3 duplicated ACKs) as
> signal of congestion and hit the brakes, by halving the congestion window
> (the amount of data that could be in flight unacknowledged, which roughly
> correlates with the send rate, if averaged over long enough time windows).
> BBR explicitly does not do this unless it really is convinced that someone
> dropped multiple packets purposefully to signal congestion.
> In practice it works rather well, in theory it could do with at
> least an rfc3168 compliant response to ECN marks (which an AQM uses to
> explicitly signal congestion, unlike a drop an ECN mark is really
> unambiguous, some hop on the way "told" the flow slow down).
>
>
> >
> > Also,
> >
> > I have UBNT "Amplifi" HD wifi units in my house. (HD units only; none
> > of the "mesh" units. Just HD units connected either via wifi or
> > wired.) Empirically, I've found that in order to reduce latency, I
> > need to set cake to about 1/4 of the total possible wifi speed;
> > otherwise if a large download comes down from my internet link, that
> > flow causes latency.
> >
> > That is, if I'm using 5ghz at 20mhz channel width, I need to set
> > cake's bandwidth argument to 40mbits to prevent video streams /
> > downloads from impacting latency for any other stream. This is w/o any
> > categorization at all; no packet marking based on port or anything
> > else; cake set to "best effort".
> >
> > Anything higher and when a large amount of data comes thru, something
> > (assumedly the buffer in the Amplifi HD units) causes 100s of
> > milliseconds of latency.
> >
> > Can anyone speak to how BBR would react to this? My ISP is full
> > gigabit; but cake is going to drop a lot of packets as it throttles
> > that down to 40mbit before it sends the packets to the wifi AP.
> >
> > Thanks,
> > Dan
> > _______________________________________________
> > Bloat mailing list
> > Bloat@lists.bufferbloat.net
> > https://lists.bufferbloat.net/listinfo/bloat
>
>
[-- Attachment #2: Type: text/html, Size: 7535 bytes --]
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] [Bloat] the future belongs to pacing
2020-07-05 6:10 ` Matt Mathis
@ 2020-07-05 12:01 ` Sebastian Moeller
2020-07-05 17:07 ` Matt Mathis
2020-07-05 17:43 ` Michael Richardson
0 siblings, 2 replies; 21+ messages in thread
From: Sebastian Moeller @ 2020-07-05 12:01 UTC (permalink / raw)
To: Matt Mathis
Cc: Daniel Sterling, Make-Wifi-fast, Carlo Augusto Grazia, bloat,
Jamshid Mahdavi
Hi Matt,
> On Jul 5, 2020, at 08:10, Matt Mathis <mattmathis@google.com> wrote:
>
> I strongly suggest that people (re)read VJ88 - I do every couple of years, and still discover things that I overlooked on previous readings.
I promise to read it. And before I give the wrong impression and for what it is worth*, I consider BBR (even v1) an interesting and important evolutionary step and agree that "pacing" is a gentler approach then bursting a full CWN into a link.
>
> All of the negative comments about BBR and loss, ECN marks,
As far as I can tell, BBRv2 aims for a decidedly non-rfc3168 response to CE-marks. This IMHO is not a clear cut case of meaningfully addressing my ECN comment. In the light of efficiently using TOR? switch buffers efficiently, that kind of response might be defensible but it does not really address my remark about it being unfortunate that BBR ignores both immediate signals of congestion, (sparse) packet drops AND explicit CE marks, the proposed (dctcp-like) CE-response seems rather weak compared to the naive expectation of halving/80%-ing of the sending rate, no? BBRv2 as I understand it will happily run roughshod over any true rfc3168 AQM on the path, I do not have the numbers, but I am not fully convinced that typically the most significant throttling on a CDN to end-user path happens still inside the CDN's data center...
> or unfairness to cubic were correct for BBRv1 but have been addressed in BBRv2.
I am not sure that unfairness was brought up as an issue in this thread.
>
> My paper has a synopsis of BBR, which is intended to get people started. See the references in the paper for more info:
I will have a look at these as well... Thanks
Best Regards
Sebastian
*) Being from outside the field, probably not much...
>
> [12] Neal Cardwell, Yuchung Cheng, C. Stephen Gunn, Soheil Hassas Yeganeh, and Van Jacobson. 2016. BBR: Congestion-Based Congestion Control. Queue 14, 5, Pages 50 (October 2016). DOI: https://doi.org/10.1145/3012426.3022184
> [13] Neal Cardwell, Yuchung Cheng, C. Stephen Gunn, Soheil Hassas Yeganeh, and Van Jacobson. 2017. BBR: Congestion-Based Congestion Control. Commun. ACM 60, 2 (January 2017), 58-66. DOI: https://doi.org/10.1145/3009824
> [22] google/bbr. 2019. GitHub repository, retrieved https://github.com/google/bbr
>
> Key definitions: self clocked: data is triggered by ACKs. All screwy packet and ACK scheduling in the network is reflected back into the network on the next RTT.
>
> Paced: data is transmitted on a timer, independent of ACK arrivals (as long as the ACKs take less than twice the measured minRTT). Thus in bulk transport there is little or no correlation between data transmissions and events elsewhere in the network.
>
> Clarification about my earlier WiFi comment: The BBRv1 WiFi fix missed 4.19 LTS, so bad results are "expected" for many distros. If you want to do useful experiments, you must read https://groups.google.com/g/bbr-dev/ and start from BBRv2 in [22].
>
> Thanks,
> --MM--
> The best way to predict the future is to create it. - Alan Kay
>
> We must not tolerate intolerance;
> however our response must be carefully measured:
> too strong would be hypocritical and risks spiraling out of control;
> too weak risks being mistaken for tacit approval.
>
>
> On Sat, Jul 4, 2020 at 11:29 AM Sebastian Moeller <moeller0@gmx.de> wrote:
>
>
> > On Jul 4, 2020, at 19:52, Daniel Sterling <sterling.daniel@gmail.com> wrote:
> >
> > On Sat, Jul 4, 2020 at 1:29 PM Matt Mathis via Bloat
> > <bloat@lists.bufferbloat.net> wrote:
> > "pacing is inevitable, because it saves large content providers money
> > (more efficient use of the most expensive silicon in the data center,
> > the switch buffer memory), however to use pacing we walk away from 30
> > years of experience with TCP self clock"
> >
> > at the risk of asking w/o doing any research,
> >
> > could someone explain this to a lay person or point to a doc talking
> > about this more?
> >
> > What does BBR do that's different from other algorithms?
>
> Well, it does not believe the network (blindly), that is currently it ignores both ECN marks and (sparse) drops as signs of congestion, instead it uses its own rate estimates to set its send rate and cyclically will re-assess its rate estimate. Sufficiently severe drops will be honored. IMHO a somewhat risky approach, that works reasonably well, as often sparse drops are not real signs of congestion but just random drops of say a wifi link (that said, these drops on wifi typically also cause painful latency spikes as wifi often takes heroic measures in attempting retransmitting for several 100s of milliseconds).
>
>
> > Why does it
> > break the clock?
>
> One can argue that there is no real clock to break. TCP gates the release on new packets on the reception of ACK signals from the receiver, this is only a clock, if one does not really care for the equi-temporal period property of a real clock. But for better or worse that is the term that is used. IMHO (and I really am calling this from way out in the left-field) gating would be a better term, but changing the nomenclature probably is not an option at this point.
>
> > Before BBR, was the clock the only way TCP did CC?
>
> No, TCP also interpreted a drop (or rather 3 duplicated ACKs) as signal of congestion and hit the brakes, by halving the congestion window (the amount of data that could be in flight unacknowledged, which roughly correlates with the send rate, if averaged over long enough time windows). BBR explicitly does not do this unless it really is convinced that someone dropped multiple packets purposefully to signal congestion.
> In practice it works rather well, in theory it could do with at least an rfc3168 compliant response to ECN marks (which an AQM uses to explicitly signal congestion, unlike a drop an ECN mark is really unambiguous, some hop on the way "told" the flow slow down).
>
>
> >
> > Also,
> >
> > I have UBNT "Amplifi" HD wifi units in my house. (HD units only; none
> > of the "mesh" units. Just HD units connected either via wifi or
> > wired.) Empirically, I've found that in order to reduce latency, I
> > need to set cake to about 1/4 of the total possible wifi speed;
> > otherwise if a large download comes down from my internet link, that
> > flow causes latency.
> >
> > That is, if I'm using 5ghz at 20mhz channel width, I need to set
> > cake's bandwidth argument to 40mbits to prevent video streams /
> > downloads from impacting latency for any other stream. This is w/o any
> > categorization at all; no packet marking based on port or anything
> > else; cake set to "best effort".
> >
> > Anything higher and when a large amount of data comes thru, something
> > (assumedly the buffer in the Amplifi HD units) causes 100s of
> > milliseconds of latency.
> >
> > Can anyone speak to how BBR would react to this? My ISP is full
> > gigabit; but cake is going to drop a lot of packets as it throttles
> > that down to 40mbit before it sends the packets to the wifi AP.
> >
> > Thanks,
> > Dan
> > _______________________________________________
> > Bloat mailing list
> > Bloat@lists.bufferbloat.net
> > https://lists.bufferbloat.net/listinfo/bloat
>
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Bloat] the future belongs to pacing
2020-07-05 12:01 ` [Make-wifi-fast] " Sebastian Moeller
@ 2020-07-05 17:07 ` Matt Mathis
2020-07-05 17:29 ` [Make-wifi-fast] " Sebastian Moeller
2020-07-05 17:43 ` Michael Richardson
1 sibling, 1 reply; 21+ messages in thread
From: Matt Mathis @ 2020-07-05 17:07 UTC (permalink / raw)
To: Sebastian Moeller
Cc: Daniel Sterling, Make-Wifi-fast, Carlo Augusto Grazia, bloat,
Jamshid Mahdavi
[-- Attachment #1: Type: text/plain, Size: 8716 bytes --]
The consensus in the standards community is that 3168 ECN is not so useful
- too late to protect small queues, too much signal (gain) to use it
to hint at future congestion. The point of non-3168 ECN is to permit
earlier gentle signalling. I am not following the ECN conversation, but
as stated at recent IETFs, the ECN code in BBRv2 is really a placeholder,
and when the ECN community comes to consensus on a standard, I would expect
BBR to do the standard.
Tor has its own special challenge with traffic management. Easy solutions
leak information, secure solutions are very hard. Remember to be useful,
the ECN bits need to be in the clear.
Thanks,
--MM--
The best way to predict the future is to create it. - Alan Kay
We must not tolerate intolerance;
however our response must be carefully measured:
too strong would be hypocritical and risks spiraling out of
control;
too weak risks being mistaken for tacit approval.
On Sun, Jul 5, 2020 at 5:01 AM Sebastian Moeller <moeller0@gmx.de> wrote:
> Hi Matt,
>
>
>
> > On Jul 5, 2020, at 08:10, Matt Mathis <mattmathis@google.com> wrote:
> >
> > I strongly suggest that people (re)read VJ88 - I do every couple of
> years, and still discover things that I overlooked on previous readings.
>
> I promise to read it. And before I give the wrong impression and
> for what it is worth*, I consider BBR (even v1) an interesting and
> important evolutionary step and agree that "pacing" is a gentler approach
> then bursting a full CWN into a link.
>
>
> >
> > All of the negative comments about BBR and loss, ECN marks,
>
> As far as I can tell, BBRv2 aims for a decidedly non-rfc3168
> response to CE-marks. This IMHO is not a clear cut case of meaningfully
> addressing my ECN comment. In the light of efficiently using TOR? switch
> buffers efficiently, that kind of response might be defensible but it does
> not really address my remark about it being unfortunate that BBR ignores
> both immediate signals of congestion, (sparse) packet drops AND explicit CE
> marks, the proposed (dctcp-like) CE-response seems rather weak compared to
> the naive expectation of halving/80%-ing of the sending rate, no? BBRv2 as
> I understand it will happily run roughshod over any true rfc3168 AQM on the
> path, I do not have the numbers, but I am not fully convinced that
> typically the most significant throttling on a CDN to end-user path happens
> still inside the CDN's data center...
>
>
> > or unfairness to cubic were correct for BBRv1 but have been addressed in
> BBRv2.
>
> I am not sure that unfairness was brought up as an issue in this
> thread.
>
>
> >
> > My paper has a synopsis of BBR, which is intended to get people
> started. See the references in the paper for more info:
>
> I will have a look at these as well... Thanks
>
> Best Regards
> Sebastian
>
> *) Being from outside the field, probably not much...
>
> >
> > [12] Neal Cardwell, Yuchung Cheng, C. Stephen Gunn, Soheil Hassas
> Yeganeh, and Van Jacobson. 2016. BBR: Congestion-Based Congestion Control.
> Queue 14, 5, Pages 50 (October 2016). DOI:
> https://doi.org/10.1145/3012426.3022184
> > [13] Neal Cardwell, Yuchung Cheng, C. Stephen Gunn, Soheil Hassas
> Yeganeh, and Van Jacobson. 2017. BBR: Congestion-Based Congestion Control.
> Commun. ACM 60, 2 (January 2017), 58-66. DOI:
> https://doi.org/10.1145/3009824
> > [22] google/bbr. 2019. GitHub repository, retrieved
> https://github.com/google/bbr
> >
> > Key definitions: self clocked: data is triggered by ACKs. All screwy
> packet and ACK scheduling in the network is reflected back into the network
> on the next RTT.
> >
> > Paced: data is transmitted on a timer, independent of ACK arrivals (as
> long as the ACKs take less than twice the measured minRTT). Thus in bulk
> transport there is little or no correlation between data transmissions and
> events elsewhere in the network.
> >
> > Clarification about my earlier WiFi comment: The BBRv1 WiFi fix missed
> 4.19 LTS, so bad results are "expected" for many distros. If you want to
> do useful experiments, you must read https://groups.google.com/g/bbr-dev/
> and start from BBRv2 in [22].
> >
> > Thanks,
> > --MM--
> > The best way to predict the future is to create it. - Alan Kay
> >
> > We must not tolerate intolerance;
> > however our response must be carefully measured:
> > too strong would be hypocritical and risks spiraling out of
> control;
> > too weak risks being mistaken for tacit approval.
> >
> >
> > On Sat, Jul 4, 2020 at 11:29 AM Sebastian Moeller <moeller0@gmx.de>
> wrote:
> >
> >
> > > On Jul 4, 2020, at 19:52, Daniel Sterling <sterling.daniel@gmail.com>
> wrote:
> > >
> > > On Sat, Jul 4, 2020 at 1:29 PM Matt Mathis via Bloat
> > > <bloat@lists.bufferbloat.net> wrote:
> > > "pacing is inevitable, because it saves large content providers money
> > > (more efficient use of the most expensive silicon in the data center,
> > > the switch buffer memory), however to use pacing we walk away from 30
> > > years of experience with TCP self clock"
> > >
> > > at the risk of asking w/o doing any research,
> > >
> > > could someone explain this to a lay person or point to a doc talking
> > > about this more?
> > >
> > > What does BBR do that's different from other algorithms?
> >
> > Well, it does not believe the network (blindly), that is
> currently it ignores both ECN marks and (sparse) drops as signs of
> congestion, instead it uses its own rate estimates to set its send rate and
> cyclically will re-assess its rate estimate. Sufficiently severe drops will
> be honored. IMHO a somewhat risky approach, that works reasonably well, as
> often sparse drops are not real signs of congestion but just random drops
> of say a wifi link (that said, these drops on wifi typically also cause
> painful latency spikes as wifi often takes heroic measures in attempting
> retransmitting for several 100s of milliseconds).
> >
> >
> > > Why does it
> > > break the clock?
> >
> > One can argue that there is no real clock to break. TCP gates
> the release on new packets on the reception of ACK signals from the
> receiver, this is only a clock, if one does not really care for the
> equi-temporal period property of a real clock. But for better or worse that
> is the term that is used. IMHO (and I really am calling this from way out
> in the left-field) gating would be a better term, but changing the
> nomenclature probably is not an option at this point.
> >
> > > Before BBR, was the clock the only way TCP did CC?
> >
> > No, TCP also interpreted a drop (or rather 3 duplicated ACKs) as
> signal of congestion and hit the brakes, by halving the congestion window
> (the amount of data that could be in flight unacknowledged, which roughly
> correlates with the send rate, if averaged over long enough time windows).
> BBR explicitly does not do this unless it really is convinced that someone
> dropped multiple packets purposefully to signal congestion.
> > In practice it works rather well, in theory it could do with at
> least an rfc3168 compliant response to ECN marks (which an AQM uses to
> explicitly signal congestion, unlike a drop an ECN mark is really
> unambiguous, some hop on the way "told" the flow slow down).
> >
> >
> > >
> > > Also,
> > >
> > > I have UBNT "Amplifi" HD wifi units in my house. (HD units only; none
> > > of the "mesh" units. Just HD units connected either via wifi or
> > > wired.) Empirically, I've found that in order to reduce latency, I
> > > need to set cake to about 1/4 of the total possible wifi speed;
> > > otherwise if a large download comes down from my internet link, that
> > > flow causes latency.
> > >
> > > That is, if I'm using 5ghz at 20mhz channel width, I need to set
> > > cake's bandwidth argument to 40mbits to prevent video streams /
> > > downloads from impacting latency for any other stream. This is w/o any
> > > categorization at all; no packet marking based on port or anything
> > > else; cake set to "best effort".
> > >
> > > Anything higher and when a large amount of data comes thru, something
> > > (assumedly the buffer in the Amplifi HD units) causes 100s of
> > > milliseconds of latency.
> > >
> > > Can anyone speak to how BBR would react to this? My ISP is full
> > > gigabit; but cake is going to drop a lot of packets as it throttles
> > > that down to 40mbit before it sends the packets to the wifi AP.
> > >
> > > Thanks,
> > > Dan
> > > _______________________________________________
> > > Bloat mailing list
> > > Bloat@lists.bufferbloat.net
> > > https://lists.bufferbloat.net/listinfo/bloat
> >
>
>
[-- Attachment #2: Type: text/html, Size: 10935 bytes --]
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] [Bloat] the future belongs to pacing
2020-07-05 17:07 ` Matt Mathis
@ 2020-07-05 17:29 ` Sebastian Moeller
0 siblings, 0 replies; 21+ messages in thread
From: Sebastian Moeller @ 2020-07-05 17:29 UTC (permalink / raw)
To: Matt Mathis
Cc: Daniel Sterling, Make-Wifi-fast, Carlo Augusto Grazia, bloat,
Jamshid Mahdavi
Hi Matt,
> On Jul 5, 2020, at 19:07, Matt Mathis <mattmathis@google.com> wrote:
>
> The consensus in the standards community is that 3168 ECN is not so useful - too late to protect small queues, too much signal (gain) to use it to hint at future congestion.
I follow the discussion in the tsw working group and believe I have a good overview of the state of the discussion. I also have gathered some experience in the bufferbloat effort to be able to realize that the L4S proposal is mostly based on wishful thinking than on solid engineering. But yes, the time seems ripe for 1/p-type congestion signaling, but how to do this seems an open question.
> The point of non-3168 ECN is to permit earlier gentle signalling. I am not following the ECN conversation, but as stated at recent IETFs, the ECN code in BBRv2 is really a placeholder, and when the ECN community comes to consensus on a standard, I would expect BBR to do the standard.
I respectfully argue that this is the wrong way around, first implement the current RFC standard aka rfc3168 and only if there is a new standard switch over to that. ATM BBRv seems to bank on the L4S proposals to sail through the IETF completely ignoring the lack of critical testing the L4S design has been treated to.
>
> Tor has its own special challenge with traffic management.
Sorry, TOR was intended to expand to Top Of Rack, which I assumed to be a common way to call the devices that house "the most expensive silicon in the data center, the switch buffer memory". I apologize for not being clear.
> Easy solutions leak information, secure solutions are very hard. Remember to be useful, the ECN bits need to be in the clear.
All good points, thanks, but more applicable to the onion router and to top of rack switches...
Best Regards
Sebastian
>
> Thanks,
> --MM--
> The best way to predict the future is to create it. - Alan Kay
>
> We must not tolerate intolerance;
> however our response must be carefully measured:
> too strong would be hypocritical and risks spiraling out of control;
> too weak risks being mistaken for tacit approval.
>
>
> On Sun, Jul 5, 2020 at 5:01 AM Sebastian Moeller <moeller0@gmx.de> wrote:
> Hi Matt,
>
>
>
> > On Jul 5, 2020, at 08:10, Matt Mathis <mattmathis@google.com> wrote:
> >
> > I strongly suggest that people (re)read VJ88 - I do every couple of years, and still discover things that I overlooked on previous readings.
>
> I promise to read it. And before I give the wrong impression and for what it is worth*, I consider BBR (even v1) an interesting and important evolutionary step and agree that "pacing" is a gentler approach then bursting a full CWN into a link.
>
>
> >
> > All of the negative comments about BBR and loss, ECN marks,
>
> As far as I can tell, BBRv2 aims for a decidedly non-rfc3168 response to CE-marks. This IMHO is not a clear cut case of meaningfully addressing my ECN comment. In the light of efficiently using TOR? switch buffers efficiently, that kind of response might be defensible but it does not really address my remark about it being unfortunate that BBR ignores both immediate signals of congestion, (sparse) packet drops AND explicit CE marks, the proposed (dctcp-like) CE-response seems rather weak compared to the naive expectation of halving/80%-ing of the sending rate, no? BBRv2 as I understand it will happily run roughshod over any true rfc3168 AQM on the path, I do not have the numbers, but I am not fully convinced that typically the most significant throttling on a CDN to end-user path happens still inside the CDN's data center...
>
>
> > or unfairness to cubic were correct for BBRv1 but have been addressed in BBRv2.
>
> I am not sure that unfairness was brought up as an issue in this thread.
>
>
> >
> > My paper has a synopsis of BBR, which is intended to get people started. See the references in the paper for more info:
>
> I will have a look at these as well... Thanks
>
> Best Regards
> Sebastian
>
> *) Being from outside the field, probably not much...
>
> >
> > [12] Neal Cardwell, Yuchung Cheng, C. Stephen Gunn, Soheil Hassas Yeganeh, and Van Jacobson. 2016. BBR: Congestion-Based Congestion Control. Queue 14, 5, Pages 50 (October 2016). DOI: https://doi.org/10.1145/3012426.3022184
> > [13] Neal Cardwell, Yuchung Cheng, C. Stephen Gunn, Soheil Hassas Yeganeh, and Van Jacobson. 2017. BBR: Congestion-Based Congestion Control. Commun. ACM 60, 2 (January 2017), 58-66. DOI: https://doi.org/10.1145/3009824
> > [22] google/bbr. 2019. GitHub repository, retrieved https://github.com/google/bbr
> >
> > Key definitions: self clocked: data is triggered by ACKs. All screwy packet and ACK scheduling in the network is reflected back into the network on the next RTT.
> >
> > Paced: data is transmitted on a timer, independent of ACK arrivals (as long as the ACKs take less than twice the measured minRTT). Thus in bulk transport there is little or no correlation between data transmissions and events elsewhere in the network.
> >
> > Clarification about my earlier WiFi comment: The BBRv1 WiFi fix missed 4.19 LTS, so bad results are "expected" for many distros. If you want to do useful experiments, you must read https://groups.google.com/g/bbr-dev/ and start from BBRv2 in [22].
> >
> > Thanks,
> > --MM--
> > The best way to predict the future is to create it. - Alan Kay
> >
> > We must not tolerate intolerance;
> > however our response must be carefully measured:
> > too strong would be hypocritical and risks spiraling out of control;
> > too weak risks being mistaken for tacit approval.
> >
> >
> > On Sat, Jul 4, 2020 at 11:29 AM Sebastian Moeller <moeller0@gmx.de> wrote:
> >
> >
> > > On Jul 4, 2020, at 19:52, Daniel Sterling <sterling.daniel@gmail.com> wrote:
> > >
> > > On Sat, Jul 4, 2020 at 1:29 PM Matt Mathis via Bloat
> > > <bloat@lists.bufferbloat.net> wrote:
> > > "pacing is inevitable, because it saves large content providers money
> > > (more efficient use of the most expensive silicon in the data center,
> > > the switch buffer memory), however to use pacing we walk away from 30
> > > years of experience with TCP self clock"
> > >
> > > at the risk of asking w/o doing any research,
> > >
> > > could someone explain this to a lay person or point to a doc talking
> > > about this more?
> > >
> > > What does BBR do that's different from other algorithms?
> >
> > Well, it does not believe the network (blindly), that is currently it ignores both ECN marks and (sparse) drops as signs of congestion, instead it uses its own rate estimates to set its send rate and cyclically will re-assess its rate estimate. Sufficiently severe drops will be honored. IMHO a somewhat risky approach, that works reasonably well, as often sparse drops are not real signs of congestion but just random drops of say a wifi link (that said, these drops on wifi typically also cause painful latency spikes as wifi often takes heroic measures in attempting retransmitting for several 100s of milliseconds).
> >
> >
> > > Why does it
> > > break the clock?
> >
> > One can argue that there is no real clock to break. TCP gates the release on new packets on the reception of ACK signals from the receiver, this is only a clock, if one does not really care for the equi-temporal period property of a real clock. But for better or worse that is the term that is used. IMHO (and I really am calling this from way out in the left-field) gating would be a better term, but changing the nomenclature probably is not an option at this point.
> >
> > > Before BBR, was the clock the only way TCP did CC?
> >
> > No, TCP also interpreted a drop (or rather 3 duplicated ACKs) as signal of congestion and hit the brakes, by halving the congestion window (the amount of data that could be in flight unacknowledged, which roughly correlates with the send rate, if averaged over long enough time windows). BBR explicitly does not do this unless it really is convinced that someone dropped multiple packets purposefully to signal congestion.
> > In practice it works rather well, in theory it could do with at least an rfc3168 compliant response to ECN marks (which an AQM uses to explicitly signal congestion, unlike a drop an ECN mark is really unambiguous, some hop on the way "told" the flow slow down).
> >
> >
> > >
> > > Also,
> > >
> > > I have UBNT "Amplifi" HD wifi units in my house. (HD units only; none
> > > of the "mesh" units. Just HD units connected either via wifi or
> > > wired.) Empirically, I've found that in order to reduce latency, I
> > > need to set cake to about 1/4 of the total possible wifi speed;
> > > otherwise if a large download comes down from my internet link, that
> > > flow causes latency.
> > >
> > > That is, if I'm using 5ghz at 20mhz channel width, I need to set
> > > cake's bandwidth argument to 40mbits to prevent video streams /
> > > downloads from impacting latency for any other stream. This is w/o any
> > > categorization at all; no packet marking based on port or anything
> > > else; cake set to "best effort".
> > >
> > > Anything higher and when a large amount of data comes thru, something
> > > (assumedly the buffer in the Amplifi HD units) causes 100s of
> > > milliseconds of latency.
> > >
> > > Can anyone speak to how BBR would react to this? My ISP is full
> > > gigabit; but cake is going to drop a lot of packets as it throttles
> > > that down to 40mbit before it sends the packets to the wifi AP.
> > >
> > > Thanks,
> > > Dan
> > > _______________________________________________
> > > Bloat mailing list
> > > Bloat@lists.bufferbloat.net
> > > https://lists.bufferbloat.net/listinfo/bloat
> >
>
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] [Bloat] the future belongs to pacing
2020-07-05 12:01 ` [Make-wifi-fast] " Sebastian Moeller
2020-07-05 17:07 ` Matt Mathis
@ 2020-07-05 17:43 ` Michael Richardson
2020-07-05 18:09 ` Stephen Hemminger
1 sibling, 1 reply; 21+ messages in thread
From: Michael Richardson @ 2020-07-05 17:43 UTC (permalink / raw)
To: Sebastian Moeller, Matt Mathis, Carlo Augusto Grazia,
Jamshid Mahdavi, bloat, Make-Wifi-fast
[-- Attachment #1: Type: text/plain, Size: 1115 bytes --]
Sebastian Moeller <moeller0@gmx.de> wrote:
> of the sending rate, no? BBRv2 as I understand it will happily run
> roughshod over any true rfc3168 AQM on the path, I do not have the
> numbers, but I am not fully convinced that typically the most
> significant throttling on a CDN to end-user path happens still inside
> the CDN's data center...
That's an interesting claim. I'm in no position to defend or refute it.
If it's true, though, it suggests some interesting solutions, because one can
more easily establish trust relationships within the data-center.
I'm specifically imagining a clock signal from the Top-of-Rack switch to the
senders.
Actually, it's not a clock, so much as a automotive style timing shaft
running down through all the 1-U servers, with fine vernier adjustments :-)
I'm also certain we've seen such technology before.
--
] Never tell me the odds! | ipv6 mesh networks [
] Michael Richardson, Sandelman Software Works | IoT architect [
] mcr@sandelman.ca http://www.sandelman.ca/ | ruby on rails [
[-- Attachment #2: signature.asc --]
[-- Type: application/pgp-signature, Size: 487 bytes --]
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] [Bloat] the future belongs to pacing
2020-07-05 17:43 ` Michael Richardson
@ 2020-07-05 18:09 ` Stephen Hemminger
2020-07-05 18:13 ` Jonathan Morton
0 siblings, 1 reply; 21+ messages in thread
From: Stephen Hemminger @ 2020-07-05 18:09 UTC (permalink / raw)
To: Michael Richardson
Cc: Sebastian Moeller, Matt Mathis, Carlo Augusto Grazia,
Jamshid Mahdavi, bloat, Make-Wifi-fast
[-- Attachment #1: Type: text/plain, Size: 1517 bytes --]
On Sun, 05 Jul 2020 13:43:27 -0400
Michael Richardson <mcr@sandelman.ca> wrote:
> Sebastian Moeller <moeller0@gmx.de> wrote:
> > of the sending rate, no? BBRv2 as I understand it will happily run
> > roughshod over any true rfc3168 AQM on the path, I do not have the
> > numbers, but I am not fully convinced that typically the most
> > significant throttling on a CDN to end-user path happens still inside
> > the CDN's data center...
>
> That's an interesting claim. I'm in no position to defend or refute it.
>
> If it's true, though, it suggests some interesting solutions, because one can
> more easily establish trust relationships within the data-center.
>
> I'm specifically imagining a clock signal from the Top-of-Rack switch to the
> senders.
>
> Actually, it's not a clock, so much as a automotive style timing shaft
> running down through all the 1-U servers, with fine vernier adjustments :-)
> I'm also certain we've seen such technology before.
>
> --
> ] Never tell me the odds! | ipv6 mesh networks [
> ] Michael Richardson, Sandelman Software Works | IoT architect [
> ] mcr@sandelman.ca http://www.sandelman.ca/ | ruby on rails [
>
>
>
>
I keep wondering how BBR will respond to intermediaries that aggregate packets.
At higher speeds, won't packet trains happen and would it not get confused
by this? Or is its measurement interval long enough that it doesn't matter.
[-- Attachment #2: OpenPGP digital signature --]
[-- Type: application/pgp-signature, Size: 833 bytes --]
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] [Bloat] the future belongs to pacing
2020-07-05 18:09 ` Stephen Hemminger
@ 2020-07-05 18:13 ` Jonathan Morton
2020-07-05 23:06 ` Matt Mathis
0 siblings, 1 reply; 21+ messages in thread
From: Jonathan Morton @ 2020-07-05 18:13 UTC (permalink / raw)
To: Stephen Hemminger
Cc: Michael Richardson, Carlo Augusto Grazia, Make-Wifi-fast,
Jamshid Mahdavi, bloat
> On 5 Jul, 2020, at 9:09 pm, Stephen Hemminger <stephen@networkplumber.org> wrote:
>
> I keep wondering how BBR will respond to intermediaries that aggregate packets.
> At higher speeds, won't packet trains happen and would it not get confused
> by this? Or is its measurement interval long enough that it doesn't matter.
Up-thread, there was mention of patches related to wifi. Aggregation is precisely one of the things that would address. I should note that the brief description I gave glossed over a lot of fine details of BBR's implementation, which include careful filtering and conditioning of the data it gathers about the network path.
I'm not altogether a fan of such complexity.
- Jonathan Morton
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Bloat] the future belongs to pacing
2020-07-05 18:13 ` Jonathan Morton
@ 2020-07-05 23:06 ` Matt Mathis
0 siblings, 0 replies; 21+ messages in thread
From: Matt Mathis @ 2020-07-05 23:06 UTC (permalink / raw)
To: Jonathan Morton
Cc: Stephen Hemminger, Make-Wifi-fast, Carlo Augusto Grazia, bloat,
Jamshid Mahdavi
[-- Attachment #1: Type: text/plain, Size: 1806 bytes --]
What the complexity buys you is that BBRs metrics max_BW, min_RTT, and the
ACK aggregation/batching metrics are actual parameters of the network, and
observable with passive instrumentation of the packet streams. Traditional
CC is a collection of heuristics to estimate cwnd, which has a clear
interpretation in terms of action (when to send), but the optimal cwnd
can't easily be observed from the packet stream.
I think this alone will have impact, in terms of being able to reason about
CC behaviors.
Thanks,
--MM--
The best way to predict the future is to create it. - Alan Kay
We must not tolerate intolerance;
however our response must be carefully measured:
too strong would be hypocritical and risks spiraling out of
control;
too weak risks being mistaken for tacit approval.
On Sun, Jul 5, 2020 at 11:13 AM Jonathan Morton <chromatix99@gmail.com>
wrote:
> > On 5 Jul, 2020, at 9:09 pm, Stephen Hemminger <
> stephen@networkplumber.org> wrote:
> >
> > I keep wondering how BBR will respond to intermediaries that aggregate
> packets.
> > At higher speeds, won't packet trains happen and would it not get
> confused
> > by this? Or is its measurement interval long enough that it doesn't
> matter.
>
> Up-thread, there was mention of patches related to wifi. Aggregation is
> precisely one of the things that would address. I should note that the
> brief description I gave glossed over a lot of fine details of BBR's
> implementation, which include careful filtering and conditioning of the
> data it gathers about the network path.
>
> I'm not altogether a fan of such complexity.
>
> - Jonathan Morton
>
> _______________________________________________
> Bloat mailing list
> Bloat@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/bloat
>
[-- Attachment #2: Type: text/html, Size: 2699 bytes --]
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] the future belongs to pacing
2019-12-13 21:25 ` [Make-wifi-fast] the future belongs to pacing Dave Taht
2020-07-04 17:29 ` [Bloat] " Matt Mathis
[not found] ` <mailman.763.1593883755.24343.bloat@lists.bufferbloat.net>
@ 2020-07-06 14:08 ` Luca Muscariello
2020-07-06 14:14 ` [Make-wifi-fast] [Bloat] " Daniel Sterling
2 siblings, 1 reply; 21+ messages in thread
From: Luca Muscariello @ 2020-07-06 14:08 UTC (permalink / raw)
To: Dave Taht; +Cc: Carlo Augusto Grazia, Make-Wifi-fast, jamshid, bloat
[-- Attachment #1: Type: text/plain, Size: 3137 bytes --]
It is not surprising that BBR comes from Van who's also designed and
implemented pathchar.
I liked reading the paper when it was published and it has the merit to be
simple to read
for a large audience.
I agree very much on the title as bang-bang congestion control (not only
AIMD) could be
deprecated entirely by measurement based approaches like BBR.
In bang-bang cc the sending rate is obtained by a root-finding algorithm
(gradient based) that
is fed by measurements of congestion (queue, loss, latency), whereas in BBR
the sending rate is
an (almost?) explicit function of the measured quantities.
In theory both approaches work, but for the former we have seen a
proliferation of root-finding algorithms
for wireless, large BDP networks, small BDP network, satellite, cellular,
shared-media, non-shared media and more.
Selection of the right one is a question of tuning, which is extremely
complex and static.
If BBR can fix that by having a unique model for all these cases that would
make deprecation, as intended in the paper,
likely to happen.
On Fri, Dec 13, 2019 at 10:25 PM Dave Taht <dave.taht@gmail.com> wrote:
> and everything we know about the tcp macroscopic model, is obsolete,
> according to a provocative paper by matt mathis and Jamshid Mahdavi
> in sigcomm.
>
> https://ccronline.sigcomm.org/wp-content/uploads/2019/10/acmdl19-323.pdf
>
>
>
>
> On Fri, Dec 13, 2019 at 1:05 PM Carlo Augusto Grazia
> <carloaugusto.grazia@unimore.it> wrote:
> >
> > Hi Dave,
> > thank you for your email!
> > Toke told me about AQL a couple of weeks ago, I definitely want to test
> it ASAP.
> > BBR struggles a lot on Wi-Fi interfaces (ones with aggregation) with
> kernel 4.14 & 4.19.
> > Anyway, it seems that with BBRv2 on new kernels this problem does not
> exist anymore.
> >
> > Best regards
> > Carlo
> >
> > Il giorno ven 13 dic 2019 alle 20:54 Dave Taht <dave.taht@gmail.com> ha
> scritto:
> >>
> >> https://sci-hub.tw/10.1109/WiMOB.2019.8923418
> >>
> >> It predates the aql work, but the bbr result is puzzling.
> >>
> >>
> >> --
> >> Make Music, Not War
> >>
> >> Dave Täht
> >> CTO, TekLibre, LLC
> >> http://www.teklibre.com
> >> Tel: 1-831-435-0729
> >
> > --
> > --------------------------------------------------------------------
> > Carlo Augusto Grazia, Ph. D.
> > Assistant Professor
> > --------------------------------------------------------------------
> > Dept. of Engineering "Enzo Ferrari"
> > University of Modena and Reggio Emilia
> > Via Pietro Vivarelli, 10/1 - 41125 - Modena - Italy
> > Building 26, floor 2, room 28
> > Tel.: +39-059-2056323
> > email: carloaugusto.grazia@unimore.it
> > Link to my personal home page here
> > --------------------------------------------------------------------
>
>
>
> --
> Make Music, Not War
>
> Dave Täht
> CTO, TekLibre, LLC
> http://www.teklibre.com
> Tel: 1-831-435-0729
> _______________________________________________
> Make-wifi-fast mailing list
> Make-wifi-fast@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/make-wifi-fast
[-- Attachment #2: Type: text/html, Size: 5621 bytes --]
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] [Bloat] the future belongs to pacing
2020-07-06 14:08 ` [Make-wifi-fast] " Luca Muscariello
@ 2020-07-06 14:14 ` Daniel Sterling
2020-07-06 17:01 ` Toke Høiland-Jørgensen
0 siblings, 1 reply; 21+ messages in thread
From: Daniel Sterling @ 2020-07-06 14:14 UTC (permalink / raw)
To: Luca Muscariello
Cc: Dave Taht, Make-Wifi-fast, Carlo Augusto Grazia, jamshid, bloat
On Mon, Jul 6, 2020 at 10:09 AM Luca Muscariello <muscariello@ieee.org> wrote:
> If BBR can fix that by having a unique model for all these cases that would make deprecation, as intended in the paper,
> likely to happen.
Interesting! Thank you all for helping a layperson like me understand.
Obviously getting CC / latency control "correct" under wifi is a
difficult problem.
I am wondering if you (the experts) have confidence we can solve it --
that is, can end-users eventually see low latency by default with
standard gear?
Or are shared transmission mediums like wifi doomed to require large
buffers for throughput, which means low latency can't be something we
can have "out of the box" -- ? Is sacrificing throughput for latency
required for "always low" latency on wifi?
Thanks,
Dan
^ permalink raw reply [flat|nested] 21+ messages in thread
* Re: [Make-wifi-fast] [Bloat] the future belongs to pacing
2020-07-06 14:14 ` [Make-wifi-fast] [Bloat] " Daniel Sterling
@ 2020-07-06 17:01 ` Toke Høiland-Jørgensen
0 siblings, 0 replies; 21+ messages in thread
From: Toke Høiland-Jørgensen @ 2020-07-06 17:01 UTC (permalink / raw)
To: Daniel Sterling, Luca Muscariello
Cc: Make-Wifi-fast, Carlo Augusto Grazia, bloat, jamshid
Daniel Sterling <sterling.daniel@gmail.com> writes:
> On Mon, Jul 6, 2020 at 10:09 AM Luca Muscariello <muscariello@ieee.org> wrote:
>> If BBR can fix that by having a unique model for all these cases that would make deprecation, as intended in the paper,
>> likely to happen.
>
> Interesting! Thank you all for helping a layperson like me understand.
>
> Obviously getting CC / latency control "correct" under wifi is a
> difficult problem.
>
> I am wondering if you (the experts) have confidence we can solve it --
> that is, can end-users eventually see low latency by default with
> standard gear?
>
> Or are shared transmission mediums like wifi doomed to require large
> buffers for throughput, which means low latency can't be something we
> can have "out of the box" -- ? Is sacrificing throughput for latency
> required for "always low" latency on wifi?
To a certain extent, yes. However, this is orthogonal to the congestion
control being used: WiFi gets its high throughput due to large
aggregates (i.e., 802.11ac significantly increases the maximum allowed
aggregation size compared to 802.11n). Because there's a fixed overhead
for each transmission, the only way you can achieve the maximum
theoretical throughput is by filling the aggregates, and if you do that
while there are a lot of users contending for the medium, you will end
up hurting latency.
So really, the right thing to do in a busy network is to lower the
maximum aggregate size: if you have 20 stations waiting to transmit and
each only transmits for 1ms each instead of the maximum 4ms, you only
add 20ms of delay while waiting for the other stations, instead of 80ms
(best-case, not counting any backoff from collisions, queueing delay,
etc).
-Toke
^ permalink raw reply [flat|nested] 21+ messages in thread
end of thread, other threads:[~2020-07-06 17:01 UTC | newest]
Thread overview: 21+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2019-12-13 19:53 [Make-wifi-fast] IEEE 802.11n/ac Wireless Network Efficiency under different TCP Congestion Controls Dave Taht
2019-12-13 21:05 ` Carlo Augusto Grazia
2019-12-13 21:25 ` [Make-wifi-fast] the future belongs to pacing Dave Taht
2020-07-04 17:29 ` [Bloat] " Matt Mathis
[not found] ` <mailman.763.1593883755.24343.bloat@lists.bufferbloat.net>
2020-07-04 17:52 ` [Make-wifi-fast] " Daniel Sterling
2020-07-04 18:02 ` Jonathan Morton
2020-07-04 18:29 ` Sebastian Moeller
2020-07-05 6:10 ` Matt Mathis
2020-07-05 12:01 ` [Make-wifi-fast] " Sebastian Moeller
2020-07-05 17:07 ` Matt Mathis
2020-07-05 17:29 ` [Make-wifi-fast] " Sebastian Moeller
2020-07-05 17:43 ` Michael Richardson
2020-07-05 18:09 ` Stephen Hemminger
2020-07-05 18:13 ` Jonathan Morton
2020-07-05 23:06 ` Matt Mathis
2020-07-06 14:08 ` [Make-wifi-fast] " Luca Muscariello
2020-07-06 14:14 ` [Make-wifi-fast] [Bloat] " Daniel Sterling
2020-07-06 17:01 ` Toke Høiland-Jørgensen
2019-12-13 21:09 ` [Make-wifi-fast] IEEE 802.11n/ac Wireless Network Efficiency under different TCP Congestion Controls Holland, Jake
2019-12-13 21:18 ` Carlo Augusto Grazia
2019-12-13 23:06 ` Dave Taht
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox