<html><head></head><body>Hi Dave,<br><br>I thought it was accepted knowledge that inter-packet delays across the internet are not reliable? Why are paced chirps immune from that problem? Or asked differently after all noise filtering required for robust and reliable operation over the existing internet, is this still going to be noticeably faster than current slow-start? In spite of the slow in the name doubling every RTT is IMHO already a pretty aggressive growth function....<br><br>Sidenote, you really think the second paper nailed PCs coffin shut?<br><br><div class="gmail_quote">On 29 July 2022 16:55:42 CEST, Dave Taht <dave.taht@gmail.com> wrote:<blockquote class="gmail_quote" style="margin: 0pt 0pt 0pt 0.8ex; border-left: 1px solid rgb(204, 204, 204); padding-left: 1ex;">
<pre dir="auto" class="k9mail">To give credit where credit is due, "packet chirping" had been<br>explored before in the context<br>of the l4s early marking ecn effort:<br><br><a href="https://www.bobbriscoe.net/presents/1802paced-chirps/1802paced-chirps.pdf">https://www.bobbriscoe.net/presents/1802paced-chirps/1802paced-chirps.pdf</a><br><br>It died here: <a href="https://bobbriscoe.net/projects/netsvc_i-f/chirp_pfldnet10.pdf">https://bobbriscoe.net/projects/netsvc_i-f/chirp_pfldnet10.pdf</a><br>For now.<br><br>On Thu, Jul 28, 2022 at 6:50 AM Dave Taht <dave.taht@gmail.com> wrote:<br><blockquote class="gmail_quote" style="margin: 0pt 0pt 1ex 0.8ex; border-left: 1px solid #729fcf; padding-left: 1ex;"><br> thx for the comments everyone!<br><br> On Thu, Jul 28, 2022 at 3:16 AM Bjørn Ivar Teigen via Starlink<br> <starlink@lists.bufferbloat.net> wrote:<br><blockquote class="gmail_quote" style="margin: 0pt 0pt 1ex 0.8ex; border-left: 1px solid #ad7fa8; padding-left: 1ex;"><br> Very good point. Perhaps we can think of it as "at what point does delay equal loss?". As you say, too much latency (causing reordering for instance, or triggering an algorithm to smooth over missing data), is functionally equivalent to loss, and therefore buffering beyond that point is making things worse by delaying more traffic. The point at which this kicks in varies a lot between applications though, so some kind of classification might still make sense.<br><br> In a way, I think FQ_Codel does this classification implicitly by treating sparse and non-sparse flows differently.<br></blockquote><br> the implicit flow analysis of fq_codel paper toke did is here:<br> <a href="http://www.diva-portal.org/smash/get/diva2:1251687/FULLTEXT01.pdf">http://www.diva-portal.org/smash/get/diva2:1251687/FULLTEXT01.pdf</a><br> It's a really nice feature!, and helps a lot when also applied to wifi<br> station scheduling.<br><br> I have sometimes thought that increasing to quantum to account for two<br> paced packets in a row (at high rates) was a good idea,<br> other times having paced transports analyze the "beat frequency" of<br> sending packets through fq_codel vs a vs the ack flow characteristics<br> (for example, filtering) might be useful.<br><br> Imagine that instead of sending packets on a fixed but increasing<br> pacing schedule within an RTT thusly<br><br> PPPPPPPPPP # IW10 burst<br> PP PP PP PP PP # often about 24 packets in what we<br> think the RTT is<br><br> PP PP PP PP PP PP PP<br><br> PPPPPPPPPPPPPPPPPP<br><br> PPPPPPPPPPPPPPPPPPPPPPP stready buffering and ultimately a drop (and<br> yes this is inaccurate a model in a zillion ways, forgive me for<br> purposes of extrapolation in ascii text)<br><br> If instead...<br><br> You broke up the pacing within an RTT on an actual curve, selecting<br> some random segment out of PI as your actual starting point, say, at<br> 3.14596 here.<br><br> PPPPPP PPPPP PPP<br> PPPPP PPPPPPPP<br> PPPPPPPPP PPP PP<br><br> 3.14159265358979323846264338327950288419716939937510<br> 58209749445923078164062862089986280348253421170679<br> 82148086513282306647093844609550582231725359408128<br> 48111745028410270193852110555964462294895493038196<br> 44288109756659334461284756482337867831652712019091<br> 45648566923460348610454326648213393607260249141273<br> 72458700660631558817488152092096282925409171536436<br> 78925903600113305305488204665213841469519415116094<br> 33057270365759591953092186117381932611793105118548<br> 07446237996274956735188575272489122793818301194912<br> 98336733624406566430860213949463952247371907021798<br> 60943702770539217176293176752384674818467669405132<br> 00056812714526356082778577134275778960917363717872<br> 14684409012249534301465495853710507922796892589235<br> 42019956112129021960864034418159813629774771309960<br> 51870721134999999837297804995105973173281609631859<br> 50244594553469083026425223082533446850352619311881<br> 71010003137838752886587533208381420617177669147303<br> 59825349042875546873115956286388235378759375195778<br> 18577805321712268066130019278766111959092164201989<br><br> what could you learn?<br><br><br><blockquote class="gmail_quote" style="margin: 0pt 0pt 1ex 0.8ex; border-left: 1px solid #ad7fa8; padding-left: 1ex;"> - Bjørn Ivar<br><br> On Thu, 28 Jul 2022 at 11:55, Sebastian Moeller <moeller0@gmx.de> wrote:<br><blockquote class="gmail_quote" style="margin: 0pt 0pt 1ex 0.8ex; border-left: 1px solid #8ae234; padding-left: 1ex;"><br> Hi all,<br><br><br><blockquote class="gmail_quote" style="margin: 0pt 0pt 1ex 0.8ex; border-left: 1px solid #fcaf3e; padding-left: 1ex;"> On Jul 28, 2022, at 11:26, Bjørn Ivar Teigen via Starlink <starlink@lists.bufferbloat.net> wrote:<br><br> Hi everyone,<br><br> Interesting paper Dave, I've got a few thoughts:<br><br> I like the split into delay-sensitive and loss-sensitive data.<br></blockquote><br> However often real data is slightly different (e.g. not nicely either delay- or loss-sensitive)... e.g. for "real-time" games you have both delay and loss sensitivity (similarly for VoIP), however both can deal with occasional lost or delayed packets (if the delay is large enough to say be re-ordered with the temporally next data packet (voice sample in VoIP, server-tick update in games), that packet's data will likely not be evaluated at all). And large scale bulk downloads are both tolerant to delay and occasional loss. So if we think about a state space spanned by a delay and a loss-sensitivity axis, I predict most real traffic types will cluster somewhat around the diagonal (more or less closely).<br><br> About the rest of the paper I have nothing to contribute, since I did not spend the time to work though it.<br><br> Regards<br> Sebastian<br><br><br><br><blockquote class="gmail_quote" style="margin: 0pt 0pt 1ex 0.8ex; border-left: 1px solid #fcaf3e; padding-left: 1ex;"> Different applications can have different needs and this split allows a queuing algorithm to take those differences into account. Not the first time I've seen this kind of split, but the other one I've seen used M/M/1/k queues (document here: <a href="https://www.researchgate.net/publication/2452029_A_Queueing_Theory_Model_that_Enables_Control_of_Loss_and_Delay_of_Traffic_at_a_Network_Switch)">https://www.researchgate.net/publication/2452029_A_Queueing_Theory_Model_that_Enables_Control_of_Loss_and_Delay_of_Traffic_at_a_Network_Switch)</a><br><br> That said, the performance metrics are derived from the embedded Markov chain of the queuing system. This means the metrics are averages over *all of time*, and thus there can be shorter periods (seconds, minutes, hours) of much worse than average performance. Therefore the conclusions of the paper should be taken with a grain of salt in my opinion.<br><br> On Thu, 28 Jul 2022 at 10:45, Bless, Roland (TM) via Starlink <starlink@lists.bufferbloat.net> wrote:<br> Hi Dave,<br><br> IMHO the problem w.r.t the applicability of most models from<br> queueing theory is that they only work for load < 1, whereas<br> we are using the network with load values ~1 (i.e., around one) due to<br> congestion control feedback loops that drive the bottleneck link<br> to saturation (unless you consider application limited traffic sources).<br><br> To be fair there are queuing theory models that include packet loss (which is the case for the paper Dave is asking about here), and these can work perfectly well for load > 1. Agree about the CC feedback loops affecting the results though. Even if the distributions are general in the paper, they still assume samples are IID which is not true for real networks. Feedback loops make real traffic self-correlated, which makes the short periods of worse than average performance worse and more frequent than IID models might suggest.<br><br> Regards,<br> Bjørn Ivar<br><br><br> Regards,<br> Roland<br><br> On 27.07.22 at 17:34 Dave Taht via Starlink wrote:<br><blockquote class="gmail_quote" style="margin: 0pt 0pt 1ex 0.8ex; border-left: 1px solid #e9b96e; padding-left: 1ex;"> Occasionally I pass along a recent paper that I don't understand in<br> the hope that someone can enlighten me.<br> This is one of those occasions, where I am trying to leverage what I<br> understand of existing FQ-codel behaviors against real traffic.<br><br> <a href="https://www.hindawi.com/journals/mpe/2022/4539940/">https://www.hindawi.com/journals/mpe/2022/4539940/</a><br><br> Compared to the previous study on finite-buffer M/M/1 priority queues<br> with time and space priority, where service times are identical and<br> exponentially distributed for both types of traffic, in our model we<br> assume that service times are different and are generally distributed<br> for different types of traffic. As a result, our model is more<br> suitable for the performance analysis of communication systems<br> accommodating multiple types of traffic with different service-time<br> distributions. For the proposed queueing model, we derive the<br> queue-length distributions, loss probabilities, and mean waiting times<br> of both types of traffic, as well as the push-out probability of<br> delay-sensitive traffic.<br></blockquote><hr> Starlink mailing list<br> Starlink@lists.bufferbloat.net<br> <a href="https://lists.bufferbloat.net/listinfo/starlink">https://lists.bufferbloat.net/listinfo/starlink</a><br><br><br> --<br> Bjørn Ivar Teigen<br> Head of Research<br> +47 47335952 | bjorn@domos.no | www.domos.no<hr> Starlink mailing list<br> Starlink@lists.bufferbloat.net<br> <a href="https://lists.bufferbloat.net/listinfo/starlink">https://lists.bufferbloat.net/listinfo/starlink</a><br></blockquote><br></blockquote><br><br> --<br> Bjørn Ivar Teigen<br> Head of Research<br> +47 47335952 | bjorn@domos.no | www.domos.no<hr> Starlink mailing list<br> Starlink@lists.bufferbloat.net<br> <a href="https://lists.bufferbloat.net/listinfo/starlink">https://lists.bufferbloat.net/listinfo/starlink</a><br></blockquote><br><br><br> --<br> FQ World Domination pending: <a href="https://blog.cerowrt.org/post/state_of_fq_codel/">https://blog.cerowrt.org/post/state_of_fq_codel/</a><br> Dave Täht CEO, TekLibre, LLC<br></blockquote><br><br><br><div class="k9mail-signature">-- <br>FQ World Domination pending: <a href="https://blog.cerowrt.org/post/state_of_fq_codel/">https://blog.cerowrt.org/post/state_of_fq_codel/</a><br>Dave Täht CEO, TekLibre, LLC<br></div></pre></blockquote></div><div style='white-space: pre-wrap'><div class='k9mail-signature'>-- <br>Sent from my Android device with K-9 Mail. Please excuse my brevity.</div></div></body></html>