[Bloat] DSLReports Speed Test has latency measurement built-in

Sebastian Moeller moeller0 at gmx.de
Fri Apr 24 01:18:18 PDT 2015


Hi jb,

this looks great!

On Apr 23, 2015, at 12:08 , jb <justin at dslr.net> wrote:

> This is how I've changed the graph of latency under load per input from you guys.
> 
> Taken away log axis.
> 
> Put in two bands. Yellow starts at double the idle latency, and goes to 4x the idle latency
> red starts there, and goes to the top. No red shows if no bars reach into it.
> And no yellow band shows if no bars get into that zone.
> 
> Is it more descriptive?

	Mmmh, so the delay we see consists out of the delay caused by the distance to the server and the delay of the access technology, meaning the un-loaded latency can range from a few milliseconds to several 100s of milliseconds (for the poor sods behind a satellite link…). Any further latency developing under load should be independent of distance and access technology as those are already factored in the bade latency. In both the extreme cases multiples of the base-latency do not seem to be relevant measures of bloat, so I would like to argue that the yellow and the red zones should be based on fixed increments and not as a ratio of the base-latency. This is relevant as people on a slow/high-access-latency link have a much smaller tolerance for additional latency than people on a fast link if certain latency guarantees need to be met, and thresholds as a function of base-latency do not reflect this.
	Now ideally the colors should not be based on the base-latency at all but should be at fixed total values, like 200 to 300 ms for voip (according to ITU-T G.114 for voip one-way delay <= 150 ms is recommended) in yellow, and say 400 to 600 ms in orange, 400ms is upper bound for good voip and 600ms for decent voip (according to ITU-T G.114,users are very satisfied up to 200 ms one way delay and satisfied up to roughly 300ms) so anything above 600 in deep red?
	I know this is not perfect and the numbers will probably require severe "bike-shedding” (and I am not sure that ITU-T G.114 really iOS good source for the thresholds), but to get a discussion started here are the numbers again:
0 	to 100 ms 	no color
101 	to 200 ms		green
201	to 400 ms		yellow
401	to 600 ms		orange
601 	to 1000 ms	red
1001 to infinity		purple (or better marina red?)

Best Regards
	Sebastian


> 
> (sorry to the list moderator, gmail keeps sending under the wrong email and I get a moderator message)
> 
> On Thu, Apr 23, 2015 at 8:05 PM, jb <justinbeech at gmail.com> wrote:
> This is how I've changed the graph of latency under load per input from you guys.
> 
> Taken away log axis.
> 
> Put in two bands. Yellow starts at double the idle latency, and goes to 4x the idle latency
> red starts there, and goes to the top. No red shows if no bars reach into it.
> And no yellow band shows if no bars get into that zone.
> 
> Is it more descriptive?
> 
> 
> On Thu, Apr 23, 2015 at 4:48 PM, Eric Dumazet <eric.dumazet at gmail.com> wrote:
> Wait, this is a 15 years old experiment using Reno and a single test
> bed, using ns simulator.
> 
> Naive TCP pacing implementations were tried, and probably failed.
> 
> Pacing individual packet is quite bad, this is the first lesson one
> learns when implementing TCP pacing, especially if you try to drive a
> 40Gbps NIC.
> 
> https://lwn.net/Articles/564978/
> 
> Also note we use usec based rtt samples, and nanosec high resolution
> timers in fq. I suspect the ns simulator experiment had sync issues
> because of using low resolution timers or simulation artifact, without
> any jitter source.
> 
> Billions of flows are now 'paced', but keep in mind most packets are not
> paced. We do not pace in slow start, and we do not pace when tcp is ACK
> clocked.
> 
> Only when someones sets SO_MAX_PACING_RATE below the TCP rate, we can
> eventually have all packets being paced, using TSO 'clusters' for TCP.
> 
> 
> 
> On Thu, 2015-04-23 at 07:27 +0200, MUSCARIELLO Luca IMT/OLN wrote:
> > one reference with pdf publicly available. On the website there are
> > various papers
> > on this topic. Others might me more relevant but I did not check all of
> > them.
> 
> > Understanding the Performance of TCP Pacing,
> > Amit Aggarwal, Stefan Savage, and Tom Anderson,
> > IEEE INFOCOM 2000 Tel-Aviv, Israel, March 2000, pages 1157-1165.
> >
> > http://www.cs.ucsd.edu/~savage/papers/Infocom2000pacing.pdf
> 
> 
> _______________________________________________
> Bloat mailing list
> Bloat at lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/bloat
> 
> 
> _______________________________________________
> Bloat mailing list
> Bloat at lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/bloat



More information about the Bloat mailing list