From: dpreed@reed.com
To: "Jonathan Morton" <chromatix99@gmail.com>
Cc: Brian Trammell <ietf@trammell.ch>,
bloat <bloat@lists.bufferbloat.net>,
"aqm@ietf.org" <aqm@ietf.org>,
"cerowrt-devel@lists.bufferbloat.net"
<cerowrt-devel@lists.bufferbloat.net>
Subject: Re: [Cerowrt-devel] [Bloat] [aqm] ping loss "considered harmful"
Date: Mon, 2 Mar 2015 07:44:18 -0500 (EST) [thread overview]
Message-ID: <1425300258.188832362@apps.rackspace.com> (raw)
In-Reply-To: <802AFC8C-B59B-4971-A4ED-5C0375E683B1@gmail.com>
On my own web server (running nginx) I provide an HTTP 1.1 accessible statistics service. It returns a single JSON structure with the underlying packet statistics for the server and the current connection. Since this packet is inserted into the HTTP 1.1 stream upon request, it provides both the statistics and a single-round-trip "ping" that can measure queue delay on demand.
Changing server code is a lot easier than changing browser code. Anyone who would like to develop a "standards-track" option for HTTP 1.1 is quite welcome to do so... right now it is a "RESTful" interface, so measurements of the client side must be obtained indirectly, but simple JavaScript code in a web page can access this with standard HTML5 code.
I'm not looking to support this code or enhance it. I've thought about putting it up on GitHub.
In any case, it's easy to add a little widget to the corner of all the web pages from a site that displays latency and net statistics at various levels in a very pretty user-understandable form.
To see how this works, just go to the alt.reed.com site (preferably using an iPhone or Android phone) to see a trivial test app I put out a number of years ago for my personal zeroth-order diagnostics.
nginx (as most know) is an incredibly tight, low overhead HTTP 1.1 server that lets you add plugins for lots of things other than retrieving web pages from disk. It's not the pile of spaghetti that Apache has become. It supports WebSockets and HTTPS quite well, too.
On Monday, March 2, 2015 5:54am, "Jonathan Morton" <chromatix99@gmail.com> said:
>
>> On 2 Mar, 2015, at 12:17, Mikael Abrahamsson <swmike@swm.pp.se> wrote:
>>
>> On Mon, 2 Mar 2015, Brian Trammell wrote:
>>
>>> Gaming protocols do this right - latency measurement is built into the protocol.
>>
>> I believe this is the only way to do it properly, and the most likely easiest way
>> to get this deployed would be to use the TCP stack.
>>
>> We need to give users an easy-to-understand metric on how well their Internet
>> traffic is working. So the problem here is that the users can't tell how well
>> it's working without resorting to ICMP PING to try to figure out what's going on.
>>
>> For instance, if their web browser had insight into what the TCP stack was doing
>> then it could present information a lot better to the user. Instead of telling
>> the user "time to first byte" (which is L4 information), it could tell the less
>> novice user about packet loss, PDV, reordering, RTT, how well concurrent
>> connections to the same IP address are doing, tell more about *why* some
>> connections are slow instead of just saying "it took 5.3 seconds to load this
>> webpage and here are the connections and how long each took". For the novice user
>> there should be some kind of expert system that collects data that you can send
>> to the ISP that also has an expert system to say "it seems your local connection
>> delays packets", please connect to a wired connection and try again". It would
>> know if the problem was excessive delay, excessive delay that varied a lot,
>> packet loss, reordering, or whatever.
>>
>> We have a huge amount of information in our TCP stacks that either are locked in
>> there and not used properly to help users figure out what's going on, and there
>> is basically zero information flow between the applications using TCP and the TCP
>> stack itself. Each just tries to do its best on its own layer.
>
> This seems like an actually good idea. Several of those statistics, at least,
> could be exposed to userspace without incurring any additional overhead in the
> stack (except for the queries themselves), which is important for high-performance
> server users. TCP stacks already track RTT, and sometimes MinRTT - the difference
> between these values is a reasonable lower-bound estimate of induced latency.
>
> For stacks which don’t already track all the desirable data, a socket option
> could be used to turn that on, allocating extra space to do so. To maximise
> portability, therefore, it might be necessary to require that option before
> statistics requests will be valid, even on stacks which do collect it all anyway.
>
> Recent versions of Windows, even, have a semi-magic system which gives a little
> indicator of whether your connection has functioning Internet connectivity or not.
> This could be extended, if Microsoft saw fit, to interpret these statistics and
> notify the user that their connection was behaving badly in the ways we now find
> interesting. Whether Microsoft will do such a thing (which would undoubtedly piss
> off every major ISP on the planet) is another matter, but it’s a concept
> that can be used by Linux desktops as well, and with less political fallout.
>
> Now, who’s going to knuckle down and implement it?
>
> - Jonathan Morton
>
> _______________________________________________
> Cerowrt-devel mailing list
> Cerowrt-devel@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/cerowrt-devel
>
next prev parent reply other threads:[~2015-03-02 12:44 UTC|newest]
Thread overview: 34+ messages / expand[flat|nested] mbox.gz Atom feed top
2015-03-02 3:57 [Cerowrt-devel] " Dave Taht
2015-03-02 4:00 ` [Cerowrt-devel] [aqm] " Andrew Mcgregor
2015-03-02 4:05 ` Mikael Abrahamsson
2015-03-02 4:06 ` [Cerowrt-devel] " David Lang
2015-03-02 9:40 ` [Cerowrt-devel] [aqm] " Brian Trammell
2015-03-02 10:17 ` Mikael Abrahamsson
2015-03-02 10:54 ` [Cerowrt-devel] [Bloat] " Jonathan Morton
2015-03-02 12:44 ` dpreed [this message]
2015-03-02 14:45 ` Brian Trammell
2015-03-02 18:41 ` David Lang
2015-03-03 6:45 ` [Cerowrt-devel] " Valdis.Kletnieks
2015-03-04 8:14 ` Mikael Abrahamsson
2015-03-02 21:53 ` Joe Touch
2015-03-02 23:14 ` David Lang
2015-03-02 23:25 ` Joe Touch
2015-03-02 23:34 ` David Lang
2015-03-03 0:07 ` Andrew Mcgregor
2015-03-02 10:47 ` Dave Dolson
2015-03-02 10:49 ` Andrew Mcgregor
2015-03-02 18:36 ` David Lang
[not found] ` <md2fsa$o1s$1@ger.gmane.org>
2015-03-02 20:33 ` Dave Dolson
2015-03-02 20:39 ` David Lang
2015-03-02 20:38 ` Dave Taht
2015-03-04 8:12 ` Mikael Abrahamsson
2015-03-03 17:20 ` Fred Baker (fred)
2015-03-03 17:29 ` [Cerowrt-devel] [Bloat] " Wesley Eddy
2015-03-03 18:00 ` [Cerowrt-devel] [aqm] [Bloat] " Fred Baker (fred)
2015-03-04 5:24 ` Dave Taht
2015-03-05 18:56 ` Curtis Villamizar
2015-03-05 19:50 ` [Cerowrt-devel] [Bloat] [aqm] " Rich Brown
2015-03-04 17:34 ` [Cerowrt-devel] [aqm] [Bloat] " dpreed
2015-03-04 19:45 ` Mikael Abrahamsson
2015-03-05 20:38 ` [Cerowrt-devel] " Matt Taggart
2015-03-05 20:53 ` Dave Taht
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
List information: https://lists.bufferbloat.net/postorius/lists/cerowrt-devel.lists.bufferbloat.net/
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1425300258.188832362@apps.rackspace.com \
--to=dpreed@reed.com \
--cc=aqm@ietf.org \
--cc=bloat@lists.bufferbloat.net \
--cc=cerowrt-devel@lists.bufferbloat.net \
--cc=chromatix99@gmail.com \
--cc=ietf@trammell.ch \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox