Starlink has bufferbloat. Bad.
 help / color / mirror / Atom feed
* [Starlink] the grinch meets cloudflare's christmas present
@ 2023-01-04 17:26 Dave Taht
  2023-01-04 19:20 ` [Starlink] [Rpm] " jf
  2023-01-06 16:38 ` [Starlink] [LibreQoS] " MORTON JR., AL
  0 siblings, 2 replies; 44+ messages in thread
From: Dave Taht @ 2023-01-04 17:26 UTC (permalink / raw)
  To: bloat, libreqos, Cake List, Dave Taht via Starlink, Rpm, IETF IPPM WG

[-- Attachment #1: Type: text/plain, Size: 1383 bytes --]

Please try the new, the shiny, the really wonderful test here:
https://speed.cloudflare.com/

I would really appreciate some independent verification of
measurements using this tool. In my brief experiments it appears - as
all the commercial tools to date - to dramatically understate the
bufferbloat, on my LTE, (and my starlink terminal is out being
hacked^H^H^H^H^H^Hworked on, so I can't measure that)

My test of their test reports 223ms 5G latency under load , where
flent reports over 2seconds. See comparison attached.

My guess is that this otherwise lovely new  tool, like too many,
doesn't run for long enough. Admittedly, most web objects (their
target market) are small, and so long as they remain small and not
heavily pipelined this test is a very good start... but I'm pretty
sure cloudflare is used for bigger uploads and downloads than that.
There's no way to change the test to run longer either.

I'd love to get some results from other networks (compared as usual to
flent), especially ones with cake on it. I'd love to know if they
measured more minimum rtts that can be obtained with fq_codel or cake,
correctly.

Love Always,
The Grinch

-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

[-- Attachment #2: image.png --]
[-- Type: image/png, Size: 256990 bytes --]

[-- Attachment #3: tcp_nup-2023-01-04T090937.211620.LTE.flent.gz --]
[-- Type: application/gzip, Size: 25192 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
  2023-01-04 17:26 [Starlink] the grinch meets cloudflare's christmas present Dave Taht
@ 2023-01-04 19:20 ` jf
  2023-01-04 20:02   ` rjmcmahon
  2023-01-06 16:38 ` [Starlink] [LibreQoS] " MORTON JR., AL
  1 sibling, 1 reply; 44+ messages in thread
From: jf @ 2023-01-04 19:20 UTC (permalink / raw)
  To: Dave Taht
  Cc: bloat, libreqos, Cake List, Dave Taht via Starlink, Rpm, IETF IPPM WG

[-- Attachment #1: Type: text/plain, Size: 489 bytes --]

HNY Dave and all the rest,

Great to see yet another capacity test add latency metrics to the results. This one looks like a good start.

Results from my Windstream DOCSIS 3.1 line (3.1 on download only, up is 3.0) Gigabit down / 35Mbps up provisioning. Using an IQrouter Pro (an i5 x86) with Cake set for 710/31 as this ISP can’t deliver reliable low-latency unless you shave a good bit off the targets. My local loop is pretty congested.

Here’s the latest Cloudflare test:


[-- Attachment #2: CFSpeedTest_Gig35_20230104.png --]
[-- Type: image/png, Size: 379539 bytes --]

[-- Attachment #3: Type: text/plain, Size: 41 bytes --]



And an Ookla test run just afterward:


[-- Attachment #4: Speedtest_net_Gig35_20230104.png --]
[-- Type: image/png, Size: 40589 bytes --]

[-- Attachment #5: Type: text/plain, Size: 1897 bytes --]



They are definitely both in the ballpark and correspond to other tests run from the router itself or my (wired) MacBook Pro.

Cheers,

Jonathan


> On Jan 4, 2023, at 12:26 PM, Dave Taht via Rpm <rpm@lists.bufferbloat.net> wrote:
> 
> Please try the new, the shiny, the really wonderful test here:
> https://speed.cloudflare.com/
> 
> I would really appreciate some independent verification of
> measurements using this tool. In my brief experiments it appears - as
> all the commercial tools to date - to dramatically understate the
> bufferbloat, on my LTE, (and my starlink terminal is out being
> hacked^H^H^H^H^H^Hworked on, so I can't measure that)
> 
> My test of their test reports 223ms 5G latency under load , where
> flent reports over 2seconds. See comparison attached.
> 
> My guess is that this otherwise lovely new  tool, like too many,
> doesn't run for long enough. Admittedly, most web objects (their
> target market) are small, and so long as they remain small and not
> heavily pipelined this test is a very good start... but I'm pretty
> sure cloudflare is used for bigger uploads and downloads than that.
> There's no way to change the test to run longer either.
> 
> I'd love to get some results from other networks (compared as usual to
> flent), especially ones with cake on it. I'd love to know if they
> measured more minimum rtts that can be obtained with fq_codel or cake,
> correctly.
> 
> Love Always,
> The Grinch
> 
> -- 
> This song goes out to all the folk that thought Stadia would work:
> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
> Dave Täht CEO, TekLibre, LLC
> <image.png><tcp_nup-2023-01-04T090937.211620.LTE.flent.gz>_______________________________________________
> Rpm mailing list
> Rpm@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/rpm


^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
  2023-01-04 19:20 ` [Starlink] [Rpm] " jf
@ 2023-01-04 20:02   ` rjmcmahon
  2023-01-04 21:16     ` Ulrich Speidel
  2023-01-05 11:11     ` [Starlink] [Bloat] " Sebastian Moeller
  0 siblings, 2 replies; 44+ messages in thread
From: rjmcmahon @ 2023-01-04 20:02 UTC (permalink / raw)
  To: jf
  Cc: Dave Taht, Dave Taht via Starlink, IETF IPPM WG, libreqos,
	Cake List, Rpm, bloat

Curious to why people keep calling capacity tests speed tests? A semi at 
55 mph isn't faster than a porsche at 141 mph because its load volume is 
larger.

Bob
> HNY Dave and all the rest,
> 
> Great to see yet another capacity test add latency metrics to the
> results. This one looks like a good start.
> 
> Results from my Windstream DOCSIS 3.1 line (3.1 on download only, up
> is 3.0) Gigabit down / 35Mbps up provisioning. Using an IQrouter Pro
> (an i5 x86) with Cake set for 710/31 as this ISP can’t deliver
> reliable low-latency unless you shave a good bit off the targets. My
> local loop is pretty congested.
> 
> Here’s the latest Cloudflare test:
> 
> 
> 
> 
> And an Ookla test run just afterward:
> 
> 
> 
> 
> They are definitely both in the ballpark and correspond to other tests
> run from the router itself or my (wired) MacBook Pro.
> 
> Cheers,
> 
> Jonathan
> 
> 
>> On Jan 4, 2023, at 12:26 PM, Dave Taht via Rpm 
>> <rpm@lists.bufferbloat.net> wrote:
>> 
>> Please try the new, the shiny, the really wonderful test here:
>> https://speed.cloudflare.com/
>> 
>> I would really appreciate some independent verification of
>> measurements using this tool. In my brief experiments it appears - as
>> all the commercial tools to date - to dramatically understate the
>> bufferbloat, on my LTE, (and my starlink terminal is out being
>> hacked^H^H^H^H^H^Hworked on, so I can't measure that)
>> 
>> My test of their test reports 223ms 5G latency under load , where
>> flent reports over 2seconds. See comparison attached.
>> 
>> My guess is that this otherwise lovely new  tool, like too many,
>> doesn't run for long enough. Admittedly, most web objects (their
>> target market) are small, and so long as they remain small and not
>> heavily pipelined this test is a very good start... but I'm pretty
>> sure cloudflare is used for bigger uploads and downloads than that.
>> There's no way to change the test to run longer either.
>> 
>> I'd love to get some results from other networks (compared as usual to
>> flent), especially ones with cake on it. I'd love to know if they
>> measured more minimum rtts that can be obtained with fq_codel or cake,
>> correctly.
>> 
>> Love Always,
>> The Grinch
>> 
>> --
>> This song goes out to all the folk that thought Stadia would work:
>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>> Dave Täht CEO, TekLibre, LLC
>> <image.png><tcp_nup-2023-01-04T090937.211620.LTE.flent.gz>_______________________________________________
>> Rpm mailing list
>> Rpm@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/rpm
> 
> 
> _______________________________________________
> Rpm mailing list
> Rpm@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/rpm

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
  2023-01-04 20:02   ` rjmcmahon
@ 2023-01-04 21:16     ` Ulrich Speidel
  2023-01-04 23:54       ` Bruce Perens
  2023-01-05 11:11     ` [Starlink] [Bloat] " Sebastian Moeller
  1 sibling, 1 reply; 44+ messages in thread
From: Ulrich Speidel @ 2023-01-04 21:16 UTC (permalink / raw)
  To: jf, rjmcmahon; +Cc: Dave Taht via Starlink, bloat

[-- Attachment #1: Type: text/plain, Size: 4507 bytes --]

The use of the term "speed" in communications used to be restricted to the speed of light (or whatever propagation speed one happened to be dealing with. Everything else was a "rate". Maybe I'm old-fashioned but I think talking about "speed tests" muddies the waters rather a lot.

--
****************************************************************
Dr. Ulrich Speidel

Department of Computer Science

Room 303S.594
Ph: (+64-9)-373-7599 ext. 85282

The University of Auckland
u.speidel@auckland.ac.nz<mailto:u.speidel@auckland.ac.nz>
http://www.cs.auckland.ac.nz/~ulrich/<http://www.cs.auckland.ac.nz/%7Eulrich/>
****************************************************************
________________________________
From: Starlink <starlink-bounces@lists.bufferbloat.net> on behalf of rjmcmahon via Starlink <starlink@lists.bufferbloat.net>
Sent: Thursday, January 5, 2023 9:02 AM
To: jf@jonathanfoulkes.com <jf@jonathanfoulkes.com>
Cc: Cake List <cake@lists.bufferbloat.net>; IETF IPPM WG <ippm@ietf.org>; libreqos <libreqos@lists.bufferbloat.net>; Dave Taht via Starlink <starlink@lists.bufferbloat.net>; Rpm <rpm@lists.bufferbloat.net>; bloat <bloat@lists.bufferbloat.net>
Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present

Curious to why people keep calling capacity tests speed tests? A semi at
55 mph isn't faster than a porsche at 141 mph because its load volume is
larger.

Bob
> HNY Dave and all the rest,
>
> Great to see yet another capacity test add latency metrics to the
> results. This one looks like a good start.
>
> Results from my Windstream DOCSIS 3.1 line (3.1 on download only, up
> is 3.0) Gigabit down / 35Mbps up provisioning. Using an IQrouter Pro
> (an i5 x86) with Cake set for 710/31 as this ISP can’t deliver
> reliable low-latency unless you shave a good bit off the targets. My
> local loop is pretty congested.
>
> Here’s the latest Cloudflare test:
>
>
>
>
> And an Ookla test run just afterward:
>
>
>
>
> They are definitely both in the ballpark and correspond to other tests
> run from the router itself or my (wired) MacBook Pro.
>
> Cheers,
>
> Jonathan
>
>
>> On Jan 4, 2023, at 12:26 PM, Dave Taht via Rpm
>> <rpm@lists.bufferbloat.net> wrote:
>>
>> Please try the new, the shiny, the really wonderful test here:
>> https://speed.cloudflare.com/<https://speed.cloudflare.com>
>>
>> I would really appreciate some independent verification of
>> measurements using this tool. In my brief experiments it appears - as
>> all the commercial tools to date - to dramatically understate the
>> bufferbloat, on my LTE, (and my starlink terminal is out being
>> hacked^H^H^H^H^H^Hworked on, so I can't measure that)
>>
>> My test of their test reports 223ms 5G latency under load , where
>> flent reports over 2seconds. See comparison attached.
>>
>> My guess is that this otherwise lovely new tool, like too many,
>> doesn't run for long enough. Admittedly, most web objects (their
>> target market) are small, and so long as they remain small and not
>> heavily pipelined this test is a very good start... but I'm pretty
>> sure cloudflare is used for bigger uploads and downloads than that.
>> There's no way to change the test to run longer either.
>>
>> I'd love to get some results from other networks (compared as usual to
>> flent), especially ones with cake on it. I'd love to know if they
>> measured more minimum rtts that can be obtained with fq_codel or cake,
>> correctly.
>>
>> Love Always,
>> The Grinch
>>
>> --
>> This song goes out to all the folk that thought Stadia would work:
>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz<https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz>
>> Dave Täht CEO, TekLibre, LLC
>> <image.png><tcp_nup-2023-01-04T090937.211620.LTE.flent.gz>_______________________________________________
>> Rpm mailing list
>> Rpm@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/rpm<https://lists.bufferbloat.net/listinfo/rpm>
>
>
> _______________________________________________
> Rpm mailing list
> Rpm@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/rpm<https://lists.bufferbloat.net/listinfo/rpm>
_______________________________________________
Starlink mailing list
Starlink@lists.bufferbloat.net
https://lists.bufferbloat.net/listinfo/starlink<https://lists.bufferbloat.net/listinfo/starlink>

[-- Attachment #2: Type: text/html, Size: 6806 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
  2023-01-04 21:16     ` Ulrich Speidel
@ 2023-01-04 23:54       ` Bruce Perens
  2023-01-05  2:48         ` Dave Collier-Brown
  2023-01-05  6:11         ` rjmcmahon
  0 siblings, 2 replies; 44+ messages in thread
From: Bruce Perens @ 2023-01-04 23:54 UTC (permalink / raw)
  To: Ulrich Speidel; +Cc: jf, rjmcmahon, Dave Taht via Starlink, bloat

[-- Attachment #1: Type: text/plain, Size: 4941 bytes --]

On the other hand, we would like to be comprehensible to normal users,
especially when we want them to press their providers to deal with
bufferbloat. Differences like speed and rate would go right over their
heads.

On Wed, Jan 4, 2023 at 1:16 PM Ulrich Speidel via Starlink <
starlink@lists.bufferbloat.net> wrote:

> The use of the term "speed" in communications used to be restricted to the
> speed of light (or whatever propagation speed one happened to be dealing
> with. Everything else was a "rate". Maybe I'm old-fashioned but I think
> talking about "speed tests" muddies the waters rather a lot.
>
> --
> ****************************************************************
> Dr. Ulrich Speidel
>
> Department of Computer Science
>
> Room 303S.594
> Ph: (+64-9)-373-7599 ext. 85282
>
> The University of Auckland
> u.speidel@auckland.ac.nz
> http://www.cs.auckland.ac.nz/~ulrich/
> ****************************************************************
> ------------------------------
> *From:* Starlink <starlink-bounces@lists.bufferbloat.net> on behalf of
> rjmcmahon via Starlink <starlink@lists.bufferbloat.net>
> *Sent:* Thursday, January 5, 2023 9:02 AM
> *To:* jf@jonathanfoulkes.com <jf@jonathanfoulkes.com>
> *Cc:* Cake List <cake@lists.bufferbloat.net>; IETF IPPM WG <ippm@ietf.org>;
> libreqos <libreqos@lists.bufferbloat.net>; Dave Taht via Starlink <
> starlink@lists.bufferbloat.net>; Rpm <rpm@lists.bufferbloat.net>; bloat <
> bloat@lists.bufferbloat.net>
> *Subject:* Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas
> present
>
> Curious to why people keep calling capacity tests speed tests? A semi at
> 55 mph isn't faster than a porsche at 141 mph because its load volume is
> larger.
>
> Bob
> > HNY Dave and all the rest,
> >
> > Great to see yet another capacity test add latency metrics to the
> > results. This one looks like a good start.
> >
> > Results from my Windstream DOCSIS 3.1 line (3.1 on download only, up
> > is 3.0) Gigabit down / 35Mbps up provisioning. Using an IQrouter Pro
> > (an i5 x86) with Cake set for 710/31 as this ISP can’t deliver
> > reliable low-latency unless you shave a good bit off the targets. My
> > local loop is pretty congested.
> >
> > Here’s the latest Cloudflare test:
> >
> >
> >
> >
> > And an Ookla test run just afterward:
> >
> >
> >
> >
> > They are definitely both in the ballpark and correspond to other tests
> > run from the router itself or my (wired) MacBook Pro.
> >
> > Cheers,
> >
> > Jonathan
> >
> >
> >> On Jan 4, 2023, at 12:26 PM, Dave Taht via Rpm
> >> <rpm@lists.bufferbloat.net> wrote:
> >>
> >> Please try the new, the shiny, the really wonderful test here:
> >> https://speed.cloudflare.com/
> >>
> >> I would really appreciate some independent verification of
> >> measurements using this tool. In my brief experiments it appears - as
> >> all the commercial tools to date - to dramatically understate the
> >> bufferbloat, on my LTE, (and my starlink terminal is out being
> >> hacked^H^H^H^H^H^Hworked on, so I can't measure that)
> >>
> >> My test of their test reports 223ms 5G latency under load , where
> >> flent reports over 2seconds. See comparison attached.
> >>
> >> My guess is that this otherwise lovely new tool, like too many,
> >> doesn't run for long enough. Admittedly, most web objects (their
> >> target market) are small, and so long as they remain small and not
> >> heavily pipelined this test is a very good start... but I'm pretty
> >> sure cloudflare is used for bigger uploads and downloads than that.
> >> There's no way to change the test to run longer either.
> >>
> >> I'd love to get some results from other networks (compared as usual to
> >> flent), especially ones with cake on it. I'd love to know if they
> >> measured more minimum rtts that can be obtained with fq_codel or cake,
> >> correctly.
> >>
> >> Love Always,
> >> The Grinch
> >>
> >> --
> >> This song goes out to all the folk that thought Stadia would work:
> >>
> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
> >> Dave Täht CEO, TekLibre, LLC
> >>
> <image.png><tcp_nup-2023-01-04T090937.211620.LTE.flent.gz>_______________________________________________
> >> Rpm mailing list
> >> Rpm@lists.bufferbloat.net
> >> https://lists.bufferbloat.net/listinfo/rpm
> >
> >
> > _______________________________________________
> > Rpm mailing list
> > Rpm@lists.bufferbloat.net
> > https://lists.bufferbloat.net/listinfo/rpm
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink
>


-- 
Bruce Perens K6BP

[-- Attachment #2: Type: text/html, Size: 8553 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
  2023-01-04 23:54       ` Bruce Perens
@ 2023-01-05  2:48         ` Dave Collier-Brown
  2023-01-05  3:11           ` Dick Roy
  2023-01-05  6:11         ` rjmcmahon
  1 sibling, 1 reply; 44+ messages in thread
From: Dave Collier-Brown @ 2023-01-05  2:48 UTC (permalink / raw)
  To: starlink

[-- Attachment #1: Type: text/plain, Size: 7002 bytes --]

I think using "speed" for "the inverse of delay" is pretty normal English, if technically erroneous when speaking nerd or physicist.

Using it for volume? Arguably more like fraudulent...

--dave

On 1/4/23 18:54, Bruce Perens via Starlink wrote:
On the other hand, we would like to be comprehensible to normal users, especially when we want them to press their providers to deal with bufferbloat. Differences like speed and rate would go right over their heads.

On Wed, Jan 4, 2023 at 1:16 PM Ulrich Speidel via Starlink <starlink@lists.bufferbloat.net<mailto:starlink@lists.bufferbloat.net>> wrote:
The use of the term "speed" in communications used to be restricted to the speed of light (or whatever propagation speed one happened to be dealing with. Everything else was a "rate". Maybe I'm old-fashioned but I think talking about "speed tests" muddies the waters rather a lot.

--
****************************************************************
Dr. Ulrich Speidel

Department of Computer Science

Room 303S.594
Ph: (+64-9)-373-7599 ext. 85282

The University of Auckland
u.speidel@auckland.ac.nz<mailto:u.speidel@auckland.ac.nz>
http://www.cs.auckland.ac.nz/~ulrich/<http://www.cs.auckland.ac.nz/%7Eulrich/>
****************************************************************
________________________________
From: Starlink <starlink-bounces@lists.bufferbloat.net<mailto:starlink-bounces@lists.bufferbloat.net>> on behalf of rjmcmahon via Starlink <starlink@lists.bufferbloat.net<mailto:starlink@lists.bufferbloat.net>>
Sent: Thursday, January 5, 2023 9:02 AM
To: jf@jonathanfoulkes.com<mailto:jf@jonathanfoulkes.com> <jf@jonathanfoulkes.com<mailto:jf@jonathanfoulkes.com>>
Cc: Cake List <cake@lists.bufferbloat.net<mailto:cake@lists.bufferbloat.net>>; IETF IPPM WG <ippm@ietf.org<mailto:ippm@ietf.org>>; libreqos <libreqos@lists.bufferbloat.net<mailto:libreqos@lists.bufferbloat.net>>; Dave Taht via Starlink <starlink@lists.bufferbloat.net<mailto:starlink@lists.bufferbloat.net>>; Rpm <rpm@lists.bufferbloat.net<mailto:rpm@lists.bufferbloat.net>>; bloat <bloat@lists.bufferbloat.net<mailto:bloat@lists.bufferbloat.net>>
Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present

Curious to why people keep calling capacity tests speed tests? A semi at
55 mph isn't faster than a porsche at 141 mph because its load volume is
larger.

Bob
> HNY Dave and all the rest,
>
> Great to see yet another capacity test add latency metrics to the
> results. This one looks like a good start.
>
> Results from my Windstream DOCSIS 3.1 line (3.1 on download only, up
> is 3.0) Gigabit down / 35Mbps up provisioning. Using an IQrouter Pro
> (an i5 x86) with Cake set for 710/31 as this ISP can’t deliver
> reliable low-latency unless you shave a good bit off the targets. My
> local loop is pretty congested.
>
> Here’s the latest Cloudflare test:
>
>
>
>
> And an Ookla test run just afterward:
>
>
>
>
> They are definitely both in the ballpark and correspond to other tests
> run from the router itself or my (wired) MacBook Pro.
>
> Cheers,
>
> Jonathan
>
>
>> On Jan 4, 2023, at 12:26 PM, Dave Taht via Rpm
>> <rpm@lists.bufferbloat.net<mailto:rpm@lists.bufferbloat.net>> wrote:
>>
>> Please try the new, the shiny, the really wonderful test here:
>> https://speed.cloudflare.com/<https://speed.cloudflare.com>
>>
>> I would really appreciate some independent verification of
>> measurements using this tool. In my brief experiments it appears - as
>> all the commercial tools to date - to dramatically understate the
>> bufferbloat, on my LTE, (and my starlink terminal is out being
>> hacked^H^H^H^H^H^Hworked on, so I can't measure that)
>>
>> My test of their test reports 223ms 5G latency under load , where
>> flent reports over 2seconds. See comparison attached.
>>
>> My guess is that this otherwise lovely new tool, like too many,
>> doesn't run for long enough. Admittedly, most web objects (their
>> target market) are small, and so long as they remain small and not
>> heavily pipelined this test is a very good start... but I'm pretty
>> sure cloudflare is used for bigger uploads and downloads than that.
>> There's no way to change the test to run longer either.
>>
>> I'd love to get some results from other networks (compared as usual to
>> flent), especially ones with cake on it. I'd love to know if they
>> measured more minimum rtts that can be obtained with fq_codel or cake,
>> correctly.
>>
>> Love Always,
>> The Grinch
>>
>> --
>> This song goes out to all the folk that thought Stadia would work:
>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>> Dave Täht CEO, TekLibre, LLC
>> <image.png><tcp_nup-2023-01-04T090937.211620.LTE.flent.gz>_______________________________________________
>> Rpm mailing list
>> Rpm@lists.bufferbloat.net<mailto:Rpm@lists.bufferbloat.net>
>> https://lists.bufferbloat.net/listinfo/rpm
>
>
> _______________________________________________
> Rpm mailing list
> Rpm@lists.bufferbloat.net<mailto:Rpm@lists.bufferbloat.net>
> https://lists.bufferbloat.net/listinfo/rpm
_______________________________________________
Starlink mailing list
Starlink@lists.bufferbloat.net<mailto:Starlink@lists.bufferbloat.net>
https://lists.bufferbloat.net/listinfo/starlink
_______________________________________________
Starlink mailing list
Starlink@lists.bufferbloat.net<mailto:Starlink@lists.bufferbloat.net>
https://lists.bufferbloat.net/listinfo/starlink


--
Bruce Perens K6BP



_______________________________________________
Starlink mailing list
Starlink@lists.bufferbloat.net<mailto:Starlink@lists.bufferbloat.net>
https://lists.bufferbloat.net/listinfo/starlink


--
David Collier-Brown,         | Always do right. This will gratify
System Programmer and Author | some people and astonish the rest
dave.collier-brown@indexexchange.com<mailto:dave.collier-brown@indexexchange.com> |              -- Mark Twain


CONFIDENTIALITY NOTICE AND DISCLAIMER : This telecommunication, including any and all attachments, contains confidential information intended only for the person(s) to whom it is addressed. Any dissemination, distribution, copying or disclosure is strictly prohibited and is not a waiver of confidentiality. If you have received this telecommunication in error, please notify the sender immediately by return electronic mail and delete the message from your inbox and deleted items folders. This telecommunication does not constitute an express or implied agreement to conduct transactions by electronic means, nor does it constitute a contract offer, a contract amendment or an acceptance of a contract offer. Contract terms contained in this telecommunication are subject to legal review and the completion of formal documentation and are not binding until same is confirmed in writing and has been signed by an authorized signatory.

[-- Attachment #2: Type: text/html, Size: 12185 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
  2023-01-05  2:48         ` Dave Collier-Brown
@ 2023-01-05  3:11           ` Dick Roy
  2023-01-05 11:25             ` Sebastian Moeller
  0 siblings, 1 reply; 44+ messages in thread
From: Dick Roy @ 2023-01-05  3:11 UTC (permalink / raw)
  To: 'Dave Collier-Brown'; +Cc: starlink

[-- Attachment #1: Type: text/plain, Size: 7240 bytes --]

 

 

  _____  

From: Starlink [mailto:starlink-bounces@lists.bufferbloat.net] On Behalf Of
Dave Collier-Brown via Starlink
Sent: Wednesday, January 4, 2023 6:48 PM
To: starlink@lists.bufferbloat.net
Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas
present

 

I think using "speed" for "the inverse of delay" is pretty normal English,
if technically erroneous when speaking nerd or physicist.

[RR] I’ve not heard of that usage before.  The units aren’t commensurate
either. 

Using it for volume? Arguably more like fraudulent...

[RR] I don’t think that was Bob’s intent.  I think “load volume” was meant
to be a metaphor for “number of bits/bytes” being transported (“by the
semi”).  

That said, aren’t users these days educated on “gigs” which they intuitively
understand to be Gigabits per second (or Gbps)?  Oddly enough, that is an
expression of “data/information/communication rate” in the appropriate units
with the nominal technically correct meaning.  

RR

--dave

On 1/4/23 18:54, Bruce Perens via Starlink wrote:

On the other hand, we would like to be comprehensible to normal users,
especially when we want them to press their providers to deal with
bufferbloat. Differences like speed and rate would go right over their
heads.

 

On Wed, Jan 4, 2023 at 1:16 PM Ulrich Speidel via Starlink
<starlink@lists.bufferbloat.net> wrote:

The use of the term "speed" in communications used to be restricted to the
speed of light (or whatever propagation speed one happened to be dealing
with. Everything else was a "rate". Maybe I'm old-fashioned but I think
talking about "speed tests" muddies the waters rather a lot.

 

-- 
**************************************************************** 
Dr. Ulrich Speidel 

Department of Computer Science 

Room 303S.594
Ph: (+64-9)-373-7599 ext. 85282 

The University of Auckland 
u.speidel@auckland.ac.nz
http://www.cs.auckland.ac.nz/~ulrich/
<http://www.cs.auckland.ac.nz/%7Eulrich/> 
**************************************************************** 


  _____  


From: Starlink <starlink-bounces@lists.bufferbloat.net> on behalf of
rjmcmahon via Starlink <starlink@lists.bufferbloat.net>
Sent: Thursday, January 5, 2023 9:02 AM
To: jf@jonathanfoulkes.com <jf@jonathanfoulkes.com>
Cc: Cake List <cake@lists.bufferbloat.net>; IETF IPPM WG <ippm@ietf.org>;
libreqos <libreqos@lists.bufferbloat.net>; Dave Taht via Starlink
<starlink@lists.bufferbloat.net>; Rpm <rpm@lists.bufferbloat.net>; bloat
<bloat@lists.bufferbloat.net>
Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas
present 

 

Curious to why people keep calling capacity tests speed tests? A semi at 
55 mph isn't faster than a porsche at 141 mph because its load volume is 
larger.

Bob
> HNY Dave and all the rest,
> 
> Great to see yet another capacity test add latency metrics to the
> results. This one looks like a good start.
> 
> Results from my Windstream DOCSIS 3.1 line (3.1 on download only, up
> is 3.0) Gigabit down / 35Mbps up provisioning. Using an IQrouter Pro
> (an i5 x86) with Cake set for 710/31 as this ISP can’t deliver
> reliable low-latency unless you shave a good bit off the targets. My
> local loop is pretty congested.
> 
> Here’s the latest Cloudflare test:
> 
> 
> 
> 
> And an Ookla test run just afterward:
> 
> 
> 
> 
> They are definitely both in the ballpark and correspond to other tests
> run from the router itself or my (wired) MacBook Pro.
> 
> Cheers,
> 
> Jonathan
> 
> 
>> On Jan 4, 2023, at 12:26 PM, Dave Taht via Rpm 
>> <rpm@lists.bufferbloat.net> wrote:
>> 
>> Please try the new, the shiny, the really wonderful test here:
>> https://speed.cloudflare.com/ <https://speed.cloudflare.com> 
>> 
>> I would really appreciate some independent verification of
>> measurements using this tool. In my brief experiments it appears - as
>> all the commercial tools to date - to dramatically understate the
>> bufferbloat, on my LTE, (and my starlink terminal is out being
>> hacked^H^H^H^H^H^Hworked on, so I can't measure that)
>> 
>> My test of their test reports 223ms 5G latency under load , where
>> flent reports over 2seconds. See comparison attached.
>> 
>> My guess is that this otherwise lovely new tool, like too many,
>> doesn't run for long enough. Admittedly, most web objects (their
>> target market) are small, and so long as they remain small and not
>> heavily pipelined this test is a very good start... but I'm pretty
>> sure cloudflare is used for bigger uploads and downloads than that.
>> There's no way to change the test to run longer either.
>> 
>> I'd love to get some results from other networks (compared as usual to
>> flent), especially ones with cake on it. I'd love to know if they
>> measured more minimum rtts that can be obtained with fq_codel or cake,
>> correctly.
>> 
>> Love Always,
>> The Grinch
>> 
>> --
>> This song goes out to all the folk that thought Stadia would work:
>>
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-698136666560
7352320-FXtz
>> Dave Täht CEO, TekLibre, LLC
>>
<image.png><tcp_nup-2023-01-04T090937.211620.LTE.flent.gz>__________________
_____________________________
>> Rpm mailing list
>> Rpm@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/rpm
> 
> 
> _______________________________________________
> Rpm mailing list
> Rpm@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/rpm
_______________________________________________
Starlink mailing list
Starlink@lists.bufferbloat.net
https://lists.bufferbloat.net/listinfo/starlink

_______________________________________________
Starlink mailing list
Starlink@lists.bufferbloat.net
https://lists.bufferbloat.net/listinfo/starlink




 

-- 

Bruce Perens K6BP






_______________________________________________
Starlink mailing list
Starlink@lists.bufferbloat.net
https://lists.bufferbloat.net/listinfo/starlink

-- 
David Collier-Brown,         | Always do right. This will gratify
System Programmer and Author | some people and astonish the rest
dave.collier-brown@indexexchange.com |              -- Mark Twain

 

CONFIDENTIALITY NOTICE AND DISCLAIMER : This telecommunication, including
any and all attachments, contains confidential information intended only for
the person(s) to whom it is addressed. Any dissemination, distribution,
copying or disclosure is strictly prohibited and is not a waiver of
confidentiality. If you have received this telecommunication in error,
please notify the sender immediately by return electronic mail and delete
the message from your inbox and deleted items folders. This
telecommunication does not constitute an express or implied agreement to
conduct transactions by electronic means, nor does it constitute a contract
offer, a contract amendment or an acceptance of a contract offer. Contract
terms contained in this telecommunication are subject to legal review and
the completion of formal documentation and are not binding until same is
confirmed in writing and has been signed by an authorized signatory.


[-- Attachment #2: Type: text/html, Size: 19394 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
  2023-01-04 23:54       ` Bruce Perens
  2023-01-05  2:48         ` Dave Collier-Brown
@ 2023-01-05  6:11         ` rjmcmahon
  1 sibling, 0 replies; 44+ messages in thread
From: rjmcmahon @ 2023-01-05  6:11 UTC (permalink / raw)
  To: Bruce Perens; +Cc: Ulrich Speidel, jf, Dave Taht via Starlink, bloat

The thing that works for gamers are colors, e.g. green, yellow and red. 
Basically, if the game slows down to a bothersome experience the 
"latency indicator" goes from green to yellow. If the game slows down to 
be unplayable it goes to red and the "phone" mfg gets lots of 
complaints. Why we call a handheld computer a phone is a whole other 
discussion.

Bob
> On the other hand, we would like to be comprehensible to normal users,
> especially when we want them to press their providers to deal with
> bufferbloat. Differences like speed and rate would go right over their
> heads.
> 
> On Wed, Jan 4, 2023 at 1:16 PM Ulrich Speidel via Starlink
> <starlink@lists.bufferbloat.net> wrote:
> 
>> The use of the term "speed" in communications used to be restricted
>> to the speed of light (or whatever propagation speed one happened to
>> be dealing with. Everything else was a "rate". Maybe I'm
>> old-fashioned but I think talking about "speed tests" muddies the
>> waters rather a lot.
>> 
>> --
>> ****************************************************************
>> Dr. Ulrich Speidel
>> 
>> Department of Computer Science
>> 
>> Room 303S.594
>> Ph: (+64-9)-373-7599 ext. 85282
>> 
>> The University of Auckland
>> u.speidel@auckland.ac.nz
>> http://www.cs.auckland.ac.nz/~ulrich/ [1]
>> ****************************************************************
>> 
>> -------------------------
>> 
>> From: Starlink <starlink-bounces@lists.bufferbloat.net> on behalf of
>> rjmcmahon via Starlink <starlink@lists.bufferbloat.net>
>> Sent: Thursday, January 5, 2023 9:02 AM
>> To: jf@jonathanfoulkes.com <jf@jonathanfoulkes.com>
>> Cc: Cake List <cake@lists.bufferbloat.net>; IETF IPPM WG
>> <ippm@ietf.org>; libreqos <libreqos@lists.bufferbloat.net>; Dave
>> Taht via Starlink <starlink@lists.bufferbloat.net>; Rpm
>> <rpm@lists.bufferbloat.net>; bloat <bloat@lists.bufferbloat.net>
>> Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's
>> christmas present
>> 
>> Curious to why people keep calling capacity tests speed tests? A
>> semi at
>> 55 mph isn't faster than a porsche at 141 mph because its load
>> volume is
>> larger.
>> 
>> Bob
>>> HNY Dave and all the rest,
>>> 
>>> Great to see yet another capacity test add latency metrics to the
>>> results. This one looks like a good start.
>>> 
>>> Results from my Windstream DOCSIS 3.1 line (3.1 on download only,
>> up
>>> is 3.0) Gigabit down / 35Mbps up provisioning. Using an IQrouter
>> Pro
>>> (an i5 x86) with Cake set for 710/31 as this ISP can’t deliver
>>> reliable low-latency unless you shave a good bit off the targets.
>> My
>>> local loop is pretty congested.
>>> 
>>> Here’s the latest Cloudflare test:
>>> 
>>> 
>>> 
>>> 
>>> And an Ookla test run just afterward:
>>> 
>>> 
>>> 
>>> 
>>> They are definitely both in the ballpark and correspond to other
>> tests
>>> run from the router itself or my (wired) MacBook Pro.
>>> 
>>> Cheers,
>>> 
>>> Jonathan
>>> 
>>> 
>>>> On Jan 4, 2023, at 12:26 PM, Dave Taht via Rpm
>>>> <rpm@lists.bufferbloat.net> wrote:
>>>> 
>>>> Please try the new, the shiny, the really wonderful test here:
>>>> https://speed.cloudflare.com/ [2]
>>>> 
>>>> I would really appreciate some independent verification of
>>>> measurements using this tool. In my brief experiments it appears
>> - as
>>>> all the commercial tools to date - to dramatically understate the
>>>> bufferbloat, on my LTE, (and my starlink terminal is out being
>>>> hacked^H^H^H^H^H^Hworked on, so I can't measure that)
>>>> 
>>>> My test of their test reports 223ms 5G latency under load , where
>>>> flent reports over 2seconds. See comparison attached.
>>>> 
>>>> My guess is that this otherwise lovely new tool, like too many,
>>>> doesn't run for long enough. Admittedly, most web objects (their
>>>> target market) are small, and so long as they remain small and
>> not
>>>> heavily pipelined this test is a very good start... but I'm
>> pretty
>>>> sure cloudflare is used for bigger uploads and downloads than
>> that.
>>>> There's no way to change the test to run longer either.
>>>> 
>>>> I'd love to get some results from other networks (compared as
>> usual to
>>>> flent), especially ones with cake on it. I'd love to know if they
>>>> measured more minimum rtts that can be obtained with fq_codel or
>> cake,
>>>> correctly.
>>>> 
>>>> Love Always,
>>>> The Grinch
>>>> 
>>>> --
>>>> This song goes out to all the folk that thought Stadia would
>> work:
>>>> 
>> 
> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>> [3]
>>>> Dave Täht CEO, TekLibre, LLC
>>>> 
>> 
> <image.png><tcp_nup-2023-01-04T090937.211620.LTE.flent.gz>_______________________________________________
>>>> Rpm mailing list
>>>> Rpm@lists.bufferbloat.net
>>>> https://lists.bufferbloat.net/listinfo/rpm [4]
>>> 
>>> 
>>> _______________________________________________
>>> Rpm mailing list
>>> Rpm@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/rpm [4]
>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink
>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink
> 
> --
> 
> Bruce Perens K6BP
> 
> Links:
> ------
> [1] http://www.cs.auckland.ac.nz/%7Eulrich/
> [2] https://speed.cloudflare.com
> [3] 
> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
> [4] https://lists.bufferbloat.net/listinfo/rpm

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Bloat] [Rpm] the grinch meets cloudflare's christmas present
  2023-01-04 20:02   ` rjmcmahon
  2023-01-04 21:16     ` Ulrich Speidel
@ 2023-01-05 11:11     ` Sebastian Moeller
  1 sibling, 0 replies; 44+ messages in thread
From: Sebastian Moeller @ 2023-01-05 11:11 UTC (permalink / raw)
  To: rjmcmahon
  Cc: jf, Cake List, IETF IPPM WG, libreqos, Dave Taht via Starlink,
	Rpm, bloat

Hi Bob,


> On Jan 4, 2023, at 21:02, rjmcmahon via Bloat <bloat@lists.bufferbloat.net> wrote:
> 
> Curious to why people keep calling capacity tests speed tests? A semi at 55 mph isn't faster than a porsche at 141 mph because its load volume is larger.

	[SM] I am not sure whether answering the "why" is likely to getting us closer to remedy the situation. IMHO we are unlikely to change that just as we are unlikely to change the equally debatable use of "bandwidth" as synonym for "maximal capacity"... These two ships have sailed no matter how much shouting at clouds is going to happen ;)

My theory about the way is, this is entirely marketing driven, both device manufacturers/ISPs and end-users desire to keep things simple so ideally a single number and a catchy name... "Speed" as in top-speed was already a well known quantity for motor vehicles that consumers as a group had accepted to correlate with price. Now purist will say that "speed" is already well-defined as distance/time and "amount of data" is not a viable distance measure (how many bits are there in a meter?), but since when has marketing and the desire for simply single-number "quality indicators" ever cared much for the complexities of the real world?
	Also when remembering the old analog modem and ISDN days, at that time additional capacity truly was my main desirable, so marketing by max capacity was relevant to me independent of how this was called, I would not be amazed if I was not alone with that view. I guess that single measure and the wrong name simply stuck...

Personally I try to use rate instead of speed or bandwidth, but I note that I occasionally fail without even noticing it.

Technically I agree that one way latency is more closely related to "speed" as between any two end-points there is always a path the information travels that has a "true" length, so speed could be defined as network-path-length/OWD, but that would only be the average speed over the path... I am not sure how informative or marketable this wuld be for end-users though ;)


Regards
	Sebastian



> 
> Bob
>> HNY Dave and all the rest,
>> Great to see yet another capacity test add latency metrics to the
>> results. This one looks like a good start.
>> Results from my Windstream DOCSIS 3.1 line (3.1 on download only, up
>> is 3.0) Gigabit down / 35Mbps up provisioning. Using an IQrouter Pro
>> (an i5 x86) with Cake set for 710/31 as this ISP can’t deliver
>> reliable low-latency unless you shave a good bit off the targets. My
>> local loop is pretty congested.
>> Here’s the latest Cloudflare test:
>> And an Ookla test run just afterward:
>> They are definitely both in the ballpark and correspond to other tests
>> run from the router itself or my (wired) MacBook Pro.
>> Cheers,
>> Jonathan
>>> On Jan 4, 2023, at 12:26 PM, Dave Taht via Rpm <rpm@lists.bufferbloat.net> wrote:
>>> Please try the new, the shiny, the really wonderful test here:
>>> https://speed.cloudflare.com/
>>> I would really appreciate some independent verification of
>>> measurements using this tool. In my brief experiments it appears - as
>>> all the commercial tools to date - to dramatically understate the
>>> bufferbloat, on my LTE, (and my starlink terminal is out being
>>> hacked^H^H^H^H^H^Hworked on, so I can't measure that)
>>> My test of their test reports 223ms 5G latency under load , where
>>> flent reports over 2seconds. See comparison attached.
>>> My guess is that this otherwise lovely new  tool, like too many,
>>> doesn't run for long enough. Admittedly, most web objects (their
>>> target market) are small, and so long as they remain small and not
>>> heavily pipelined this test is a very good start... but I'm pretty
>>> sure cloudflare is used for bigger uploads and downloads than that.
>>> There's no way to change the test to run longer either.
>>> I'd love to get some results from other networks (compared as usual to
>>> flent), especially ones with cake on it. I'd love to know if they
>>> measured more minimum rtts that can be obtained with fq_codel or cake,
>>> correctly.
>>> Love Always,
>>> The Grinch
>>> --
>>> This song goes out to all the folk that thought Stadia would work:
>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>> Dave Täht CEO, TekLibre, LLC
>>> <image.png><tcp_nup-2023-01-04T090937.211620.LTE.flent.gz>_______________________________________________
>>> Rpm mailing list
>>> Rpm@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/rpm
>> _______________________________________________
>> Rpm mailing list
>> Rpm@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/rpm
> _______________________________________________
> Bloat mailing list
> Bloat@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/bloat


^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
  2023-01-05  3:11           ` Dick Roy
@ 2023-01-05 11:25             ` Sebastian Moeller
  2023-01-06  0:01               ` Dick Roy
  0 siblings, 1 reply; 44+ messages in thread
From: Sebastian Moeller @ 2023-01-05 11:25 UTC (permalink / raw)
  To: Dick Roy; +Cc: Dave Collier-Brown, starlink

Hi RR,


> On Jan 5, 2023, at 04:11, Dick Roy via Starlink <starlink@lists.bufferbloat.net> wrote:
> 
>  
>  
> From: Starlink [mailto:starlink-bounces@lists.bufferbloat.net] On Behalf Of Dave Collier-Brown via Starlink
> Sent: Wednesday, January 4, 2023 6:48 PM
> To: starlink@lists.bufferbloat.net
> Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
>  
> I think using "speed" for "the inverse of delay" is pretty normal English, if technically erroneous when speaking nerd or physicist.
> 
> [RR] I’ve not heard of that usage before.  The units aren’t commensurate either. 
> 
> Using it for volume? Arguably more like fraudulent...
> 
> [RR] I don’t think that was Bob’s intent.  I think “load volume” was meant to be a metaphor for “number of bits/bytes” being transported (“by the semi”).  
> 
> That said, aren’t users these days educated on “gigs” which they intuitively understand to be Gigabits per second (or Gbps)?  Oddly enough, that is an expression of “data/information/communication rate” in the appropriate units with the nominal technically correct meaning.  

	[SM] Gigs would have the following confounds if used without a proper definition:
a) base10 or base2^10?
b) giga-what? Bit or Byte
c) Volume or capacity
d) if capacity, minimal, average, or maximal?

I note (again, sorry to sound like a broken record) that the national regulatory agency for networks (Bundes-Netzagentur, short BNetzA) in Germany has some detailed instructions about what information ISPs need to supply to their potential customers pre-sale (see https://www.bundesnetzagentur.de/SharedDocs/Downloads/DE/Sachgebiete/Telekommunikation/Unternehmen_Institutionen/Anbieterpflichten/Kundenschutz/Transparenzmaßnahmen/Instruction_for_drawing_up_PIS.pdf?__blob=publicationFile&v=1) where the headlines talk correctly about "data transmission rates" but in the text they occasionally fall back to "speed". They also state: "Data transmission rates must be given in megabits per second (Mbit/s)."
This is both in response to our "speed" discussion, but also one potential way to clarify b) c) and d) above... given that is official this probably also answers a) (base10 otherwise the text would be "Data transmission rates must be given in mebibits per second (Mibit/s).")

--Sebastian

> 
> RR
> 
> --dave
> 
> On 1/4/23 18:54, Bruce Perens via Starlink wrote:
>> On the other hand, we would like to be comprehensible to normal users, especially when we want them to press their providers to deal with bufferbloat. Differences like speed and rate would go right over their heads.
>>  
>> On Wed, Jan 4, 2023 at 1:16 PM Ulrich Speidel via Starlink <starlink@lists.bufferbloat.net> wrote:
>>> The use of the term "speed" in communications used to be restricted to the speed of light (or whatever propagation speed one happened to be dealing with. Everything else was a "rate". Maybe I'm old-fashioned but I think talking about "speed tests" muddies the waters rather a lot.
>>>  
>>> -- 
>>> **************************************************************** 
>>> Dr. Ulrich Speidel 
>>> 
>>> Department of Computer Science 
>>> 
>>> Room 303S.594
>>> Ph: (+64-9)-373-7599 ext. 85282 
>>> 
>>> The University of Auckland 
>>> u.speidel@auckland.ac.nz
>>> http://www.cs.auckland.ac.nz/~ulrich/
>>> **************************************************************** 
>>> From: Starlink <starlink-bounces@lists.bufferbloat.net> on behalf of rjmcmahon via Starlink <starlink@lists.bufferbloat.net>
>>> Sent: Thursday, January 5, 2023 9:02 AM
>>> To: jf@jonathanfoulkes.com <jf@jonathanfoulkes.com>
>>> Cc: Cake List <cake@lists.bufferbloat.net>; IETF IPPM WG <ippm@ietf.org>; libreqos <libreqos@lists.bufferbloat.net>; Dave Taht via Starlink <starlink@lists.bufferbloat.net>; Rpm <rpm@lists.bufferbloat.net>; bloat <bloat@lists.bufferbloat.net>
>>> Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
>>>  
>>> Curious to why people keep calling capacity tests speed tests? A semi at 
>>> 55 mph isn't faster than a porsche at 141 mph because its load volume is 
>>> larger.
>>> 
>>> Bob
>>> > HNY Dave and all the rest,
>>> > 
>>> > Great to see yet another capacity test add latency metrics to the
>>> > results. This one looks like a good start.
>>> > 
>>> > Results from my Windstream DOCSIS 3.1 line (3.1 on download only, up
>>> > is 3.0) Gigabit down / 35Mbps up provisioning. Using an IQrouter Pro
>>> > (an i5 x86) with Cake set for 710/31 as this ISP can’t deliver
>>> > reliable low-latency unless you shave a good bit off the targets. My
>>> > local loop is pretty congested.
>>> > 
>>> > Here’s the latest Cloudflare test:
>>> > 
>>> > 
>>> > 
>>> > 
>>> > And an Ookla test run just afterward:
>>> > 
>>> > 
>>> > 
>>> > 
>>> > They are definitely both in the ballpark and correspond to other tests
>>> > run from the router itself or my (wired) MacBook Pro.
>>> > 
>>> > Cheers,
>>> > 
>>> > Jonathan
>>> > 
>>> > 
>>> >> On Jan 4, 2023, at 12:26 PM, Dave Taht via Rpm 
>>> >> <rpm@lists.bufferbloat.net> wrote:
>>> >> 
>>> >> Please try the new, the shiny, the really wonderful test here:
>>> >> https://speed.cloudflare.com/
>>> >> 
>>> >> I would really appreciate some independent verification of
>>> >> measurements using this tool. In my brief experiments it appears - as
>>> >> all the commercial tools to date - to dramatically understate the
>>> >> bufferbloat, on my LTE, (and my starlink terminal is out being
>>> >> hacked^H^H^H^H^H^Hworked on, so I can't measure that)
>>> >> 
>>> >> My test of their test reports 223ms 5G latency under load , where
>>> >> flent reports over 2seconds. See comparison attached.
>>> >> 
>>> >> My guess is that this otherwise lovely new tool, like too many,
>>> >> doesn't run for long enough. Admittedly, most web objects (their
>>> >> target market) are small, and so long as they remain small and not
>>> >> heavily pipelined this test is a very good start... but I'm pretty
>>> >> sure cloudflare is used for bigger uploads and downloads than that.
>>> >> There's no way to change the test to run longer either.
>>> >> 
>>> >> I'd love to get some results from other networks (compared as usual to
>>> >> flent), especially ones with cake on it. I'd love to know if they
>>> >> measured more minimum rtts that can be obtained with fq_codel or cake,
>>> >> correctly.
>>> >> 
>>> >> Love Always,
>>> >> The Grinch
>>> >> 
>>> >> --
>>> >> This song goes out to all the folk that thought Stadia would work:
>>> >> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>> >> Dave Täht CEO, TekLibre, LLC
>>> >> <image.png><tcp_nup-2023-01-04T090937.211620.LTE.flent.gz>_______________________________________________
>>> >> Rpm mailing list
>>> >> Rpm@lists.bufferbloat.net
>>> >> https://lists.bufferbloat.net/listinfo/rpm
>>> > 
>>> > 
>>> > _______________________________________________
>>> > Rpm mailing list
>>> > Rpm@lists.bufferbloat.net
>>> > https://lists.bufferbloat.net/listinfo/rpm
>>> _______________________________________________
>>> Starlink mailing list
>>> Starlink@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/starlink
>>> _______________________________________________
>>> Starlink mailing list
>>> Starlink@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/starlink
>> 
>>  
>> -- 
>> Bruce Perens K6BP
>> 
>> 
>> 
>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink
> -- 
> David Collier-Brown,         | Always do right. This will gratify
> System Programmer and Author | some people and astonish the rest
> dave.collier-brown@indexexchange.com |              -- Mark Twain
>  
> CONFIDENTIALITY NOTICE AND DISCLAIMER : This telecommunication, including any and all attachments, contains confidential information intended only for the person(s) to whom it is addressed. Any dissemination, distribution, copying or disclosure is strictly prohibited and is not a waiver of confidentiality. If you have received this telecommunication in error, please notify the sender immediately by return electronic mail and delete the message from your inbox and deleted items folders. This telecommunication does not constitute an express or implied agreement to conduct transactions by electronic means, nor does it constitute a contract offer, a contract amendment or an acceptance of a contract offer. Contract terms contained in this telecommunication are subject to legal review and the completion of formal documentation and are not binding until same is confirmed in writing and has been signed by an authorized signatory.
> 
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink


^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
  2023-01-05 11:25             ` Sebastian Moeller
@ 2023-01-06  0:01               ` Dick Roy
  2023-01-06  9:43                 ` Sebastian Moeller
  0 siblings, 1 reply; 44+ messages in thread
From: Dick Roy @ 2023-01-06  0:01 UTC (permalink / raw)
  To: 'Sebastian Moeller'; +Cc: 'Dave Collier-Brown', starlink

[-- Attachment #1: Type: text/plain, Size: 11785 bytes --]

Hi Sebastian,

 

See below …

 

-----Original Message-----
From: Sebastian Moeller [mailto:moeller0@gmx.de] 
Sent: Thursday, January 5, 2023 3:26 AM
To: Dick Roy
Cc: Dave Collier-Brown; starlink@lists.bufferbloat.net
Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas
present

 

Hi RR,

 

 

> On Jan 5, 2023, at 04:11, Dick Roy via Starlink
<starlink@lists.bufferbloat.net> wrote:

> 

>  

>  

> From: Starlink [mailto:starlink-bounces@lists.bufferbloat.net] On Behalf
Of Dave Collier-Brown via Starlink

> Sent: Wednesday, January 4, 2023 6:48 PM

> To: starlink@lists.bufferbloat.net

> Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas
present

>  

> I think using "speed" for "the inverse of delay" is pretty normal English,
if technically erroneous when speaking nerd or physicist.

> 

> [RR] I’ve not heard of that usage before.  The units aren’t commensurate
either. 

> 

> Using it for volume? Arguably more like fraudulent...

> 

> [RR] I don’t think that was Bob’s intent.  I think “load volume” was meant
to be a metaphor for “number of bits/bytes” being transported (“by the
semi”).  

> 

> That said, aren’t users these days educated on “gigs” which they
intuitively understand to be Gigabits per second (or Gbps)?  Oddly enough,
that is an expression of “data/information/communication rate” in the
appropriate units with the nominal technically correct meaning.  

 

      [SM] Gigs would have the following confounds if used without a proper
definition:

a) base10 or base2^10?

b) giga-what? Bit or Byte

c) Volume or capacity

d) if capacity, minimal, average, or maximal?

 

I note (again, sorry to sound like a broken record) that the national
regulatory agency for networks (Bundes-Netzagentur, short BNetzA) in Germany
has some detailed instructions about what information ISPs need to supply to
their potential customers pre-sale (see
https://www.bundesnetzagentur.de/SharedDocs/Downloads/DE/Sachgebiete/Telekom
munikation/Unternehmen_Institutionen/Anbieterpflichten/Kundenschutz/Transpar
enzmaßnahmen/Instruction_for_drawing_up_PIS.pdf?__blob=publicationFile&v=1)
where the headlines talk correctly about "data transmission rates" but in
the text they occasionally fall back to "speed". They also state: "Data
transmission rates must be given in megabits per second (Mbit/s)."

This is both in response to our "speed" discussion, but also one potential
way to clarify b) c) and d) above... given that is official this probably
also answers a) (base10 otherwise the text would be "Data transmission rates
must be given in mebibits per second (Mibit/s).")

[RR] My reference to “gigs” was to the ads out nowadays from AT&T about
becoming Gagillionaires (“Yes, I am Jurgous. … We know!”) that “now have gig
speed wireless from AT&T” so they can play all kinds of VR games. :-) That
said, not sure why BNetzA mandates a particular unit for information rates,
but that’s their prerogative I guess.  Given that the fundamental unit of
information is the answer to a YES/NO question (aka a “bit”), it makes sense
to measure information in bits (although trits or any other higher order
concept could be used as long as the system accounted for fractions
thereof:-)) (and sets of bits (aka bytes or really octets) because of
ancient computer arcitectures:-)).  Since we have pretty much settled on the
SI second as the accepted unit of time (and multiples thereof e.g. msec,
usec, nsec, etc.), it makes sense to measure information flow in bits/sec or
some multiples thereof such as Gbps, Mbps, Kbps, etc. and their byte (really
octet) versions GBps, MBps, KBps, etc.. Not sure why BNetzA mandates ONLY
one of these, but whatever … :-)

 

As for capacity, remember capacity is not something that is measured. It is
a fundamental property (an information rate!) of a communication channel
which has no other attributes such as minimal, average, or maximal (unless
one is talking about time-varying channels and is wanting to characterize
the capacity of the channel over time, but that’s another story).  As such,
comparing volume and capacity is comparing apples and oranges; one is a size
of something (e.g. number of megabytes) and the other is a rate (e.g. MBps)
so I am not sure what “Volume or capacity” really means. I suspect the
concept you may be looking for is “achievable rate” rather than “capacity”.
Achievable rate IS something that is measureable, and varies with load when
channels are shared, etc..  Loosely speaking, achievable rate is always less
than or equal to the capacity of a channel. 

 

HNY,

 

RR

 

--Sebastian

 

> 

> RR

> 

> --dave

> 

> On 1/4/23 18:54, Bruce Perens via Starlink wrote:

>> On the other hand, we would like to be comprehensible to normal users,
especially when we want them to press their providers to deal with
bufferbloat. Differences like speed and rate would go right over their
heads.

>>  

>> On Wed, Jan 4, 2023 at 1:16 PM Ulrich Speidel via Starlink
<starlink@lists.bufferbloat.net> wrote:

>>> The use of the term "speed" in communications used to be restricted to
the speed of light (or whatever propagation speed one happened to be dealing
with. Everything else was a "rate". Maybe I'm old-fashioned but I think
talking about "speed tests" muddies the waters rather a lot.

>>>  

>>> -- 

>>> **************************************************************** 

>>> Dr. Ulrich Speidel 

>>> 

>>> Department of Computer Science 

>>> 

>>> Room 303S.594

>>> Ph: (+64-9)-373-7599 ext. 85282 

>>> 

>>> The University of Auckland 

>>> u.speidel@auckland.ac.nz

>>> http://www.cs.auckland.ac.nz/~ulrich/

>>> **************************************************************** 

>>> From: Starlink <starlink-bounces@lists.bufferbloat.net> on behalf of
rjmcmahon via Starlink <starlink@lists.bufferbloat.net>

>>> Sent: Thursday, January 5, 2023 9:02 AM

>>> To: jf@jonathanfoulkes.com <jf@jonathanfoulkes.com>

>>> Cc: Cake List <cake@lists.bufferbloat.net>; IETF IPPM WG
<ippm@ietf.org>; libreqos <libreqos@lists.bufferbloat.net>; Dave Taht via
Starlink <starlink@lists.bufferbloat.net>; Rpm <rpm@lists.bufferbloat.net>;
bloat <bloat@lists.bufferbloat.net>

>>> Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas
present

>>>  

>>> Curious to why people keep calling capacity tests speed tests? A semi at


>>> 55 mph isn't faster than a porsche at 141 mph because its load volume is


>>> larger.

>>> 

>>> Bob

>>> > HNY Dave and all the rest,

>>> > 

>>> > Great to see yet another capacity test add latency metrics to the

>>> > results. This one looks like a good start.

>>> > 

>>> > Results from my Windstream DOCSIS 3.1 line (3.1 on download only, up

>>> > is 3.0) Gigabit down / 35Mbps up provisioning. Using an IQrouter Pro

>>> > (an i5 x86) with Cake set for 710/31 as this ISP can’t deliver

>>> > reliable low-latency unless you shave a good bit off the targets. My

>>> > local loop is pretty congested.

>>> > 

>>> > Here’s the latest Cloudflare test:

>>> > 

>>> > 

>>> > 

>>> > 

>>> > And an Ookla test run just afterward:

>>> > 

>>> > 

>>> > 

>>> > 

>>> > They are definitely both in the ballpark and correspond to other tests

>>> > run from the router itself or my (wired) MacBook Pro.

>>> > 

>>> > Cheers,

>>> > 

>>> > Jonathan

>>> > 

>>> > 

>>> >> On Jan 4, 2023, at 12:26 PM, Dave Taht via Rpm 

>>> >> <rpm@lists.bufferbloat.net> wrote:

>>> >> 

>>> >> Please try the new, the shiny, the really wonderful test here:

>>> >> https://speed.cloudflare.com/

>>> >> 

>>> >> I would really appreciate some independent verification of

>>> >> measurements using this tool. In my brief experiments it appears - as

>>> >> all the commercial tools to date - to dramatically understate the

>>> >> bufferbloat, on my LTE, (and my starlink terminal is out being

>>> >> hacked^H^H^H^H^H^Hworked on, so I can't measure that)

>>> >> 

>>> >> My test of their test reports 223ms 5G latency under load , where

>>> >> flent reports over 2seconds. See comparison attached.

>>> >> 

>>> >> My guess is that this otherwise lovely new tool, like too many,

>>> >> doesn't run for long enough. Admittedly, most web objects (their

>>> >> target market) are small, and so long as they remain small and not

>>> >> heavily pipelined this test is a very good start... but I'm pretty

>>> >> sure cloudflare is used for bigger uploads and downloads than that.

>>> >> There's no way to change the test to run longer either.

>>> >> 

>>> >> I'd love to get some results from other networks (compared as usual
to

>>> >> flent), especially ones with cake on it. I'd love to know if they

>>> >> measured more minimum rtts that can be obtained with fq_codel or
cake,

>>> >> correctly.

>>> >> 

>>> >> Love Always,

>>> >> The Grinch

>>> >> 

>>> >> --

>>> >> This song goes out to all the folk that thought Stadia would work:

>>> >>
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-698136666560
7352320-FXtz

>>> >> Dave Täht CEO, TekLibre, LLC

>>> >>
<image.png><tcp_nup-2023-01-04T090937.211620.LTE.flent.gz>__________________
_____________________________

>>> >> Rpm mailing list

>>> >> Rpm@lists.bufferbloat.net

>>> >> https://lists.bufferbloat.net/listinfo/rpm

>>> > 

>>> > 

>>> > _______________________________________________

>>> > Rpm mailing list

>>> > Rpm@lists.bufferbloat.net

>>> > https://lists.bufferbloat.net/listinfo/rpm

>>> _______________________________________________

>>> Starlink mailing list

>>> Starlink@lists.bufferbloat.net

>>> https://lists.bufferbloat.net/listinfo/starlink

>>> _______________________________________________

>>> Starlink mailing list

>>> Starlink@lists.bufferbloat.net

>>> https://lists.bufferbloat.net/listinfo/starlink

>> 

>>  

>> -- 

>> Bruce Perens K6BP

>> 

>> 

>> 

>> _______________________________________________

>> Starlink mailing list

>> Starlink@lists.bufferbloat.net

>> https://lists.bufferbloat.net/listinfo/starlink

> -- 

> David Collier-Brown,         | Always do right. This will gratify

> System Programmer and Author | some people and astonish the rest

> dave.collier-brown@indexexchange.com |              -- Mark Twain

>  

> CONFIDENTIALITY NOTICE AND DISCLAIMER : This telecommunication, including
any and all attachments, contains confidential information intended only for
the person(s) to whom it is addressed. Any dissemination, distribution,
copying or disclosure is strictly prohibited and is not a waiver of
confidentiality. If you have received this telecommunication in error,
please notify the sender immediately by return electronic mail and delete
the message from your inbox and deleted items folders. This
telecommunication does not constitute an express or implied agreement to
conduct transactions by electronic means, nor does it constitute a contract
offer, a contract amendment or an acceptance of a contract offer. Contract
terms contained in this telecommunication are subject to legal review and
the completion of formal documentation and are not binding until same is
confirmed in writing and has been signed by an authorized signatory.

> 

> _______________________________________________

> Starlink mailing list

> Starlink@lists.bufferbloat.net

> https://lists.bufferbloat.net/listinfo/starlink


[-- Attachment #2: Type: text/html, Size: 38418 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
  2023-01-06  0:01               ` Dick Roy
@ 2023-01-06  9:43                 ` Sebastian Moeller
  0 siblings, 0 replies; 44+ messages in thread
From: Sebastian Moeller @ 2023-01-06  9:43 UTC (permalink / raw)
  To: Dick Roy; +Cc: Dave Collier-Brown, starlink

Hi RR,


> On Jan 6, 2023, at 01:01, Dick Roy <dickroy@alum.mit.edu> wrote:
> 
> Hi Sebastian,
>  
> See below …
>  
> -----Original Message-----
> From: Sebastian Moeller [mailto:moeller0@gmx.de] 
> Sent: Thursday, January 5, 2023 3:26 AM
> To: Dick Roy
> Cc: Dave Collier-Brown; starlink@lists.bufferbloat.net
> Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
>  
> Hi RR,
>  
>  
> > On Jan 5, 2023, at 04:11, Dick Roy via Starlink <starlink@lists.bufferbloat.net> wrote:
> > 
> >  
> >  
> > From: Starlink [mailto:starlink-bounces@lists.bufferbloat.net] On Behalf Of Dave Collier-Brown via Starlink
> > Sent: Wednesday, January 4, 2023 6:48 PM
> > To: starlink@lists.bufferbloat.net
> > Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
> >  
> > I think using "speed" for "the inverse of delay" is pretty normal English, if technically erroneous when speaking nerd or physicist.
> > 
> > [RR] I’ve not heard of that usage before.  The units aren’t commensurate either. 
> > 
> > Using it for volume? Arguably more like fraudulent...
> > 
> > [RR] I don’t think that was Bob’s intent.  I think “load volume” was meant to be a metaphor for “number of bits/bytes” being transported (“by the semi”).  
> > 
> > That said, aren’t users these days educated on “gigs” which they intuitively understand to be Gigabits per second (or Gbps)?  Oddly enough, that is an expression of “data/information/communication rate” in the appropriate units with the nominal technically correct meaning.  
>  
>       [SM] Gigs would have the following confounds if used without a proper definition:
> a) base10 or base2^10?
> b) giga-what? Bit or Byte
> c) Volume or capacity
> d) if capacity, minimal, average, or maximal?
>  
> I note (again, sorry to sound like a broken record) that the national regulatory agency for networks (Bundes-Netzagentur, short BNetzA) in Germany has some detailed instructions about what information ISPs need to supply to their potential customers pre-sale (seehttps://www.bundesnetzagentur.de/SharedDocs/Downloads/DE/Sachgebiete/Telekommunikation/Unternehmen_Institutionen/Anbieterpflichten/Kundenschutz/Transparenzmaßnahmen/Instruction_for_drawing_up_PIS.pdf?__blob=publicationFile&v=1) where the headlines talk correctly about "data transmission rates" but in the text they occasionally fall back to "speed". They also state: "Data transmission rates must be given in megabits per second (Mbit/s)."
> This is both in response to our "speed" discussion, but also one potential way to clarify b) c) and d) above... given that is official this probably also answers a) (base10 otherwise the text would be "Data transmission rates must be given in mebibits per second (Mibit/s).")

> [RR] My reference to “gigs” was to the ads out nowadays from AT&T about becoming Gagillionaires (“Yes, I am Jurgous. … We know!”) that “now have gig speed wireless from AT&T” so they can play all kinds of VR games. 

	[SM2] Ah, not being in the U.S. that campaign has completely evaded my attention.


> J That said, not sure why BNetzA mandates a particular unit for information rates, but that’s their prerogative I guess.

	[SM2] My bigger point was actually that they use the term "data transmission rates" instead of speed... But to your point the product information sheets are designed specifically to make it easy for consumers to compare different offers (for all values of consumer knowledge of networking and SI-prefixes) and requiring a single unit makes comparison simple and gives less leeway to play games. This BNetzA initiative basically removed the previous "up to XX Mbps" promise of ISPs and especially the interpretation that YY with YY << XX is still the contractually fine as "up to" implies not guarantee.


>  Given that the fundamental unit of information is the answer to a YES/NO question (aka a “bit”), it makes sense to measure information in bits (although trits or any other higher order concept could be used as long as the system accounted for fractions thereofJ) (and sets of bits (aka bytes or really octets) because of ancient computer arcitecturesJ).

	[SM2] I agree, but prefer this to be inherent in the term used, and "gigs" did not carry any of that information for me, but again I had not encountered the AT&T campaign...


>  Since we have pretty much settled on the SI second as the accepted unit of time (and multiples thereof e.g. msec, usec, nsec, etc.), it makes sense to measure information flow in bits/sec or some multiples thereof such as Gbps, Mbps, Kbps, etc. and their byte (really octet) versions GBps, MBps, KBps, etc.. Not sure why BNetzA mandates ONLY one of these, but whatever … J

	[SM2] Again I speculate that this is to allow easy comparison even by consumers that are not "fluent" in interpreting SI-prefixes and that might e.g. not know by heart whether Mb is larger or smaller than Gb, let alone why mB is not equal to Mb... this is a transparency measure foremost aimed at all end-customers (business contract do not fall under that regulation as far as I know).

> As for capacity, remember capacity is not something that is measured. It is a fundamental property (an information rate!) of a communication channel which has no other attributes such as minimal, average, or maximal (unless one is talking about time-varying channels and is wanting to characterize the capacity of the channel over time, but that’s another story).

	[SM] On a shared medium (and let's face it head on sooner or later "internet-access" will become a shared medium) the capacity share each user can relay on is rarely static, most often ISPs oversubscribe links and use traffic shapers to limit the maximum capacity share a user can use, but that share is not guaranteed to be available. BNetzA went as far as defining three different data rates with different expected (temporal) availability as well as a (rather complicated) method how end-users can confirm whether ISPs actually deliver the promised rates.


>  As such, comparing volume and capacity is comparing apples and oranges; one is a size of something (e.g. number of megabytes) and the other is a rate (e.g. MBps) so I am not sure what “Volume or capacity” really means.

	[SM] The term Gigs unlike e.g. Mb/s or Mbps (for Megabits per second) does not intuitively define whether time is taken into account, and over here mobile contracts typically come with a "high-speed Volume" after the consumption of which mobile links fall back to truly atrocious rates like 32Kbps, for these kind of contracts ISPs tend to primarily market the Volume (mostly in GB) and only mention the maximal data rates as secondary information.


> I suspect the concept you may be looking for is “achievable rate” rather than “capacity”.

	[SM2] Actually I always try to first calculate the theoretical upper limit of my links (which I then dub the link's "capacity") and I compare the actual measured rates with the theoretical limit... This works well enough for me, since the consumer links I use typically are dominated by a predictable traffic shaper on the ISP side...


>  Achievable rate IS something that is measureable, and varies with load when channels are shared, etc..  Loosely speaking, achievable rate is always less than or equal to the capacity of a channel. 

	[SM2] Yepp, that is why I try to calculate a capacity estimate (and calculate the resulting throughput on different levels).

Regards
	Sebastian


>  
> HNY,
>  
> RR
>  
> --Sebastian
>  
> > 
> > RR
> > 
> > --dave
> > 
> > On 1/4/23 18:54, Bruce Perens via Starlink wrote:
> >> On the other hand, we would like to be comprehensible to normal users, especially when we want them to press their providers to deal with bufferbloat. Differences like speed and rate would go right over their heads.
> >>  
> >> On Wed, Jan 4, 2023 at 1:16 PM Ulrich Speidel via Starlink <starlink@lists.bufferbloat.net> wrote:
> >>> The use of the term "speed" in communications used to be restricted to the speed of light (or whatever propagation speed one happened to be dealing with. Everything else was a "rate". Maybe I'm old-fashioned but I think talking about "speed tests" muddies the waters rather a lot.
> >>>  
> >>> -- 
> >>> **************************************************************** 
> >>> Dr. Ulrich Speidel
> >>> 
> >>> Department of Computer Science 
> >>> 
> >>> Room 303S.594
> >>> Ph: (+64-9)-373-7599 ext. 85282 
> >>> 
> >>> The University of Auckland
> >>> u.speidel@auckland.ac.nz
> >>> http://www.cs.auckland.ac.nz/~ulrich/
> >>> **************************************************************** 
> >>> From: Starlink <starlink-bounces@lists.bufferbloat.net> on behalf of rjmcmahon via Starlink <starlink@lists.bufferbloat.net>
> >>> Sent: Thursday, January 5, 2023 9:02 AM
> >>> To: jf@jonathanfoulkes.com <jf@jonathanfoulkes.com>
> >>> Cc: Cake List <cake@lists.bufferbloat.net>; IETF IPPM WG <ippm@ietf.org>; libreqos <libreqos@lists.bufferbloat.net>; Dave Taht via Starlink <starlink@lists.bufferbloat.net>; Rpm <rpm@lists.bufferbloat.net>; bloat <bloat@lists.bufferbloat.net>
> >>> Subject: Re: [Starlink] [Rpm] the grinch meets cloudflare's christmas present
> >>>  
> >>> Curious to why people keep calling capacity tests speed tests? A semi at 
> >>> 55 mph isn't faster than a porsche at 141 mph because its load volume is 
> >>> larger.
> >>> 
> >>> Bob
> >>> > HNY Dave and all the rest,
> >>> > 
> >>> > Great to see yet another capacity test add latency metrics to the
> >>> > results. This one looks like a good start.
> >>> > 
> >>> > Results from my Windstream DOCSIS 3.1 line (3.1 on download only, up
> >>> > is 3.0) Gigabit down / 35Mbps up provisioning. Using an IQrouter Pro
> >>> > (an i5 x86) with Cake set for 710/31 as this ISP can’t deliver
> >>> > reliable low-latency unless you shave a good bit off the targets. My
> >>> > local loop is pretty congested.
> >>> > 
> >>> > Here’s the latest Cloudflare test:
> >>> > 
> >>> > 
> >>> > 
> >>> > 
> >>> > And an Ookla test run just afterward:
> >>> > 
> >>> > 
> >>> > 
> >>> > 
> >>> > They are definitely both in the ballpark and correspond to other tests
> >>> > run from the router itself or my (wired) MacBook Pro.
> >>> > 
> >>> > Cheers,
> >>> > 
> >>> > Jonathan
> >>> > 
> >>> > 
> >>> >> On Jan 4, 2023, at 12:26 PM, Dave Taht via Rpm 
> >>> >> <rpm@lists.bufferbloat.net> wrote:
> >>> >> 
> >>> >> Please try the new, the shiny, the really wonderful test here:
> >>> >> https://speed.cloudflare.com/
> >>> >> 
> >>> >> I would really appreciate some independent verification of
> >>> >> measurements using this tool. In my brief experiments it appears - as
> >>> >> all the commercial tools to date - to dramatically understate the
> >>> >> bufferbloat, on my LTE, (and my starlink terminal is out being
> >>> >> hacked^H^H^H^H^H^Hworked on, so I can't measure that)
> >>> >> 
> >>> >> My test of their test reports 223ms 5G latency under load , where
> >>> >> flent reports over 2seconds. See comparison attached.
> >>> >> 
> >>> >> My guess is that this otherwise lovely new tool, like too many,
> >>> >> doesn't run for long enough. Admittedly, most web objects (their
> >>> >> target market) are small, and so long as they remain small and not
> >>> >> heavily pipelined this test is a very good start... but I'm pretty
> >>> >> sure cloudflare is used for bigger uploads and downloads than that.
> >>> >> There's no way to change the test to run longer either.
> >>> >> 
> >>> >> I'd love to get some results from other networks (compared as usual to
> >>> >> flent), especially ones with cake on it. I'd love to know if they
> >>> >> measured more minimum rtts that can be obtained with fq_codel or cake,
> >>> >> correctly.
> >>> >> 
> >>> >> Love Always,
> >>> >> The Grinch
> >>> >> 
> >>> >> --
> >>> >> This song goes out to all the folk that thought Stadia would work:
> >>> >> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
> >>> >> Dave Täht CEO, TekLibre, LLC
> >>> >> <image.png><tcp_nup-2023-01-04T090937.211620.LTE.flent.gz>_______________________________________________
> >>> >> Rpm mailing list
> >>> >> Rpm@lists.bufferbloat.net
> >>> >> https://lists.bufferbloat.net/listinfo/rpm
> >>> > 
> >>> > 
> >>> > _______________________________________________
> >>> > Rpm mailing list
> >>> > Rpm@lists.bufferbloat.net
> >>> > https://lists.bufferbloat.net/listinfo/rpm
> >>> _______________________________________________
> >>> Starlink mailing list
> >>> Starlink@lists.bufferbloat.net
> >>> https://lists.bufferbloat.net/listinfo/starlink
> >>> _______________________________________________
> >>> Starlink mailing list
> >>> Starlink@lists.bufferbloat.net
> >>> https://lists.bufferbloat.net/listinfo/starlink
> >> 
> >>  
> >> -- 
> >> Bruce Perens K6BP
> >> 
> >> 
> >> 
> >> _______________________________________________
> >> Starlink mailing list
> >> Starlink@lists.bufferbloat.net
> >> https://lists.bufferbloat.net/listinfo/starlink
> > -- 
> > David Collier-Brown,         | Always do right. This will gratify
> > System Programmer and Author | some people and astonish the rest
> > dave.collier-brown@indexexchange.com |              -- Mark Twain
> >  
> > CONFIDENTIALITY NOTICE AND DISCLAIMER : This telecommunication, including any and all attachments, contains confidential information intended only for the person(s) to whom it is addressed. Any dissemination, distribution, copying or disclosure is strictly prohibited and is not a waiver of confidentiality. If you have received this telecommunication in error, please notify the sender immediately by return electronic mail and delete the message from your inbox and deleted items folders. This telecommunication does not constitute an express or implied agreement to conduct transactions by electronic means, nor does it constitute a contract offer, a contract amendment or an acceptance of a contract offer. Contract terms contained in this telecommunication are subject to legal review and the completion of formal documentation and are not binding until same is confirmed in writing and has been signed by an authorized signatory.
> > 
> > _______________________________________________
> > Starlink mailing list
> > Starlink@lists.bufferbloat.net
> > https://lists.bufferbloat.net/listinfo/starlink


^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [LibreQoS] the grinch meets cloudflare's christmas present
  2023-01-04 17:26 [Starlink] the grinch meets cloudflare's christmas present Dave Taht
  2023-01-04 19:20 ` [Starlink] [Rpm] " jf
@ 2023-01-06 16:38 ` MORTON JR., AL
  2023-01-06 20:38   ` [Starlink] [Rpm] " rjmcmahon
  1 sibling, 1 reply; 44+ messages in thread
From: MORTON JR., AL @ 2023-01-06 16:38 UTC (permalink / raw)
  To: Dave Taht, bloat, libreqos, Cake List, Dave Taht via Starlink,
	Rpm, IETF IPPM WG


[-- Attachment #1.1: Type: text/plain, Size: 2347 bytes --]

> -----Original Message-----

> From: LibreQoS <libreqos-bounces@lists.bufferbloat.net> On Behalf Of Dave Taht

> via LibreQoS

> Sent: Wednesday, January 4, 2023 12:26 PM

> Subject: [LibreQoS] the grinch meets cloudflare's christmas present

>

> Please try the new, the shiny, the really wonderful test here:

> https://urldefense.com/v3/__https://speed.cloudflare.com/__;!!BhdT!iZcFJ8WVU9S<https://urldefense.com/v3/__https:/speed.cloudflare.com/__;!!BhdT!iZcFJ8WVU9S9zz5t456oxkfObrC5Xb9j5AG8UO3DqD5x4GAJkawZr0iGwEUtF0_09U8mCDnAkrJ9QEMHGbCMKVw$>

> 9zz5t456oxkfObrC5Xb9j5AG8UO3DqD5x4GAJkawZr0iGwEUtF0_09U8mCDnAkrJ9QEMHGbCMKVw$<https://urldefense.com/v3/__https:/speed.cloudflare.com/__;!!BhdT!iZcFJ8WVU9S9zz5t456oxkfObrC5Xb9j5AG8UO3DqD5x4GAJkawZr0iGwEUtF0_09U8mCDnAkrJ9QEMHGbCMKVw$>

>

> I would really appreciate some independent verification of

> measurements using this tool. In my brief experiments it appears - as

> all the commercial tools to date - to dramatically understate the

> bufferbloat, on my LTE, (and my starlink terminal is out being

> hacked^H^H^H^H^H^Hworked on, so I can't measure that)

[acm]



Hi Dave, I made some time to test "cloudflare's christmas present" yesterday.



I'm on DOCSIS 3.1 service with 1Gbps Down. The Upstream has a "turbo" mode with 40-50Mbps for the first ~3 sec, then steady-state about 23Mbps.



When I saw the ~620Mbps Downstream measurement, I was ready to complain that even the IP-Layer Capacity was grossly underestimated.  In addition, the Latency measurements seem very low (as you asserted), although the cloud server was “nearby”.



However, I found that Ookla and the ISP-provided measurement were also reporting ~600Mbps! So the cloudflare Downstream capacity (or throughput?) measurement was consistent with others. Our UDPST server was unreachable, otherwise I would have added that measurement, too.



The Upstream measurement graph seems to illustrate the “turbo” mode, with the dip after attaining 44.5Mbps.



UDPST saturates the uplink and we measure the full 250ms of the Upstream buffer. Cloudflare’s latency measurements don’t even come close.



Al






[Screen Shot 2023-01-05 at 5.54.26 PM.png][Screen Shot 2023-01-05 at 5.54.53 PM.png][Screen Shot 2023-01-05 at 5.55.39 PM.png]



[-- Attachment #1.2: Type: text/html, Size: 42872 bytes --]

[-- Attachment #2: image001.png --]
[-- Type: image/png, Size: 176230 bytes --]

[-- Attachment #3: image002.png --]
[-- Type: image/png, Size: 461849 bytes --]

[-- Attachment #4: image003.png --]
[-- Type: image/png, Size: 225012 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Rpm] [LibreQoS] the grinch meets cloudflare's christmas present
  2023-01-06 16:38 ` [Starlink] [LibreQoS] " MORTON JR., AL
@ 2023-01-06 20:38   ` rjmcmahon
  2023-01-06 20:47     ` rjmcmahon
       [not found]     ` <89D796E75967416B9723211C183A8396@SRA6>
  0 siblings, 2 replies; 44+ messages in thread
From: rjmcmahon @ 2023-01-06 20:38 UTC (permalink / raw)
  To: MORTON JR., AL
  Cc: Dave Taht, bloat, libreqos, Cake List, Dave Taht via Starlink,
	Rpm, IETF IPPM WG

Some thoughts are not to use UDP for testing here. Also, these speed 
tests have little to no information for network engineers about what's 
going on. Iperf 2 may better assist network engineers but then I'm 
biased ;)

Running iperf 2 https://sourceforge.net/projects/iperf2/ with 
--trip-times. Though the sampling and central limit theorem averaging is 
hiding the real distributions (use --histograms to get those)

Below are 4 parallel TCP streams from my home to one of my servers in 
the cloud. First where TCP is limited per CCA. Second is source side 
write rate limiting. Things to note:

o) connect times for both at 10-15 ms
o) multiple TCP retries on a few rites - one case is 4 with 5 writes. 
Source side pacing eliminates retries
o) Fairness with CCA isn't great but quite good with source side write 
pacing
o) Queue depth with CCA is about 150 Kbytes about 100K byte with source 
side pacing
o) min write to read is about 80 ms for both
o) max is 220 ms vs 97 ms
o) stdev for CCA write/read is 30 ms vs 3 ms
o) TCP RTT is 20ms w/CCA and 90 ms with ssp - seems odd here as 
TCP_QUICACK and TCP_NODELAY are both enabled.

[ CT] final connect times (min/avg/max/stdev) = 
10.326/13.522/14.986/2150.329 ms (tot/err) = 4/0
[rjmcmahon@ryzen3950 iperf2-code]$ iperf -c *** --hide-ips -e 
--trip-times -i 1 -P 4 -t 10 -w 4m --tcp-quickack -N
------------------------------------------------------------
Client connecting to (**hidden**), TCP port 5001 with pid 107678 (4 
flows)
Write buffer size: 131072 Byte
TOS set to 0x0 and nodelay (Nagle off)
TCP window size: 7.63 MByte (WARNING: requested 3.81 MByte)
Event based writes (pending queue watermark at 16384 bytes)
------------------------------------------------------------
[  1] local *.*.*.85%enp4s0 port 42480 connected with *.*.*.123 port 
5001 (prefetch=16384) (trip-times) (sock=3) (qack) 
(icwnd/mss/irtt=14/1448/10534) (ct=10.63 ms) on 2023-01-06 12:17:56 
(PST)
[  4] local *.*.*.85%enp4s0 port 42488 connected with *.*.*.123 port 
5001 (prefetch=16384) (trip-times) (sock=5) (qack) 
(icwnd/mss/irtt=14/1448/14023) (ct=14.08 ms) on 2023-01-06 12:17:56 
(PST)
[  3] local *.*.*.85%enp4s0 port 42502 connected with *.*.*.123 port 
5001 (prefetch=16384) (trip-times) (sock=6) (qack) 
(icwnd/mss/irtt=14/1448/14642) (ct=14.70 ms) on 2023-01-06 12:17:56 
(PST)
[  2] local *.*.*.85%enp4s0 port 42484 connected with *.*.*.123 port 
5001 (prefetch=16384) (trip-times) (sock=4) (qack) 
(icwnd/mss/irtt=14/1448/14728) (ct=14.79 ms) on 2023-01-06 12:17:56 
(PST)
[ ID] Interval        Transfer    Bandwidth       Write/Err  Rtry     
Cwnd/RTT(var)        NetPwr
...
[  4] 4.00-5.00 sec  1.38 MBytes  11.5 Mbits/sec  11/0         3       
29K/21088(1142) us  68.37
[  2] 4.00-5.00 sec  1.62 MBytes  13.6 Mbits/sec  13/0         2       
31K/19284(612) us  88.36
[  1] 4.00-5.00 sec   896 KBytes  7.34 Mbits/sec  7/0         5       
16K/18996(658) us  48.30
[  3] 4.00-5.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         5       
18K/18133(208) us  57.83
[SUM] 4.00-5.00 sec  4.88 MBytes  40.9 Mbits/sec  39/0        15
[  4] 5.00-6.00 sec  1.25 MBytes  10.5 Mbits/sec  10/0         4       
29K/14717(489) us  89.06
[  1] 5.00-6.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         4       
16K/15874(408) us  66.06
[  3] 5.00-6.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         4       
16K/15826(382) us  74.54
[  2] 5.00-6.00 sec  1.50 MBytes  12.6 Mbits/sec  12/0         6        
9K/14878(557) us  106
[SUM] 5.00-6.00 sec  4.88 MBytes  40.9 Mbits/sec  39/0        18
[  4] 6.00-7.00 sec  1.75 MBytes  14.7 Mbits/sec  14/0         4       
25K/15472(496) us  119
[  2] 6.00-7.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         2       
26K/16417(427) us  63.87
[  1] 6.00-7.00 sec  1.25 MBytes  10.5 Mbits/sec  10/0         5       
16K/16268(679) us  80.57
[  3] 6.00-7.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         6       
15K/16629(799) us  63.06
[SUM] 6.00-7.00 sec  5.00 MBytes  41.9 Mbits/sec  40/0        17
[  4] 7.00-8.00 sec  1.75 MBytes  14.7 Mbits/sec  14/0         4       
22K/13986(519) us  131
[  1] 7.00-8.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         4       
16K/12679(377) us  93.04
[  3] 7.00-8.00 sec   896 KBytes  7.34 Mbits/sec  7/0         5       
14K/12971(367) us  70.74
[  2] 7.00-8.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         6       
15K/14740(779) us  80.03
[SUM] 7.00-8.00 sec  4.88 MBytes  40.9 Mbits/sec  39/0        19

[root@bobcat iperf2-code]# iperf -s -i 1 -e --hide-ips -w 4m
------------------------------------------------------------
Server listening on TCP port 5001 with pid 233615
Read buffer size:  128 KByte (Dist bin width=16.0 KByte)
TCP window size: 7.63 MByte (WARNING: requested 3.81 MByte)
------------------------------------------------------------
[  1] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port 42480 
(trip-times) (sock=4) (peer 2.1.9-master) (qack) 
(icwnd/mss/irtt=14/1448/11636) on 2023-01-06 12:17:56 (PST)
[  2] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port 42502 
(trip-times) (sock=5) (peer 2.1.9-master) (qack) 
(icwnd/mss/irtt=14/1448/11898) on 2023-01-06 12:17:56 (PST)
[  3] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port 42484 
(trip-times) (sock=6) (peer 2.1.9-master) (qack) 
(icwnd/mss/irtt=14/1448/11938) on 2023-01-06 12:17:56 (PST)
[  4] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port 42488 
(trip-times) (sock=7) (peer 2.1.9-master) (qack) 
(icwnd/mss/irtt=14/1448/11919) on 2023-01-06 12:17:56 (PST)
[ ID] Interval        Transfer    Bandwidth    Burst Latency 
avg/min/max/stdev (cnt/size) inP NetPwr  Reads=Dist
...
[  2] 4.00-5.00 sec  1.06 MBytes  8.86 Mbits/sec  
129.819/90.391/186.075/31.346 ms (9/123080)  154 KByte 8.532803  
467=461:6:0:0:0:0:0:0
[  3] 4.00-5.00 sec  1.52 MBytes  12.8 Mbits/sec  
103.552/82.653/169.274/28.382 ms (12/132854)  149 KByte 15.40  
646=643:1:2:0:0:0:0:0
[  4] 4.00-5.00 sec  1.39 MBytes  11.6 Mbits/sec  
107.503/66.843/143.038/24.269 ms (11/132294)  149 KByte 13.54  
619=617:1:1:0:0:0:0:0
[  1] 4.00-5.00 sec   988 KBytes  8.10 Mbits/sec  
141.389/119.961/178.785/18.812 ms (7/144593)  170 KByte 7.158641  
409=404:5:0:0:0:0:0:0
[SUM] 4.00-5.00 sec  4.93 MBytes  41.4 Mbits/sec  
2141=2125:13:3:0:0:0:0:0
[  4] 5.00-6.00 sec  1.29 MBytes  10.8 Mbits/sec  
118.943/86.253/176.128/31.248 ms (10/135098)  164 KByte 11.36  
511=506:2:3:0:0:0:0:0
[  2] 5.00-6.00 sec  1.09 MBytes  9.17 Mbits/sec  
139.821/102.418/218.875/40.422 ms (9/127424)  148 KByte 8.202049  
487=484:2:1:0:0:0:0:0
[  3] 5.00-6.00 sec  1.51 MBytes  12.6 Mbits/sec  
102.146/77.085/140.893/18.441 ms (13/121520)  151 KByte 15.47  
640=636:1:3:0:0:0:0:0
[  1] 5.00-6.00 sec   981 KBytes  8.04 Mbits/sec  
161.901/105.582/219.931/36.260 ms (8/125614)  134 KByte 6.206944  
415=413:2:0:0:0:0:0:0
[SUM] 5.00-6.00 sec  4.85 MBytes  40.7 Mbits/sec  
2053=2039:7:7:0:0:0:0:0
[  4] 6.00-7.00 sec  1.74 MBytes  14.6 Mbits/sec  
88.846/74.297/101.859/7.118 ms (14/130526)  156 KByte 20.57  
711=707:3:1:0:0:0:0:0
[  1] 6.00-7.00 sec  1.22 MBytes  10.2 Mbits/sec  
120.639/100.257/157.567/21.770 ms (10/127568)  157 KByte 10.57  
494=488:5:1:0:0:0:0:0
[  2] 6.00-7.00 sec  1015 KBytes  8.32 Mbits/sec  
144.632/124.368/171.349/16.597 ms (8/129958)  143 KByte 7.188321  
408=403:5:0:0:0:0:0:0
[  3] 6.00-7.00 sec  1.02 MBytes  8.60 Mbits/sec  
143.516/102.322/173.001/24.089 ms (8/134302)  146 KByte 7.486359  
484=480:4:0:0:0:0:0:0
[SUM] 6.00-7.00 sec  4.98 MBytes  41.7 Mbits/sec  
2097=2078:17:2:0:0:0:0:0
[  4] 7.00-8.00 sec  1.77 MBytes  14.9 Mbits/sec  
85.406/65.797/103.418/12.609 ms (14/132595)  153 KByte 21.74  
692=687:2:3:0:0:0:0:0
[  2] 7.00-8.00 sec   957 KBytes  7.84 Mbits/sec  
153.936/131.452/191.464/19.361 ms (7/140042)  160 KByte 6.368199  
429=425:4:0:0:0:0:0:0
[  1] 7.00-8.00 sec  1.13 MBytes  9.44 Mbits/sec  
131.146/109.737/166.774/22.035 ms (9/131124)  146 KByte 8.998528  
520=516:4:0:0:0:0:0:0
[  3] 7.00-8.00 sec  1.13 MBytes  9.51 Mbits/sec  
126.512/88.404/220.175/42.237 ms (9/132089)  172 KByte 9.396784  
527=524:1:2:0:0:0:0:0
[SUM] 7.00-8.00 sec  4.96 MBytes  41.6 Mbits/sec  
2168=2152:11:5:0:0:0:0:0

With source side rate limiting to 9 mb/s per stream.

[rjmcmahon@ryzen3950 iperf2-code]$ iperf -c *** --hide-ips -e 
--trip-times -i 1 -P 4 -t 10 -w 4m --tcp-quickack -N -b 9m
------------------------------------------------------------
Client connecting to (**hidden**), TCP port 5001 with pid 108884 (4 
flows)
Write buffer size: 131072 Byte
TOS set to 0x0 and nodelay (Nagle off)
TCP window size: 7.63 MByte (WARNING: requested 3.81 MByte)
Event based writes (pending queue watermark at 16384 bytes)
------------------------------------------------------------
[  1] local *.*.*.85%enp4s0 port 46448 connected with *.*.*.123 port 
5001 (prefetch=16384) (trip-times) (sock=3) (qack) 
(icwnd/mss/irtt=14/1448/10666) (ct=10.70 ms) on 2023-01-06 12:27:45 
(PST)
[  3] local *.*.*.85%enp4s0 port 46460 connected with *.*.*.123 port 
5001 (prefetch=16384) (trip-times) (sock=6) (qack) 
(icwnd/mss/irtt=14/1448/16499) (ct=16.54 ms) on 2023-01-06 12:27:45 
(PST)
[  2] local *.*.*.85%enp4s0 port 46454 connected with *.*.*.123 port 
5001 (prefetch=16384) (trip-times) (sock=4) (qack) 
(icwnd/mss/irtt=14/1448/16580) (ct=16.86 ms) on 2023-01-06 12:27:45 
(PST)
[  4] local *.*.*.85%enp4s0 port 46458 connected with *.*.*.123 port 
5001 (prefetch=16384) (trip-times) (sock=5) (qack) 
(icwnd/mss/irtt=14/1448/16802) (ct=16.83 ms) on 2023-01-06 12:27:45 
(PST)
[ ID] Interval        Transfer    Bandwidth       Write/Err  Rtry     
Cwnd/RTT(var)        NetPwr
...
[  2] 4.00-5.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0      
134K/88055(12329) us  11.91
[  4] 4.00-5.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0      
132K/74867(11755) us  14.01
[  1] 4.00-5.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0      
134K/89101(13134) us  11.77
[  3] 4.00-5.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0      
131K/91451(11938) us  11.47
[SUM] 4.00-5.00 sec  4.00 MBytes  33.6 Mbits/sec  32/0         0
[  2] 5.00-6.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0      
134K/85135(14580) us  13.86
[  4] 5.00-6.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0      
132K/85124(15654) us  13.86
[  1] 5.00-6.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0      
134K/91336(11335) us  12.92
[  3] 5.00-6.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0      
131K/89185(13499) us  13.23
[SUM] 5.00-6.00 sec  4.50 MBytes  37.7 Mbits/sec  36/0         0
[  2] 6.00-7.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0      
134K/85687(13489) us  13.77
[  4] 6.00-7.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0      
132K/82803(13001) us  14.25
[  1] 6.00-7.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0      
134K/86869(15186) us  13.58
[  3] 6.00-7.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0      
131K/91447(12515) us  12.90
[SUM] 6.00-7.00 sec  4.50 MBytes  37.7 Mbits/sec  36/0         0
[  2] 7.00-8.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0      
134K/81814(13168) us  12.82
[  4] 7.00-8.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0      
132K/89008(13283) us  11.78
[  1] 7.00-8.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0      
134K/89494(12151) us  11.72
[  3] 7.00-8.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0      
131K/91083(12797) us  11.51
[SUM] 7.00-8.00 sec  4.00 MBytes  33.6 Mbits/sec  32/0         0

[root@bobcat iperf2-code]# iperf -s -i 1 -e --hide-ips -w 4m
------------------------------------------------------------
Server listening on TCP port 5001 with pid 233981
Read buffer size:  128 KByte (Dist bin width=16.0 KByte)
TCP window size: 7.63 MByte (WARNING: requested 3.81 MByte)
------------------------------------------------------------
[  1] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port 46448 
(trip-times) (sock=4) (peer 2.1.9-master) (qack) 
(icwnd/mss/irtt=14/1448/11987) on 2023-01-06 12:27:45 (PST)
[  2] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port 46454 
(trip-times) (sock=5) (peer 2.1.9-master) (qack) 
(icwnd/mss/irtt=14/1448/11132) on 2023-01-06 12:27:45 (PST)
[  3] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port 46460 
(trip-times) (sock=6) (peer 2.1.9-master) (qack) 
(icwnd/mss/irtt=14/1448/11097) on 2023-01-06 12:27:45 (PST)
[  4] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port 46458 
(trip-times) (sock=7) (peer 2.1.9-master) (qack) 
(icwnd/mss/irtt=14/1448/17823) on 2023-01-06 12:27:45 (PST)
[ ID] Interval        Transfer    Bandwidth    Burst Latency 
avg/min/max/stdev (cnt/size) inP NetPwr  Reads=Dist
[  4] 0.00-1.00 sec  1.09 MBytes  9.15 Mbits/sec  
93.383/90.103/95.661/2.232 ms (8/143028)  128 KByte 12.25  
451=451:0:0:0:0:0:0:0
[  3] 0.00-1.00 sec  1.08 MBytes  9.06 Mbits/sec  
96.834/95.229/102.645/2.442 ms (8/141580)  131 KByte 11.70  
472=472:0:0:0:0:0:0:0
[  1] 0.00-1.00 sec  1.10 MBytes  9.19 Mbits/sec  
95.183/92.623/97.579/1.431 ms (8/143571)  131 KByte 12.07  
495=495:0:0:0:0:0:0:0
[  2] 0.00-1.00 sec  1.09 MBytes  9.15 Mbits/sec  
89.317/84.865/94.906/3.674 ms (8/143028)  122 KByte 12.81  
489=489:0:0:0:0:0:0:0
[ ID] Interval        Transfer    Bandwidth       Reads=Dist
[SUM] 0.00-1.00 sec  4.36 MBytes  36.6 Mbits/sec  
1907=1907:0:0:0:0:0:0:0
[  4] 1.00-2.00 sec  1.07 MBytes  8.95 Mbits/sec  
92.649/89.987/95.036/1.828 ms (9/124314) 96.5 KByte 12.08  
492=492:0:0:0:0:0:0:0
[  3] 1.00-2.00 sec  1.06 MBytes  8.93 Mbits/sec  
96.305/95.647/96.794/0.432 ms (9/123992)  100 KByte 11.59  
480=480:0:0:0:0:0:0:0
[  1] 1.00-2.00 sec  1.06 MBytes  8.89 Mbits/sec  
92.578/90.866/94.145/1.371 ms (9/123510) 95.8 KByte 12.01  
513=513:0:0:0:0:0:0:0
[  2] 1.00-2.00 sec  1.07 MBytes  8.96 Mbits/sec  
90.767/87.984/94.352/1.944 ms (9/124475) 94.7 KByte 12.34  
489=489:0:0:0:0:0:0:0
[SUM] 1.00-2.00 sec  4.26 MBytes  35.7 Mbits/sec  
1974=1974:0:0:0:0:0:0:0
[  4] 2.00-3.00 sec  1.09 MBytes  9.13 Mbits/sec  
93.977/91.795/96.561/1.693 ms (8/142656)  112 KByte 12.14  
497=497:0:0:0:0:0:0:0
[  3] 2.00-3.00 sec  1.08 MBytes  9.04 Mbits/sec  
96.544/95.815/97.798/0.693 ms (8/141208)  114 KByte 11.70  
503=503:0:0:0:0:0:0:0
[  1] 2.00-3.00 sec  1.07 MBytes  9.01 Mbits/sec  
93.970/91.193/96.325/1.796 ms (8/140846)  111 KByte 11.99  
509=509:0:0:0:0:0:0:0
[  2] 2.00-3.00 sec  1.08 MBytes  9.10 Mbits/sec  
92.843/90.216/96.355/2.040 ms (8/142113)  111 KByte 12.25  
509=509:0:0:0:0:0:0:0
[SUM] 2.00-3.00 sec  4.32 MBytes  36.3 Mbits/sec  
2018=2018:0:0:0:0:0:0:0
[  4] 3.00-4.00 sec  1.06 MBytes  8.86 Mbits/sec  
93.222/89.063/96.104/2.346 ms (9/123027) 96.1 KByte 11.88  
487=487:0:0:0:0:0:0:0
[  3] 3.00-4.00 sec  1.07 MBytes  8.97 Mbits/sec  
96.277/95.051/97.230/0.767 ms (9/124636)  101 KByte 11.65  
489=489:0:0:0:0:0:0:0
[  1] 3.00-4.00 sec  1.08 MBytes  9.02 Mbits/sec  
93.899/88.732/96.972/2.737 ms (9/125280) 98.6 KByte 12.01  
493=493:0:0:0:0:0:0:0
[  2] 3.00-4.00 sec  1.07 MBytes  8.97 Mbits/sec  
92.490/89.862/95.265/1.796 ms (9/124636) 96.6 KByte 12.13  
493=493:0:0:0:0:0:0:0
[SUM] 3.00-4.00 sec  4.27 MBytes  35.8 Mbits/sec  
1962=1962:0:0:0:0:0:0:0
[  4] 4.00-5.00 sec  1.07 MBytes  9.00 Mbits/sec  
92.431/81.888/96.221/4.524 ms (9/124958) 96.8 KByte 12.17  
498=498:0:0:0:0:0:0:0
[  1] 4.00-5.00 sec  1.07 MBytes  8.97 Mbits/sec  
95.018/93.445/96.200/0.957 ms (9/124636) 99.3 KByte 11.81  
490=490:0:0:0:0:0:0:0
[  2] 4.00-5.00 sec  1.06 MBytes  8.90 Mbits/sec  
93.874/86.485/95.672/2.810 ms (9/123671) 97.3 KByte 11.86  
481=481:0:0:0:0:0:0:0
[  3] 4.00-5.00 sec  1.08 MBytes  9.09 Mbits/sec  
95.737/93.881/97.197/0.972 ms (9/126245)  101 KByte 11.87  
484=484:0:0:0:0:0:0:0
[SUM] 4.00-5.00 sec  4.29 MBytes  36.0 Mbits/sec  
1953=1953:0:0:0:0:0:0:0
[  4] 5.00-6.00 sec  1.09 MBytes  9.13 Mbits/sec  
92.908/86.844/95.994/3.012 ms (8/142656)  111 KByte 12.28  
467=467:0:0:0:0:0:0:0
[  3] 5.00-6.00 sec  1.07 MBytes  8.94 Mbits/sec  
96.593/95.343/97.660/0.876 ms (8/139760)  113 KByte 11.58  
478=478:0:0:0:0:0:0:0
[  1] 5.00-6.00 sec  1.08 MBytes  9.03 Mbits/sec  
95.021/91.421/97.167/1.893 ms (8/141027)  112 KByte 11.87  
491=491:0:0:0:0:0:0:0
[  2] 5.00-6.00 sec  1.08 MBytes  9.06 Mbits/sec  
92.162/82.720/97.692/5.060 ms (8/141570)  109 KByte 12.29  
488=488:0:0:0:0:0:0:0
[SUM] 5.00-6.00 sec  4.31 MBytes  36.2 Mbits/sec  
1924=1924:0:0:0:0:0:0:0
[  4] 6.00-7.00 sec  1.04 MBytes  8.70 Mbits/sec  
92.793/85.343/96.967/3.552 ms (9/120775) 93.9 KByte 11.71  
485=485:0:0:0:0:0:0:0
[  2] 6.00-7.00 sec  1.05 MBytes  8.79 Mbits/sec  
91.679/84.479/96.760/3.975 ms (9/122062) 93.8 KByte 11.98  
472=472:0:0:0:0:0:0:0
[  3] 6.00-7.00 sec  1.06 MBytes  8.88 Mbits/sec  
96.982/95.933/98.371/0.680 ms (9/123349)  100 KByte 11.45  
477=477:0:0:0:0:0:0:0
[  1] 6.00-7.00 sec  1.05 MBytes  8.80 Mbits/sec  
94.342/91.660/96.025/1.660 ms (9/122223) 96.7 KByte 11.66  
494=494:0:0:0:0:0:0:0
[SUM] 6.00-7.00 sec  4.19 MBytes  35.2 Mbits/sec  
1928=1928:0:0:0:0:0:0:0
[  4] 7.00-8.00 sec  1.10 MBytes  9.25 Mbits/sec  
92.515/88.182/96.351/2.538 ms (8/144466)  112 KByte 12.49  
510=510:0:0:0:0:0:0:0
[  3] 7.00-8.00 sec  1.09 MBytes  9.13 Mbits/sec  
96.580/95.737/98.977/1.098 ms (8/142656)  115 KByte 11.82  
480=480:0:0:0:0:0:0:0
[  1] 7.00-8.00 sec  1.10 MBytes  9.21 Mbits/sec  
95.269/91.719/97.514/2.126 ms (8/143923)  115 KByte 12.09  
515=515:0:0:0:0:0:0:0
[  2] 7.00-8.00 sec  1.11 MBytes  9.29 Mbits/sec  
90.073/84.700/96.176/4.324 ms (8/145190)  110 KByte 12.90  
508=508:0:0:0:0:0:0:0
[SUM] 7.00-8.00 sec  4.40 MBytes  36.9 Mbits/sec  
2013=2013:0:0:0:0:0:0:0

Bob

>> -----Original Message-----
> 
>> From: LibreQoS <libreqos-bounces@lists.bufferbloat.net> On Behalf Of
> Dave Taht
> 
>> via LibreQoS
> 
>> Sent: Wednesday, January 4, 2023 12:26 PM
> 
>> Subject: [LibreQoS] the grinch meets cloudflare's christmas present
> 
>> 
> 
>> Please try the new, the shiny, the really wonderful test here:
> 
>> 
> https://urldefense.com/v3/__https://speed.cloudflare.com/__;!!BhdT!iZcFJ8WVU9S
> [1]
> 
>> 
> 9zz5t456oxkfObrC5Xb9j5AG8UO3DqD5x4GAJkawZr0iGwEUtF0_09U8mCDnAkrJ9QEMHGbCMKVw$
> [1]
> 
>> 
> 
>> I would really appreciate some independent verification of
> 
>> measurements using this tool. In my brief experiments it appears -
> as
> 
>> all the commercial tools to date - to dramatically understate the
> 
>> bufferbloat, on my LTE, (and my starlink terminal is out being
> 
>> hacked^H^H^H^H^H^Hworked on, so I can't measure that)
> 
> [acm]
> 
> Hi Dave, I made some time to test "cloudflare's christmas present"
> yesterday.
> 
> I'm on DOCSIS 3.1 service with 1Gbps Down. The Upstream has a "turbo"
> mode with 40-50Mbps for the first ~3 sec, then steady-state about
> 23Mbps.
> 
> When I saw the ~620Mbps Downstream measurement, I was ready to
> complain that even the IP-Layer Capacity was grossly underestimated.
> In addition, the Latency measurements seem very low (as you asserted),
> although the cloud server was “nearby”.
> 
> However, I found that Ookla and the ISP-provided measurement were also
> reporting ~600Mbps! So the cloudflare Downstream capacity (or
> throughput?) measurement was consistent with others. Our UDPST server
> was unreachable, otherwise I would have added that measurement, too.
> 
> The Upstream measurement graph seems to illustrate the “turbo”
> mode, with the dip after attaining 44.5Mbps.
> 
> UDPST saturates the uplink and we measure the full 250ms of the
> Upstream buffer. Cloudflare’s latency measurements don’t even come
> close.
> 
> Al
> 
> 
> 
> Links:
> ------
> [1] 
> https://urldefense.com/v3/__https:/speed.cloudflare.com/__;!!BhdT!iZcFJ8WVU9S9zz5t456oxkfObrC5Xb9j5AG8UO3DqD5x4GAJkawZr0iGwEUtF0_09U8mCDnAkrJ9QEMHGbCMKVw$
> _______________________________________________
> Rpm mailing list
> Rpm@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/rpm

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] [Rpm] [LibreQoS] the grinch meets cloudflare's christmas present
  2023-01-06 20:38   ` [Starlink] [Rpm] " rjmcmahon
@ 2023-01-06 20:47     ` rjmcmahon
       [not found]     ` <89D796E75967416B9723211C183A8396@SRA6>
  1 sibling, 0 replies; 44+ messages in thread
From: rjmcmahon @ 2023-01-06 20:47 UTC (permalink / raw)
  To: MORTON JR., AL
  Cc: Dave Taht, bloat, libreqos, Cake List, Dave Taht via Starlink,
	Rpm, IETF IPPM WG

For responsiveness, the bounceback seems reasonable even with upstream 
competition. Bunch more TCP retries though.

[rjmcmahon@ryzen3950 iperf2-code]$ iperf -c *** --hide-ips -e 
--trip-times -i 1 --bounceback -t 3
------------------------------------------------------------
Client connecting to (**hidden**), TCP port 5001 with pid 111022 (1 
flows)
Write buffer size:  100 Byte
Bursting:  100 Byte writes 10 times every 1.00 second(s)
Bounce-back test (size= 100 Byte) (server hold req=0 usecs & 
tcp_quickack)
TOS set to 0x0 and nodelay (Nagle off)
TCP window size: 16.0 KByte (default)
Event based writes (pending queue watermark at 16384 bytes)
------------------------------------------------------------
[  1] local *.*.*.86%enp7s0 port 36976 connected with *.*.*.123 port 
5001 (prefetch=16384) (bb w/quickack len/hold=100/0) (trip-times) 
(sock=3) (icwnd/mss/irtt=14/1448/9862) (ct=9.90 ms) on 2023-01-06 
12:42:18 (PST)
[ ID] Interval        Transfer    Bandwidth         BB 
cnt=avg/min/max/stdev         Rtry  Cwnd/RTT    RPS
[  1] 0.00-1.00 sec  1.95 KBytes  16.0 Kbits/sec    
10=12.195/9.298/16.457/2.679 ms    0   14K/11327 us    82 rps
[  1] 1.00-2.00 sec  1.95 KBytes  16.0 Kbits/sec    
10=12.613/9.271/15.489/2.788 ms    0   14K/12165 us    79 rps
[  1] 2.00-3.00 sec  1.95 KBytes  16.0 Kbits/sec    
10=13.390/9.376/15.986/2.520 ms    0   14K/13164 us    75 rps
[  1] 0.00-3.03 sec  5.86 KBytes  15.8 Kbits/sec    
30=12.733/9.271/16.457/2.620 ms    0   14K/15138 us    79 rps
[  1] 0.00-3.03 sec  OWD Delays (ms) Cnt=30 To=7.937/4.634/11.327/2.457 
 From=4.778/4.401/5.350/0.258 Asymmetry=3.166/0.097/6.311/2.318    79 rps
[  1] 0.00-3.03 sec BB8(f)-PDF: 
bin(w=100us):cnt(30)=93:2,94:3,95:2,97:1,100:1,102:1,105:1,114:2,142:1,143:1,144:2,145:3,146:1,147:1,148:1,151:1,152:1,154:1,155:1,156:1,160:1,165:1 
(5.00/95.00/99.7%=93/160/165,Outliers=0,obl/obu=0/0)

[rjmcmahon@ryzen3950 iperf2-code]$ iperf -c *** --hide-ips -e 
--trip-times -i 1 --bounceback -t 3 --bounceback-congest=up,4
------------------------------------------------------------
Client connecting to (**hidden**), TCP port 5001 with pid 111069 (1 
flows)
Write buffer size:  100 Byte
Bursting:  100 Byte writes 10 times every 1.00 second(s)
Bounce-back test (size= 100 Byte) (server hold req=0 usecs & 
tcp_quickack)
TOS set to 0x0 and nodelay (Nagle off)
TCP window size: 16.0 KByte (default)
Event based writes (pending queue watermark at 16384 bytes)
------------------------------------------------------------
[  2] local *.*.*.85%enp4s0 port 38342 connected with *.*.*.123 port 
5001 (prefetch=16384) (trip-times) (sock=3) (qack) 
(icwnd/mss/irtt=14/1448/10613) (ct=10.66 ms) on 2023-01-06 12:42:36 
(PST)
[  1] local *.*.*.85%enp4s0 port 38360 connected with *.*.*.123 port 
5001 (prefetch=16384) (bb w/quickack len/hold=100/0) (trip-times) 
(sock=4) (icwnd/mss/irtt=14/1448/14901) (ct=14.96 ms) on 2023-01-06 
12:42:36 (PST)
[  3] local *.*.*.85%enp4s0 port 38386 connected with *.*.*.123 port 
5001 (prefetch=16384) (trip-times) (sock=7) (qack) 
(icwnd/mss/irtt=14/1448/15295) (ct=15.31 ms) on 2023-01-06 12:42:36 
(PST)
[  4] local *.*.*.85%enp4s0 port 38348 connected with *.*.*.123 port 
5001 (prefetch=16384) (trip-times) (sock=5) (qack) 
(icwnd/mss/irtt=14/1448/14901) (ct=14.95 ms) on 2023-01-06 12:42:36 
(PST)
[  5] local *.*.*.85%enp4s0 port 38372 connected with *.*.*.123 port 
5001 (prefetch=16384) (trip-times) (sock=6) (qack) 
(icwnd/mss/irtt=14/1448/15371) (ct=15.42 ms) on 2023-01-06 12:42:36 
(PST)
[ ID] Interval        Transfer    Bandwidth       Write/Err  Rtry     
Cwnd/RTT(var)        NetPwr
[  3] 0.00-1.00 sec  1.29 MBytes  10.8 Mbits/sec  13502/0       115      
  28K/22594(904) us  59.76
[  4] 0.00-1.00 sec  1.63 MBytes  13.6 Mbits/sec  17048/0       140      
  42K/22728(568) us  75.01
[ ID] Interval        Transfer    Bandwidth         BB 
cnt=avg/min/max/stdev         Rtry  Cwnd/RTT    RPS
[  1] 0.00-1.00 sec  1.95 KBytes  16.0 Kbits/sec    
10=76.140/17.224/123.195/43.168 ms    0   14K/68136 us    13 rps
[  5] 0.00-1.00 sec  1.04 MBytes  8.72 Mbits/sec  10893/0        82      
  25K/23400(644) us  46.55
[SUM] 0.00-1.00 sec  3.95 MBytes  33.2 Mbits/sec  41443/0       337
[  2] 0.00-1.00 sec  1.10 MBytes  9.25 Mbits/sec  11566/0        77      
  22K/23557(432) us  49.10
[  3] 1.00-2.00 sec  1.24 MBytes  10.4 Mbits/sec  13037/0        20      
  28K/14427(503) us  90.37
[  4] 1.00-2.00 sec  1.43 MBytes  12.0 Mbits/sec  14954/0        31      
  12K/13348(407) us  112
[  1] 1.00-2.00 sec  1.95 KBytes  16.0 Kbits/sec    
10=14.581/10.801/20.356/3.599 ms    0   14K/27791 us    69 rps
[  5] 1.00-2.00 sec  1.26 MBytes  10.6 Mbits/sec  13191/0        16      
  12K/14749(675) us  89.44
[SUM] 1.00-2.00 sec  3.93 MBytes  32.9 Mbits/sec  41182/0        67
[  2] 1.00-2.00 sec  1000 KBytes  8.19 Mbits/sec  10237/0        13      
  19K/14467(1068) us  70.76
[  3] 2.00-3.00 sec  1.33 MBytes  11.2 Mbits/sec  13994/0         4      
  24K/20749(495) us  67.44
[  4] 2.00-3.00 sec  1.20 MBytes  10.1 Mbits/sec  12615/0         3      
  31K/20877(718) us  60.43
[  1] 2.00-3.00 sec  1.95 KBytes  16.0 Kbits/sec    
10=11.298/9.407/14.245/1.330 ms    0   14K/15474 us    89 rps
[  5] 2.00-3.00 sec  1.08 MBytes  9.03 Mbits/sec  11284/0         3      
  28K/21031(430) us  53.65
[SUM] 2.00-3.00 sec  3.61 MBytes  30.3 Mbits/sec  37893/0        10
[  2] 2.00-3.00 sec  1.29 MBytes  10.8 Mbits/sec  13492/0         3      
  29K/20409(688) us  66.11
[  3] 0.00-3.03 sec  3.87 MBytes  10.7 Mbits/sec  40534/0       139      
  25K/20645(557) us  64.85
[  5] 0.00-3.03 sec  3.37 MBytes  9.35 Mbits/sec  35369/0       101      
  29K/20489(668) us  57.02
[  4] 0.00-3.03 sec  4.26 MBytes  11.8 Mbits/sec  44618/0       174      
  32K/21240(961) us  69.40
[  2] 0.00-3.03 sec  3.37 MBytes  9.31 Mbits/sec  35296/0        94      
  19K/21504(948) us  54.13
[  1] 0.00-3.14 sec  7.81 KBytes  20.4 Kbits/sec    
40=28.332/5.611/123.195/34.940 ms    0   14K/14000 us    35 rps
[  1] 0.00-3.14 sec  OWD Delays (ms) Cnt=40 
To=23.730/1.110/118.744/34.957 From=4.567/4.356/5.171/0.141 
Asymmetry=19.332/0.189/114.294/34.869    35 rps
[  1] 0.00-3.14 sec BB8(f)-PDF: 
bin(w=100us):cnt(40)=57:1,94:2,95:2,96:2,98:1,101:1,106:1,109:2,111:1,112:2,113:1,115:1,118:1,119:2,143:2,145:1,146:1,152:1,158:1,173:1,176:1,194:1,195:1,204:1,205:1,274:1,554:1,790:1,925:1,1125:1,1126:1,1225:1,1232:1 
(5.00/95.00/99.7%=94/1225/1232,Outliers=0,obl/obu=0/0)
[SUM] 0.00-3.11 sec  11.5 MBytes  31.0 Mbits/sec  120521/0       414
[ CT] final connect times (min/avg/max/stdev) = 
10.661/14.261/15.423/2023.369 ms (tot/err) = 5/0

> Some thoughts are not to use UDP for testing here. Also, these speed
> tests have little to no information for network engineers about what's
> going on. Iperf 2 may better assist network engineers but then I'm
> biased ;)
> 
> Running iperf 2 https://sourceforge.net/projects/iperf2/ with
> --trip-times. Though the sampling and central limit theorem averaging
> is hiding the real distributions (use --histograms to get those)
> 
> Below are 4 parallel TCP streams from my home to one of my servers in
> the cloud. First where TCP is limited per CCA. Second is source side
> write rate limiting. Things to note:
> 
> o) connect times for both at 10-15 ms
> o) multiple TCP retries on a few rites - one case is 4 with 5 writes.
> Source side pacing eliminates retries
> o) Fairness with CCA isn't great but quite good with source side write 
> pacing
> o) Queue depth with CCA is about 150 Kbytes about 100K byte with
> source side pacing
> o) min write to read is about 80 ms for both
> o) max is 220 ms vs 97 ms
> o) stdev for CCA write/read is 30 ms vs 3 ms
> o) TCP RTT is 20ms w/CCA and 90 ms with ssp - seems odd here as
> TCP_QUICACK and TCP_NODELAY are both enabled.
> 
> [ CT] final connect times (min/avg/max/stdev) =
> 10.326/13.522/14.986/2150.329 ms (tot/err) = 4/0
> [rjmcmahon@ryzen3950 iperf2-code]$ iperf -c *** --hide-ips -e
> --trip-times -i 1 -P 4 -t 10 -w 4m --tcp-quickack -N
> ------------------------------------------------------------
> Client connecting to (**hidden**), TCP port 5001 with pid 107678 (4 
> flows)
> Write buffer size: 131072 Byte
> TOS set to 0x0 and nodelay (Nagle off)
> TCP window size: 7.63 MByte (WARNING: requested 3.81 MByte)
> Event based writes (pending queue watermark at 16384 bytes)
> ------------------------------------------------------------
> [  1] local *.*.*.85%enp4s0 port 42480 connected with *.*.*.123 port
> 5001 (prefetch=16384) (trip-times) (sock=3) (qack)
> (icwnd/mss/irtt=14/1448/10534) (ct=10.63 ms) on 2023-01-06 12:17:56
> (PST)
> [  4] local *.*.*.85%enp4s0 port 42488 connected with *.*.*.123 port
> 5001 (prefetch=16384) (trip-times) (sock=5) (qack)
> (icwnd/mss/irtt=14/1448/14023) (ct=14.08 ms) on 2023-01-06 12:17:56
> (PST)
> [  3] local *.*.*.85%enp4s0 port 42502 connected with *.*.*.123 port
> 5001 (prefetch=16384) (trip-times) (sock=6) (qack)
> (icwnd/mss/irtt=14/1448/14642) (ct=14.70 ms) on 2023-01-06 12:17:56
> (PST)
> [  2] local *.*.*.85%enp4s0 port 42484 connected with *.*.*.123 port
> 5001 (prefetch=16384) (trip-times) (sock=4) (qack)
> (icwnd/mss/irtt=14/1448/14728) (ct=14.79 ms) on 2023-01-06 12:17:56
> (PST)
> [ ID] Interval        Transfer    Bandwidth       Write/Err  Rtry
> Cwnd/RTT(var)        NetPwr
> ...
> [  4] 4.00-5.00 sec  1.38 MBytes  11.5 Mbits/sec  11/0         3
> 29K/21088(1142) us  68.37
> [  2] 4.00-5.00 sec  1.62 MBytes  13.6 Mbits/sec  13/0         2
> 31K/19284(612) us  88.36
> [  1] 4.00-5.00 sec   896 KBytes  7.34 Mbits/sec  7/0         5
> 16K/18996(658) us  48.30
> [  3] 4.00-5.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         5
> 18K/18133(208) us  57.83
> [SUM] 4.00-5.00 sec  4.88 MBytes  40.9 Mbits/sec  39/0        15
> [  4] 5.00-6.00 sec  1.25 MBytes  10.5 Mbits/sec  10/0         4
> 29K/14717(489) us  89.06
> [  1] 5.00-6.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         4
> 16K/15874(408) us  66.06
> [  3] 5.00-6.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         4
> 16K/15826(382) us  74.54
> [  2] 5.00-6.00 sec  1.50 MBytes  12.6 Mbits/sec  12/0         6
>  9K/14878(557) us  106
> [SUM] 5.00-6.00 sec  4.88 MBytes  40.9 Mbits/sec  39/0        18
> [  4] 6.00-7.00 sec  1.75 MBytes  14.7 Mbits/sec  14/0         4
> 25K/15472(496) us  119
> [  2] 6.00-7.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         2
> 26K/16417(427) us  63.87
> [  1] 6.00-7.00 sec  1.25 MBytes  10.5 Mbits/sec  10/0         5
> 16K/16268(679) us  80.57
> [  3] 6.00-7.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         6
> 15K/16629(799) us  63.06
> [SUM] 6.00-7.00 sec  5.00 MBytes  41.9 Mbits/sec  40/0        17
> [  4] 7.00-8.00 sec  1.75 MBytes  14.7 Mbits/sec  14/0         4
> 22K/13986(519) us  131
> [  1] 7.00-8.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         4
> 16K/12679(377) us  93.04
> [  3] 7.00-8.00 sec   896 KBytes  7.34 Mbits/sec  7/0         5
> 14K/12971(367) us  70.74
> [  2] 7.00-8.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         6
> 15K/14740(779) us  80.03
> [SUM] 7.00-8.00 sec  4.88 MBytes  40.9 Mbits/sec  39/0        19
> 
> [root@bobcat iperf2-code]# iperf -s -i 1 -e --hide-ips -w 4m
> ------------------------------------------------------------
> Server listening on TCP port 5001 with pid 233615
> Read buffer size:  128 KByte (Dist bin width=16.0 KByte)
> TCP window size: 7.63 MByte (WARNING: requested 3.81 MByte)
> ------------------------------------------------------------
> [  1] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port
> 42480 (trip-times) (sock=4) (peer 2.1.9-master) (qack)
> (icwnd/mss/irtt=14/1448/11636) on 2023-01-06 12:17:56 (PST)
> [  2] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port
> 42502 (trip-times) (sock=5) (peer 2.1.9-master) (qack)
> (icwnd/mss/irtt=14/1448/11898) on 2023-01-06 12:17:56 (PST)
> [  3] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port
> 42484 (trip-times) (sock=6) (peer 2.1.9-master) (qack)
> (icwnd/mss/irtt=14/1448/11938) on 2023-01-06 12:17:56 (PST)
> [  4] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port
> 42488 (trip-times) (sock=7) (peer 2.1.9-master) (qack)
> (icwnd/mss/irtt=14/1448/11919) on 2023-01-06 12:17:56 (PST)
> [ ID] Interval        Transfer    Bandwidth    Burst Latency
> avg/min/max/stdev (cnt/size) inP NetPwr  Reads=Dist
> ...
> [  2] 4.00-5.00 sec  1.06 MBytes  8.86 Mbits/sec
> 129.819/90.391/186.075/31.346 ms (9/123080)  154 KByte 8.532803
> 467=461:6:0:0:0:0:0:0
> [  3] 4.00-5.00 sec  1.52 MBytes  12.8 Mbits/sec
> 103.552/82.653/169.274/28.382 ms (12/132854)  149 KByte 15.40
> 646=643:1:2:0:0:0:0:0
> [  4] 4.00-5.00 sec  1.39 MBytes  11.6 Mbits/sec
> 107.503/66.843/143.038/24.269 ms (11/132294)  149 KByte 13.54
> 619=617:1:1:0:0:0:0:0
> [  1] 4.00-5.00 sec   988 KBytes  8.10 Mbits/sec
> 141.389/119.961/178.785/18.812 ms (7/144593)  170 KByte 7.158641
> 409=404:5:0:0:0:0:0:0
> [SUM] 4.00-5.00 sec  4.93 MBytes  41.4 Mbits/sec  
> 2141=2125:13:3:0:0:0:0:0
> [  4] 5.00-6.00 sec  1.29 MBytes  10.8 Mbits/sec
> 118.943/86.253/176.128/31.248 ms (10/135098)  164 KByte 11.36
> 511=506:2:3:0:0:0:0:0
> [  2] 5.00-6.00 sec  1.09 MBytes  9.17 Mbits/sec
> 139.821/102.418/218.875/40.422 ms (9/127424)  148 KByte 8.202049
> 487=484:2:1:0:0:0:0:0
> [  3] 5.00-6.00 sec  1.51 MBytes  12.6 Mbits/sec
> 102.146/77.085/140.893/18.441 ms (13/121520)  151 KByte 15.47
> 640=636:1:3:0:0:0:0:0
> [  1] 5.00-6.00 sec   981 KBytes  8.04 Mbits/sec
> 161.901/105.582/219.931/36.260 ms (8/125614)  134 KByte 6.206944
> 415=413:2:0:0:0:0:0:0
> [SUM] 5.00-6.00 sec  4.85 MBytes  40.7 Mbits/sec  
> 2053=2039:7:7:0:0:0:0:0
> [  4] 6.00-7.00 sec  1.74 MBytes  14.6 Mbits/sec
> 88.846/74.297/101.859/7.118 ms (14/130526)  156 KByte 20.57
> 711=707:3:1:0:0:0:0:0
> [  1] 6.00-7.00 sec  1.22 MBytes  10.2 Mbits/sec
> 120.639/100.257/157.567/21.770 ms (10/127568)  157 KByte 10.57
> 494=488:5:1:0:0:0:0:0
> [  2] 6.00-7.00 sec  1015 KBytes  8.32 Mbits/sec
> 144.632/124.368/171.349/16.597 ms (8/129958)  143 KByte 7.188321
> 408=403:5:0:0:0:0:0:0
> [  3] 6.00-7.00 sec  1.02 MBytes  8.60 Mbits/sec
> 143.516/102.322/173.001/24.089 ms (8/134302)  146 KByte 7.486359
> 484=480:4:0:0:0:0:0:0
> [SUM] 6.00-7.00 sec  4.98 MBytes  41.7 Mbits/sec  
> 2097=2078:17:2:0:0:0:0:0
> [  4] 7.00-8.00 sec  1.77 MBytes  14.9 Mbits/sec
> 85.406/65.797/103.418/12.609 ms (14/132595)  153 KByte 21.74
> 692=687:2:3:0:0:0:0:0
> [  2] 7.00-8.00 sec   957 KBytes  7.84 Mbits/sec
> 153.936/131.452/191.464/19.361 ms (7/140042)  160 KByte 6.368199
> 429=425:4:0:0:0:0:0:0
> [  1] 7.00-8.00 sec  1.13 MBytes  9.44 Mbits/sec
> 131.146/109.737/166.774/22.035 ms (9/131124)  146 KByte 8.998528
> 520=516:4:0:0:0:0:0:0
> [  3] 7.00-8.00 sec  1.13 MBytes  9.51 Mbits/sec
> 126.512/88.404/220.175/42.237 ms (9/132089)  172 KByte 9.396784
> 527=524:1:2:0:0:0:0:0
> [SUM] 7.00-8.00 sec  4.96 MBytes  41.6 Mbits/sec  
> 2168=2152:11:5:0:0:0:0:0
> 
> With source side rate limiting to 9 mb/s per stream.
> 
> [rjmcmahon@ryzen3950 iperf2-code]$ iperf -c *** --hide-ips -e
> --trip-times -i 1 -P 4 -t 10 -w 4m --tcp-quickack -N -b 9m
> ------------------------------------------------------------
> Client connecting to (**hidden**), TCP port 5001 with pid 108884 (4 
> flows)
> Write buffer size: 131072 Byte
> TOS set to 0x0 and nodelay (Nagle off)
> TCP window size: 7.63 MByte (WARNING: requested 3.81 MByte)
> Event based writes (pending queue watermark at 16384 bytes)
> ------------------------------------------------------------
> [  1] local *.*.*.85%enp4s0 port 46448 connected with *.*.*.123 port
> 5001 (prefetch=16384) (trip-times) (sock=3) (qack)
> (icwnd/mss/irtt=14/1448/10666) (ct=10.70 ms) on 2023-01-06 12:27:45
> (PST)
> [  3] local *.*.*.85%enp4s0 port 46460 connected with *.*.*.123 port
> 5001 (prefetch=16384) (trip-times) (sock=6) (qack)
> (icwnd/mss/irtt=14/1448/16499) (ct=16.54 ms) on 2023-01-06 12:27:45
> (PST)
> [  2] local *.*.*.85%enp4s0 port 46454 connected with *.*.*.123 port
> 5001 (prefetch=16384) (trip-times) (sock=4) (qack)
> (icwnd/mss/irtt=14/1448/16580) (ct=16.86 ms) on 2023-01-06 12:27:45
> (PST)
> [  4] local *.*.*.85%enp4s0 port 46458 connected with *.*.*.123 port
> 5001 (prefetch=16384) (trip-times) (sock=5) (qack)
> (icwnd/mss/irtt=14/1448/16802) (ct=16.83 ms) on 2023-01-06 12:27:45
> (PST)
> [ ID] Interval        Transfer    Bandwidth       Write/Err  Rtry
> Cwnd/RTT(var)        NetPwr
> ...
> [  2] 4.00-5.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0
> 134K/88055(12329) us  11.91
> [  4] 4.00-5.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0
> 132K/74867(11755) us  14.01
> [  1] 4.00-5.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0
> 134K/89101(13134) us  11.77
> [  3] 4.00-5.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0
> 131K/91451(11938) us  11.47
> [SUM] 4.00-5.00 sec  4.00 MBytes  33.6 Mbits/sec  32/0         0
> [  2] 5.00-6.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0
> 134K/85135(14580) us  13.86
> [  4] 5.00-6.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0
> 132K/85124(15654) us  13.86
> [  1] 5.00-6.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0
> 134K/91336(11335) us  12.92
> [  3] 5.00-6.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0
> 131K/89185(13499) us  13.23
> [SUM] 5.00-6.00 sec  4.50 MBytes  37.7 Mbits/sec  36/0         0
> [  2] 6.00-7.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0
> 134K/85687(13489) us  13.77
> [  4] 6.00-7.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0
> 132K/82803(13001) us  14.25
> [  1] 6.00-7.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0
> 134K/86869(15186) us  13.58
> [  3] 6.00-7.00 sec  1.12 MBytes  9.44 Mbits/sec  9/0         0
> 131K/91447(12515) us  12.90
> [SUM] 6.00-7.00 sec  4.50 MBytes  37.7 Mbits/sec  36/0         0
> [  2] 7.00-8.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0
> 134K/81814(13168) us  12.82
> [  4] 7.00-8.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0
> 132K/89008(13283) us  11.78
> [  1] 7.00-8.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0
> 134K/89494(12151) us  11.72
> [  3] 7.00-8.00 sec  1.00 MBytes  8.39 Mbits/sec  8/0         0
> 131K/91083(12797) us  11.51
> [SUM] 7.00-8.00 sec  4.00 MBytes  33.6 Mbits/sec  32/0         0
> 
> [root@bobcat iperf2-code]# iperf -s -i 1 -e --hide-ips -w 4m
> ------------------------------------------------------------
> Server listening on TCP port 5001 with pid 233981
> Read buffer size:  128 KByte (Dist bin width=16.0 KByte)
> TCP window size: 7.63 MByte (WARNING: requested 3.81 MByte)
> ------------------------------------------------------------
> [  1] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port
> 46448 (trip-times) (sock=4) (peer 2.1.9-master) (qack)
> (icwnd/mss/irtt=14/1448/11987) on 2023-01-06 12:27:45 (PST)
> [  2] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port
> 46454 (trip-times) (sock=5) (peer 2.1.9-master) (qack)
> (icwnd/mss/irtt=14/1448/11132) on 2023-01-06 12:27:45 (PST)
> [  3] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port
> 46460 (trip-times) (sock=6) (peer 2.1.9-master) (qack)
> (icwnd/mss/irtt=14/1448/11097) on 2023-01-06 12:27:45 (PST)
> [  4] local *.*.*.123%eth0 port 5001 connected with *.*.*.171 port
> 46458 (trip-times) (sock=7) (peer 2.1.9-master) (qack)
> (icwnd/mss/irtt=14/1448/17823) on 2023-01-06 12:27:45 (PST)
> [ ID] Interval        Transfer    Bandwidth    Burst Latency
> avg/min/max/stdev (cnt/size) inP NetPwr  Reads=Dist
> [  4] 0.00-1.00 sec  1.09 MBytes  9.15 Mbits/sec
> 93.383/90.103/95.661/2.232 ms (8/143028)  128 KByte 12.25
> 451=451:0:0:0:0:0:0:0
> [  3] 0.00-1.00 sec  1.08 MBytes  9.06 Mbits/sec
> 96.834/95.229/102.645/2.442 ms (8/141580)  131 KByte 11.70
> 472=472:0:0:0:0:0:0:0
> [  1] 0.00-1.00 sec  1.10 MBytes  9.19 Mbits/sec
> 95.183/92.623/97.579/1.431 ms (8/143571)  131 KByte 12.07
> 495=495:0:0:0:0:0:0:0
> [  2] 0.00-1.00 sec  1.09 MBytes  9.15 Mbits/sec
> 89.317/84.865/94.906/3.674 ms (8/143028)  122 KByte 12.81
> 489=489:0:0:0:0:0:0:0
> [ ID] Interval        Transfer    Bandwidth       Reads=Dist
> [SUM] 0.00-1.00 sec  4.36 MBytes  36.6 Mbits/sec  
> 1907=1907:0:0:0:0:0:0:0
> [  4] 1.00-2.00 sec  1.07 MBytes  8.95 Mbits/sec
> 92.649/89.987/95.036/1.828 ms (9/124314) 96.5 KByte 12.08
> 492=492:0:0:0:0:0:0:0
> [  3] 1.00-2.00 sec  1.06 MBytes  8.93 Mbits/sec
> 96.305/95.647/96.794/0.432 ms (9/123992)  100 KByte 11.59
> 480=480:0:0:0:0:0:0:0
> [  1] 1.00-2.00 sec  1.06 MBytes  8.89 Mbits/sec
> 92.578/90.866/94.145/1.371 ms (9/123510) 95.8 KByte 12.01
> 513=513:0:0:0:0:0:0:0
> [  2] 1.00-2.00 sec  1.07 MBytes  8.96 Mbits/sec
> 90.767/87.984/94.352/1.944 ms (9/124475) 94.7 KByte 12.34
> 489=489:0:0:0:0:0:0:0
> [SUM] 1.00-2.00 sec  4.26 MBytes  35.7 Mbits/sec  
> 1974=1974:0:0:0:0:0:0:0
> [  4] 2.00-3.00 sec  1.09 MBytes  9.13 Mbits/sec
> 93.977/91.795/96.561/1.693 ms (8/142656)  112 KByte 12.14
> 497=497:0:0:0:0:0:0:0
> [  3] 2.00-3.00 sec  1.08 MBytes  9.04 Mbits/sec
> 96.544/95.815/97.798/0.693 ms (8/141208)  114 KByte 11.70
> 503=503:0:0:0:0:0:0:0
> [  1] 2.00-3.00 sec  1.07 MBytes  9.01 Mbits/sec
> 93.970/91.193/96.325/1.796 ms (8/140846)  111 KByte 11.99
> 509=509:0:0:0:0:0:0:0
> [  2] 2.00-3.00 sec  1.08 MBytes  9.10 Mbits/sec
> 92.843/90.216/96.355/2.040 ms (8/142113)  111 KByte 12.25
> 509=509:0:0:0:0:0:0:0
> [SUM] 2.00-3.00 sec  4.32 MBytes  36.3 Mbits/sec  
> 2018=2018:0:0:0:0:0:0:0
> [  4] 3.00-4.00 sec  1.06 MBytes  8.86 Mbits/sec
> 93.222/89.063/96.104/2.346 ms (9/123027) 96.1 KByte 11.88
> 487=487:0:0:0:0:0:0:0
> [  3] 3.00-4.00 sec  1.07 MBytes  8.97 Mbits/sec
> 96.277/95.051/97.230/0.767 ms (9/124636)  101 KByte 11.65
> 489=489:0:0:0:0:0:0:0
> [  1] 3.00-4.00 sec  1.08 MBytes  9.02 Mbits/sec
> 93.899/88.732/96.972/2.737 ms (9/125280) 98.6 KByte 12.01
> 493=493:0:0:0:0:0:0:0
> [  2] 3.00-4.00 sec  1.07 MBytes  8.97 Mbits/sec
> 92.490/89.862/95.265/1.796 ms (9/124636) 96.6 KByte 12.13
> 493=493:0:0:0:0:0:0:0
> [SUM] 3.00-4.00 sec  4.27 MBytes  35.8 Mbits/sec  
> 1962=1962:0:0:0:0:0:0:0
> [  4] 4.00-5.00 sec  1.07 MBytes  9.00 Mbits/sec
> 92.431/81.888/96.221/4.524 ms (9/124958) 96.8 KByte 12.17
> 498=498:0:0:0:0:0:0:0
> [  1] 4.00-5.00 sec  1.07 MBytes  8.97 Mbits/sec
> 95.018/93.445/96.200/0.957 ms (9/124636) 99.3 KByte 11.81
> 490=490:0:0:0:0:0:0:0
> [  2] 4.00-5.00 sec  1.06 MBytes  8.90 Mbits/sec
> 93.874/86.485/95.672/2.810 ms (9/123671) 97.3 KByte 11.86
> 481=481:0:0:0:0:0:0:0
> [  3] 4.00-5.00 sec  1.08 MBytes  9.09 Mbits/sec
> 95.737/93.881/97.197/0.972 ms (9/126245)  101 KByte 11.87
> 484=484:0:0:0:0:0:0:0
> [SUM] 4.00-5.00 sec  4.29 MBytes  36.0 Mbits/sec  
> 1953=1953:0:0:0:0:0:0:0
> [  4] 5.00-6.00 sec  1.09 MBytes  9.13 Mbits/sec
> 92.908/86.844/95.994/3.012 ms (8/142656)  111 KByte 12.28
> 467=467:0:0:0:0:0:0:0
> [  3] 5.00-6.00 sec  1.07 MBytes  8.94 Mbits/sec
> 96.593/95.343/97.660/0.876 ms (8/139760)  113 KByte 11.58
> 478=478:0:0:0:0:0:0:0
> [  1] 5.00-6.00 sec  1.08 MBytes  9.03 Mbits/sec
> 95.021/91.421/97.167/1.893 ms (8/141027)  112 KByte 11.87
> 491=491:0:0:0:0:0:0:0
> [  2] 5.00-6.00 sec  1.08 MBytes  9.06 Mbits/sec
> 92.162/82.720/97.692/5.060 ms (8/141570)  109 KByte 12.29
> 488=488:0:0:0:0:0:0:0
> [SUM] 5.00-6.00 sec  4.31 MBytes  36.2 Mbits/sec  
> 1924=1924:0:0:0:0:0:0:0
> [  4] 6.00-7.00 sec  1.04 MBytes  8.70 Mbits/sec
> 92.793/85.343/96.967/3.552 ms (9/120775) 93.9 KByte 11.71
> 485=485:0:0:0:0:0:0:0
> [  2] 6.00-7.00 sec  1.05 MBytes  8.79 Mbits/sec
> 91.679/84.479/96.760/3.975 ms (9/122062) 93.8 KByte 11.98
> 472=472:0:0:0:0:0:0:0
> [  3] 6.00-7.00 sec  1.06 MBytes  8.88 Mbits/sec
> 96.982/95.933/98.371/0.680 ms (9/123349)  100 KByte 11.45
> 477=477:0:0:0:0:0:0:0
> [  1] 6.00-7.00 sec  1.05 MBytes  8.80 Mbits/sec
> 94.342/91.660/96.025/1.660 ms (9/122223) 96.7 KByte 11.66
> 494=494:0:0:0:0:0:0:0
> [SUM] 6.00-7.00 sec  4.19 MBytes  35.2 Mbits/sec  
> 1928=1928:0:0:0:0:0:0:0
> [  4] 7.00-8.00 sec  1.10 MBytes  9.25 Mbits/sec
> 92.515/88.182/96.351/2.538 ms (8/144466)  112 KByte 12.49
> 510=510:0:0:0:0:0:0:0
> [  3] 7.00-8.00 sec  1.09 MBytes  9.13 Mbits/sec
> 96.580/95.737/98.977/1.098 ms (8/142656)  115 KByte 11.82
> 480=480:0:0:0:0:0:0:0
> [  1] 7.00-8.00 sec  1.10 MBytes  9.21 Mbits/sec
> 95.269/91.719/97.514/2.126 ms (8/143923)  115 KByte 12.09
> 515=515:0:0:0:0:0:0:0
> [  2] 7.00-8.00 sec  1.11 MBytes  9.29 Mbits/sec
> 90.073/84.700/96.176/4.324 ms (8/145190)  110 KByte 12.90
> 508=508:0:0:0:0:0:0:0
> [SUM] 7.00-8.00 sec  4.40 MBytes  36.9 Mbits/sec  
> 2013=2013:0:0:0:0:0:0:0
> 
> Bob
> 
>>> -----Original Message-----
>> 
>>> From: LibreQoS <libreqos-bounces@lists.bufferbloat.net> On Behalf Of
>> Dave Taht
>> 
>>> via LibreQoS
>> 
>>> Sent: Wednesday, January 4, 2023 12:26 PM
>> 
>>> Subject: [LibreQoS] the grinch meets cloudflare's christmas present
>> 
>>> 
>> 
>>> Please try the new, the shiny, the really wonderful test here:
>> 
>>> 
>> https://urldefense.com/v3/__https://speed.cloudflare.com/__;!!BhdT!iZcFJ8WVU9S
>> [1]
>> 
>>> 
>> 9zz5t456oxkfObrC5Xb9j5AG8UO3DqD5x4GAJkawZr0iGwEUtF0_09U8mCDnAkrJ9QEMHGbCMKVw$
>> [1]
>> 
>>> 
>> 
>>> I would really appreciate some independent verification of
>> 
>>> measurements using this tool. In my brief experiments it appears -
>> as
>> 
>>> all the commercial tools to date - to dramatically understate the
>> 
>>> bufferbloat, on my LTE, (and my starlink terminal is out being
>> 
>>> hacked^H^H^H^H^H^Hworked on, so I can't measure that)
>> 
>> [acm]
>> 
>> Hi Dave, I made some time to test "cloudflare's christmas present"
>> yesterday.
>> 
>> I'm on DOCSIS 3.1 service with 1Gbps Down. The Upstream has a "turbo"
>> mode with 40-50Mbps for the first ~3 sec, then steady-state about
>> 23Mbps.
>> 
>> When I saw the ~620Mbps Downstream measurement, I was ready to
>> complain that even the IP-Layer Capacity was grossly underestimated.
>> In addition, the Latency measurements seem very low (as you asserted),
>> although the cloud server was “nearby”.
>> 
>> However, I found that Ookla and the ISP-provided measurement were also
>> reporting ~600Mbps! So the cloudflare Downstream capacity (or
>> throughput?) measurement was consistent with others. Our UDPST server
>> was unreachable, otherwise I would have added that measurement, too.
>> 
>> The Upstream measurement graph seems to illustrate the “turbo”
>> mode, with the dip after attaining 44.5Mbps.
>> 
>> UDPST saturates the uplink and we measure the full 250ms of the
>> Upstream buffer. Cloudflare’s latency measurements don’t even come
>> close.
>> 
>> Al
>> 
>> 
>> 
>> Links:
>> ------
>> [1] 
>> https://urldefense.com/v3/__https:/speed.cloudflare.com/__;!!BhdT!iZcFJ8WVU9S9zz5t456oxkfObrC5Xb9j5AG8UO3DqD5x4GAJkawZr0iGwEUtF0_09U8mCDnAkrJ9QEMHGbCMKVw$
>> _______________________________________________
>> Rpm mailing list
>> Rpm@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/rpm

^ permalink raw reply	[flat|nested] 44+ messages in thread

* [Starlink] insanely great waveform result for starlink
       [not found]                       ` <CAKJdXWDOFbzsam2C_24e9DLkc18ed4uhV51hOKVjDipk1Uhc2g@mail.gmail.com>
@ 2023-01-13  4:08                         ` Dave Taht
  2023-01-13  4:26                           ` Jonathan Bennett
  0 siblings, 1 reply; 44+ messages in thread
From: Dave Taht @ 2023-01-13  4:08 UTC (permalink / raw)
  To: Dave Taht via Starlink

[-- Attachment #1: Type: text/plain, Size: 2089 bytes --]

has anyone been testing theirs this week?

---------- Forwarded message ---------
From: Luis A. Cornejo <luis.a.cornejo@gmail.com>
Date: Thu, Jan 12, 2023 at 7:30 PM
Subject: Re: [Bloat] [Rpm] [Starlink] [LibreQoS] the grinch meets
cloudflare'schristmas present
To: MORTON JR., AL <acmorton@att.com>
Cc: Dave Taht <dave.taht@gmail.com>, Jay Moran <jay@tp.org>, Cake List
<cake@lists.bufferbloat.net>, IETF IPPM WG <ippm@ietf.org>, Rpm
<rpm@lists.bufferbloat.net>, bloat <bloat@lists.bufferbloat.net>,
dickroy@alum.mit.edu <dickroy@alum.mit.edu>, libreqos
<libreqos@lists.bufferbloat.net>


Well Reddit has many posts talking about noticeable performance
increases for Starlink. Here is a primetime run:

waveform:
https://www.waveform.com/tools/bufferbloat?test-id=333f97c7-7cbd-406c-8d9a-9f850cb5de7d

cloudflare attached



On Thu, Jan 12, 2023 at 11:43 AM MORTON JR., AL <acmorton@att.com> wrote:
>
> Dave and Luis,
>
> Do you know if any of these tools are using ~random payloads, to defeat compression?
>
> UDPST has a CLI option:
> (m)    -X           Randomize datagram payload (else zeroes)
>
> When I used this option testing shipboard satellite access, download was about 115kbps.
>
> Al
>
> > -----Original Message-----
> > From: Dave Taht <dave.taht@gmail.com>
> > Sent: Thursday, January 12, 2023 11:12 AM
> > To: Luis A. Cornejo <luis.a.cornejo@gmail.com>
> > Cc: Jay Moran <jay@tp.org>; Cake List <cake@lists.bufferbloat.net>; IETF IPPM
> > WG <ippm@ietf.org>; MORTON JR., AL <acmorton@att.com>; Rpm
> > <rpm@lists.bufferbloat.net>; bloat <bloat@lists.bufferbloat.net>;
> > dickroy@alum.mit.edu; libreqos <libreqos@lists.bufferbloat.net>
> > Subject: Re: [Bloat] [Rpm] [Starlink] [LibreQoS] the grinch meets
> > cloudflare'schristmas present
> >
> > Either starlink has vastly improved, or the test is way off in this case.



-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

[-- Attachment #2: Web capture_12-1-2023_212657_speed.cloudflare.com.jpeg --]
[-- Type: image/jpeg, Size: 181599 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13  4:08                         ` [Starlink] insanely great waveform result for starlink Dave Taht
@ 2023-01-13  4:26                           ` Jonathan Bennett
  2023-01-13  5:13                             ` Ulrich Speidel
  0 siblings, 1 reply; 44+ messages in thread
From: Jonathan Bennett @ 2023-01-13  4:26 UTC (permalink / raw)
  To: Dave Taht; +Cc: Dave Taht via Starlink


[-- Attachment #1.1: Type: text/plain, Size: 2766 bytes --]

Just ran a Waveform test, and ping times are still terrible while maxing
out download speeds. Upload saturation doesn't seem to affect latency,
though. Also interesting, it's giving me a valid ipv6 block again.

[image: image.png]
[image: image.png]
Jonathan Bennett
Hackaday.com


On Thu, Jan 12, 2023 at 10:08 PM Dave Taht via Starlink <
starlink@lists.bufferbloat.net> wrote:

> has anyone been testing theirs this week?
>
> ---------- Forwarded message ---------
> From: Luis A. Cornejo <luis.a.cornejo@gmail.com>
> Date: Thu, Jan 12, 2023 at 7:30 PM
> Subject: Re: [Bloat] [Rpm] [Starlink] [LibreQoS] the grinch meets
> cloudflare'schristmas present
> To: MORTON JR., AL <acmorton@att.com>
> Cc: Dave Taht <dave.taht@gmail.com>, Jay Moran <jay@tp.org>, Cake List
> <cake@lists.bufferbloat.net>, IETF IPPM WG <ippm@ietf.org>, Rpm
> <rpm@lists.bufferbloat.net>, bloat <bloat@lists.bufferbloat.net>,
> dickroy@alum.mit.edu <dickroy@alum.mit.edu>, libreqos
> <libreqos@lists.bufferbloat.net>
>
>
> Well Reddit has many posts talking about noticeable performance
> increases for Starlink. Here is a primetime run:
>
> waveform:
>
> https://www.waveform.com/tools/bufferbloat?test-id=333f97c7-7cbd-406c-8d9a-9f850cb5de7d
>
> cloudflare attached
>
>
>
> On Thu, Jan 12, 2023 at 11:43 AM MORTON JR., AL <acmorton@att.com> wrote:
> >
> > Dave and Luis,
> >
> > Do you know if any of these tools are using ~random payloads, to defeat
> compression?
> >
> > UDPST has a CLI option:
> > (m)    -X           Randomize datagram payload (else zeroes)
> >
> > When I used this option testing shipboard satellite access, download was
> about 115kbps.
> >
> > Al
> >
> > > -----Original Message-----
> > > From: Dave Taht <dave.taht@gmail.com>
> > > Sent: Thursday, January 12, 2023 11:12 AM
> > > To: Luis A. Cornejo <luis.a.cornejo@gmail.com>
> > > Cc: Jay Moran <jay@tp.org>; Cake List <cake@lists.bufferbloat.net>;
> IETF IPPM
> > > WG <ippm@ietf.org>; MORTON JR., AL <acmorton@att.com>; Rpm
> > > <rpm@lists.bufferbloat.net>; bloat <bloat@lists.bufferbloat.net>;
> > > dickroy@alum.mit.edu; libreqos <libreqos@lists.bufferbloat.net>
> > > Subject: Re: [Bloat] [Rpm] [Starlink] [LibreQoS] the grinch meets
> > > cloudflare'schristmas present
> > >
> > > Either starlink has vastly improved, or the test is way off in this
> case.
>
>
>
> --
> This song goes out to all the folk that thought Stadia would work:
>
> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
> Dave Täht CEO, TekLibre, LLC
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink
>

[-- Attachment #1.2: Type: text/html, Size: 5499 bytes --]

[-- Attachment #2: image.png --]
[-- Type: image/png, Size: 235280 bytes --]

[-- Attachment #3: image.png --]
[-- Type: image/png, Size: 298232 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13  4:26                           ` Jonathan Bennett
@ 2023-01-13  5:13                             ` Ulrich Speidel
  2023-01-13  5:25                               ` Dave Taht
  2023-01-13 12:27                               ` Ulrich Speidel
  0 siblings, 2 replies; 44+ messages in thread
From: Ulrich Speidel @ 2023-01-13  5:13 UTC (permalink / raw)
  To: starlink

[-- Attachment #1: Type: text/plain, Size: 3984 bytes --]

 From Auckland, New Zealand, using a roaming subscription, it puts me in 
touch with a server 2000 km away. OK then:


IP address: nix six.

My thoughts shall follow later.

On 13/01/2023 5:26 pm, Jonathan Bennett via Starlink wrote:
> Just ran a Waveform test, and ping times are still terrible 
> while maxing out download speeds. Upload saturation doesn't seem to 
> affect latency, though. Also interesting, it's giving me a valid ipv6 
> block again.
>
> image.png
> image.png
> Jonathan Bennett
> Hackaday.com
>
>
> On Thu, Jan 12, 2023 at 10:08 PM Dave Taht via Starlink 
> <starlink@lists.bufferbloat.net> wrote:
>
>     has anyone been testing theirs this week?
>
>     ---------- Forwarded message ---------
>     From: Luis A. Cornejo <luis.a.cornejo@gmail.com>
>     Date: Thu, Jan 12, 2023 at 7:30 PM
>     Subject: Re: [Bloat] [Rpm] [Starlink] [LibreQoS] the grinch meets
>     cloudflare'schristmas present
>     To: MORTON JR., AL <acmorton@att.com>
>     Cc: Dave Taht <dave.taht@gmail.com>, Jay Moran <jay@tp.org>, Cake List
>     <cake@lists.bufferbloat.net>, IETF IPPM WG <ippm@ietf.org>, Rpm
>     <rpm@lists.bufferbloat.net>, bloat <bloat@lists.bufferbloat.net>,
>     dickroy@alum.mit.edu <dickroy@alum.mit.edu>, libreqos
>     <libreqos@lists.bufferbloat.net>
>
>
>     Well Reddit has many posts talking about noticeable performance
>     increases for Starlink. Here is a primetime run:
>
>     waveform:
>     https://www.waveform.com/tools/bufferbloat?test-id=333f97c7-7cbd-406c-8d9a-9f850cb5de7d
>     <https://www.waveform.com/tools/bufferbloat?test-id=333f97c7-7cbd-406c-8d9a-9f850cb5de7d>
>
>     cloudflare attached
>
>
>
>     On Thu, Jan 12, 2023 at 11:43 AM MORTON JR., AL <acmorton@att.com>
>     wrote:
>     >
>     > Dave and Luis,
>     >
>     > Do you know if any of these tools are using ~random payloads, to
>     defeat compression?
>     >
>     > UDPST has a CLI option:
>     > (m)    -X           Randomize datagram payload (else zeroes)
>     >
>     > When I used this option testing shipboard satellite access,
>     download was about 115kbps.
>     >
>     > Al
>     >
>     > > -----Original Message-----
>     > > From: Dave Taht <dave.taht@gmail.com>
>     > > Sent: Thursday, January 12, 2023 11:12 AM
>     > > To: Luis A. Cornejo <luis.a.cornejo@gmail.com>
>     > > Cc: Jay Moran <jay@tp.org>; Cake List
>     <cake@lists.bufferbloat.net>; IETF IPPM
>     > > WG <ippm@ietf.org>; MORTON JR., AL <acmorton@att.com>; Rpm
>     > > <rpm@lists.bufferbloat.net>; bloat <bloat@lists.bufferbloat.net>;
>     > > dickroy@alum.mit.edu; libreqos <libreqos@lists.bufferbloat.net>
>     > > Subject: Re: [Bloat] [Rpm] [Starlink] [LibreQoS] the grinch meets
>     > > cloudflare'schristmas present
>     > >
>     > > Either starlink has vastly improved, or the test is way off in
>     this case.
>
>
>
>     -- 
>     This song goes out to all the folk that thought Stadia would work:
>     https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>     <https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz>
>     Dave Täht CEO, TekLibre, LLC
>     _______________________________________________
>     Starlink mailing list
>     Starlink@lists.bufferbloat.net
>     https://lists.bufferbloat.net/listinfo/starlink
>     <https://lists.bufferbloat.net/listinfo/starlink>
>
>
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink
>
-- 
****************************************************************
Dr. Ulrich Speidel

School of Computer Science

Room 303S.594 (City Campus)

The University of Auckland
u.speidel@auckland.ac.nz  
http://www.cs.auckland.ac.nz/~ulrich/
****************************************************************



[-- Attachment #2.1: Type: text/html, Size: 10059 bytes --]

[-- Attachment #2.2: sPzzvwIXE4go16tN.png --]
[-- Type: image/png, Size: 109764 bytes --]

[-- Attachment #2.3: OMUOBp06rXY0I8Ws.png --]
[-- Type: image/png, Size: 92859 bytes --]

[-- Attachment #2.4: image.png --]
[-- Type: image/png, Size: 235280 bytes --]

[-- Attachment #2.5: image.png --]
[-- Type: image/png, Size: 298232 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13  5:13                             ` Ulrich Speidel
@ 2023-01-13  5:25                               ` Dave Taht
  2023-01-13 12:27                               ` Ulrich Speidel
  1 sibling, 0 replies; 44+ messages in thread
From: Dave Taht @ 2023-01-13  5:25 UTC (permalink / raw)
  To: Ulrich Speidel; +Cc: starlink


[-- Attachment #1.1: Type: text/plain, Size: 4136 bytes --]

both of you are showing insanely great upload numbers.

On Thu, Jan 12, 2023 at 9:14 PM Ulrich Speidel via Starlink <
starlink@lists.bufferbloat.net> wrote:

> From Auckland, New Zealand, using a roaming subscription, it puts me in
> touch with a server 2000 km away. OK then:
>
>
> IP address: nix six.
>
> My thoughts shall follow later.
>
> On 13/01/2023 5:26 pm, Jonathan Bennett via Starlink wrote:
>
> Just ran a Waveform test, and ping times are still terrible while maxing
> out download speeds. Upload saturation doesn't seem to affect latency,
> though. Also interesting, it's giving me a valid ipv6 block again.
>
> [image: image.png]
> [image: image.png]
> Jonathan Bennett
> Hackaday.com
>
>
> On Thu, Jan 12, 2023 at 10:08 PM Dave Taht via Starlink <
> starlink@lists.bufferbloat.net> wrote:
>
>> has anyone been testing theirs this week?
>>
>> ---------- Forwarded message ---------
>> From: Luis A. Cornejo <luis.a.cornejo@gmail.com>
>> Date: Thu, Jan 12, 2023 at 7:30 PM
>> Subject: Re: [Bloat] [Rpm] [Starlink] [LibreQoS] the grinch meets
>> cloudflare'schristmas present
>> To: MORTON JR., AL <acmorton@att.com>
>> Cc: Dave Taht <dave.taht@gmail.com>, Jay Moran <jay@tp.org>, Cake List
>> <cake@lists.bufferbloat.net>, IETF IPPM WG <ippm@ietf.org>, Rpm
>> <rpm@lists.bufferbloat.net>, bloat <bloat@lists.bufferbloat.net>,
>> dickroy@alum.mit.edu <dickroy@alum.mit.edu>, libreqos
>> <libreqos@lists.bufferbloat.net>
>>
>>
>> Well Reddit has many posts talking about noticeable performance
>> increases for Starlink. Here is a primetime run:
>>
>> waveform:
>>
>> https://www.waveform.com/tools/bufferbloat?test-id=333f97c7-7cbd-406c-8d9a-9f850cb5de7d
>>
>> cloudflare attached
>>
>>
>>
>> On Thu, Jan 12, 2023 at 11:43 AM MORTON JR., AL <acmorton@att.com> wrote:
>> >
>> > Dave and Luis,
>> >
>> > Do you know if any of these tools are using ~random payloads, to defeat
>> compression?
>> >
>> > UDPST has a CLI option:
>> > (m)    -X           Randomize datagram payload (else zeroes)
>> >
>> > When I used this option testing shipboard satellite access, download
>> was about 115kbps.
>> >
>> > Al
>> >
>> > > -----Original Message-----
>> > > From: Dave Taht <dave.taht@gmail.com>
>> > > Sent: Thursday, January 12, 2023 11:12 AM
>> > > To: Luis A. Cornejo <luis.a.cornejo@gmail.com>
>> > > Cc: Jay Moran <jay@tp.org>; Cake List <cake@lists.bufferbloat.net>;
>> IETF IPPM
>> > > WG <ippm@ietf.org>; MORTON JR., AL <acmorton@att.com>; Rpm
>> > > <rpm@lists.bufferbloat.net>; bloat <bloat@lists.bufferbloat.net>;
>> > > dickroy@alum.mit.edu; libreqos <libreqos@lists.bufferbloat.net>
>> > > Subject: Re: [Bloat] [Rpm] [Starlink] [LibreQoS] the grinch meets
>> > > cloudflare'schristmas present
>> > >
>> > > Either starlink has vastly improved, or the test is way off in this
>> case.
>>
>>
>>
>> --
>> This song goes out to all the folk that thought Stadia would work:
>>
>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>> Dave Täht CEO, TekLibre, LLC
>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink
>>
>
> _______________________________________________
> Starlink mailing listStarlink@lists.bufferbloat.nethttps://lists.bufferbloat.net/listinfo/starlink
>
> --
> ****************************************************************
> Dr. Ulrich Speidel
>
> School of Computer Science
>
> Room 303S.594 (City Campus)
>
> The University of Aucklandu.speidel@auckland.ac.nz http://www.cs.auckland.ac.nz/~ulrich/
> ****************************************************************
>
>
>
>
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink
>


-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

[-- Attachment #1.2: Type: text/html, Size: 9289 bytes --]

[-- Attachment #2: sPzzvwIXE4go16tN.png --]
[-- Type: image/png, Size: 109764 bytes --]

[-- Attachment #3: OMUOBp06rXY0I8Ws.png --]
[-- Type: image/png, Size: 92859 bytes --]

[-- Attachment #4: image.png --]
[-- Type: image/png, Size: 235280 bytes --]

[-- Attachment #5: image.png --]
[-- Type: image/png, Size: 298232 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13  5:13                             ` Ulrich Speidel
  2023-01-13  5:25                               ` Dave Taht
@ 2023-01-13 12:27                               ` Ulrich Speidel
  2023-01-13 17:02                                 ` Jonathan Bennett
  1 sibling, 1 reply; 44+ messages in thread
From: Ulrich Speidel @ 2023-01-13 12:27 UTC (permalink / raw)
  To: starlink

On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>
> From Auckland, New Zealand, using a roaming subscription, it puts me 
> in touch with a server 2000 km away. OK then:
>
>
> IP address: nix six.
>
> My thoughts shall follow later.

OK, so here we go.

I'm always a bit skeptical when it comes to speed tests - they're really 
laden with so many caveats that it's not funny. I took our new work 
Starlink kit home in December to give it a try and the other day finally 
got around to set it up. It's on a roaming subscription because our 
badly built-up campus really isn't ideal in terms of a clear view of the 
sky. Oh - and did I mention that I used the Starlink Ethernet adapter, 
not the WiFi?

Caveat 1: Location, location. I live in a place where the best Starlink 
promises is about 1/3 in terms of data rate you can actually get from 
fibre to the home at under half of Starlink's price. Read: There are few 
Starlink users around. I might be the only one in my suburb.

Caveat 2: Auckland has three Starlink gateways close by: Clevedon (which 
is at a stretch daytrip cycling distance from here), Te Hana and Puwera, 
the most distant of the three and about 130 km away from me as the crow 
flies. Read: My dishy can use any satellite that any of these three can 
see, and then depending on where I put it and how much of the southern 
sky it can see, maybe also the one in Hinds, 840 km away, although that 
is obviously stretching it a bit. Either way, that's plenty of options 
for my bits to travel without needing a lot of handovers. Why? Easy: If 
your nearest teleport is close by, then the set of satellites that the 
teleport can see and the set that you can see is almost the same, so you 
can essentially stick with the same satellite while it's in view for you 
because it'll also be in view for the teleport. Pretty much any bird 
above you will do.

And because I don't get a lot of competition from other users in my area 
vying for one of the few available satellites that can see both us and 
the teleport, this is about as good as it gets at 37S latitude. If I'd 
want it any better, I'd have to move a lot further south.

It'd be interesting to hear from Jonathan what the availability of home 
broadband is like in the Dallas area. I note that it's at a lower 
latitude (33N) than Auckland, but the difference isn't huge. I notice 
two teleports each about 160 km away, which is also not too bad. I also 
note Starlink availability in the area is restricted at the moment - 
oversubscribed? But if Jonathan gets good data rates, then that means 
that competition for bird capacity can't be too bad - for whatever reason.

Caveat 3: Backhaul. There isn't just one queue between me and whatever I 
talk to in terms of my communications. Traceroute shows about 10 hops 
between me and the University of Auckland via Starlink. That's 10 
queues, not one. Many of them will have cross traffic. So it's a bit 
hard to tell where our packets really get to wait or where they get 
dropped. The insidious bit here is that a lot of them will be between 1 
Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can all 
turn into bottlenecks. This isn't like a narrowband GEO link of a few 
Mb/s where it's obvious where the dominant long latency bottleneck in 
your TCP connection's path is. Read: It's pretty hard to tell whether a 
drop in "speed" is due to a performance issue in the Starlink system or 
somewhere between Starlink's systems and the target system.

I see RTTs here between 20 ms and 250 ms, where the physical latency 
should be under 15 ms. So there's clearly a bit of buffer here along the 
chain that occasionally fills up.

Caveat 4: Handovers. Handover between birds and teleports is inevitably 
associated with a change in RTT and in most cases also available 
bandwidth. Plus your packets now arrive at a new queue on a new 
satellite while your TCP is still trying to respond to whatever it 
thought the queue on the previous bird was doing. Read: Whatever your 
cwnd is immediately after a handover, it's probably not what it should be.

I ran a somewhat hamstrung (sky view restricted) set of four Ookla 
speedtest.net tests each to five local servers. Average upload rate was 
13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP that 
Starlink seems to be buying its local connectivity from (Vocus Group) 
varied between 3.04 and 14.38 Mb/s, download between 23.33 and 52.22 
Mb/s, with RTTs between 37 and 56 ms not correlating well to rates 
observed. In fact, they were the ISP with consistently the worst rates.

Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and 
between 106.5 and 183.8 Mb/s down, again with RTTs badly correlating 
with rates. Average RTT was the same as for Vocus.

Note the variation though: More or less a factor of two between highest 
and lowest rates for each ISP. Did MyRepublic just get lucky in my 
tests? Or is there something systematic behind this? Way too few tests 
to tell.

What these tests do is establish a ballpark.

I'm currently repeating tests with dish placed on a trestle closer to 
the heavens. This seems to have translated into fewer outages / ping 
losses (around 1/4 of what I had yesterday with dishy on the ground on 
my deck). Still good enough for a lengthy video Skype call with my folks 
in Germany, although they did comment about reduced video quality. But 
maybe that was the lighting or the different background as I wasn't in 
my usual spot with my laptop when I called them.

-- 

****************************************************************
Dr. Ulrich Speidel

School of Computer Science

Room 303S.594 (City Campus)

The University of Auckland
u.speidel@auckland.ac.nz
http://www.cs.auckland.ac.nz/~ulrich/
****************************************************************




^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 12:27                               ` Ulrich Speidel
@ 2023-01-13 17:02                                 ` Jonathan Bennett
  2023-01-13 17:26                                   ` Dave Taht
  2023-01-13 22:51                                   ` Ulrich Speidel
  0 siblings, 2 replies; 44+ messages in thread
From: Jonathan Bennett @ 2023-01-13 17:02 UTC (permalink / raw)
  To: Ulrich Speidel; +Cc: starlink

[-- Attachment #1: Type: text/plain, Size: 6764 bytes --]

On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
starlink@lists.bufferbloat.net> wrote:

> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
> >
> > From Auckland, New Zealand, using a roaming subscription, it puts me
> > in touch with a server 2000 km away. OK then:
> >
> >
> > IP address: nix six.
> >
> > My thoughts shall follow later.
>
> OK, so here we go.
>
> I'm always a bit skeptical when it comes to speed tests - they're really
> laden with so many caveats that it's not funny. I took our new work
> Starlink kit home in December to give it a try and the other day finally
> got around to set it up. It's on a roaming subscription because our
> badly built-up campus really isn't ideal in terms of a clear view of the
> sky. Oh - and did I mention that I used the Starlink Ethernet adapter,
> not the WiFi?
>
> Caveat 1: Location, location. I live in a place where the best Starlink
> promises is about 1/3 in terms of data rate you can actually get from
> fibre to the home at under half of Starlink's price. Read: There are few
> Starlink users around. I might be the only one in my suburb.
>
> Caveat 2: Auckland has three Starlink gateways close by: Clevedon (which
> is at a stretch daytrip cycling distance from here), Te Hana and Puwera,
> the most distant of the three and about 130 km away from me as the crow
> flies. Read: My dishy can use any satellite that any of these three can
> see, and then depending on where I put it and how much of the southern
> sky it can see, maybe also the one in Hinds, 840 km away, although that
> is obviously stretching it a bit. Either way, that's plenty of options
> for my bits to travel without needing a lot of handovers. Why? Easy: If
> your nearest teleport is close by, then the set of satellites that the
> teleport can see and the set that you can see is almost the same, so you
> can essentially stick with the same satellite while it's in view for you
> because it'll also be in view for the teleport. Pretty much any bird
> above you will do.
>
> And because I don't get a lot of competition from other users in my area
> vying for one of the few available satellites that can see both us and
> the teleport, this is about as good as it gets at 37S latitude. If I'd
> want it any better, I'd have to move a lot further south.
>
> It'd be interesting to hear from Jonathan what the availability of home
> broadband is like in the Dallas area. I note that it's at a lower
> latitude (33N) than Auckland, but the difference isn't huge. I notice
> two teleports each about 160 km away, which is also not too bad. I also
> note Starlink availability in the area is restricted at the moment -
> oversubscribed? But if Jonathan gets good data rates, then that means
> that competition for bird capacity can't be too bad - for whatever reason.
>
I'm in Southwest Oklahoma, but Dallas is the nearby Starlink gateway. In
cities, like Dallas, and Lawton where I live, there are good broadband
options. But there are also many people that live outside cities, and the
options are much worse. The low density userbase in rural Oklahoma and
Texas is probably ideal conditions for Starlink.

>
> Caveat 3: Backhaul. There isn't just one queue between me and whatever I
> talk to in terms of my communications. Traceroute shows about 10 hops
> between me and the University of Auckland via Starlink. That's 10
> queues, not one. Many of them will have cross traffic. So it's a bit
> hard to tell where our packets really get to wait or where they get
> dropped. The insidious bit here is that a lot of them will be between 1
> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can all
> turn into bottlenecks. This isn't like a narrowband GEO link of a few
> Mb/s where it's obvious where the dominant long latency bottleneck in
> your TCP connection's path is. Read: It's pretty hard to tell whether a
> drop in "speed" is due to a performance issue in the Starlink system or
> somewhere between Starlink's systems and the target system.
>
> I see RTTs here between 20 ms and 250 ms, where the physical latency
> should be under 15 ms. So there's clearly a bit of buffer here along the
> chain that occasionally fills up.
>
> Caveat 4: Handovers. Handover between birds and teleports is inevitably
> associated with a change in RTT and in most cases also available
> bandwidth. Plus your packets now arrive at a new queue on a new
> satellite while your TCP is still trying to respond to whatever it
> thought the queue on the previous bird was doing. Read: Whatever your
> cwnd is immediately after a handover, it's probably not what it should be.
>
> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
> speedtest.net tests each to five local servers. Average upload rate was
> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP that
> Starlink seems to be buying its local connectivity from (Vocus Group)
> varied between 3.04 and 14.38 Mb/s, download between 23.33 and 52.22
> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
> observed. In fact, they were the ISP with consistently the worst rates.
>
> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and
> between 106.5 and 183.8 Mb/s down, again with RTTs badly correlating
> with rates. Average RTT was the same as for Vocus.
>
> Note the variation though: More or less a factor of two between highest
> and lowest rates for each ISP. Did MyRepublic just get lucky in my
> tests? Or is there something systematic behind this? Way too few tests
> to tell.
>
> What these tests do is establish a ballpark.
>
> I'm currently repeating tests with dish placed on a trestle closer to
> the heavens. This seems to have translated into fewer outages / ping
> losses (around 1/4 of what I had yesterday with dishy on the ground on
> my deck). Still good enough for a lengthy video Skype call with my folks
> in Germany, although they did comment about reduced video quality. But
> maybe that was the lighting or the different background as I wasn't in
> my usual spot with my laptop when I called them.
>
Clear view of the sky is king for Starlink reliability. I've got my dishy
mounted on the back fence, looking up over an empty field, so it's pretty
much best-case scenario here.

>
> --
>
> ****************************************************************
> Dr. Ulrich Speidel
>
> School of Computer Science
>
> Room 303S.594 (City Campus)
>
> The University of Auckland
> u.speidel@auckland.ac.nz
> http://www.cs.auckland.ac.nz/~ulrich/
> ****************************************************************
>
>
>
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink
>

[-- Attachment #2: Type: text/html, Size: 8429 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 17:02                                 ` Jonathan Bennett
@ 2023-01-13 17:26                                   ` Dave Taht
  2023-01-13 17:41                                     ` Jonathan Bennett
  2023-01-13 20:25                                     ` Pat Jensen
  2023-01-13 22:51                                   ` Ulrich Speidel
  1 sibling, 2 replies; 44+ messages in thread
From: Dave Taht @ 2023-01-13 17:26 UTC (permalink / raw)
  To: Jonathan Bennett; +Cc: Ulrich Speidel, starlink

packet caps would be nice... all this is very exciting news.

I'd so love for one or more of y'all reporting such great uplink
results nowadays to duplicate and re-plot the original irtt tests we
did:

irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o whatever.json

They MUST have changed their scheduling to get such amazing uplink
results, in addition to better queue management.

(for the record, my servers are de, london, fremont, sydney, dallas,
newark, atlanta, singapore, mumbai)

There's an R and gnuplot script for plotting that output around here
somewhere (I have largely personally put down the starlink project,
loaning out mine) - that went by on this list... I should have written
a blog entry so I can find that stuff again.

On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
<starlink@lists.bufferbloat.net> wrote:
>
>
> On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <starlink@lists.bufferbloat.net> wrote:
>>
>> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>> >
>> > From Auckland, New Zealand, using a roaming subscription, it puts me
>> > in touch with a server 2000 km away. OK then:
>> >
>> >
>> > IP address: nix six.
>> >
>> > My thoughts shall follow later.
>>
>> OK, so here we go.
>>
>> I'm always a bit skeptical when it comes to speed tests - they're really
>> laden with so many caveats that it's not funny. I took our new work
>> Starlink kit home in December to give it a try and the other day finally
>> got around to set it up. It's on a roaming subscription because our
>> badly built-up campus really isn't ideal in terms of a clear view of the
>> sky. Oh - and did I mention that I used the Starlink Ethernet adapter,
>> not the WiFi?
>>
>> Caveat 1: Location, location. I live in a place where the best Starlink
>> promises is about 1/3 in terms of data rate you can actually get from
>> fibre to the home at under half of Starlink's price. Read: There are few
>> Starlink users around. I might be the only one in my suburb.
>>
>> Caveat 2: Auckland has three Starlink gateways close by: Clevedon (which
>> is at a stretch daytrip cycling distance from here), Te Hana and Puwera,
>> the most distant of the three and about 130 km away from me as the crow
>> flies. Read: My dishy can use any satellite that any of these three can
>> see, and then depending on where I put it and how much of the southern
>> sky it can see, maybe also the one in Hinds, 840 km away, although that
>> is obviously stretching it a bit. Either way, that's plenty of options
>> for my bits to travel without needing a lot of handovers. Why? Easy: If
>> your nearest teleport is close by, then the set of satellites that the
>> teleport can see and the set that you can see is almost the same, so you
>> can essentially stick with the same satellite while it's in view for you
>> because it'll also be in view for the teleport. Pretty much any bird
>> above you will do.
>>
>> And because I don't get a lot of competition from other users in my area
>> vying for one of the few available satellites that can see both us and
>> the teleport, this is about as good as it gets at 37S latitude. If I'd
>> want it any better, I'd have to move a lot further south.
>>
>> It'd be interesting to hear from Jonathan what the availability of home
>> broadband is like in the Dallas area. I note that it's at a lower
>> latitude (33N) than Auckland, but the difference isn't huge. I notice
>> two teleports each about 160 km away, which is also not too bad. I also
>> note Starlink availability in the area is restricted at the moment -
>> oversubscribed? But if Jonathan gets good data rates, then that means
>> that competition for bird capacity can't be too bad - for whatever reason.
>
> I'm in Southwest Oklahoma, but Dallas is the nearby Starlink gateway. In cities, like Dallas, and Lawton where I live, there are good broadband options. But there are also many people that live outside cities, and the options are much worse. The low density userbase in rural Oklahoma and Texas is probably ideal conditions for Starlink.
>>
>>
>> Caveat 3: Backhaul. There isn't just one queue between me and whatever I
>> talk to in terms of my communications. Traceroute shows about 10 hops
>> between me and the University of Auckland via Starlink. That's 10
>> queues, not one. Many of them will have cross traffic. So it's a bit
>> hard to tell where our packets really get to wait or where they get
>> dropped. The insidious bit here is that a lot of them will be between 1
>> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can all
>> turn into bottlenecks. This isn't like a narrowband GEO link of a few
>> Mb/s where it's obvious where the dominant long latency bottleneck in
>> your TCP connection's path is. Read: It's pretty hard to tell whether a
>> drop in "speed" is due to a performance issue in the Starlink system or
>> somewhere between Starlink's systems and the target system.
>>
>> I see RTTs here between 20 ms and 250 ms, where the physical latency
>> should be under 15 ms. So there's clearly a bit of buffer here along the
>> chain that occasionally fills up.
>>
>> Caveat 4: Handovers. Handover between birds and teleports is inevitably
>> associated with a change in RTT and in most cases also available
>> bandwidth. Plus your packets now arrive at a new queue on a new
>> satellite while your TCP is still trying to respond to whatever it
>> thought the queue on the previous bird was doing. Read: Whatever your
>> cwnd is immediately after a handover, it's probably not what it should be.
>>
>> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
>> speedtest.net tests each to five local servers. Average upload rate was
>> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP that
>> Starlink seems to be buying its local connectivity from (Vocus Group)
>> varied between 3.04 and 14.38 Mb/s, download between 23.33 and 52.22
>> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
>> observed. In fact, they were the ISP with consistently the worst rates.
>>
>> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and
>> between 106.5 and 183.8 Mb/s down, again with RTTs badly correlating
>> with rates. Average RTT was the same as for Vocus.
>>
>> Note the variation though: More or less a factor of two between highest
>> and lowest rates for each ISP. Did MyRepublic just get lucky in my
>> tests? Or is there something systematic behind this? Way too few tests
>> to tell.
>>
>> What these tests do is establish a ballpark.
>>
>> I'm currently repeating tests with dish placed on a trestle closer to
>> the heavens. This seems to have translated into fewer outages / ping
>> losses (around 1/4 of what I had yesterday with dishy on the ground on
>> my deck). Still good enough for a lengthy video Skype call with my folks
>> in Germany, although they did comment about reduced video quality. But
>> maybe that was the lighting or the different background as I wasn't in
>> my usual spot with my laptop when I called them.
>
> Clear view of the sky is king for Starlink reliability. I've got my dishy mounted on the back fence, looking up over an empty field, so it's pretty much best-case scenario here.
>>
>>
>> --
>>
>> ****************************************************************
>> Dr. Ulrich Speidel
>>
>> School of Computer Science
>>
>> Room 303S.594 (City Campus)
>>
>> The University of Auckland
>> u.speidel@auckland.ac.nz
>> http://www.cs.auckland.ac.nz/~ulrich/
>> ****************************************************************
>>
>>
>>
>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink
>
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink



-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 17:26                                   ` Dave Taht
@ 2023-01-13 17:41                                     ` Jonathan Bennett
  2023-01-13 18:09                                       ` Nathan Owens
  2023-01-13 20:25                                     ` Pat Jensen
  1 sibling, 1 reply; 44+ messages in thread
From: Jonathan Bennett @ 2023-01-13 17:41 UTC (permalink / raw)
  To: Dave Taht; +Cc: Ulrich Speidel, starlink

[-- Attachment #1: Type: text/plain, Size: 8889 bytes --]

The irtt command, run with normal, light usage:
https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link

Jonathan Bennett
Hackaday.com


On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com> wrote:

> packet caps would be nice... all this is very exciting news.
>
> I'd so love for one or more of y'all reporting such great uplink
> results nowadays to duplicate and re-plot the original irtt tests we
> did:
>
> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
> whatever.json
>
> They MUST have changed their scheduling to get such amazing uplink
> results, in addition to better queue management.
>
> (for the record, my servers are de, london, fremont, sydney, dallas,
> newark, atlanta, singapore, mumbai)
>
> There's an R and gnuplot script for plotting that output around here
> somewhere (I have largely personally put down the starlink project,
> loaning out mine) - that went by on this list... I should have written
> a blog entry so I can find that stuff again.
>
> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
> <starlink@lists.bufferbloat.net> wrote:
> >
> >
> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
> starlink@lists.bufferbloat.net> wrote:
> >>
> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
> >> >
> >> > From Auckland, New Zealand, using a roaming subscription, it puts me
> >> > in touch with a server 2000 km away. OK then:
> >> >
> >> >
> >> > IP address: nix six.
> >> >
> >> > My thoughts shall follow later.
> >>
> >> OK, so here we go.
> >>
> >> I'm always a bit skeptical when it comes to speed tests - they're really
> >> laden with so many caveats that it's not funny. I took our new work
> >> Starlink kit home in December to give it a try and the other day finally
> >> got around to set it up. It's on a roaming subscription because our
> >> badly built-up campus really isn't ideal in terms of a clear view of the
> >> sky. Oh - and did I mention that I used the Starlink Ethernet adapter,
> >> not the WiFi?
> >>
> >> Caveat 1: Location, location. I live in a place where the best Starlink
> >> promises is about 1/3 in terms of data rate you can actually get from
> >> fibre to the home at under half of Starlink's price. Read: There are few
> >> Starlink users around. I might be the only one in my suburb.
> >>
> >> Caveat 2: Auckland has three Starlink gateways close by: Clevedon (which
> >> is at a stretch daytrip cycling distance from here), Te Hana and Puwera,
> >> the most distant of the three and about 130 km away from me as the crow
> >> flies. Read: My dishy can use any satellite that any of these three can
> >> see, and then depending on where I put it and how much of the southern
> >> sky it can see, maybe also the one in Hinds, 840 km away, although that
> >> is obviously stretching it a bit. Either way, that's plenty of options
> >> for my bits to travel without needing a lot of handovers. Why? Easy: If
> >> your nearest teleport is close by, then the set of satellites that the
> >> teleport can see and the set that you can see is almost the same, so you
> >> can essentially stick with the same satellite while it's in view for you
> >> because it'll also be in view for the teleport. Pretty much any bird
> >> above you will do.
> >>
> >> And because I don't get a lot of competition from other users in my area
> >> vying for one of the few available satellites that can see both us and
> >> the teleport, this is about as good as it gets at 37S latitude. If I'd
> >> want it any better, I'd have to move a lot further south.
> >>
> >> It'd be interesting to hear from Jonathan what the availability of home
> >> broadband is like in the Dallas area. I note that it's at a lower
> >> latitude (33N) than Auckland, but the difference isn't huge. I notice
> >> two teleports each about 160 km away, which is also not too bad. I also
> >> note Starlink availability in the area is restricted at the moment -
> >> oversubscribed? But if Jonathan gets good data rates, then that means
> >> that competition for bird capacity can't be too bad - for whatever
> reason.
> >
> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink gateway. In
> cities, like Dallas, and Lawton where I live, there are good broadband
> options. But there are also many people that live outside cities, and the
> options are much worse. The low density userbase in rural Oklahoma and
> Texas is probably ideal conditions for Starlink.
> >>
> >>
> >> Caveat 3: Backhaul. There isn't just one queue between me and whatever I
> >> talk to in terms of my communications. Traceroute shows about 10 hops
> >> between me and the University of Auckland via Starlink. That's 10
> >> queues, not one. Many of them will have cross traffic. So it's a bit
> >> hard to tell where our packets really get to wait or where they get
> >> dropped. The insidious bit here is that a lot of them will be between 1
> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can all
> >> turn into bottlenecks. This isn't like a narrowband GEO link of a few
> >> Mb/s where it's obvious where the dominant long latency bottleneck in
> >> your TCP connection's path is. Read: It's pretty hard to tell whether a
> >> drop in "speed" is due to a performance issue in the Starlink system or
> >> somewhere between Starlink's systems and the target system.
> >>
> >> I see RTTs here between 20 ms and 250 ms, where the physical latency
> >> should be under 15 ms. So there's clearly a bit of buffer here along the
> >> chain that occasionally fills up.
> >>
> >> Caveat 4: Handovers. Handover between birds and teleports is inevitably
> >> associated with a change in RTT and in most cases also available
> >> bandwidth. Plus your packets now arrive at a new queue on a new
> >> satellite while your TCP is still trying to respond to whatever it
> >> thought the queue on the previous bird was doing. Read: Whatever your
> >> cwnd is immediately after a handover, it's probably not what it should
> be.
> >>
> >> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
> >> speedtest.net tests each to five local servers. Average upload rate was
> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP that
> >> Starlink seems to be buying its local connectivity from (Vocus Group)
> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and 52.22
> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
> >> observed. In fact, they were the ISP with consistently the worst rates.
> >>
> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and
> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly correlating
> >> with rates. Average RTT was the same as for Vocus.
> >>
> >> Note the variation though: More or less a factor of two between highest
> >> and lowest rates for each ISP. Did MyRepublic just get lucky in my
> >> tests? Or is there something systematic behind this? Way too few tests
> >> to tell.
> >>
> >> What these tests do is establish a ballpark.
> >>
> >> I'm currently repeating tests with dish placed on a trestle closer to
> >> the heavens. This seems to have translated into fewer outages / ping
> >> losses (around 1/4 of what I had yesterday with dishy on the ground on
> >> my deck). Still good enough for a lengthy video Skype call with my folks
> >> in Germany, although they did comment about reduced video quality. But
> >> maybe that was the lighting or the different background as I wasn't in
> >> my usual spot with my laptop when I called them.
> >
> > Clear view of the sky is king for Starlink reliability. I've got my
> dishy mounted on the back fence, looking up over an empty field, so it's
> pretty much best-case scenario here.
> >>
> >>
> >> --
> >>
> >> ****************************************************************
> >> Dr. Ulrich Speidel
> >>
> >> School of Computer Science
> >>
> >> Room 303S.594 (City Campus)
> >>
> >> The University of Auckland
> >> u.speidel@auckland.ac.nz
> >> http://www.cs.auckland.ac.nz/~ulrich/
> >> ****************************************************************
> >>
> >>
> >>
> >> _______________________________________________
> >> Starlink mailing list
> >> Starlink@lists.bufferbloat.net
> >> https://lists.bufferbloat.net/listinfo/starlink
> >
> > _______________________________________________
> > Starlink mailing list
> > Starlink@lists.bufferbloat.net
> > https://lists.bufferbloat.net/listinfo/starlink
>
>
>
> --
> This song goes out to all the folk that thought Stadia would work:
>
> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
> Dave Täht CEO, TekLibre, LLC
>

[-- Attachment #2: Type: text/html, Size: 11664 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 17:41                                     ` Jonathan Bennett
@ 2023-01-13 18:09                                       ` Nathan Owens
  2023-01-13 20:30                                         ` Nathan Owens
  0 siblings, 1 reply; 44+ messages in thread
From: Nathan Owens @ 2023-01-13 18:09 UTC (permalink / raw)
  To: Jonathan Bennett; +Cc: Dave Taht, starlink

[-- Attachment #1: Type: text/plain, Size: 9488 bytes --]

I’ll run my visualization code on this result this afternoon and report
back!

On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
starlink@lists.bufferbloat.net> wrote:

> The irtt command, run with normal, light usage:
> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>
> Jonathan Bennett
> Hackaday.com
>
>
> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com> wrote:
>
>> packet caps would be nice... all this is very exciting news.
>>
>> I'd so love for one or more of y'all reporting such great uplink
>> results nowadays to duplicate and re-plot the original irtt tests we
>> did:
>>
>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
>> whatever.json
>>
>> They MUST have changed their scheduling to get such amazing uplink
>> results, in addition to better queue management.
>>
>> (for the record, my servers are de, london, fremont, sydney, dallas,
>> newark, atlanta, singapore, mumbai)
>>
>> There's an R and gnuplot script for plotting that output around here
>> somewhere (I have largely personally put down the starlink project,
>> loaning out mine) - that went by on this list... I should have written
>> a blog entry so I can find that stuff again.
>>
>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>> <starlink@lists.bufferbloat.net> wrote:
>> >
>> >
>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>> starlink@lists.bufferbloat.net> wrote:
>> >>
>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>> >> >
>> >> > From Auckland, New Zealand, using a roaming subscription, it puts me
>> >> > in touch with a server 2000 km away. OK then:
>> >> >
>> >> >
>> >> > IP address: nix six.
>> >> >
>> >> > My thoughts shall follow later.
>> >>
>> >> OK, so here we go.
>> >>
>> >> I'm always a bit skeptical when it comes to speed tests - they're
>> really
>> >> laden with so many caveats that it's not funny. I took our new work
>> >> Starlink kit home in December to give it a try and the other day
>> finally
>> >> got around to set it up. It's on a roaming subscription because our
>> >> badly built-up campus really isn't ideal in terms of a clear view of
>> the
>> >> sky. Oh - and did I mention that I used the Starlink Ethernet adapter,
>> >> not the WiFi?
>> >>
>> >> Caveat 1: Location, location. I live in a place where the best Starlink
>> >> promises is about 1/3 in terms of data rate you can actually get from
>> >> fibre to the home at under half of Starlink's price. Read: There are
>> few
>> >> Starlink users around. I might be the only one in my suburb.
>> >>
>> >> Caveat 2: Auckland has three Starlink gateways close by: Clevedon
>> (which
>> >> is at a stretch daytrip cycling distance from here), Te Hana and
>> Puwera,
>> >> the most distant of the three and about 130 km away from me as the crow
>> >> flies. Read: My dishy can use any satellite that any of these three can
>> >> see, and then depending on where I put it and how much of the southern
>> >> sky it can see, maybe also the one in Hinds, 840 km away, although that
>> >> is obviously stretching it a bit. Either way, that's plenty of options
>> >> for my bits to travel without needing a lot of handovers. Why? Easy: If
>> >> your nearest teleport is close by, then the set of satellites that the
>> >> teleport can see and the set that you can see is almost the same, so
>> you
>> >> can essentially stick with the same satellite while it's in view for
>> you
>> >> because it'll also be in view for the teleport. Pretty much any bird
>> >> above you will do.
>> >>
>> >> And because I don't get a lot of competition from other users in my
>> area
>> >> vying for one of the few available satellites that can see both us and
>> >> the teleport, this is about as good as it gets at 37S latitude. If I'd
>> >> want it any better, I'd have to move a lot further south.
>> >>
>> >> It'd be interesting to hear from Jonathan what the availability of home
>> >> broadband is like in the Dallas area. I note that it's at a lower
>> >> latitude (33N) than Auckland, but the difference isn't huge. I notice
>> >> two teleports each about 160 km away, which is also not too bad. I also
>> >> note Starlink availability in the area is restricted at the moment -
>> >> oversubscribed? But if Jonathan gets good data rates, then that means
>> >> that competition for bird capacity can't be too bad - for whatever
>> reason.
>> >
>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink gateway.
>> In cities, like Dallas, and Lawton where I live, there are good broadband
>> options. But there are also many people that live outside cities, and the
>> options are much worse. The low density userbase in rural Oklahoma and
>> Texas is probably ideal conditions for Starlink.
>> >>
>> >>
>> >> Caveat 3: Backhaul. There isn't just one queue between me and whatever
>> I
>> >> talk to in terms of my communications. Traceroute shows about 10 hops
>> >> between me and the University of Auckland via Starlink. That's 10
>> >> queues, not one. Many of them will have cross traffic. So it's a bit
>> >> hard to tell where our packets really get to wait or where they get
>> >> dropped. The insidious bit here is that a lot of them will be between 1
>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can all
>> >> turn into bottlenecks. This isn't like a narrowband GEO link of a few
>> >> Mb/s where it's obvious where the dominant long latency bottleneck in
>> >> your TCP connection's path is. Read: It's pretty hard to tell whether a
>> >> drop in "speed" is due to a performance issue in the Starlink system or
>> >> somewhere between Starlink's systems and the target system.
>> >>
>> >> I see RTTs here between 20 ms and 250 ms, where the physical latency
>> >> should be under 15 ms. So there's clearly a bit of buffer here along
>> the
>> >> chain that occasionally fills up.
>> >>
>> >> Caveat 4: Handovers. Handover between birds and teleports is inevitably
>> >> associated with a change in RTT and in most cases also available
>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>> >> satellite while your TCP is still trying to respond to whatever it
>> >> thought the queue on the previous bird was doing. Read: Whatever your
>> >> cwnd is immediately after a handover, it's probably not what it should
>> be.
>> >>
>> >> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
>> >> speedtest.net tests each to five local servers. Average upload rate
>> was
>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP that
>> >> Starlink seems to be buying its local connectivity from (Vocus Group)
>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and 52.22
>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
>> >> observed. In fact, they were the ISP with consistently the worst rates.
>> >>
>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and
>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly correlating
>> >> with rates. Average RTT was the same as for Vocus.
>> >>
>> >> Note the variation though: More or less a factor of two between highest
>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in my
>> >> tests? Or is there something systematic behind this? Way too few tests
>> >> to tell.
>> >>
>> >> What these tests do is establish a ballpark.
>> >>
>> >> I'm currently repeating tests with dish placed on a trestle closer to
>> >> the heavens. This seems to have translated into fewer outages / ping
>> >> losses (around 1/4 of what I had yesterday with dishy on the ground on
>> >> my deck). Still good enough for a lengthy video Skype call with my
>> folks
>> >> in Germany, although they did comment about reduced video quality. But
>> >> maybe that was the lighting or the different background as I wasn't in
>> >> my usual spot with my laptop when I called them.
>> >
>> > Clear view of the sky is king for Starlink reliability. I've got my
>> dishy mounted on the back fence, looking up over an empty field, so it's
>> pretty much best-case scenario here.
>> >>
>> >>
>> >> --
>> >>
>> >> ****************************************************************
>> >> Dr. Ulrich Speidel
>> >>
>> >> School of Computer Science
>> >>
>> >> Room 303S.594 (City Campus)
>> >>
>> >> The University of Auckland
>> >> u.speidel@auckland.ac.nz
>> >> http://www.cs.auckland.ac.nz/~ulrich/
>> >> ****************************************************************
>> >>
>> >>
>> >>
>> >> _______________________________________________
>> >> Starlink mailing list
>> >> Starlink@lists.bufferbloat.net
>> >> https://lists.bufferbloat.net/listinfo/starlink
>> >
>> > _______________________________________________
>> > Starlink mailing list
>> > Starlink@lists.bufferbloat.net
>> > https://lists.bufferbloat.net/listinfo/starlink
>>
>>
>>
>> --
>> This song goes out to all the folk that thought Stadia would work:
>>
>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>> Dave Täht CEO, TekLibre, LLC
>>
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink
>

[-- Attachment #2: Type: text/html, Size: 12576 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 17:26                                   ` Dave Taht
  2023-01-13 17:41                                     ` Jonathan Bennett
@ 2023-01-13 20:25                                     ` Pat Jensen
  2023-01-13 20:40                                       ` Dave Taht
  1 sibling, 1 reply; 44+ messages in thread
From: Pat Jensen @ 2023-01-13 20:25 UTC (permalink / raw)
  To: Dave Taht; +Cc: Jonathan Bennett, starlink

Dave,

 From Central California (using Starlink Los Angeles POP w/ live IPv6)
via fremont.starlink.taht.net at 12:20PM PST non-peak
Wired via 1000baseT with no other users on the network

                           Min      Mean    Median      Max   Stddev
                           ---      ----    ------      ---   ------
                 RTT    24.5ms   51.22ms   48.75ms  262.5ms  14.92ms
          send delay   76.44ms   94.79ms   92.69ms  288.7ms  12.06ms
       receive delay  -54.57ms  -43.58ms  -44.87ms   7.67ms   6.21ms

       IPDV (jitter)    92.5µs    4.39ms    2.98ms  80.71ms   4.26ms
           send IPDV    2.06µs    3.85ms    2.99ms  80.76ms   3.25ms
        receive IPDV        0s     1.5ms    36.5µs  49.72ms   3.18ms

      send call time    6.46µs    34.9µs             1.32ms   15.9µs
         timer error        0s    72.1µs             4.97ms   69.2µs
   server proc. time     620ns    4.19µs              647µs   6.34µs

                 duration: 5m1s (wait 787.6ms)
    packets sent/received: 99879/83786 (16.11% loss)
  server packets received: 83897/99879 (16.00%/0.13% loss up/down)
late (out-of-order) pkts: 1 (0.00%)
      bytes sent/received: 5992740/5027160
        send/receive rate: 159.8 Kbps / 134.1 Kbps
            packet length: 60 bytes
              timer stats: 121/100000 (0.12%) missed, 2.40% error

patj@air-2 ~ % curl ipinfo.io
{
   "ip": "98.97.140.145",
   "hostname": "customer.lsancax1.pop.starlinkisp.net",
   "city": "Los Angeles",
   "region": "California",
   "country": "US",
   "loc": "34.0522,-118.2437",
   "org": "AS14593 Space Exploration Technologies Corporation",
   "postal": "90009",
   "timezone": "America/Los_Angeles",
   "readme": "https://ipinfo.io/missingauth"
}%

Pat

On 2023-01-13 09:26, Dave Taht via Starlink wrote:
> packet caps would be nice... all this is very exciting news.
> 
> I'd so love for one or more of y'all reporting such great uplink
> results nowadays to duplicate and re-plot the original irtt tests we
> did:
> 
> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o 
> whatever.json
> 
> They MUST have changed their scheduling to get such amazing uplink
> results, in addition to better queue management.
> 
> (for the record, my servers are de, london, fremont, sydney, dallas,
> newark, atlanta, singapore, mumbai)
> 
> There's an R and gnuplot script for plotting that output around here
> somewhere (I have largely personally put down the starlink project,
> loaning out mine) - that went by on this list... I should have written
> a blog entry so I can find that stuff again.
> 
> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
> <starlink@lists.bufferbloat.net> wrote:
>> 
>> 
>> On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink 
>> <starlink@lists.bufferbloat.net> wrote:
>>> 
>>> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>> >
>>> > From Auckland, New Zealand, using a roaming subscription, it puts me
>>> > in touch with a server 2000 km away. OK then:
>>> >
>>> >
>>> > IP address: nix six.
>>> >
>>> > My thoughts shall follow later.
>>> 
>>> OK, so here we go.
>>> 
>>> I'm always a bit skeptical when it comes to speed tests - they're 
>>> really
>>> laden with so many caveats that it's not funny. I took our new work
>>> Starlink kit home in December to give it a try and the other day 
>>> finally
>>> got around to set it up. It's on a roaming subscription because our
>>> badly built-up campus really isn't ideal in terms of a clear view of 
>>> the
>>> sky. Oh - and did I mention that I used the Starlink Ethernet 
>>> adapter,
>>> not the WiFi?
>>> 
>>> Caveat 1: Location, location. I live in a place where the best 
>>> Starlink
>>> promises is about 1/3 in terms of data rate you can actually get from
>>> fibre to the home at under half of Starlink's price. Read: There are 
>>> few
>>> Starlink users around. I might be the only one in my suburb.
>>> 
>>> Caveat 2: Auckland has three Starlink gateways close by: Clevedon 
>>> (which
>>> is at a stretch daytrip cycling distance from here), Te Hana and 
>>> Puwera,
>>> the most distant of the three and about 130 km away from me as the 
>>> crow
>>> flies. Read: My dishy can use any satellite that any of these three 
>>> can
>>> see, and then depending on where I put it and how much of the 
>>> southern
>>> sky it can see, maybe also the one in Hinds, 840 km away, although 
>>> that
>>> is obviously stretching it a bit. Either way, that's plenty of 
>>> options
>>> for my bits to travel without needing a lot of handovers. Why? Easy: 
>>> If
>>> your nearest teleport is close by, then the set of satellites that 
>>> the
>>> teleport can see and the set that you can see is almost the same, so 
>>> you
>>> can essentially stick with the same satellite while it's in view for 
>>> you
>>> because it'll also be in view for the teleport. Pretty much any bird
>>> above you will do.
>>> 
>>> And because I don't get a lot of competition from other users in my 
>>> area
>>> vying for one of the few available satellites that can see both us 
>>> and
>>> the teleport, this is about as good as it gets at 37S latitude. If 
>>> I'd
>>> want it any better, I'd have to move a lot further south.
>>> 
>>> It'd be interesting to hear from Jonathan what the availability of 
>>> home
>>> broadband is like in the Dallas area. I note that it's at a lower
>>> latitude (33N) than Auckland, but the difference isn't huge. I notice
>>> two teleports each about 160 km away, which is also not too bad. I 
>>> also
>>> note Starlink availability in the area is restricted at the moment -
>>> oversubscribed? But if Jonathan gets good data rates, then that means
>>> that competition for bird capacity can't be too bad - for whatever 
>>> reason.
>> 
>> I'm in Southwest Oklahoma, but Dallas is the nearby Starlink gateway. 
>> In cities, like Dallas, and Lawton where I live, there are good 
>> broadband options. But there are also many people that live outside 
>> cities, and the options are much worse. The low density userbase in 
>> rural Oklahoma and Texas is probably ideal conditions for Starlink.
>>> 
>>> 
>>> Caveat 3: Backhaul. There isn't just one queue between me and 
>>> whatever I
>>> talk to in terms of my communications. Traceroute shows about 10 hops
>>> between me and the University of Auckland via Starlink. That's 10
>>> queues, not one. Many of them will have cross traffic. So it's a bit
>>> hard to tell where our packets really get to wait or where they get
>>> dropped. The insidious bit here is that a lot of them will be between 
>>> 1
>>> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can all
>>> turn into bottlenecks. This isn't like a narrowband GEO link of a few
>>> Mb/s where it's obvious where the dominant long latency bottleneck in
>>> your TCP connection's path is. Read: It's pretty hard to tell whether 
>>> a
>>> drop in "speed" is due to a performance issue in the Starlink system 
>>> or
>>> somewhere between Starlink's systems and the target system.
>>> 
>>> I see RTTs here between 20 ms and 250 ms, where the physical latency
>>> should be under 15 ms. So there's clearly a bit of buffer here along 
>>> the
>>> chain that occasionally fills up.
>>> 
>>> Caveat 4: Handovers. Handover between birds and teleports is 
>>> inevitably
>>> associated with a change in RTT and in most cases also available
>>> bandwidth. Plus your packets now arrive at a new queue on a new
>>> satellite while your TCP is still trying to respond to whatever it
>>> thought the queue on the previous bird was doing. Read: Whatever your
>>> cwnd is immediately after a handover, it's probably not what it 
>>> should be.
>>> 
>>> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
>>> speedtest.net tests each to five local servers. Average upload rate 
>>> was
>>> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP that
>>> Starlink seems to be buying its local connectivity from (Vocus Group)
>>> varied between 3.04 and 14.38 Mb/s, download between 23.33 and 52.22
>>> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
>>> observed. In fact, they were the ISP with consistently the worst 
>>> rates.
>>> 
>>> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and
>>> between 106.5 and 183.8 Mb/s down, again with RTTs badly correlating
>>> with rates. Average RTT was the same as for Vocus.
>>> 
>>> Note the variation though: More or less a factor of two between 
>>> highest
>>> and lowest rates for each ISP. Did MyRepublic just get lucky in my
>>> tests? Or is there something systematic behind this? Way too few 
>>> tests
>>> to tell.
>>> 
>>> What these tests do is establish a ballpark.
>>> 
>>> I'm currently repeating tests with dish placed on a trestle closer to
>>> the heavens. This seems to have translated into fewer outages / ping
>>> losses (around 1/4 of what I had yesterday with dishy on the ground 
>>> on
>>> my deck). Still good enough for a lengthy video Skype call with my 
>>> folks
>>> in Germany, although they did comment about reduced video quality. 
>>> But
>>> maybe that was the lighting or the different background as I wasn't 
>>> in
>>> my usual spot with my laptop when I called them.
>> 
>> Clear view of the sky is king for Starlink reliability. I've got my 
>> dishy mounted on the back fence, looking up over an empty field, so 
>> it's pretty much best-case scenario here.
>>> 
>>> 
>>> --
>>> 
>>> ****************************************************************
>>> Dr. Ulrich Speidel
>>> 
>>> School of Computer Science
>>> 
>>> Room 303S.594 (City Campus)
>>> 
>>> The University of Auckland
>>> u.speidel@auckland.ac.nz
>>> http://www.cs.auckland.ac.nz/~ulrich/
>>> ****************************************************************
>>> 
>>> 
>>> 
>>> _______________________________________________
>>> Starlink mailing list
>>> Starlink@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/starlink
>> 
>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 18:09                                       ` Nathan Owens
@ 2023-01-13 20:30                                         ` Nathan Owens
  2023-01-13 20:37                                           ` Dave Taht
  2023-01-13 20:49                                           ` Dave Taht
  0 siblings, 2 replies; 44+ messages in thread
From: Nathan Owens @ 2023-01-13 20:30 UTC (permalink / raw)
  To: Jonathan Bennett; +Cc: Dave Taht, starlink


[-- Attachment #1.1.1: Type: text/plain, Size: 10110 bytes --]

Here's the data visualization for Johnathan's Data

[image: Screenshot 2023-01-13 at 12.29.15 PM.png]

You can see the path change at :12, :27, :42, :57 after the minute. Some
paths are clearly busier than others with increased loss, latency, and
jitter.

On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io> wrote:

> I’ll run my visualization code on this result this afternoon and report
> back!
>
> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
> starlink@lists.bufferbloat.net> wrote:
>
>> The irtt command, run with normal, light usage:
>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>
>> Jonathan Bennett
>> Hackaday.com
>>
>>
>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com> wrote:
>>
>>> packet caps would be nice... all this is very exciting news.
>>>
>>> I'd so love for one or more of y'all reporting such great uplink
>>> results nowadays to duplicate and re-plot the original irtt tests we
>>> did:
>>>
>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
>>> whatever.json
>>>
>>> They MUST have changed their scheduling to get such amazing uplink
>>> results, in addition to better queue management.
>>>
>>> (for the record, my servers are de, london, fremont, sydney, dallas,
>>> newark, atlanta, singapore, mumbai)
>>>
>>> There's an R and gnuplot script for plotting that output around here
>>> somewhere (I have largely personally put down the starlink project,
>>> loaning out mine) - that went by on this list... I should have written
>>> a blog entry so I can find that stuff again.
>>>
>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>> <starlink@lists.bufferbloat.net> wrote:
>>> >
>>> >
>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>> starlink@lists.bufferbloat.net> wrote:
>>> >>
>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>> >> >
>>> >> > From Auckland, New Zealand, using a roaming subscription, it puts me
>>> >> > in touch with a server 2000 km away. OK then:
>>> >> >
>>> >> >
>>> >> > IP address: nix six.
>>> >> >
>>> >> > My thoughts shall follow later.
>>> >>
>>> >> OK, so here we go.
>>> >>
>>> >> I'm always a bit skeptical when it comes to speed tests - they're
>>> really
>>> >> laden with so many caveats that it's not funny. I took our new work
>>> >> Starlink kit home in December to give it a try and the other day
>>> finally
>>> >> got around to set it up. It's on a roaming subscription because our
>>> >> badly built-up campus really isn't ideal in terms of a clear view of
>>> the
>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet adapter,
>>> >> not the WiFi?
>>> >>
>>> >> Caveat 1: Location, location. I live in a place where the best
>>> Starlink
>>> >> promises is about 1/3 in terms of data rate you can actually get from
>>> >> fibre to the home at under half of Starlink's price. Read: There are
>>> few
>>> >> Starlink users around. I might be the only one in my suburb.
>>> >>
>>> >> Caveat 2: Auckland has three Starlink gateways close by: Clevedon
>>> (which
>>> >> is at a stretch daytrip cycling distance from here), Te Hana and
>>> Puwera,
>>> >> the most distant of the three and about 130 km away from me as the
>>> crow
>>> >> flies. Read: My dishy can use any satellite that any of these three
>>> can
>>> >> see, and then depending on where I put it and how much of the southern
>>> >> sky it can see, maybe also the one in Hinds, 840 km away, although
>>> that
>>> >> is obviously stretching it a bit. Either way, that's plenty of options
>>> >> for my bits to travel without needing a lot of handovers. Why? Easy:
>>> If
>>> >> your nearest teleport is close by, then the set of satellites that the
>>> >> teleport can see and the set that you can see is almost the same, so
>>> you
>>> >> can essentially stick with the same satellite while it's in view for
>>> you
>>> >> because it'll also be in view for the teleport. Pretty much any bird
>>> >> above you will do.
>>> >>
>>> >> And because I don't get a lot of competition from other users in my
>>> area
>>> >> vying for one of the few available satellites that can see both us and
>>> >> the teleport, this is about as good as it gets at 37S latitude. If I'd
>>> >> want it any better, I'd have to move a lot further south.
>>> >>
>>> >> It'd be interesting to hear from Jonathan what the availability of
>>> home
>>> >> broadband is like in the Dallas area. I note that it's at a lower
>>> >> latitude (33N) than Auckland, but the difference isn't huge. I notice
>>> >> two teleports each about 160 km away, which is also not too bad. I
>>> also
>>> >> note Starlink availability in the area is restricted at the moment -
>>> >> oversubscribed? But if Jonathan gets good data rates, then that means
>>> >> that competition for bird capacity can't be too bad - for whatever
>>> reason.
>>> >
>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink gateway.
>>> In cities, like Dallas, and Lawton where I live, there are good broadband
>>> options. But there are also many people that live outside cities, and the
>>> options are much worse. The low density userbase in rural Oklahoma and
>>> Texas is probably ideal conditions for Starlink.
>>> >>
>>> >>
>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>> whatever I
>>> >> talk to in terms of my communications. Traceroute shows about 10 hops
>>> >> between me and the University of Auckland via Starlink. That's 10
>>> >> queues, not one. Many of them will have cross traffic. So it's a bit
>>> >> hard to tell where our packets really get to wait or where they get
>>> >> dropped. The insidious bit here is that a lot of them will be between
>>> 1
>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can all
>>> >> turn into bottlenecks. This isn't like a narrowband GEO link of a few
>>> >> Mb/s where it's obvious where the dominant long latency bottleneck in
>>> >> your TCP connection's path is. Read: It's pretty hard to tell whether
>>> a
>>> >> drop in "speed" is due to a performance issue in the Starlink system
>>> or
>>> >> somewhere between Starlink's systems and the target system.
>>> >>
>>> >> I see RTTs here between 20 ms and 250 ms, where the physical latency
>>> >> should be under 15 ms. So there's clearly a bit of buffer here along
>>> the
>>> >> chain that occasionally fills up.
>>> >>
>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>> inevitably
>>> >> associated with a change in RTT and in most cases also available
>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>> >> satellite while your TCP is still trying to respond to whatever it
>>> >> thought the queue on the previous bird was doing. Read: Whatever your
>>> >> cwnd is immediately after a handover, it's probably not what it
>>> should be.
>>> >>
>>> >> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
>>> >> speedtest.net tests each to five local servers. Average upload rate
>>> was
>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP that
>>> >> Starlink seems to be buying its local connectivity from (Vocus Group)
>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and 52.22
>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
>>> >> observed. In fact, they were the ISP with consistently the worst
>>> rates.
>>> >>
>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and
>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly correlating
>>> >> with rates. Average RTT was the same as for Vocus.
>>> >>
>>> >> Note the variation though: More or less a factor of two between
>>> highest
>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in my
>>> >> tests? Or is there something systematic behind this? Way too few tests
>>> >> to tell.
>>> >>
>>> >> What these tests do is establish a ballpark.
>>> >>
>>> >> I'm currently repeating tests with dish placed on a trestle closer to
>>> >> the heavens. This seems to have translated into fewer outages / ping
>>> >> losses (around 1/4 of what I had yesterday with dishy on the ground on
>>> >> my deck). Still good enough for a lengthy video Skype call with my
>>> folks
>>> >> in Germany, although they did comment about reduced video quality. But
>>> >> maybe that was the lighting or the different background as I wasn't in
>>> >> my usual spot with my laptop when I called them.
>>> >
>>> > Clear view of the sky is king for Starlink reliability. I've got my
>>> dishy mounted on the back fence, looking up over an empty field, so it's
>>> pretty much best-case scenario here.
>>> >>
>>> >>
>>> >> --
>>> >>
>>> >> ****************************************************************
>>> >> Dr. Ulrich Speidel
>>> >>
>>> >> School of Computer Science
>>> >>
>>> >> Room 303S.594 (City Campus)
>>> >>
>>> >> The University of Auckland
>>> >> u.speidel@auckland.ac.nz
>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>> >> ****************************************************************
>>> >>
>>> >>
>>> >>
>>> >> _______________________________________________
>>> >> Starlink mailing list
>>> >> Starlink@lists.bufferbloat.net
>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>> >
>>> > _______________________________________________
>>> > Starlink mailing list
>>> > Starlink@lists.bufferbloat.net
>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>
>>>
>>>
>>> --
>>> This song goes out to all the folk that thought Stadia would work:
>>>
>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>> Dave Täht CEO, TekLibre, LLC
>>>
>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink
>>
>

[-- Attachment #1.1.2: Type: text/html, Size: 13177 bytes --]

[-- Attachment #1.2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #2: jp_starlink_irtt.pdf --]
[-- Type: application/pdf, Size: 697557 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 20:30                                         ` Nathan Owens
@ 2023-01-13 20:37                                           ` Dave Taht
  2023-01-13 21:24                                             ` Nathan Owens
  2023-01-13 20:49                                           ` Dave Taht
  1 sibling, 1 reply; 44+ messages in thread
From: Dave Taht @ 2023-01-13 20:37 UTC (permalink / raw)
  To: Nathan Owens; +Cc: Jonathan Bennett, starlink


[-- Attachment #1.1: Type: text/plain, Size: 11256 bytes --]

That is amazingly better than what I'd seen before.

(can you repost your script? What's the difference between the red and blue
dots?)

Jonathan (or all? ulrich?), could you fire up the irtt for that long
interval, and fire off a waveform test, pause, then a cloudflare test,
then, if you have it?

flent -H oneofmyclosestservers.starlink.taht.net -t starlink_vs_irtt
--step-size=.05 --socket-stats --test-parameter=upload_streams=4 tcp_nup

Going the full monty and enabling ecn and taking a packet cap of this last
would additionally make my day...

On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io> wrote:

> Here's the data visualization for Johnathan's Data
>
> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>
> You can see the path change at :12, :27, :42, :57 after the minute. Some
> paths are clearly busier than others with increased loss, latency, and
> jitter.
>
> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io> wrote:
>
>> I’ll run my visualization code on this result this afternoon and report
>> back!
>>
>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>> starlink@lists.bufferbloat.net> wrote:
>>
>>> The irtt command, run with normal, light usage:
>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>
>>> Jonathan Bennett
>>> Hackaday.com
>>>
>>>
>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com> wrote:
>>>
>>>> packet caps would be nice... all this is very exciting news.
>>>>
>>>> I'd so love for one or more of y'all reporting such great uplink
>>>> results nowadays to duplicate and re-plot the original irtt tests we
>>>> did:
>>>>
>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
>>>> whatever.json
>>>>
>>>> They MUST have changed their scheduling to get such amazing uplink
>>>> results, in addition to better queue management.
>>>>
>>>> (for the record, my servers are de, london, fremont, sydney, dallas,
>>>> newark, atlanta, singapore, mumbai)
>>>>
>>>> There's an R and gnuplot script for plotting that output around here
>>>> somewhere (I have largely personally put down the starlink project,
>>>> loaning out mine) - that went by on this list... I should have written
>>>> a blog entry so I can find that stuff again.
>>>>
>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>> <starlink@lists.bufferbloat.net> wrote:
>>>> >
>>>> >
>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>> starlink@lists.bufferbloat.net> wrote:
>>>> >>
>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>> >> >
>>>> >> > From Auckland, New Zealand, using a roaming subscription, it puts
>>>> me
>>>> >> > in touch with a server 2000 km away. OK then:
>>>> >> >
>>>> >> >
>>>> >> > IP address: nix six.
>>>> >> >
>>>> >> > My thoughts shall follow later.
>>>> >>
>>>> >> OK, so here we go.
>>>> >>
>>>> >> I'm always a bit skeptical when it comes to speed tests - they're
>>>> really
>>>> >> laden with so many caveats that it's not funny. I took our new work
>>>> >> Starlink kit home in December to give it a try and the other day
>>>> finally
>>>> >> got around to set it up. It's on a roaming subscription because our
>>>> >> badly built-up campus really isn't ideal in terms of a clear view of
>>>> the
>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>> adapter,
>>>> >> not the WiFi?
>>>> >>
>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>> Starlink
>>>> >> promises is about 1/3 in terms of data rate you can actually get from
>>>> >> fibre to the home at under half of Starlink's price. Read: There are
>>>> few
>>>> >> Starlink users around. I might be the only one in my suburb.
>>>> >>
>>>> >> Caveat 2: Auckland has three Starlink gateways close by: Clevedon
>>>> (which
>>>> >> is at a stretch daytrip cycling distance from here), Te Hana and
>>>> Puwera,
>>>> >> the most distant of the three and about 130 km away from me as the
>>>> crow
>>>> >> flies. Read: My dishy can use any satellite that any of these three
>>>> can
>>>> >> see, and then depending on where I put it and how much of the
>>>> southern
>>>> >> sky it can see, maybe also the one in Hinds, 840 km away, although
>>>> that
>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>> options
>>>> >> for my bits to travel without needing a lot of handovers. Why? Easy:
>>>> If
>>>> >> your nearest teleport is close by, then the set of satellites that
>>>> the
>>>> >> teleport can see and the set that you can see is almost the same, so
>>>> you
>>>> >> can essentially stick with the same satellite while it's in view for
>>>> you
>>>> >> because it'll also be in view for the teleport. Pretty much any bird
>>>> >> above you will do.
>>>> >>
>>>> >> And because I don't get a lot of competition from other users in my
>>>> area
>>>> >> vying for one of the few available satellites that can see both us
>>>> and
>>>> >> the teleport, this is about as good as it gets at 37S latitude. If
>>>> I'd
>>>> >> want it any better, I'd have to move a lot further south.
>>>> >>
>>>> >> It'd be interesting to hear from Jonathan what the availability of
>>>> home
>>>> >> broadband is like in the Dallas area. I note that it's at a lower
>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I notice
>>>> >> two teleports each about 160 km away, which is also not too bad. I
>>>> also
>>>> >> note Starlink availability in the area is restricted at the moment -
>>>> >> oversubscribed? But if Jonathan gets good data rates, then that means
>>>> >> that competition for bird capacity can't be too bad - for whatever
>>>> reason.
>>>> >
>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink gateway.
>>>> In cities, like Dallas, and Lawton where I live, there are good broadband
>>>> options. But there are also many people that live outside cities, and the
>>>> options are much worse. The low density userbase in rural Oklahoma and
>>>> Texas is probably ideal conditions for Starlink.
>>>> >>
>>>> >>
>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>> whatever I
>>>> >> talk to in terms of my communications. Traceroute shows about 10 hops
>>>> >> between me and the University of Auckland via Starlink. That's 10
>>>> >> queues, not one. Many of them will have cross traffic. So it's a bit
>>>> >> hard to tell where our packets really get to wait or where they get
>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>> between 1
>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can all
>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link of a few
>>>> >> Mb/s where it's obvious where the dominant long latency bottleneck in
>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>> whether a
>>>> >> drop in "speed" is due to a performance issue in the Starlink system
>>>> or
>>>> >> somewhere between Starlink's systems and the target system.
>>>> >>
>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical latency
>>>> >> should be under 15 ms. So there's clearly a bit of buffer here along
>>>> the
>>>> >> chain that occasionally fills up.
>>>> >>
>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>> inevitably
>>>> >> associated with a change in RTT and in most cases also available
>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>> >> satellite while your TCP is still trying to respond to whatever it
>>>> >> thought the queue on the previous bird was doing. Read: Whatever your
>>>> >> cwnd is immediately after a handover, it's probably not what it
>>>> should be.
>>>> >>
>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
>>>> >> speedtest.net tests each to five local servers. Average upload rate
>>>> was
>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP that
>>>> >> Starlink seems to be buying its local connectivity from (Vocus Group)
>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and 52.22
>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
>>>> >> observed. In fact, they were the ISP with consistently the worst
>>>> rates.
>>>> >>
>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and
>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly correlating
>>>> >> with rates. Average RTT was the same as for Vocus.
>>>> >>
>>>> >> Note the variation though: More or less a factor of two between
>>>> highest
>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in my
>>>> >> tests? Or is there something systematic behind this? Way too few
>>>> tests
>>>> >> to tell.
>>>> >>
>>>> >> What these tests do is establish a ballpark.
>>>> >>
>>>> >> I'm currently repeating tests with dish placed on a trestle closer to
>>>> >> the heavens. This seems to have translated into fewer outages / ping
>>>> >> losses (around 1/4 of what I had yesterday with dishy on the ground
>>>> on
>>>> >> my deck). Still good enough for a lengthy video Skype call with my
>>>> folks
>>>> >> in Germany, although they did comment about reduced video quality.
>>>> But
>>>> >> maybe that was the lighting or the different background as I wasn't
>>>> in
>>>> >> my usual spot with my laptop when I called them.
>>>> >
>>>> > Clear view of the sky is king for Starlink reliability. I've got my
>>>> dishy mounted on the back fence, looking up over an empty field, so it's
>>>> pretty much best-case scenario here.
>>>> >>
>>>> >>
>>>> >> --
>>>> >>
>>>> >> ****************************************************************
>>>> >> Dr. Ulrich Speidel
>>>> >>
>>>> >> School of Computer Science
>>>> >>
>>>> >> Room 303S.594 (City Campus)
>>>> >>
>>>> >> The University of Auckland
>>>> >> u.speidel@auckland.ac.nz
>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>> >> ****************************************************************
>>>> >>
>>>> >>
>>>> >>
>>>> >> _______________________________________________
>>>> >> Starlink mailing list
>>>> >> Starlink@lists.bufferbloat.net
>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>> >
>>>> > _______________________________________________
>>>> > Starlink mailing list
>>>> > Starlink@lists.bufferbloat.net
>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>
>>>>
>>>>
>>>> --
>>>> This song goes out to all the folk that thought Stadia would work:
>>>>
>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>> Dave Täht CEO, TekLibre, LLC
>>>>
>>> _______________________________________________
>>> Starlink mailing list
>>> Starlink@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/starlink
>>>
>>

-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

[-- Attachment #1.2: Type: text/html, Size: 14708 bytes --]

[-- Attachment #2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 20:25                                     ` Pat Jensen
@ 2023-01-13 20:40                                       ` Dave Taht
  0 siblings, 0 replies; 44+ messages in thread
From: Dave Taht @ 2023-01-13 20:40 UTC (permalink / raw)
  To: Pat Jensen; +Cc: Jonathan Bennett, starlink

On Fri, Jan 13, 2023 at 12:26 PM Pat Jensen <patj@jensencloud.net> wrote:
>
> Dave,
>
>  From Central California (using Starlink Los Angeles POP w/ live IPv6)
> via fremont.starlink.taht.net at 12:20PM PST non-peak
> Wired via 1000baseT with no other users on the network
>
>                            Min      Mean    Median      Max   Stddev
>                            ---      ----    ------      ---   ------
>                  RTT    24.5ms   51.22ms   48.75ms  262.5ms  14.92ms
>           send delay   76.44ms   94.79ms   92.69ms  288.7ms  12.06ms
>        receive delay  -54.57ms  -43.58ms  -44.87ms   7.67ms   6.21ms
>
>        IPDV (jitter)    92.5µs    4.39ms    2.98ms  80.71ms   4.26ms
>            send IPDV    2.06µs    3.85ms    2.99ms  80.76ms   3.25ms
>         receive IPDV        0s     1.5ms    36.5µs  49.72ms   3.18ms
>
>       send call time    6.46µs    34.9µs             1.32ms   15.9µs
>          timer error        0s    72.1µs             4.97ms   69.2µs
>    server proc. time     620ns    4.19µs              647µs   6.34µs

Thank you pat. In general, I am almost never interested in summary
statistics like these, but in plotting the long term behaviors,
as nathan just did, starting with idle baseline, and then adding
various loads. Could you slide him the full output of this test, since
he's got the magic plotting script (my json-fu is non-existent)

And then, same test, against the waveform, cloudflare, and flent loads?

>                  duration: 5m1s (wait 787.6ms)
>     packets sent/received: 99879/83786 (16.11% loss)
>   server packets received: 83897/99879 (16.00%/0.13% loss up/down)
> late (out-of-order) pkts: 1 (0.00%)
>       bytes sent/received: 5992740/5027160
>         send/receive rate: 159.8 Kbps / 134.1 Kbps
>             packet length: 60 bytes
>               timer stats: 121/100000 (0.12%) missed, 2.40% error
>
> patj@air-2 ~ % curl ipinfo.io
> {
>    "ip": "98.97.140.145",
>    "hostname": "customer.lsancax1.pop.starlinkisp.net",
>    "city": "Los Angeles",
>    "region": "California",
>    "country": "US",
>    "loc": "34.0522,-118.2437",
>    "org": "AS14593 Space Exploration Technologies Corporation",
>    "postal": "90009",
>    "timezone": "America/Los_Angeles",
>    "readme": "https://ipinfo.io/missingauth"
> }%
>
> Pat
>
> On 2023-01-13 09:26, Dave Taht via Starlink wrote:
> > packet caps would be nice... all this is very exciting news.
> >
> > I'd so love for one or more of y'all reporting such great uplink
> > results nowadays to duplicate and re-plot the original irtt tests we
> > did:
> >
> > irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
> > whatever.json
> >
> > They MUST have changed their scheduling to get such amazing uplink
> > results, in addition to better queue management.
> >
> > (for the record, my servers are de, london, fremont, sydney, dallas,
> > newark, atlanta, singapore, mumbai)
> >
> > There's an R and gnuplot script for plotting that output around here
> > somewhere (I have largely personally put down the starlink project,
> > loaning out mine) - that went by on this list... I should have written
> > a blog entry so I can find that stuff again.
> >
> > On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
> > <starlink@lists.bufferbloat.net> wrote:
> >>
> >>
> >> On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink
> >> <starlink@lists.bufferbloat.net> wrote:
> >>>
> >>> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
> >>> >
> >>> > From Auckland, New Zealand, using a roaming subscription, it puts me
> >>> > in touch with a server 2000 km away. OK then:
> >>> >
> >>> >
> >>> > IP address: nix six.
> >>> >
> >>> > My thoughts shall follow later.
> >>>
> >>> OK, so here we go.
> >>>
> >>> I'm always a bit skeptical when it comes to speed tests - they're
> >>> really
> >>> laden with so many caveats that it's not funny. I took our new work
> >>> Starlink kit home in December to give it a try and the other day
> >>> finally
> >>> got around to set it up. It's on a roaming subscription because our
> >>> badly built-up campus really isn't ideal in terms of a clear view of
> >>> the
> >>> sky. Oh - and did I mention that I used the Starlink Ethernet
> >>> adapter,
> >>> not the WiFi?
> >>>
> >>> Caveat 1: Location, location. I live in a place where the best
> >>> Starlink
> >>> promises is about 1/3 in terms of data rate you can actually get from
> >>> fibre to the home at under half of Starlink's price. Read: There are
> >>> few
> >>> Starlink users around. I might be the only one in my suburb.
> >>>
> >>> Caveat 2: Auckland has three Starlink gateways close by: Clevedon
> >>> (which
> >>> is at a stretch daytrip cycling distance from here), Te Hana and
> >>> Puwera,
> >>> the most distant of the three and about 130 km away from me as the
> >>> crow
> >>> flies. Read: My dishy can use any satellite that any of these three
> >>> can
> >>> see, and then depending on where I put it and how much of the
> >>> southern
> >>> sky it can see, maybe also the one in Hinds, 840 km away, although
> >>> that
> >>> is obviously stretching it a bit. Either way, that's plenty of
> >>> options
> >>> for my bits to travel without needing a lot of handovers. Why? Easy:
> >>> If
> >>> your nearest teleport is close by, then the set of satellites that
> >>> the
> >>> teleport can see and the set that you can see is almost the same, so
> >>> you
> >>> can essentially stick with the same satellite while it's in view for
> >>> you
> >>> because it'll also be in view for the teleport. Pretty much any bird
> >>> above you will do.
> >>>
> >>> And because I don't get a lot of competition from other users in my
> >>> area
> >>> vying for one of the few available satellites that can see both us
> >>> and
> >>> the teleport, this is about as good as it gets at 37S latitude. If
> >>> I'd
> >>> want it any better, I'd have to move a lot further south.
> >>>
> >>> It'd be interesting to hear from Jonathan what the availability of
> >>> home
> >>> broadband is like in the Dallas area. I note that it's at a lower
> >>> latitude (33N) than Auckland, but the difference isn't huge. I notice
> >>> two teleports each about 160 km away, which is also not too bad. I
> >>> also
> >>> note Starlink availability in the area is restricted at the moment -
> >>> oversubscribed? But if Jonathan gets good data rates, then that means
> >>> that competition for bird capacity can't be too bad - for whatever
> >>> reason.
> >>
> >> I'm in Southwest Oklahoma, but Dallas is the nearby Starlink gateway.
> >> In cities, like Dallas, and Lawton where I live, there are good
> >> broadband options. But there are also many people that live outside
> >> cities, and the options are much worse. The low density userbase in
> >> rural Oklahoma and Texas is probably ideal conditions for Starlink.
> >>>
> >>>
> >>> Caveat 3: Backhaul. There isn't just one queue between me and
> >>> whatever I
> >>> talk to in terms of my communications. Traceroute shows about 10 hops
> >>> between me and the University of Auckland via Starlink. That's 10
> >>> queues, not one. Many of them will have cross traffic. So it's a bit
> >>> hard to tell where our packets really get to wait or where they get
> >>> dropped. The insidious bit here is that a lot of them will be between
> >>> 1
> >>> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can all
> >>> turn into bottlenecks. This isn't like a narrowband GEO link of a few
> >>> Mb/s where it's obvious where the dominant long latency bottleneck in
> >>> your TCP connection's path is. Read: It's pretty hard to tell whether
> >>> a
> >>> drop in "speed" is due to a performance issue in the Starlink system
> >>> or
> >>> somewhere between Starlink's systems and the target system.
> >>>
> >>> I see RTTs here between 20 ms and 250 ms, where the physical latency
> >>> should be under 15 ms. So there's clearly a bit of buffer here along
> >>> the
> >>> chain that occasionally fills up.
> >>>
> >>> Caveat 4: Handovers. Handover between birds and teleports is
> >>> inevitably
> >>> associated with a change in RTT and in most cases also available
> >>> bandwidth. Plus your packets now arrive at a new queue on a new
> >>> satellite while your TCP is still trying to respond to whatever it
> >>> thought the queue on the previous bird was doing. Read: Whatever your
> >>> cwnd is immediately after a handover, it's probably not what it
> >>> should be.
> >>>
> >>> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
> >>> speedtest.net tests each to five local servers. Average upload rate
> >>> was
> >>> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP that
> >>> Starlink seems to be buying its local connectivity from (Vocus Group)
> >>> varied between 3.04 and 14.38 Mb/s, download between 23.33 and 52.22
> >>> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
> >>> observed. In fact, they were the ISP with consistently the worst
> >>> rates.
> >>>
> >>> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and
> >>> between 106.5 and 183.8 Mb/s down, again with RTTs badly correlating
> >>> with rates. Average RTT was the same as for Vocus.
> >>>
> >>> Note the variation though: More or less a factor of two between
> >>> highest
> >>> and lowest rates for each ISP. Did MyRepublic just get lucky in my
> >>> tests? Or is there something systematic behind this? Way too few
> >>> tests
> >>> to tell.
> >>>
> >>> What these tests do is establish a ballpark.
> >>>
> >>> I'm currently repeating tests with dish placed on a trestle closer to
> >>> the heavens. This seems to have translated into fewer outages / ping
> >>> losses (around 1/4 of what I had yesterday with dishy on the ground
> >>> on
> >>> my deck). Still good enough for a lengthy video Skype call with my
> >>> folks
> >>> in Germany, although they did comment about reduced video quality.
> >>> But
> >>> maybe that was the lighting or the different background as I wasn't
> >>> in
> >>> my usual spot with my laptop when I called them.
> >>
> >> Clear view of the sky is king for Starlink reliability. I've got my
> >> dishy mounted on the back fence, looking up over an empty field, so
> >> it's pretty much best-case scenario here.
> >>>
> >>>
> >>> --
> >>>
> >>> ****************************************************************
> >>> Dr. Ulrich Speidel
> >>>
> >>> School of Computer Science
> >>>
> >>> Room 303S.594 (City Campus)
> >>>
> >>> The University of Auckland
> >>> u.speidel@auckland.ac.nz
> >>> http://www.cs.auckland.ac.nz/~ulrich/
> >>> ****************************************************************
> >>>
> >>>
> >>>
> >>> _______________________________________________
> >>> Starlink mailing list
> >>> Starlink@lists.bufferbloat.net
> >>> https://lists.bufferbloat.net/listinfo/starlink
> >>
> >> _______________________________________________
> >> Starlink mailing list
> >> Starlink@lists.bufferbloat.net
> >> https://lists.bufferbloat.net/listinfo/starlink



-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 20:30                                         ` Nathan Owens
  2023-01-13 20:37                                           ` Dave Taht
@ 2023-01-13 20:49                                           ` Dave Taht
  2023-01-13 21:25                                             ` Luis A. Cornejo
  1 sibling, 1 reply; 44+ messages in thread
From: Dave Taht @ 2023-01-13 20:49 UTC (permalink / raw)
  To: Nathan Owens; +Cc: Jonathan Bennett, starlink


[-- Attachment #1.1: Type: text/plain, Size: 11254 bytes --]

On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io> wrote:

> Here's the data visualization for Johnathan's Data
>
> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>
> You can see the path change at :12, :27, :42, :57 after the minute. Some
> paths are clearly busier than others with increased loss, latency, and
> jitter.
>

I am so glad to see loss and bounded delay here. Also a bit of rigor
regarding what traffic was active locally vs on the path would be nice,
although it seems to line up with the known 15s starlink switchover thing
(need a name for this), in this case, doing a few speedtests
while that irtt is running would show the impact(s) of whatever else they
are up to.

It would also be my hope that the loss distribution in the middle portion
of this data is good, not bursty, but we don't have a tool to take apart
that. (I am so hopeless at json)



>
>
> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io> wrote:
>
>> I’ll run my visualization code on this result this afternoon and report
>> back!
>>
>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>> starlink@lists.bufferbloat.net> wrote:
>>
>>> The irtt command, run with normal, light usage:
>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>
>>> Jonathan Bennett
>>> Hackaday.com
>>>
>>>
>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com> wrote:
>>>
>>>> packet caps would be nice... all this is very exciting news.
>>>>
>>>> I'd so love for one or more of y'all reporting such great uplink
>>>> results nowadays to duplicate and re-plot the original irtt tests we
>>>> did:
>>>>
>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
>>>> whatever.json
>>>>
>>>> They MUST have changed their scheduling to get such amazing uplink
>>>> results, in addition to better queue management.
>>>>
>>>> (for the record, my servers are de, london, fremont, sydney, dallas,
>>>> newark, atlanta, singapore, mumbai)
>>>>
>>>> There's an R and gnuplot script for plotting that output around here
>>>> somewhere (I have largely personally put down the starlink project,
>>>> loaning out mine) - that went by on this list... I should have written
>>>> a blog entry so I can find that stuff again.
>>>>
>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>> <starlink@lists.bufferbloat.net> wrote:
>>>> >
>>>> >
>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>> starlink@lists.bufferbloat.net> wrote:
>>>> >>
>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>> >> >
>>>> >> > From Auckland, New Zealand, using a roaming subscription, it puts
>>>> me
>>>> >> > in touch with a server 2000 km away. OK then:
>>>> >> >
>>>> >> >
>>>> >> > IP address: nix six.
>>>> >> >
>>>> >> > My thoughts shall follow later.
>>>> >>
>>>> >> OK, so here we go.
>>>> >>
>>>> >> I'm always a bit skeptical when it comes to speed tests - they're
>>>> really
>>>> >> laden with so many caveats that it's not funny. I took our new work
>>>> >> Starlink kit home in December to give it a try and the other day
>>>> finally
>>>> >> got around to set it up. It's on a roaming subscription because our
>>>> >> badly built-up campus really isn't ideal in terms of a clear view of
>>>> the
>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>> adapter,
>>>> >> not the WiFi?
>>>> >>
>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>> Starlink
>>>> >> promises is about 1/3 in terms of data rate you can actually get from
>>>> >> fibre to the home at under half of Starlink's price. Read: There are
>>>> few
>>>> >> Starlink users around. I might be the only one in my suburb.
>>>> >>
>>>> >> Caveat 2: Auckland has three Starlink gateways close by: Clevedon
>>>> (which
>>>> >> is at a stretch daytrip cycling distance from here), Te Hana and
>>>> Puwera,
>>>> >> the most distant of the three and about 130 km away from me as the
>>>> crow
>>>> >> flies. Read: My dishy can use any satellite that any of these three
>>>> can
>>>> >> see, and then depending on where I put it and how much of the
>>>> southern
>>>> >> sky it can see, maybe also the one in Hinds, 840 km away, although
>>>> that
>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>> options
>>>> >> for my bits to travel without needing a lot of handovers. Why? Easy:
>>>> If
>>>> >> your nearest teleport is close by, then the set of satellites that
>>>> the
>>>> >> teleport can see and the set that you can see is almost the same, so
>>>> you
>>>> >> can essentially stick with the same satellite while it's in view for
>>>> you
>>>> >> because it'll also be in view for the teleport. Pretty much any bird
>>>> >> above you will do.
>>>> >>
>>>> >> And because I don't get a lot of competition from other users in my
>>>> area
>>>> >> vying for one of the few available satellites that can see both us
>>>> and
>>>> >> the teleport, this is about as good as it gets at 37S latitude. If
>>>> I'd
>>>> >> want it any better, I'd have to move a lot further south.
>>>> >>
>>>> >> It'd be interesting to hear from Jonathan what the availability of
>>>> home
>>>> >> broadband is like in the Dallas area. I note that it's at a lower
>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I notice
>>>> >> two teleports each about 160 km away, which is also not too bad. I
>>>> also
>>>> >> note Starlink availability in the area is restricted at the moment -
>>>> >> oversubscribed? But if Jonathan gets good data rates, then that means
>>>> >> that competition for bird capacity can't be too bad - for whatever
>>>> reason.
>>>> >
>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink gateway.
>>>> In cities, like Dallas, and Lawton where I live, there are good broadband
>>>> options. But there are also many people that live outside cities, and the
>>>> options are much worse. The low density userbase in rural Oklahoma and
>>>> Texas is probably ideal conditions for Starlink.
>>>> >>
>>>> >>
>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>> whatever I
>>>> >> talk to in terms of my communications. Traceroute shows about 10 hops
>>>> >> between me and the University of Auckland via Starlink. That's 10
>>>> >> queues, not one. Many of them will have cross traffic. So it's a bit
>>>> >> hard to tell where our packets really get to wait or where they get
>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>> between 1
>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can all
>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link of a few
>>>> >> Mb/s where it's obvious where the dominant long latency bottleneck in
>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>> whether a
>>>> >> drop in "speed" is due to a performance issue in the Starlink system
>>>> or
>>>> >> somewhere between Starlink's systems and the target system.
>>>> >>
>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical latency
>>>> >> should be under 15 ms. So there's clearly a bit of buffer here along
>>>> the
>>>> >> chain that occasionally fills up.
>>>> >>
>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>> inevitably
>>>> >> associated with a change in RTT and in most cases also available
>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>> >> satellite while your TCP is still trying to respond to whatever it
>>>> >> thought the queue on the previous bird was doing. Read: Whatever your
>>>> >> cwnd is immediately after a handover, it's probably not what it
>>>> should be.
>>>> >>
>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
>>>> >> speedtest.net tests each to five local servers. Average upload rate
>>>> was
>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP that
>>>> >> Starlink seems to be buying its local connectivity from (Vocus Group)
>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and 52.22
>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
>>>> >> observed. In fact, they were the ISP with consistently the worst
>>>> rates.
>>>> >>
>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and
>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly correlating
>>>> >> with rates. Average RTT was the same as for Vocus.
>>>> >>
>>>> >> Note the variation though: More or less a factor of two between
>>>> highest
>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in my
>>>> >> tests? Or is there something systematic behind this? Way too few
>>>> tests
>>>> >> to tell.
>>>> >>
>>>> >> What these tests do is establish a ballpark.
>>>> >>
>>>> >> I'm currently repeating tests with dish placed on a trestle closer to
>>>> >> the heavens. This seems to have translated into fewer outages / ping
>>>> >> losses (around 1/4 of what I had yesterday with dishy on the ground
>>>> on
>>>> >> my deck). Still good enough for a lengthy video Skype call with my
>>>> folks
>>>> >> in Germany, although they did comment about reduced video quality.
>>>> But
>>>> >> maybe that was the lighting or the different background as I wasn't
>>>> in
>>>> >> my usual spot with my laptop when I called them.
>>>> >
>>>> > Clear view of the sky is king for Starlink reliability. I've got my
>>>> dishy mounted on the back fence, looking up over an empty field, so it's
>>>> pretty much best-case scenario here.
>>>> >>
>>>> >>
>>>> >> --
>>>> >>
>>>> >> ****************************************************************
>>>> >> Dr. Ulrich Speidel
>>>> >>
>>>> >> School of Computer Science
>>>> >>
>>>> >> Room 303S.594 (City Campus)
>>>> >>
>>>> >> The University of Auckland
>>>> >> u.speidel@auckland.ac.nz
>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>> >> ****************************************************************
>>>> >>
>>>> >>
>>>> >>
>>>> >> _______________________________________________
>>>> >> Starlink mailing list
>>>> >> Starlink@lists.bufferbloat.net
>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>> >
>>>> > _______________________________________________
>>>> > Starlink mailing list
>>>> > Starlink@lists.bufferbloat.net
>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>
>>>>
>>>>
>>>> --
>>>> This song goes out to all the folk that thought Stadia would work:
>>>>
>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>> Dave Täht CEO, TekLibre, LLC
>>>>
>>> _______________________________________________
>>> Starlink mailing list
>>> Starlink@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/starlink
>>>
>>

-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

[-- Attachment #1.2: Type: text/html, Size: 14812 bytes --]

[-- Attachment #2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 20:37                                           ` Dave Taht
@ 2023-01-13 21:24                                             ` Nathan Owens
  0 siblings, 0 replies; 44+ messages in thread
From: Nathan Owens @ 2023-01-13 21:24 UTC (permalink / raw)
  To: Dave Taht; +Cc: Jonathan Bennett, starlink


[-- Attachment #1.1: Type: text/plain, Size: 11756 bytes --]

They are just alternating colors to make it look pretty.

On Fri, Jan 13, 2023 at 12:37 PM Dave Taht <dave.taht@gmail.com> wrote:

> That is amazingly better than what I'd seen before.
>
> (can you repost your script? What's the difference between the red and
> blue dots?)
>
> Jonathan (or all? ulrich?), could you fire up the irtt for that long
> interval, and fire off a waveform test, pause, then a cloudflare test,
> then, if you have it?
>
> flent -H oneofmyclosestservers.starlink.taht.net -t starlink_vs_irtt
> --step-size=.05 --socket-stats --test-parameter=upload_streams=4 tcp_nup
>
> Going the full monty and enabling ecn and taking a packet cap of this last
> would additionally make my day...
>
> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io> wrote:
>
>> Here's the data visualization for Johnathan's Data
>>
>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>
>> You can see the path change at :12, :27, :42, :57 after the minute. Some
>> paths are clearly busier than others with increased loss, latency, and
>> jitter.
>>
>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io> wrote:
>>
>>> I’ll run my visualization code on this result this afternoon and report
>>> back!
>>>
>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>> starlink@lists.bufferbloat.net> wrote:
>>>
>>>> The irtt command, run with normal, light usage:
>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>
>>>> Jonathan Bennett
>>>> Hackaday.com
>>>>
>>>>
>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com> wrote:
>>>>
>>>>> packet caps would be nice... all this is very exciting news.
>>>>>
>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>> results nowadays to duplicate and re-plot the original irtt tests we
>>>>> did:
>>>>>
>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
>>>>> whatever.json
>>>>>
>>>>> They MUST have changed their scheduling to get such amazing uplink
>>>>> results, in addition to better queue management.
>>>>>
>>>>> (for the record, my servers are de, london, fremont, sydney, dallas,
>>>>> newark, atlanta, singapore, mumbai)
>>>>>
>>>>> There's an R and gnuplot script for plotting that output around here
>>>>> somewhere (I have largely personally put down the starlink project,
>>>>> loaning out mine) - that went by on this list... I should have written
>>>>> a blog entry so I can find that stuff again.
>>>>>
>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>> >
>>>>> >
>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>> >>
>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>> >> >
>>>>> >> > From Auckland, New Zealand, using a roaming subscription, it puts
>>>>> me
>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>> >> >
>>>>> >> >
>>>>> >> > IP address: nix six.
>>>>> >> >
>>>>> >> > My thoughts shall follow later.
>>>>> >>
>>>>> >> OK, so here we go.
>>>>> >>
>>>>> >> I'm always a bit skeptical when it comes to speed tests - they're
>>>>> really
>>>>> >> laden with so many caveats that it's not funny. I took our new work
>>>>> >> Starlink kit home in December to give it a try and the other day
>>>>> finally
>>>>> >> got around to set it up. It's on a roaming subscription because our
>>>>> >> badly built-up campus really isn't ideal in terms of a clear view
>>>>> of the
>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>> adapter,
>>>>> >> not the WiFi?
>>>>> >>
>>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>>> Starlink
>>>>> >> promises is about 1/3 in terms of data rate you can actually get
>>>>> from
>>>>> >> fibre to the home at under half of Starlink's price. Read: There
>>>>> are few
>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>> >>
>>>>> >> Caveat 2: Auckland has three Starlink gateways close by: Clevedon
>>>>> (which
>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana and
>>>>> Puwera,
>>>>> >> the most distant of the three and about 130 km away from me as the
>>>>> crow
>>>>> >> flies. Read: My dishy can use any satellite that any of these three
>>>>> can
>>>>> >> see, and then depending on where I put it and how much of the
>>>>> southern
>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away, although
>>>>> that
>>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>>> options
>>>>> >> for my bits to travel without needing a lot of handovers. Why?
>>>>> Easy: If
>>>>> >> your nearest teleport is close by, then the set of satellites that
>>>>> the
>>>>> >> teleport can see and the set that you can see is almost the same,
>>>>> so you
>>>>> >> can essentially stick with the same satellite while it's in view
>>>>> for you
>>>>> >> because it'll also be in view for the teleport. Pretty much any bird
>>>>> >> above you will do.
>>>>> >>
>>>>> >> And because I don't get a lot of competition from other users in my
>>>>> area
>>>>> >> vying for one of the few available satellites that can see both us
>>>>> and
>>>>> >> the teleport, this is about as good as it gets at 37S latitude. If
>>>>> I'd
>>>>> >> want it any better, I'd have to move a lot further south.
>>>>> >>
>>>>> >> It'd be interesting to hear from Jonathan what the availability of
>>>>> home
>>>>> >> broadband is like in the Dallas area. I note that it's at a lower
>>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I
>>>>> notice
>>>>> >> two teleports each about 160 km away, which is also not too bad. I
>>>>> also
>>>>> >> note Starlink availability in the area is restricted at the moment -
>>>>> >> oversubscribed? But if Jonathan gets good data rates, then that
>>>>> means
>>>>> >> that competition for bird capacity can't be too bad - for whatever
>>>>> reason.
>>>>> >
>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>> broadband options. But there are also many people that live outside cities,
>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>> and Texas is probably ideal conditions for Starlink.
>>>>> >>
>>>>> >>
>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>> whatever I
>>>>> >> talk to in terms of my communications. Traceroute shows about 10
>>>>> hops
>>>>> >> between me and the University of Auckland via Starlink. That's 10
>>>>> >> queues, not one. Many of them will have cross traffic. So it's a bit
>>>>> >> hard to tell where our packets really get to wait or where they get
>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>> between 1
>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can
>>>>> all
>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link of a
>>>>> few
>>>>> >> Mb/s where it's obvious where the dominant long latency bottleneck
>>>>> in
>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>> whether a
>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>> system or
>>>>> >> somewhere between Starlink's systems and the target system.
>>>>> >>
>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical latency
>>>>> >> should be under 15 ms. So there's clearly a bit of buffer here
>>>>> along the
>>>>> >> chain that occasionally fills up.
>>>>> >>
>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>> inevitably
>>>>> >> associated with a change in RTT and in most cases also available
>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>>> >> satellite while your TCP is still trying to respond to whatever it
>>>>> >> thought the queue on the previous bird was doing. Read: Whatever
>>>>> your
>>>>> >> cwnd is immediately after a handover, it's probably not what it
>>>>> should be.
>>>>> >>
>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
>>>>> >> speedtest.net tests each to five local servers. Average upload
>>>>> rate was
>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP
>>>>> that
>>>>> >> Starlink seems to be buying its local connectivity from (Vocus
>>>>> Group)
>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and 52.22
>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
>>>>> >> observed. In fact, they were the ISP with consistently the worst
>>>>> rates.
>>>>> >>
>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and
>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly correlating
>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>> >>
>>>>> >> Note the variation though: More or less a factor of two between
>>>>> highest
>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in my
>>>>> >> tests? Or is there something systematic behind this? Way too few
>>>>> tests
>>>>> >> to tell.
>>>>> >>
>>>>> >> What these tests do is establish a ballpark.
>>>>> >>
>>>>> >> I'm currently repeating tests with dish placed on a trestle closer
>>>>> to
>>>>> >> the heavens. This seems to have translated into fewer outages / ping
>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the ground
>>>>> on
>>>>> >> my deck). Still good enough for a lengthy video Skype call with my
>>>>> folks
>>>>> >> in Germany, although they did comment about reduced video quality.
>>>>> But
>>>>> >> maybe that was the lighting or the different background as I wasn't
>>>>> in
>>>>> >> my usual spot with my laptop when I called them.
>>>>> >
>>>>> > Clear view of the sky is king for Starlink reliability. I've got my
>>>>> dishy mounted on the back fence, looking up over an empty field, so it's
>>>>> pretty much best-case scenario here.
>>>>> >>
>>>>> >>
>>>>> >> --
>>>>> >>
>>>>> >> ****************************************************************
>>>>> >> Dr. Ulrich Speidel
>>>>> >>
>>>>> >> School of Computer Science
>>>>> >>
>>>>> >> Room 303S.594 (City Campus)
>>>>> >>
>>>>> >> The University of Auckland
>>>>> >> u.speidel@auckland.ac.nz
>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>> >> ****************************************************************
>>>>> >>
>>>>> >>
>>>>> >>
>>>>> >> _______________________________________________
>>>>> >> Starlink mailing list
>>>>> >> Starlink@lists.bufferbloat.net
>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>> >
>>>>> > _______________________________________________
>>>>> > Starlink mailing list
>>>>> > Starlink@lists.bufferbloat.net
>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>
>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>
>>>> _______________________________________________
>>>> Starlink mailing list
>>>> Starlink@lists.bufferbloat.net
>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>
>>>
>
> --
> This song goes out to all the folk that thought Stadia would work:
>
> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
> Dave Täht CEO, TekLibre, LLC
>

[-- Attachment #1.2: Type: text/html, Size: 15131 bytes --]

[-- Attachment #2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 20:49                                           ` Dave Taht
@ 2023-01-13 21:25                                             ` Luis A. Cornejo
  2023-01-13 21:30                                               ` Nathan Owens
  0 siblings, 1 reply; 44+ messages in thread
From: Luis A. Cornejo @ 2023-01-13 21:25 UTC (permalink / raw)
  To: Dave Taht; +Cc: Nathan Owens, starlink


[-- Attachment #1.1: Type: text/plain, Size: 12332 bytes --]

Dave,

Here is a run the way I think you wanted it.

irtt running for 5 min to your dallas server, followed by a waveform test,
then a few seconds of inactivity, cloudflare test, a few more secs of
nothing, flent test to dallas. Packet capture is currently uploading (will
be done in 20 min or so), irtt JSON also in there (.zip file):

https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link

-Luis

On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
starlink@lists.bufferbloat.net> wrote:

>
>
> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io> wrote:
>
>> Here's the data visualization for Johnathan's Data
>>
>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>
>> You can see the path change at :12, :27, :42, :57 after the minute. Some
>> paths are clearly busier than others with increased loss, latency, and
>> jitter.
>>
>
> I am so glad to see loss and bounded delay here. Also a bit of rigor
> regarding what traffic was active locally vs on the path would be nice,
> although it seems to line up with the known 15s starlink switchover thing
> (need a name for this), in this case, doing a few speedtests
> while that irtt is running would show the impact(s) of whatever else they
> are up to.
>
> It would also be my hope that the loss distribution in the middle portion
> of this data is good, not bursty, but we don't have a tool to take apart
> that. (I am so hopeless at json)
>
>
>
>>
>>
>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io> wrote:
>>
>>> I’ll run my visualization code on this result this afternoon and report
>>> back!
>>>
>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>> starlink@lists.bufferbloat.net> wrote:
>>>
>>>> The irtt command, run with normal, light usage:
>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>
>>>> Jonathan Bennett
>>>> Hackaday.com
>>>>
>>>>
>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com> wrote:
>>>>
>>>>> packet caps would be nice... all this is very exciting news.
>>>>>
>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>> results nowadays to duplicate and re-plot the original irtt tests we
>>>>> did:
>>>>>
>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
>>>>> whatever.json
>>>>>
>>>>> They MUST have changed their scheduling to get such amazing uplink
>>>>> results, in addition to better queue management.
>>>>>
>>>>> (for the record, my servers are de, london, fremont, sydney, dallas,
>>>>> newark, atlanta, singapore, mumbai)
>>>>>
>>>>> There's an R and gnuplot script for plotting that output around here
>>>>> somewhere (I have largely personally put down the starlink project,
>>>>> loaning out mine) - that went by on this list... I should have written
>>>>> a blog entry so I can find that stuff again.
>>>>>
>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>> >
>>>>> >
>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>> >>
>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>> >> >
>>>>> >> > From Auckland, New Zealand, using a roaming subscription, it puts
>>>>> me
>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>> >> >
>>>>> >> >
>>>>> >> > IP address: nix six.
>>>>> >> >
>>>>> >> > My thoughts shall follow later.
>>>>> >>
>>>>> >> OK, so here we go.
>>>>> >>
>>>>> >> I'm always a bit skeptical when it comes to speed tests - they're
>>>>> really
>>>>> >> laden with so many caveats that it's not funny. I took our new work
>>>>> >> Starlink kit home in December to give it a try and the other day
>>>>> finally
>>>>> >> got around to set it up. It's on a roaming subscription because our
>>>>> >> badly built-up campus really isn't ideal in terms of a clear view
>>>>> of the
>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>> adapter,
>>>>> >> not the WiFi?
>>>>> >>
>>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>>> Starlink
>>>>> >> promises is about 1/3 in terms of data rate you can actually get
>>>>> from
>>>>> >> fibre to the home at under half of Starlink's price. Read: There
>>>>> are few
>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>> >>
>>>>> >> Caveat 2: Auckland has three Starlink gateways close by: Clevedon
>>>>> (which
>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana and
>>>>> Puwera,
>>>>> >> the most distant of the three and about 130 km away from me as the
>>>>> crow
>>>>> >> flies. Read: My dishy can use any satellite that any of these three
>>>>> can
>>>>> >> see, and then depending on where I put it and how much of the
>>>>> southern
>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away, although
>>>>> that
>>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>>> options
>>>>> >> for my bits to travel without needing a lot of handovers. Why?
>>>>> Easy: If
>>>>> >> your nearest teleport is close by, then the set of satellites that
>>>>> the
>>>>> >> teleport can see and the set that you can see is almost the same,
>>>>> so you
>>>>> >> can essentially stick with the same satellite while it's in view
>>>>> for you
>>>>> >> because it'll also be in view for the teleport. Pretty much any bird
>>>>> >> above you will do.
>>>>> >>
>>>>> >> And because I don't get a lot of competition from other users in my
>>>>> area
>>>>> >> vying for one of the few available satellites that can see both us
>>>>> and
>>>>> >> the teleport, this is about as good as it gets at 37S latitude. If
>>>>> I'd
>>>>> >> want it any better, I'd have to move a lot further south.
>>>>> >>
>>>>> >> It'd be interesting to hear from Jonathan what the availability of
>>>>> home
>>>>> >> broadband is like in the Dallas area. I note that it's at a lower
>>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I
>>>>> notice
>>>>> >> two teleports each about 160 km away, which is also not too bad. I
>>>>> also
>>>>> >> note Starlink availability in the area is restricted at the moment -
>>>>> >> oversubscribed? But if Jonathan gets good data rates, then that
>>>>> means
>>>>> >> that competition for bird capacity can't be too bad - for whatever
>>>>> reason.
>>>>> >
>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>> broadband options. But there are also many people that live outside cities,
>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>> and Texas is probably ideal conditions for Starlink.
>>>>> >>
>>>>> >>
>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>> whatever I
>>>>> >> talk to in terms of my communications. Traceroute shows about 10
>>>>> hops
>>>>> >> between me and the University of Auckland via Starlink. That's 10
>>>>> >> queues, not one. Many of them will have cross traffic. So it's a bit
>>>>> >> hard to tell where our packets really get to wait or where they get
>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>> between 1
>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can
>>>>> all
>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link of a
>>>>> few
>>>>> >> Mb/s where it's obvious where the dominant long latency bottleneck
>>>>> in
>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>> whether a
>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>> system or
>>>>> >> somewhere between Starlink's systems and the target system.
>>>>> >>
>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical latency
>>>>> >> should be under 15 ms. So there's clearly a bit of buffer here
>>>>> along the
>>>>> >> chain that occasionally fills up.
>>>>> >>
>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>> inevitably
>>>>> >> associated with a change in RTT and in most cases also available
>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>>> >> satellite while your TCP is still trying to respond to whatever it
>>>>> >> thought the queue on the previous bird was doing. Read: Whatever
>>>>> your
>>>>> >> cwnd is immediately after a handover, it's probably not what it
>>>>> should be.
>>>>> >>
>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
>>>>> >> speedtest.net tests each to five local servers. Average upload
>>>>> rate was
>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP
>>>>> that
>>>>> >> Starlink seems to be buying its local connectivity from (Vocus
>>>>> Group)
>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and 52.22
>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
>>>>> >> observed. In fact, they were the ISP with consistently the worst
>>>>> rates.
>>>>> >>
>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and
>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly correlating
>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>> >>
>>>>> >> Note the variation though: More or less a factor of two between
>>>>> highest
>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in my
>>>>> >> tests? Or is there something systematic behind this? Way too few
>>>>> tests
>>>>> >> to tell.
>>>>> >>
>>>>> >> What these tests do is establish a ballpark.
>>>>> >>
>>>>> >> I'm currently repeating tests with dish placed on a trestle closer
>>>>> to
>>>>> >> the heavens. This seems to have translated into fewer outages / ping
>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the ground
>>>>> on
>>>>> >> my deck). Still good enough for a lengthy video Skype call with my
>>>>> folks
>>>>> >> in Germany, although they did comment about reduced video quality.
>>>>> But
>>>>> >> maybe that was the lighting or the different background as I wasn't
>>>>> in
>>>>> >> my usual spot with my laptop when I called them.
>>>>> >
>>>>> > Clear view of the sky is king for Starlink reliability. I've got my
>>>>> dishy mounted on the back fence, looking up over an empty field, so it's
>>>>> pretty much best-case scenario here.
>>>>> >>
>>>>> >>
>>>>> >> --
>>>>> >>
>>>>> >> ****************************************************************
>>>>> >> Dr. Ulrich Speidel
>>>>> >>
>>>>> >> School of Computer Science
>>>>> >>
>>>>> >> Room 303S.594 (City Campus)
>>>>> >>
>>>>> >> The University of Auckland
>>>>> >> u.speidel@auckland.ac.nz
>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>> >> ****************************************************************
>>>>> >>
>>>>> >>
>>>>> >>
>>>>> >> _______________________________________________
>>>>> >> Starlink mailing list
>>>>> >> Starlink@lists.bufferbloat.net
>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>> >
>>>>> > _______________________________________________
>>>>> > Starlink mailing list
>>>>> > Starlink@lists.bufferbloat.net
>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>
>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>
>>>> _______________________________________________
>>>> Starlink mailing list
>>>> Starlink@lists.bufferbloat.net
>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>
>>>
>
> --
> This song goes out to all the folk that thought Stadia would work:
>
> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
> Dave Täht CEO, TekLibre, LLC
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink
>

[-- Attachment #1.2: Type: text/html, Size: 16168 bytes --]

[-- Attachment #2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 21:25                                             ` Luis A. Cornejo
@ 2023-01-13 21:30                                               ` Nathan Owens
  2023-01-13 22:09                                                 ` Jonathan Bennett
  0 siblings, 1 reply; 44+ messages in thread
From: Nathan Owens @ 2023-01-13 21:30 UTC (permalink / raw)
  To: Luis A. Cornejo; +Cc: Dave Taht, starlink


[-- Attachment #1.1.1: Type: text/plain, Size: 12927 bytes --]

Here's Luis's run -- the top line below the edge of the graph is 200ms
[image: Screenshot 2023-01-13 at 1.30.03 PM.png]


On Fri, Jan 13, 2023 at 1:25 PM Luis A. Cornejo <luis.a.cornejo@gmail.com>
wrote:

> Dave,
>
> Here is a run the way I think you wanted it.
>
> irtt running for 5 min to your dallas server, followed by a waveform test,
> then a few seconds of inactivity, cloudflare test, a few more secs of
> nothing, flent test to dallas. Packet capture is currently uploading (will
> be done in 20 min or so), irtt JSON also in there (.zip file):
>
>
> https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link
>
> -Luis
>
> On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
> starlink@lists.bufferbloat.net> wrote:
>
>>
>>
>> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io> wrote:
>>
>>> Here's the data visualization for Johnathan's Data
>>>
>>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>>
>>> You can see the path change at :12, :27, :42, :57 after the minute. Some
>>> paths are clearly busier than others with increased loss, latency, and
>>> jitter.
>>>
>>
>> I am so glad to see loss and bounded delay here. Also a bit of rigor
>> regarding what traffic was active locally vs on the path would be nice,
>> although it seems to line up with the known 15s starlink switchover thing
>> (need a name for this), in this case, doing a few speedtests
>> while that irtt is running would show the impact(s) of whatever else they
>> are up to.
>>
>> It would also be my hope that the loss distribution in the middle portion
>> of this data is good, not bursty, but we don't have a tool to take apart
>> that. (I am so hopeless at json)
>>
>>
>>
>>>
>>>
>>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io> wrote:
>>>
>>>> I’ll run my visualization code on this result this afternoon and report
>>>> back!
>>>>
>>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>>> starlink@lists.bufferbloat.net> wrote:
>>>>
>>>>> The irtt command, run with normal, light usage:
>>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>>
>>>>> Jonathan Bennett
>>>>> Hackaday.com
>>>>>
>>>>>
>>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> packet caps would be nice... all this is very exciting news.
>>>>>>
>>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>>> results nowadays to duplicate and re-plot the original irtt tests we
>>>>>> did:
>>>>>>
>>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
>>>>>> whatever.json
>>>>>>
>>>>>> They MUST have changed their scheduling to get such amazing uplink
>>>>>> results, in addition to better queue management.
>>>>>>
>>>>>> (for the record, my servers are de, london, fremont, sydney, dallas,
>>>>>> newark, atlanta, singapore, mumbai)
>>>>>>
>>>>>> There's an R and gnuplot script for plotting that output around here
>>>>>> somewhere (I have largely personally put down the starlink project,
>>>>>> loaning out mine) - that went by on this list... I should have written
>>>>>> a blog entry so I can find that stuff again.
>>>>>>
>>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>>> >
>>>>>> >
>>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>> >>
>>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>>> >> >
>>>>>> >> > From Auckland, New Zealand, using a roaming subscription, it
>>>>>> puts me
>>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>>> >> >
>>>>>> >> >
>>>>>> >> > IP address: nix six.
>>>>>> >> >
>>>>>> >> > My thoughts shall follow later.
>>>>>> >>
>>>>>> >> OK, so here we go.
>>>>>> >>
>>>>>> >> I'm always a bit skeptical when it comes to speed tests - they're
>>>>>> really
>>>>>> >> laden with so many caveats that it's not funny. I took our new work
>>>>>> >> Starlink kit home in December to give it a try and the other day
>>>>>> finally
>>>>>> >> got around to set it up. It's on a roaming subscription because our
>>>>>> >> badly built-up campus really isn't ideal in terms of a clear view
>>>>>> of the
>>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>>> adapter,
>>>>>> >> not the WiFi?
>>>>>> >>
>>>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>>>> Starlink
>>>>>> >> promises is about 1/3 in terms of data rate you can actually get
>>>>>> from
>>>>>> >> fibre to the home at under half of Starlink's price. Read: There
>>>>>> are few
>>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>>> >>
>>>>>> >> Caveat 2: Auckland has three Starlink gateways close by: Clevedon
>>>>>> (which
>>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana and
>>>>>> Puwera,
>>>>>> >> the most distant of the three and about 130 km away from me as the
>>>>>> crow
>>>>>> >> flies. Read: My dishy can use any satellite that any of these
>>>>>> three can
>>>>>> >> see, and then depending on where I put it and how much of the
>>>>>> southern
>>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away, although
>>>>>> that
>>>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>>>> options
>>>>>> >> for my bits to travel without needing a lot of handovers. Why?
>>>>>> Easy: If
>>>>>> >> your nearest teleport is close by, then the set of satellites that
>>>>>> the
>>>>>> >> teleport can see and the set that you can see is almost the same,
>>>>>> so you
>>>>>> >> can essentially stick with the same satellite while it's in view
>>>>>> for you
>>>>>> >> because it'll also be in view for the teleport. Pretty much any
>>>>>> bird
>>>>>> >> above you will do.
>>>>>> >>
>>>>>> >> And because I don't get a lot of competition from other users in
>>>>>> my area
>>>>>> >> vying for one of the few available satellites that can see both us
>>>>>> and
>>>>>> >> the teleport, this is about as good as it gets at 37S latitude. If
>>>>>> I'd
>>>>>> >> want it any better, I'd have to move a lot further south.
>>>>>> >>
>>>>>> >> It'd be interesting to hear from Jonathan what the availability of
>>>>>> home
>>>>>> >> broadband is like in the Dallas area. I note that it's at a lower
>>>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I
>>>>>> notice
>>>>>> >> two teleports each about 160 km away, which is also not too bad. I
>>>>>> also
>>>>>> >> note Starlink availability in the area is restricted at the moment
>>>>>> -
>>>>>> >> oversubscribed? But if Jonathan gets good data rates, then that
>>>>>> means
>>>>>> >> that competition for bird capacity can't be too bad - for whatever
>>>>>> reason.
>>>>>> >
>>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>>> broadband options. But there are also many people that live outside cities,
>>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>>> and Texas is probably ideal conditions for Starlink.
>>>>>> >>
>>>>>> >>
>>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>>> whatever I
>>>>>> >> talk to in terms of my communications. Traceroute shows about 10
>>>>>> hops
>>>>>> >> between me and the University of Auckland via Starlink. That's 10
>>>>>> >> queues, not one. Many of them will have cross traffic. So it's a
>>>>>> bit
>>>>>> >> hard to tell where our packets really get to wait or where they get
>>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>>> between 1
>>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can
>>>>>> all
>>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link of a
>>>>>> few
>>>>>> >> Mb/s where it's obvious where the dominant long latency bottleneck
>>>>>> in
>>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>>> whether a
>>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>>> system or
>>>>>> >> somewhere between Starlink's systems and the target system.
>>>>>> >>
>>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical
>>>>>> latency
>>>>>> >> should be under 15 ms. So there's clearly a bit of buffer here
>>>>>> along the
>>>>>> >> chain that occasionally fills up.
>>>>>> >>
>>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>>> inevitably
>>>>>> >> associated with a change in RTT and in most cases also available
>>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>>>> >> satellite while your TCP is still trying to respond to whatever it
>>>>>> >> thought the queue on the previous bird was doing. Read: Whatever
>>>>>> your
>>>>>> >> cwnd is immediately after a handover, it's probably not what it
>>>>>> should be.
>>>>>> >>
>>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
>>>>>> >> speedtest.net tests each to five local servers. Average upload
>>>>>> rate was
>>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP
>>>>>> that
>>>>>> >> Starlink seems to be buying its local connectivity from (Vocus
>>>>>> Group)
>>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and
>>>>>> 52.22
>>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
>>>>>> >> observed. In fact, they were the ISP with consistently the worst
>>>>>> rates.
>>>>>> >>
>>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up and
>>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly
>>>>>> correlating
>>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>>> >>
>>>>>> >> Note the variation though: More or less a factor of two between
>>>>>> highest
>>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in my
>>>>>> >> tests? Or is there something systematic behind this? Way too few
>>>>>> tests
>>>>>> >> to tell.
>>>>>> >>
>>>>>> >> What these tests do is establish a ballpark.
>>>>>> >>
>>>>>> >> I'm currently repeating tests with dish placed on a trestle closer
>>>>>> to
>>>>>> >> the heavens. This seems to have translated into fewer outages /
>>>>>> ping
>>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the
>>>>>> ground on
>>>>>> >> my deck). Still good enough for a lengthy video Skype call with my
>>>>>> folks
>>>>>> >> in Germany, although they did comment about reduced video quality.
>>>>>> But
>>>>>> >> maybe that was the lighting or the different background as I
>>>>>> wasn't in
>>>>>> >> my usual spot with my laptop when I called them.
>>>>>> >
>>>>>> > Clear view of the sky is king for Starlink reliability. I've got my
>>>>>> dishy mounted on the back fence, looking up over an empty field, so it's
>>>>>> pretty much best-case scenario here.
>>>>>> >>
>>>>>> >>
>>>>>> >> --
>>>>>> >>
>>>>>> >> ****************************************************************
>>>>>> >> Dr. Ulrich Speidel
>>>>>> >>
>>>>>> >> School of Computer Science
>>>>>> >>
>>>>>> >> Room 303S.594 (City Campus)
>>>>>> >>
>>>>>> >> The University of Auckland
>>>>>> >> u.speidel@auckland.ac.nz
>>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>>> >> ****************************************************************
>>>>>> >>
>>>>>> >>
>>>>>> >>
>>>>>> >> _______________________________________________
>>>>>> >> Starlink mailing list
>>>>>> >> Starlink@lists.bufferbloat.net
>>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>>> >
>>>>>> > _______________________________________________
>>>>>> > Starlink mailing list
>>>>>> > Starlink@lists.bufferbloat.net
>>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>
>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>
>>>>> _______________________________________________
>>>>> Starlink mailing list
>>>>> Starlink@lists.bufferbloat.net
>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>
>>>>
>>
>> --
>> This song goes out to all the folk that thought Stadia would work:
>>
>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>> Dave Täht CEO, TekLibre, LLC
>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink
>>
>

[-- Attachment #1.1.2: Type: text/html, Size: 16759 bytes --]

[-- Attachment #1.2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #1.3: Screenshot 2023-01-13 at 1.30.03 PM.png --]
[-- Type: image/png, Size: 855840 bytes --]

[-- Attachment #2: lc_starlink_irtt.pdf --]
[-- Type: application/pdf, Size: 697557 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 21:30                                               ` Nathan Owens
@ 2023-01-13 22:09                                                 ` Jonathan Bennett
  2023-01-13 22:30                                                   ` Luis A. Cornejo
                                                                     ` (2 more replies)
  0 siblings, 3 replies; 44+ messages in thread
From: Jonathan Bennett @ 2023-01-13 22:09 UTC (permalink / raw)
  To: Nathan Owens; +Cc: Luis A. Cornejo, starlink


[-- Attachment #1.1: Type: text/plain, Size: 13994 bytes --]

The irtt run finished a few seconds before the flent run, but here are the
results:

https://drive.google.com/file/d/1FKve13ssUMW1LLWOXLM2931Yx6uMHw8K/view?usp=share_link
https://drive.google.com/file/d/1ZXd64A0pfUedLr3FyhDNTHA7vxv8S2Gk/view?usp=share_link
https://drive.google.com/file/d/1rx64UPQHHz3IMNiJtb1oFqtqw2DvhvEE/view?usp=share_link


[image: image.png]
[image: image.png]


Jonathan Bennett
Hackaday.com


On Fri, Jan 13, 2023 at 3:30 PM Nathan Owens via Starlink <
starlink@lists.bufferbloat.net> wrote:

> Here's Luis's run -- the top line below the edge of the graph is 200ms
> [image: Screenshot 2023-01-13 at 1.30.03 PM.png]
>
>
> On Fri, Jan 13, 2023 at 1:25 PM Luis A. Cornejo <luis.a.cornejo@gmail.com>
> wrote:
>
>> Dave,
>>
>> Here is a run the way I think you wanted it.
>>
>> irtt running for 5 min to your dallas server, followed by a waveform
>> test, then a few seconds of inactivity, cloudflare test, a few more secs of
>> nothing, flent test to dallas. Packet capture is currently uploading (will
>> be done in 20 min or so), irtt JSON also in there (.zip file):
>>
>>
>> https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link
>>
>> -Luis
>>
>> On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
>> starlink@lists.bufferbloat.net> wrote:
>>
>>>
>>>
>>> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io> wrote:
>>>
>>>> Here's the data visualization for Johnathan's Data
>>>>
>>>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>>>
>>>> You can see the path change at :12, :27, :42, :57 after the minute.
>>>> Some paths are clearly busier than others with increased loss, latency, and
>>>> jitter.
>>>>
>>>
>>> I am so glad to see loss and bounded delay here. Also a bit of rigor
>>> regarding what traffic was active locally vs on the path would be nice,
>>> although it seems to line up with the known 15s starlink switchover thing
>>> (need a name for this), in this case, doing a few speedtests
>>> while that irtt is running would show the impact(s) of whatever else
>>> they are up to.
>>>
>>> It would also be my hope that the loss distribution in the middle
>>> portion of this data is good, not bursty, but we don't have a tool to take
>>> apart that. (I am so hopeless at json)
>>>
>>>
>>>
>>>>
>>>>
>>>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io> wrote:
>>>>
>>>>> I’ll run my visualization code on this result this afternoon and
>>>>> report back!
>>>>>
>>>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>
>>>>>> The irtt command, run with normal, light usage:
>>>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>>>
>>>>>> Jonathan Bennett
>>>>>> Hackaday.com
>>>>>>
>>>>>>
>>>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> packet caps would be nice... all this is very exciting news.
>>>>>>>
>>>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>>>> results nowadays to duplicate and re-plot the original irtt tests we
>>>>>>> did:
>>>>>>>
>>>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
>>>>>>> whatever.json
>>>>>>>
>>>>>>> They MUST have changed their scheduling to get such amazing uplink
>>>>>>> results, in addition to better queue management.
>>>>>>>
>>>>>>> (for the record, my servers are de, london, fremont, sydney, dallas,
>>>>>>> newark, atlanta, singapore, mumbai)
>>>>>>>
>>>>>>> There's an R and gnuplot script for plotting that output around here
>>>>>>> somewhere (I have largely personally put down the starlink project,
>>>>>>> loaning out mine) - that went by on this list... I should have
>>>>>>> written
>>>>>>> a blog entry so I can find that stuff again.
>>>>>>>
>>>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>>>> >
>>>>>>> >
>>>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>> >>
>>>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>>>> >> >
>>>>>>> >> > From Auckland, New Zealand, using a roaming subscription, it
>>>>>>> puts me
>>>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>>>> >> >
>>>>>>> >> >
>>>>>>> >> > IP address: nix six.
>>>>>>> >> >
>>>>>>> >> > My thoughts shall follow later.
>>>>>>> >>
>>>>>>> >> OK, so here we go.
>>>>>>> >>
>>>>>>> >> I'm always a bit skeptical when it comes to speed tests - they're
>>>>>>> really
>>>>>>> >> laden with so many caveats that it's not funny. I took our new
>>>>>>> work
>>>>>>> >> Starlink kit home in December to give it a try and the other day
>>>>>>> finally
>>>>>>> >> got around to set it up. It's on a roaming subscription because
>>>>>>> our
>>>>>>> >> badly built-up campus really isn't ideal in terms of a clear view
>>>>>>> of the
>>>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>>>> adapter,
>>>>>>> >> not the WiFi?
>>>>>>> >>
>>>>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>>>>> Starlink
>>>>>>> >> promises is about 1/3 in terms of data rate you can actually get
>>>>>>> from
>>>>>>> >> fibre to the home at under half of Starlink's price. Read: There
>>>>>>> are few
>>>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>>>> >>
>>>>>>> >> Caveat 2: Auckland has three Starlink gateways close by: Clevedon
>>>>>>> (which
>>>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana and
>>>>>>> Puwera,
>>>>>>> >> the most distant of the three and about 130 km away from me as
>>>>>>> the crow
>>>>>>> >> flies. Read: My dishy can use any satellite that any of these
>>>>>>> three can
>>>>>>> >> see, and then depending on where I put it and how much of the
>>>>>>> southern
>>>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away,
>>>>>>> although that
>>>>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>>>>> options
>>>>>>> >> for my bits to travel without needing a lot of handovers. Why?
>>>>>>> Easy: If
>>>>>>> >> your nearest teleport is close by, then the set of satellites
>>>>>>> that the
>>>>>>> >> teleport can see and the set that you can see is almost the same,
>>>>>>> so you
>>>>>>> >> can essentially stick with the same satellite while it's in view
>>>>>>> for you
>>>>>>> >> because it'll also be in view for the teleport. Pretty much any
>>>>>>> bird
>>>>>>> >> above you will do.
>>>>>>> >>
>>>>>>> >> And because I don't get a lot of competition from other users in
>>>>>>> my area
>>>>>>> >> vying for one of the few available satellites that can see both
>>>>>>> us and
>>>>>>> >> the teleport, this is about as good as it gets at 37S latitude.
>>>>>>> If I'd
>>>>>>> >> want it any better, I'd have to move a lot further south.
>>>>>>> >>
>>>>>>> >> It'd be interesting to hear from Jonathan what the availability
>>>>>>> of home
>>>>>>> >> broadband is like in the Dallas area. I note that it's at a lower
>>>>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I
>>>>>>> notice
>>>>>>> >> two teleports each about 160 km away, which is also not too bad.
>>>>>>> I also
>>>>>>> >> note Starlink availability in the area is restricted at the
>>>>>>> moment -
>>>>>>> >> oversubscribed? But if Jonathan gets good data rates, then that
>>>>>>> means
>>>>>>> >> that competition for bird capacity can't be too bad - for
>>>>>>> whatever reason.
>>>>>>> >
>>>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>>>> broadband options. But there are also many people that live outside cities,
>>>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>>>> and Texas is probably ideal conditions for Starlink.
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>>>> whatever I
>>>>>>> >> talk to in terms of my communications. Traceroute shows about 10
>>>>>>> hops
>>>>>>> >> between me and the University of Auckland via Starlink. That's 10
>>>>>>> >> queues, not one. Many of them will have cross traffic. So it's a
>>>>>>> bit
>>>>>>> >> hard to tell where our packets really get to wait or where they
>>>>>>> get
>>>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>>>> between 1
>>>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they can
>>>>>>> all
>>>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link of a
>>>>>>> few
>>>>>>> >> Mb/s where it's obvious where the dominant long latency
>>>>>>> bottleneck in
>>>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>>>> whether a
>>>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>>>> system or
>>>>>>> >> somewhere between Starlink's systems and the target system.
>>>>>>> >>
>>>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical
>>>>>>> latency
>>>>>>> >> should be under 15 ms. So there's clearly a bit of buffer here
>>>>>>> along the
>>>>>>> >> chain that occasionally fills up.
>>>>>>> >>
>>>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>>>> inevitably
>>>>>>> >> associated with a change in RTT and in most cases also available
>>>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>>>>> >> satellite while your TCP is still trying to respond to whatever it
>>>>>>> >> thought the queue on the previous bird was doing. Read: Whatever
>>>>>>> your
>>>>>>> >> cwnd is immediately after a handover, it's probably not what it
>>>>>>> should be.
>>>>>>> >>
>>>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four Ookla
>>>>>>> >> speedtest.net tests each to five local servers. Average upload
>>>>>>> rate was
>>>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP
>>>>>>> that
>>>>>>> >> Starlink seems to be buying its local connectivity from (Vocus
>>>>>>> Group)
>>>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and
>>>>>>> 52.22
>>>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to rates
>>>>>>> >> observed. In fact, they were the ISP with consistently the worst
>>>>>>> rates.
>>>>>>> >>
>>>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up
>>>>>>> and
>>>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly
>>>>>>> correlating
>>>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>>>> >>
>>>>>>> >> Note the variation though: More or less a factor of two between
>>>>>>> highest
>>>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in my
>>>>>>> >> tests? Or is there something systematic behind this? Way too few
>>>>>>> tests
>>>>>>> >> to tell.
>>>>>>> >>
>>>>>>> >> What these tests do is establish a ballpark.
>>>>>>> >>
>>>>>>> >> I'm currently repeating tests with dish placed on a trestle
>>>>>>> closer to
>>>>>>> >> the heavens. This seems to have translated into fewer outages /
>>>>>>> ping
>>>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the
>>>>>>> ground on
>>>>>>> >> my deck). Still good enough for a lengthy video Skype call with
>>>>>>> my folks
>>>>>>> >> in Germany, although they did comment about reduced video
>>>>>>> quality. But
>>>>>>> >> maybe that was the lighting or the different background as I
>>>>>>> wasn't in
>>>>>>> >> my usual spot with my laptop when I called them.
>>>>>>> >
>>>>>>> > Clear view of the sky is king for Starlink reliability. I've got
>>>>>>> my dishy mounted on the back fence, looking up over an empty field, so it's
>>>>>>> pretty much best-case scenario here.
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> --
>>>>>>> >>
>>>>>>> >> ****************************************************************
>>>>>>> >> Dr. Ulrich Speidel
>>>>>>> >>
>>>>>>> >> School of Computer Science
>>>>>>> >>
>>>>>>> >> Room 303S.594 (City Campus)
>>>>>>> >>
>>>>>>> >> The University of Auckland
>>>>>>> >> u.speidel@auckland.ac.nz
>>>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>>>> >> ****************************************************************
>>>>>>> >>
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> _______________________________________________
>>>>>>> >> Starlink mailing list
>>>>>>> >> Starlink@lists.bufferbloat.net
>>>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>> >
>>>>>>> > _______________________________________________
>>>>>>> > Starlink mailing list
>>>>>>> > Starlink@lists.bufferbloat.net
>>>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>>
>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>>
>>>>>> _______________________________________________
>>>>>> Starlink mailing list
>>>>>> Starlink@lists.bufferbloat.net
>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>
>>>>>
>>>
>>> --
>>> This song goes out to all the folk that thought Stadia would work:
>>>
>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>> Dave Täht CEO, TekLibre, LLC
>>> _______________________________________________
>>> Starlink mailing list
>>> Starlink@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/starlink
>>>
>> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink
>

[-- Attachment #1.2: Type: text/html, Size: 18652 bytes --]

[-- Attachment #2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #3: Screenshot 2023-01-13 at 1.30.03 PM.png --]
[-- Type: image/png, Size: 855840 bytes --]

[-- Attachment #4: image.png --]
[-- Type: image/png, Size: 296989 bytes --]

[-- Attachment #5: image.png --]
[-- Type: image/png, Size: 222655 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 22:09                                                 ` Jonathan Bennett
@ 2023-01-13 22:30                                                   ` Luis A. Cornejo
  2023-01-13 22:32                                                   ` Dave Taht
  2023-01-13 23:44                                                   ` Dave Taht
  2 siblings, 0 replies; 44+ messages in thread
From: Luis A. Cornejo @ 2023-01-13 22:30 UTC (permalink / raw)
  To: Jonathan Bennett; +Cc: Nathan Owens, starlink


[-- Attachment #1.1: Type: text/plain, Size: 14590 bytes --]

Johnathan,

Is your IPv6 address routable and are you accepting pings?

-Luis

On Fri, Jan 13, 2023 at 4:09 PM Jonathan Bennett <
jonathanbennett@hackaday.com> wrote:

> The irtt run finished a few seconds before the flent run, but here are the
> results:
>
>
> https://drive.google.com/file/d/1FKve13ssUMW1LLWOXLM2931Yx6uMHw8K/view?usp=share_link
>
> https://drive.google.com/file/d/1ZXd64A0pfUedLr3FyhDNTHA7vxv8S2Gk/view?usp=share_link
>
> https://drive.google.com/file/d/1rx64UPQHHz3IMNiJtb1oFqtqw2DvhvEE/view?usp=share_link
>
>
> [image: image.png]
> [image: image.png]
>
>
> Jonathan Bennett
> Hackaday.com
>
>
> On Fri, Jan 13, 2023 at 3:30 PM Nathan Owens via Starlink <
> starlink@lists.bufferbloat.net> wrote:
>
>> Here's Luis's run -- the top line below the edge of the graph is 200ms
>> [image: Screenshot 2023-01-13 at 1.30.03 PM.png]
>>
>>
>> On Fri, Jan 13, 2023 at 1:25 PM Luis A. Cornejo <luis.a.cornejo@gmail.com>
>> wrote:
>>
>>> Dave,
>>>
>>> Here is a run the way I think you wanted it.
>>>
>>> irtt running for 5 min to your dallas server, followed by a waveform
>>> test, then a few seconds of inactivity, cloudflare test, a few more secs of
>>> nothing, flent test to dallas. Packet capture is currently uploading (will
>>> be done in 20 min or so), irtt JSON also in there (.zip file):
>>>
>>>
>>> https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link
>>>
>>> -Luis
>>>
>>> On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
>>> starlink@lists.bufferbloat.net> wrote:
>>>
>>>>
>>>>
>>>> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io> wrote:
>>>>
>>>>> Here's the data visualization for Johnathan's Data
>>>>>
>>>>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>>>>
>>>>> You can see the path change at :12, :27, :42, :57 after the minute.
>>>>> Some paths are clearly busier than others with increased loss, latency, and
>>>>> jitter.
>>>>>
>>>>
>>>> I am so glad to see loss and bounded delay here. Also a bit of rigor
>>>> regarding what traffic was active locally vs on the path would be nice,
>>>> although it seems to line up with the known 15s starlink switchover thing
>>>> (need a name for this), in this case, doing a few speedtests
>>>> while that irtt is running would show the impact(s) of whatever else
>>>> they are up to.
>>>>
>>>> It would also be my hope that the loss distribution in the middle
>>>> portion of this data is good, not bursty, but we don't have a tool to take
>>>> apart that. (I am so hopeless at json)
>>>>
>>>>
>>>>
>>>>>
>>>>>
>>>>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io>
>>>>> wrote:
>>>>>
>>>>>> I’ll run my visualization code on this result this afternoon and
>>>>>> report back!
>>>>>>
>>>>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>
>>>>>>> The irtt command, run with normal, light usage:
>>>>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>>>>
>>>>>>> Jonathan Bennett
>>>>>>> Hackaday.com
>>>>>>>
>>>>>>>
>>>>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> packet caps would be nice... all this is very exciting news.
>>>>>>>>
>>>>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>>>>> results nowadays to duplicate and re-plot the original irtt tests we
>>>>>>>> did:
>>>>>>>>
>>>>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
>>>>>>>> whatever.json
>>>>>>>>
>>>>>>>> They MUST have changed their scheduling to get such amazing uplink
>>>>>>>> results, in addition to better queue management.
>>>>>>>>
>>>>>>>> (for the record, my servers are de, london, fremont, sydney, dallas,
>>>>>>>> newark, atlanta, singapore, mumbai)
>>>>>>>>
>>>>>>>> There's an R and gnuplot script for plotting that output around here
>>>>>>>> somewhere (I have largely personally put down the starlink project,
>>>>>>>> loaning out mine) - that went by on this list... I should have
>>>>>>>> written
>>>>>>>> a blog entry so I can find that stuff again.
>>>>>>>>
>>>>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>>>>> >
>>>>>>>> >
>>>>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>> >>
>>>>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>>>>> >> >
>>>>>>>> >> > From Auckland, New Zealand, using a roaming subscription, it
>>>>>>>> puts me
>>>>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>>>>> >> >
>>>>>>>> >> >
>>>>>>>> >> > IP address: nix six.
>>>>>>>> >> >
>>>>>>>> >> > My thoughts shall follow later.
>>>>>>>> >>
>>>>>>>> >> OK, so here we go.
>>>>>>>> >>
>>>>>>>> >> I'm always a bit skeptical when it comes to speed tests -
>>>>>>>> they're really
>>>>>>>> >> laden with so many caveats that it's not funny. I took our new
>>>>>>>> work
>>>>>>>> >> Starlink kit home in December to give it a try and the other day
>>>>>>>> finally
>>>>>>>> >> got around to set it up. It's on a roaming subscription because
>>>>>>>> our
>>>>>>>> >> badly built-up campus really isn't ideal in terms of a clear
>>>>>>>> view of the
>>>>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>>>>> adapter,
>>>>>>>> >> not the WiFi?
>>>>>>>> >>
>>>>>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>>>>>> Starlink
>>>>>>>> >> promises is about 1/3 in terms of data rate you can actually get
>>>>>>>> from
>>>>>>>> >> fibre to the home at under half of Starlink's price. Read: There
>>>>>>>> are few
>>>>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>>>>> >>
>>>>>>>> >> Caveat 2: Auckland has three Starlink gateways close by:
>>>>>>>> Clevedon (which
>>>>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana and
>>>>>>>> Puwera,
>>>>>>>> >> the most distant of the three and about 130 km away from me as
>>>>>>>> the crow
>>>>>>>> >> flies. Read: My dishy can use any satellite that any of these
>>>>>>>> three can
>>>>>>>> >> see, and then depending on where I put it and how much of the
>>>>>>>> southern
>>>>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away,
>>>>>>>> although that
>>>>>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>>>>>> options
>>>>>>>> >> for my bits to travel without needing a lot of handovers. Why?
>>>>>>>> Easy: If
>>>>>>>> >> your nearest teleport is close by, then the set of satellites
>>>>>>>> that the
>>>>>>>> >> teleport can see and the set that you can see is almost the
>>>>>>>> same, so you
>>>>>>>> >> can essentially stick with the same satellite while it's in view
>>>>>>>> for you
>>>>>>>> >> because it'll also be in view for the teleport. Pretty much any
>>>>>>>> bird
>>>>>>>> >> above you will do.
>>>>>>>> >>
>>>>>>>> >> And because I don't get a lot of competition from other users in
>>>>>>>> my area
>>>>>>>> >> vying for one of the few available satellites that can see both
>>>>>>>> us and
>>>>>>>> >> the teleport, this is about as good as it gets at 37S latitude.
>>>>>>>> If I'd
>>>>>>>> >> want it any better, I'd have to move a lot further south.
>>>>>>>> >>
>>>>>>>> >> It'd be interesting to hear from Jonathan what the availability
>>>>>>>> of home
>>>>>>>> >> broadband is like in the Dallas area. I note that it's at a lower
>>>>>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I
>>>>>>>> notice
>>>>>>>> >> two teleports each about 160 km away, which is also not too bad.
>>>>>>>> I also
>>>>>>>> >> note Starlink availability in the area is restricted at the
>>>>>>>> moment -
>>>>>>>> >> oversubscribed? But if Jonathan gets good data rates, then that
>>>>>>>> means
>>>>>>>> >> that competition for bird capacity can't be too bad - for
>>>>>>>> whatever reason.
>>>>>>>> >
>>>>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>>>>> broadband options. But there are also many people that live outside cities,
>>>>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>>>>> and Texas is probably ideal conditions for Starlink.
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>>>>> whatever I
>>>>>>>> >> talk to in terms of my communications. Traceroute shows about 10
>>>>>>>> hops
>>>>>>>> >> between me and the University of Auckland via Starlink. That's 10
>>>>>>>> >> queues, not one. Many of them will have cross traffic. So it's a
>>>>>>>> bit
>>>>>>>> >> hard to tell where our packets really get to wait or where they
>>>>>>>> get
>>>>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>>>>> between 1
>>>>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they
>>>>>>>> can all
>>>>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link of
>>>>>>>> a few
>>>>>>>> >> Mb/s where it's obvious where the dominant long latency
>>>>>>>> bottleneck in
>>>>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>>>>> whether a
>>>>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>>>>> system or
>>>>>>>> >> somewhere between Starlink's systems and the target system.
>>>>>>>> >>
>>>>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical
>>>>>>>> latency
>>>>>>>> >> should be under 15 ms. So there's clearly a bit of buffer here
>>>>>>>> along the
>>>>>>>> >> chain that occasionally fills up.
>>>>>>>> >>
>>>>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>>>>> inevitably
>>>>>>>> >> associated with a change in RTT and in most cases also available
>>>>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>>>>>> >> satellite while your TCP is still trying to respond to whatever
>>>>>>>> it
>>>>>>>> >> thought the queue on the previous bird was doing. Read: Whatever
>>>>>>>> your
>>>>>>>> >> cwnd is immediately after a handover, it's probably not what it
>>>>>>>> should be.
>>>>>>>> >>
>>>>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four
>>>>>>>> Ookla
>>>>>>>> >> speedtest.net tests each to five local servers. Average upload
>>>>>>>> rate was
>>>>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP
>>>>>>>> that
>>>>>>>> >> Starlink seems to be buying its local connectivity from (Vocus
>>>>>>>> Group)
>>>>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and
>>>>>>>> 52.22
>>>>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to
>>>>>>>> rates
>>>>>>>> >> observed. In fact, they were the ISP with consistently the worst
>>>>>>>> rates.
>>>>>>>> >>
>>>>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up
>>>>>>>> and
>>>>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly
>>>>>>>> correlating
>>>>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>>>>> >>
>>>>>>>> >> Note the variation though: More or less a factor of two between
>>>>>>>> highest
>>>>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in
>>>>>>>> my
>>>>>>>> >> tests? Or is there something systematic behind this? Way too few
>>>>>>>> tests
>>>>>>>> >> to tell.
>>>>>>>> >>
>>>>>>>> >> What these tests do is establish a ballpark.
>>>>>>>> >>
>>>>>>>> >> I'm currently repeating tests with dish placed on a trestle
>>>>>>>> closer to
>>>>>>>> >> the heavens. This seems to have translated into fewer outages /
>>>>>>>> ping
>>>>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the
>>>>>>>> ground on
>>>>>>>> >> my deck). Still good enough for a lengthy video Skype call with
>>>>>>>> my folks
>>>>>>>> >> in Germany, although they did comment about reduced video
>>>>>>>> quality. But
>>>>>>>> >> maybe that was the lighting or the different background as I
>>>>>>>> wasn't in
>>>>>>>> >> my usual spot with my laptop when I called them.
>>>>>>>> >
>>>>>>>> > Clear view of the sky is king for Starlink reliability. I've got
>>>>>>>> my dishy mounted on the back fence, looking up over an empty field, so it's
>>>>>>>> pretty much best-case scenario here.
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> --
>>>>>>>> >>
>>>>>>>> >> ****************************************************************
>>>>>>>> >> Dr. Ulrich Speidel
>>>>>>>> >>
>>>>>>>> >> School of Computer Science
>>>>>>>> >>
>>>>>>>> >> Room 303S.594 (City Campus)
>>>>>>>> >>
>>>>>>>> >> The University of Auckland
>>>>>>>> >> u.speidel@auckland.ac.nz
>>>>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>>>>> >> ****************************************************************
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> _______________________________________________
>>>>>>>> >> Starlink mailing list
>>>>>>>> >> Starlink@lists.bufferbloat.net
>>>>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>> >
>>>>>>>> > _______________________________________________
>>>>>>>> > Starlink mailing list
>>>>>>>> > Starlink@lists.bufferbloat.net
>>>>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>>>
>>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Starlink mailing list
>>>>>>> Starlink@lists.bufferbloat.net
>>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>
>>>>>>
>>>>
>>>> --
>>>> This song goes out to all the folk that thought Stadia would work:
>>>>
>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>> Dave Täht CEO, TekLibre, LLC
>>>> _______________________________________________
>>>> Starlink mailing list
>>>> Starlink@lists.bufferbloat.net
>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>
>>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink
>>
>

[-- Attachment #1.2: Type: text/html, Size: 19177 bytes --]

[-- Attachment #2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #3: Screenshot 2023-01-13 at 1.30.03 PM.png --]
[-- Type: image/png, Size: 855840 bytes --]

[-- Attachment #4: image.png --]
[-- Type: image/png, Size: 296989 bytes --]

[-- Attachment #5: image.png --]
[-- Type: image/png, Size: 222655 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 22:09                                                 ` Jonathan Bennett
  2023-01-13 22:30                                                   ` Luis A. Cornejo
@ 2023-01-13 22:32                                                   ` Dave Taht
  2023-01-13 22:36                                                     ` Luis A. Cornejo
  2023-01-13 22:42                                                     ` Jonathan Bennett
  2023-01-13 23:44                                                   ` Dave Taht
  2 siblings, 2 replies; 44+ messages in thread
From: Dave Taht @ 2023-01-13 22:32 UTC (permalink / raw)
  To: Jonathan Bennett; +Cc: Nathan Owens, starlink


[-- Attachment #1.1.1: Type: text/plain, Size: 15803 bytes --]

thank you all. in both the flent cases there appears to be no tcp_rtt
statistics, did you run with --socket-stats?

(That seems to be a new bug, both with sampling correctly, and it's either
in newer linuxes or in flent itself. I hate to ask ya but could you install
the git version of flent?)

Thank you for the packet capture!!!! I'm still downloading.

Anyway, the "celabon" flent plot clearly shows the inverse relationship
between latency and throughput still in this starlink terminal, so there is
no AQM in play there, darn it. (in my fq_codeled world the latency stays
flat, only the throughput changes)

So I am incredibly puzzled now at the ostensibly awesome waveform test
result (and need to look at that capture, and/or get tcp rtt stats)

The other plot (luis's) shows incredibly consistent latency and bounded
throughput at about 6mbit.

Patiently awaiting that download to complete.


On Fri, Jan 13, 2023 at 2:10 PM Jonathan Bennett via Starlink <
starlink@lists.bufferbloat.net> wrote:

> The irtt run finished a few seconds before the flent run, but here are the
> results:
>
>
> https://drive.google.com/file/d/1FKve13ssUMW1LLWOXLM2931Yx6uMHw8K/view?usp=share_link
>
> https://drive.google.com/file/d/1ZXd64A0pfUedLr3FyhDNTHA7vxv8S2Gk/view?usp=share_link
>
> https://drive.google.com/file/d/1rx64UPQHHz3IMNiJtb1oFqtqw2DvhvEE/view?usp=share_link
>
>
> [image: image.png]
> [image: image.png]
>
>
> Jonathan Bennett
> Hackaday.com
>
>
> On Fri, Jan 13, 2023 at 3:30 PM Nathan Owens via Starlink <
> starlink@lists.bufferbloat.net> wrote:
>
>> Here's Luis's run -- the top line below the edge of the graph is 200ms
>> [image: Screenshot 2023-01-13 at 1.30.03 PM.png]
>>
>>
>> On Fri, Jan 13, 2023 at 1:25 PM Luis A. Cornejo <luis.a.cornejo@gmail.com>
>> wrote:
>>
>>> Dave,
>>>
>>> Here is a run the way I think you wanted it.
>>>
>>> irtt running for 5 min to your dallas server, followed by a waveform
>>> test, then a few seconds of inactivity, cloudflare test, a few more secs of
>>> nothing, flent test to dallas. Packet capture is currently uploading (will
>>> be done in 20 min or so), irtt JSON also in there (.zip file):
>>>
>>>
>>> https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link
>>>
>>> -Luis
>>>
>>> On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
>>> starlink@lists.bufferbloat.net> wrote:
>>>
>>>>
>>>>
>>>> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io> wrote:
>>>>
>>>>> Here's the data visualization for Johnathan's Data
>>>>>
>>>>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>>>>
>>>>> You can see the path change at :12, :27, :42, :57 after the minute.
>>>>> Some paths are clearly busier than others with increased loss, latency, and
>>>>> jitter.
>>>>>
>>>>
>>>> I am so glad to see loss and bounded delay here. Also a bit of rigor
>>>> regarding what traffic was active locally vs on the path would be nice,
>>>> although it seems to line up with the known 15s starlink switchover thing
>>>> (need a name for this), in this case, doing a few speedtests
>>>> while that irtt is running would show the impact(s) of whatever else
>>>> they are up to.
>>>>
>>>> It would also be my hope that the loss distribution in the middle
>>>> portion of this data is good, not bursty, but we don't have a tool to take
>>>> apart that. (I am so hopeless at json)
>>>>
>>>>
>>>>
>>>>>
>>>>>
>>>>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io>
>>>>> wrote:
>>>>>
>>>>>> I’ll run my visualization code on this result this afternoon and
>>>>>> report back!
>>>>>>
>>>>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>
>>>>>>> The irtt command, run with normal, light usage:
>>>>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>>>>
>>>>>>> Jonathan Bennett
>>>>>>> Hackaday.com
>>>>>>>
>>>>>>>
>>>>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> packet caps would be nice... all this is very exciting news.
>>>>>>>>
>>>>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>>>>> results nowadays to duplicate and re-plot the original irtt tests we
>>>>>>>> did:
>>>>>>>>
>>>>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
>>>>>>>> whatever.json
>>>>>>>>
>>>>>>>> They MUST have changed their scheduling to get such amazing uplink
>>>>>>>> results, in addition to better queue management.
>>>>>>>>
>>>>>>>> (for the record, my servers are de, london, fremont, sydney, dallas,
>>>>>>>> newark, atlanta, singapore, mumbai)
>>>>>>>>
>>>>>>>> There's an R and gnuplot script for plotting that output around here
>>>>>>>> somewhere (I have largely personally put down the starlink project,
>>>>>>>> loaning out mine) - that went by on this list... I should have
>>>>>>>> written
>>>>>>>> a blog entry so I can find that stuff again.
>>>>>>>>
>>>>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>>>>> >
>>>>>>>> >
>>>>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>> >>
>>>>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>>>>> >> >
>>>>>>>> >> > From Auckland, New Zealand, using a roaming subscription, it
>>>>>>>> puts me
>>>>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>>>>> >> >
>>>>>>>> >> >
>>>>>>>> >> > IP address: nix six.
>>>>>>>> >> >
>>>>>>>> >> > My thoughts shall follow later.
>>>>>>>> >>
>>>>>>>> >> OK, so here we go.
>>>>>>>> >>
>>>>>>>> >> I'm always a bit skeptical when it comes to speed tests -
>>>>>>>> they're really
>>>>>>>> >> laden with so many caveats that it's not funny. I took our new
>>>>>>>> work
>>>>>>>> >> Starlink kit home in December to give it a try and the other day
>>>>>>>> finally
>>>>>>>> >> got around to set it up. It's on a roaming subscription because
>>>>>>>> our
>>>>>>>> >> badly built-up campus really isn't ideal in terms of a clear
>>>>>>>> view of the
>>>>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>>>>> adapter,
>>>>>>>> >> not the WiFi?
>>>>>>>> >>
>>>>>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>>>>>> Starlink
>>>>>>>> >> promises is about 1/3 in terms of data rate you can actually get
>>>>>>>> from
>>>>>>>> >> fibre to the home at under half of Starlink's price. Read: There
>>>>>>>> are few
>>>>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>>>>> >>
>>>>>>>> >> Caveat 2: Auckland has three Starlink gateways close by:
>>>>>>>> Clevedon (which
>>>>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana and
>>>>>>>> Puwera,
>>>>>>>> >> the most distant of the three and about 130 km away from me as
>>>>>>>> the crow
>>>>>>>> >> flies. Read: My dishy can use any satellite that any of these
>>>>>>>> three can
>>>>>>>> >> see, and then depending on where I put it and how much of the
>>>>>>>> southern
>>>>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away,
>>>>>>>> although that
>>>>>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>>>>>> options
>>>>>>>> >> for my bits to travel without needing a lot of handovers. Why?
>>>>>>>> Easy: If
>>>>>>>> >> your nearest teleport is close by, then the set of satellites
>>>>>>>> that the
>>>>>>>> >> teleport can see and the set that you can see is almost the
>>>>>>>> same, so you
>>>>>>>> >> can essentially stick with the same satellite while it's in view
>>>>>>>> for you
>>>>>>>> >> because it'll also be in view for the teleport. Pretty much any
>>>>>>>> bird
>>>>>>>> >> above you will do.
>>>>>>>> >>
>>>>>>>> >> And because I don't get a lot of competition from other users in
>>>>>>>> my area
>>>>>>>> >> vying for one of the few available satellites that can see both
>>>>>>>> us and
>>>>>>>> >> the teleport, this is about as good as it gets at 37S latitude.
>>>>>>>> If I'd
>>>>>>>> >> want it any better, I'd have to move a lot further south.
>>>>>>>> >>
>>>>>>>> >> It'd be interesting to hear from Jonathan what the availability
>>>>>>>> of home
>>>>>>>> >> broadband is like in the Dallas area. I note that it's at a lower
>>>>>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I
>>>>>>>> notice
>>>>>>>> >> two teleports each about 160 km away, which is also not too bad.
>>>>>>>> I also
>>>>>>>> >> note Starlink availability in the area is restricted at the
>>>>>>>> moment -
>>>>>>>> >> oversubscribed? But if Jonathan gets good data rates, then that
>>>>>>>> means
>>>>>>>> >> that competition for bird capacity can't be too bad - for
>>>>>>>> whatever reason.
>>>>>>>> >
>>>>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>>>>> broadband options. But there are also many people that live outside cities,
>>>>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>>>>> and Texas is probably ideal conditions for Starlink.
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>>>>> whatever I
>>>>>>>> >> talk to in terms of my communications. Traceroute shows about 10
>>>>>>>> hops
>>>>>>>> >> between me and the University of Auckland via Starlink. That's 10
>>>>>>>> >> queues, not one. Many of them will have cross traffic. So it's a
>>>>>>>> bit
>>>>>>>> >> hard to tell where our packets really get to wait or where they
>>>>>>>> get
>>>>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>>>>> between 1
>>>>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they
>>>>>>>> can all
>>>>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link of
>>>>>>>> a few
>>>>>>>> >> Mb/s where it's obvious where the dominant long latency
>>>>>>>> bottleneck in
>>>>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>>>>> whether a
>>>>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>>>>> system or
>>>>>>>> >> somewhere between Starlink's systems and the target system.
>>>>>>>> >>
>>>>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical
>>>>>>>> latency
>>>>>>>> >> should be under 15 ms. So there's clearly a bit of buffer here
>>>>>>>> along the
>>>>>>>> >> chain that occasionally fills up.
>>>>>>>> >>
>>>>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>>>>> inevitably
>>>>>>>> >> associated with a change in RTT and in most cases also available
>>>>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>>>>>> >> satellite while your TCP is still trying to respond to whatever
>>>>>>>> it
>>>>>>>> >> thought the queue on the previous bird was doing. Read: Whatever
>>>>>>>> your
>>>>>>>> >> cwnd is immediately after a handover, it's probably not what it
>>>>>>>> should be.
>>>>>>>> >>
>>>>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four
>>>>>>>> Ookla
>>>>>>>> >> speedtest.net tests each to five local servers. Average upload
>>>>>>>> rate was
>>>>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP
>>>>>>>> that
>>>>>>>> >> Starlink seems to be buying its local connectivity from (Vocus
>>>>>>>> Group)
>>>>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and
>>>>>>>> 52.22
>>>>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to
>>>>>>>> rates
>>>>>>>> >> observed. In fact, they were the ISP with consistently the worst
>>>>>>>> rates.
>>>>>>>> >>
>>>>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up
>>>>>>>> and
>>>>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly
>>>>>>>> correlating
>>>>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>>>>> >>
>>>>>>>> >> Note the variation though: More or less a factor of two between
>>>>>>>> highest
>>>>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in
>>>>>>>> my
>>>>>>>> >> tests? Or is there something systematic behind this? Way too few
>>>>>>>> tests
>>>>>>>> >> to tell.
>>>>>>>> >>
>>>>>>>> >> What these tests do is establish a ballpark.
>>>>>>>> >>
>>>>>>>> >> I'm currently repeating tests with dish placed on a trestle
>>>>>>>> closer to
>>>>>>>> >> the heavens. This seems to have translated into fewer outages /
>>>>>>>> ping
>>>>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the
>>>>>>>> ground on
>>>>>>>> >> my deck). Still good enough for a lengthy video Skype call with
>>>>>>>> my folks
>>>>>>>> >> in Germany, although they did comment about reduced video
>>>>>>>> quality. But
>>>>>>>> >> maybe that was the lighting or the different background as I
>>>>>>>> wasn't in
>>>>>>>> >> my usual spot with my laptop when I called them.
>>>>>>>> >
>>>>>>>> > Clear view of the sky is king for Starlink reliability. I've got
>>>>>>>> my dishy mounted on the back fence, looking up over an empty field, so it's
>>>>>>>> pretty much best-case scenario here.
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> --
>>>>>>>> >>
>>>>>>>> >> ****************************************************************
>>>>>>>> >> Dr. Ulrich Speidel
>>>>>>>> >>
>>>>>>>> >> School of Computer Science
>>>>>>>> >>
>>>>>>>> >> Room 303S.594 (City Campus)
>>>>>>>> >>
>>>>>>>> >> The University of Auckland
>>>>>>>> >> u.speidel@auckland.ac.nz
>>>>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>>>>> >> ****************************************************************
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> _______________________________________________
>>>>>>>> >> Starlink mailing list
>>>>>>>> >> Starlink@lists.bufferbloat.net
>>>>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>> >
>>>>>>>> > _______________________________________________
>>>>>>>> > Starlink mailing list
>>>>>>>> > Starlink@lists.bufferbloat.net
>>>>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>>>
>>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Starlink mailing list
>>>>>>> Starlink@lists.bufferbloat.net
>>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>
>>>>>>
>>>>
>>>> --
>>>> This song goes out to all the folk that thought Stadia would work:
>>>>
>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>> Dave Täht CEO, TekLibre, LLC
>>>> _______________________________________________
>>>> Starlink mailing list
>>>> Starlink@lists.bufferbloat.net
>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>
>>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink
>>
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink
>


-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

[-- Attachment #1.1.2: Type: text/html, Size: 20910 bytes --]

[-- Attachment #1.2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #1.3: Screenshot 2023-01-13 at 1.30.03 PM.png --]
[-- Type: image/png, Size: 855840 bytes --]

[-- Attachment #1.4: image.png --]
[-- Type: image/png, Size: 296989 bytes --]

[-- Attachment #1.5: image.png --]
[-- Type: image/png, Size: 222655 bytes --]

[-- Attachment #2: tcp_nup_-_starlink_vs_irtt.png --]
[-- Type: image/png, Size: 205282 bytes --]

[-- Attachment #3: tcp_nup_-_2starlink_vs_irtt.png --]
[-- Type: image/png, Size: 271672 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 22:32                                                   ` Dave Taht
@ 2023-01-13 22:36                                                     ` Luis A. Cornejo
  2023-01-13 22:42                                                     ` Jonathan Bennett
  1 sibling, 0 replies; 44+ messages in thread
From: Luis A. Cornejo @ 2023-01-13 22:36 UTC (permalink / raw)
  To: Dave Taht; +Cc: Jonathan Bennett, starlink


[-- Attachment #1.1: Type: text/plain, Size: 16696 bytes --]

I ran it with:

flent -H dallas.starlink.taht.net -t starlink_vs_irtt --step-size=.05
--socket-stats --test-parameter=upload_streams=4 tcp_nup

On Fri, Jan 13, 2023 at 4:33 PM Dave Taht via Starlink <
starlink@lists.bufferbloat.net> wrote:

> thank you all. in both the flent cases there appears to be no tcp_rtt
> statistics, did you run with --socket-stats?
>
> (That seems to be a new bug, both with sampling correctly, and it's either
> in newer linuxes or in flent itself. I hate to ask ya but could you install
> the git version of flent?)
>
> Thank you for the packet capture!!!! I'm still downloading.
>
> Anyway, the "celabon" flent plot clearly shows the inverse relationship
> between latency and throughput still in this starlink terminal, so there is
> no AQM in play there, darn it. (in my fq_codeled world the latency stays
> flat, only the throughput changes)
>
> So I am incredibly puzzled now at the ostensibly awesome waveform test
> result (and need to look at that capture, and/or get tcp rtt stats)
>
> The other plot (luis's) shows incredibly consistent latency and bounded
> throughput at about 6mbit.
>
> Patiently awaiting that download to complete.
>
>
> On Fri, Jan 13, 2023 at 2:10 PM Jonathan Bennett via Starlink <
> starlink@lists.bufferbloat.net> wrote:
>
>> The irtt run finished a few seconds before the flent run, but here are
>> the results:
>>
>>
>> https://drive.google.com/file/d/1FKve13ssUMW1LLWOXLM2931Yx6uMHw8K/view?usp=share_link
>>
>> https://drive.google.com/file/d/1ZXd64A0pfUedLr3FyhDNTHA7vxv8S2Gk/view?usp=share_link
>>
>> https://drive.google.com/file/d/1rx64UPQHHz3IMNiJtb1oFqtqw2DvhvEE/view?usp=share_link
>>
>>
>> [image: image.png]
>> [image: image.png]
>>
>>
>> Jonathan Bennett
>> Hackaday.com
>>
>>
>> On Fri, Jan 13, 2023 at 3:30 PM Nathan Owens via Starlink <
>> starlink@lists.bufferbloat.net> wrote:
>>
>>> Here's Luis's run -- the top line below the edge of the graph is 200ms
>>> [image: Screenshot 2023-01-13 at 1.30.03 PM.png]
>>>
>>>
>>> On Fri, Jan 13, 2023 at 1:25 PM Luis A. Cornejo <
>>> luis.a.cornejo@gmail.com> wrote:
>>>
>>>> Dave,
>>>>
>>>> Here is a run the way I think you wanted it.
>>>>
>>>> irtt running for 5 min to your dallas server, followed by a waveform
>>>> test, then a few seconds of inactivity, cloudflare test, a few more secs of
>>>> nothing, flent test to dallas. Packet capture is currently uploading (will
>>>> be done in 20 min or so), irtt JSON also in there (.zip file):
>>>>
>>>>
>>>> https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link
>>>>
>>>> -Luis
>>>>
>>>> On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
>>>> starlink@lists.bufferbloat.net> wrote:
>>>>
>>>>>
>>>>>
>>>>> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io>
>>>>> wrote:
>>>>>
>>>>>> Here's the data visualization for Johnathan's Data
>>>>>>
>>>>>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>>>>>
>>>>>> You can see the path change at :12, :27, :42, :57 after the minute.
>>>>>> Some paths are clearly busier than others with increased loss, latency, and
>>>>>> jitter.
>>>>>>
>>>>>
>>>>> I am so glad to see loss and bounded delay here. Also a bit of rigor
>>>>> regarding what traffic was active locally vs on the path would be nice,
>>>>> although it seems to line up with the known 15s starlink switchover thing
>>>>> (need a name for this), in this case, doing a few speedtests
>>>>> while that irtt is running would show the impact(s) of whatever else
>>>>> they are up to.
>>>>>
>>>>> It would also be my hope that the loss distribution in the middle
>>>>> portion of this data is good, not bursty, but we don't have a tool to take
>>>>> apart that. (I am so hopeless at json)
>>>>>
>>>>>
>>>>>
>>>>>>
>>>>>>
>>>>>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io>
>>>>>> wrote:
>>>>>>
>>>>>>> I’ll run my visualization code on this result this afternoon and
>>>>>>> report back!
>>>>>>>
>>>>>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>
>>>>>>>> The irtt command, run with normal, light usage:
>>>>>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>>>>>
>>>>>>>> Jonathan Bennett
>>>>>>>> Hackaday.com
>>>>>>>>
>>>>>>>>
>>>>>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> packet caps would be nice... all this is very exciting news.
>>>>>>>>>
>>>>>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>>>>>> results nowadays to duplicate and re-plot the original irtt tests
>>>>>>>>> we
>>>>>>>>> did:
>>>>>>>>>
>>>>>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net
>>>>>>>>> -o whatever.json
>>>>>>>>>
>>>>>>>>> They MUST have changed their scheduling to get such amazing uplink
>>>>>>>>> results, in addition to better queue management.
>>>>>>>>>
>>>>>>>>> (for the record, my servers are de, london, fremont, sydney,
>>>>>>>>> dallas,
>>>>>>>>> newark, atlanta, singapore, mumbai)
>>>>>>>>>
>>>>>>>>> There's an R and gnuplot script for plotting that output around
>>>>>>>>> here
>>>>>>>>> somewhere (I have largely personally put down the starlink project,
>>>>>>>>> loaning out mine) - that went by on this list... I should have
>>>>>>>>> written
>>>>>>>>> a blog entry so I can find that stuff again.
>>>>>>>>>
>>>>>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>>>>>> >
>>>>>>>>> >
>>>>>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>>> >>
>>>>>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>>>>>> >> >
>>>>>>>>> >> > From Auckland, New Zealand, using a roaming subscription, it
>>>>>>>>> puts me
>>>>>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>>>>>> >> >
>>>>>>>>> >> >
>>>>>>>>> >> > IP address: nix six.
>>>>>>>>> >> >
>>>>>>>>> >> > My thoughts shall follow later.
>>>>>>>>> >>
>>>>>>>>> >> OK, so here we go.
>>>>>>>>> >>
>>>>>>>>> >> I'm always a bit skeptical when it comes to speed tests -
>>>>>>>>> they're really
>>>>>>>>> >> laden with so many caveats that it's not funny. I took our new
>>>>>>>>> work
>>>>>>>>> >> Starlink kit home in December to give it a try and the other
>>>>>>>>> day finally
>>>>>>>>> >> got around to set it up. It's on a roaming subscription because
>>>>>>>>> our
>>>>>>>>> >> badly built-up campus really isn't ideal in terms of a clear
>>>>>>>>> view of the
>>>>>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>>>>>> adapter,
>>>>>>>>> >> not the WiFi?
>>>>>>>>> >>
>>>>>>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>>>>>>> Starlink
>>>>>>>>> >> promises is about 1/3 in terms of data rate you can actually
>>>>>>>>> get from
>>>>>>>>> >> fibre to the home at under half of Starlink's price. Read:
>>>>>>>>> There are few
>>>>>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>>>>>> >>
>>>>>>>>> >> Caveat 2: Auckland has three Starlink gateways close by:
>>>>>>>>> Clevedon (which
>>>>>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana
>>>>>>>>> and Puwera,
>>>>>>>>> >> the most distant of the three and about 130 km away from me as
>>>>>>>>> the crow
>>>>>>>>> >> flies. Read: My dishy can use any satellite that any of these
>>>>>>>>> three can
>>>>>>>>> >> see, and then depending on where I put it and how much of the
>>>>>>>>> southern
>>>>>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away,
>>>>>>>>> although that
>>>>>>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>>>>>>> options
>>>>>>>>> >> for my bits to travel without needing a lot of handovers. Why?
>>>>>>>>> Easy: If
>>>>>>>>> >> your nearest teleport is close by, then the set of satellites
>>>>>>>>> that the
>>>>>>>>> >> teleport can see and the set that you can see is almost the
>>>>>>>>> same, so you
>>>>>>>>> >> can essentially stick with the same satellite while it's in
>>>>>>>>> view for you
>>>>>>>>> >> because it'll also be in view for the teleport. Pretty much any
>>>>>>>>> bird
>>>>>>>>> >> above you will do.
>>>>>>>>> >>
>>>>>>>>> >> And because I don't get a lot of competition from other users
>>>>>>>>> in my area
>>>>>>>>> >> vying for one of the few available satellites that can see both
>>>>>>>>> us and
>>>>>>>>> >> the teleport, this is about as good as it gets at 37S latitude.
>>>>>>>>> If I'd
>>>>>>>>> >> want it any better, I'd have to move a lot further south.
>>>>>>>>> >>
>>>>>>>>> >> It'd be interesting to hear from Jonathan what the availability
>>>>>>>>> of home
>>>>>>>>> >> broadband is like in the Dallas area. I note that it's at a
>>>>>>>>> lower
>>>>>>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I
>>>>>>>>> notice
>>>>>>>>> >> two teleports each about 160 km away, which is also not too
>>>>>>>>> bad. I also
>>>>>>>>> >> note Starlink availability in the area is restricted at the
>>>>>>>>> moment -
>>>>>>>>> >> oversubscribed? But if Jonathan gets good data rates, then that
>>>>>>>>> means
>>>>>>>>> >> that competition for bird capacity can't be too bad - for
>>>>>>>>> whatever reason.
>>>>>>>>> >
>>>>>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>>>>>> broadband options. But there are also many people that live outside cities,
>>>>>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>>>>>> and Texas is probably ideal conditions for Starlink.
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>>>>>> whatever I
>>>>>>>>> >> talk to in terms of my communications. Traceroute shows about
>>>>>>>>> 10 hops
>>>>>>>>> >> between me and the University of Auckland via Starlink. That's
>>>>>>>>> 10
>>>>>>>>> >> queues, not one. Many of them will have cross traffic. So it's
>>>>>>>>> a bit
>>>>>>>>> >> hard to tell where our packets really get to wait or where they
>>>>>>>>> get
>>>>>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>>>>>> between 1
>>>>>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they
>>>>>>>>> can all
>>>>>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link of
>>>>>>>>> a few
>>>>>>>>> >> Mb/s where it's obvious where the dominant long latency
>>>>>>>>> bottleneck in
>>>>>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>>>>>> whether a
>>>>>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>>>>>> system or
>>>>>>>>> >> somewhere between Starlink's systems and the target system.
>>>>>>>>> >>
>>>>>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical
>>>>>>>>> latency
>>>>>>>>> >> should be under 15 ms. So there's clearly a bit of buffer here
>>>>>>>>> along the
>>>>>>>>> >> chain that occasionally fills up.
>>>>>>>>> >>
>>>>>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>>>>>> inevitably
>>>>>>>>> >> associated with a change in RTT and in most cases also available
>>>>>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>>>>>>> >> satellite while your TCP is still trying to respond to whatever
>>>>>>>>> it
>>>>>>>>> >> thought the queue on the previous bird was doing. Read:
>>>>>>>>> Whatever your
>>>>>>>>> >> cwnd is immediately after a handover, it's probably not what it
>>>>>>>>> should be.
>>>>>>>>> >>
>>>>>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four
>>>>>>>>> Ookla
>>>>>>>>> >> speedtest.net tests each to five local servers. Average upload
>>>>>>>>> rate was
>>>>>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the
>>>>>>>>> ISP that
>>>>>>>>> >> Starlink seems to be buying its local connectivity from (Vocus
>>>>>>>>> Group)
>>>>>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and
>>>>>>>>> 52.22
>>>>>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to
>>>>>>>>> rates
>>>>>>>>> >> observed. In fact, they were the ISP with consistently the
>>>>>>>>> worst rates.
>>>>>>>>> >>
>>>>>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up
>>>>>>>>> and
>>>>>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly
>>>>>>>>> correlating
>>>>>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>>>>>> >>
>>>>>>>>> >> Note the variation though: More or less a factor of two between
>>>>>>>>> highest
>>>>>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in
>>>>>>>>> my
>>>>>>>>> >> tests? Or is there something systematic behind this? Way too
>>>>>>>>> few tests
>>>>>>>>> >> to tell.
>>>>>>>>> >>
>>>>>>>>> >> What these tests do is establish a ballpark.
>>>>>>>>> >>
>>>>>>>>> >> I'm currently repeating tests with dish placed on a trestle
>>>>>>>>> closer to
>>>>>>>>> >> the heavens. This seems to have translated into fewer outages /
>>>>>>>>> ping
>>>>>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the
>>>>>>>>> ground on
>>>>>>>>> >> my deck). Still good enough for a lengthy video Skype call with
>>>>>>>>> my folks
>>>>>>>>> >> in Germany, although they did comment about reduced video
>>>>>>>>> quality. But
>>>>>>>>> >> maybe that was the lighting or the different background as I
>>>>>>>>> wasn't in
>>>>>>>>> >> my usual spot with my laptop when I called them.
>>>>>>>>> >
>>>>>>>>> > Clear view of the sky is king for Starlink reliability. I've got
>>>>>>>>> my dishy mounted on the back fence, looking up over an empty field, so it's
>>>>>>>>> pretty much best-case scenario here.
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >> --
>>>>>>>>> >>
>>>>>>>>> >> ****************************************************************
>>>>>>>>> >> Dr. Ulrich Speidel
>>>>>>>>> >>
>>>>>>>>> >> School of Computer Science
>>>>>>>>> >>
>>>>>>>>> >> Room 303S.594 (City Campus)
>>>>>>>>> >>
>>>>>>>>> >> The University of Auckland
>>>>>>>>> >> u.speidel@auckland.ac.nz
>>>>>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>>>>>> >> ****************************************************************
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >> _______________________________________________
>>>>>>>>> >> Starlink mailing list
>>>>>>>>> >> Starlink@lists.bufferbloat.net
>>>>>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>> >
>>>>>>>>> > _______________________________________________
>>>>>>>>> > Starlink mailing list
>>>>>>>>> > Starlink@lists.bufferbloat.net
>>>>>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>>>>
>>>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Starlink mailing list
>>>>>>>> Starlink@lists.bufferbloat.net
>>>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>
>>>>>>>
>>>>>
>>>>> --
>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>
>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>> Dave Täht CEO, TekLibre, LLC
>>>>> _______________________________________________
>>>>> Starlink mailing list
>>>>> Starlink@lists.bufferbloat.net
>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>
>>>> _______________________________________________
>>> Starlink mailing list
>>> Starlink@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/starlink
>>>
>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink
>>
>
>
> --
> This song goes out to all the folk that thought Stadia would work:
>
> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
> Dave Täht CEO, TekLibre, LLC
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink
>

[-- Attachment #1.2: Type: text/html, Size: 21845 bytes --]

[-- Attachment #2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #3: Screenshot 2023-01-13 at 1.30.03 PM.png --]
[-- Type: image/png, Size: 855840 bytes --]

[-- Attachment #4: image.png --]
[-- Type: image/png, Size: 296989 bytes --]

[-- Attachment #5: image.png --]
[-- Type: image/png, Size: 222655 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 22:32                                                   ` Dave Taht
  2023-01-13 22:36                                                     ` Luis A. Cornejo
@ 2023-01-13 22:42                                                     ` Jonathan Bennett
  2023-01-13 22:49                                                       ` Dave Taht
  1 sibling, 1 reply; 44+ messages in thread
From: Jonathan Bennett @ 2023-01-13 22:42 UTC (permalink / raw)
  To: Dave Taht; +Cc: Nathan Owens, starlink


[-- Attachment #1.1.1: Type: text/plain, Size: 16562 bytes --]

Flent run using git, and ./run-flent -H atlanta.starlink.taht.net -t
starlink_vs_irtt --step-size=.05 --socket-stats
--test-parameter=upload_streams=4 tcp_nup


Jonathan Bennett
Hackaday.com


On Fri, Jan 13, 2023 at 4:33 PM Dave Taht <dave.taht@gmail.com> wrote:

> thank you all. in both the flent cases there appears to be no tcp_rtt
> statistics, did you run with --socket-stats?
>
> (That seems to be a new bug, both with sampling correctly, and it's either
> in newer linuxes or in flent itself. I hate to ask ya but could you install
> the git version of flent?)
>
> Thank you for the packet capture!!!! I'm still downloading.
>
> Anyway, the "celabon" flent plot clearly shows the inverse relationship
> between latency and throughput still in this starlink terminal, so there is
> no AQM in play there, darn it. (in my fq_codeled world the latency stays
> flat, only the throughput changes)
>
> So I am incredibly puzzled now at the ostensibly awesome waveform test
> result (and need to look at that capture, and/or get tcp rtt stats)
>
> The other plot (luis's) shows incredibly consistent latency and bounded
> throughput at about 6mbit.
>
> Patiently awaiting that download to complete.
>
>
> On Fri, Jan 13, 2023 at 2:10 PM Jonathan Bennett via Starlink <
> starlink@lists.bufferbloat.net> wrote:
>
>> The irtt run finished a few seconds before the flent run, but here are
>> the results:
>>
>>
>> https://drive.google.com/file/d/1FKve13ssUMW1LLWOXLM2931Yx6uMHw8K/view?usp=share_link
>>
>> https://drive.google.com/file/d/1ZXd64A0pfUedLr3FyhDNTHA7vxv8S2Gk/view?usp=share_link
>>
>> https://drive.google.com/file/d/1rx64UPQHHz3IMNiJtb1oFqtqw2DvhvEE/view?usp=share_link
>>
>>
>> [image: image.png]
>> [image: image.png]
>>
>>
>> Jonathan Bennett
>> Hackaday.com
>>
>>
>> On Fri, Jan 13, 2023 at 3:30 PM Nathan Owens via Starlink <
>> starlink@lists.bufferbloat.net> wrote:
>>
>>> Here's Luis's run -- the top line below the edge of the graph is 200ms
>>> [image: Screenshot 2023-01-13 at 1.30.03 PM.png]
>>>
>>>
>>> On Fri, Jan 13, 2023 at 1:25 PM Luis A. Cornejo <
>>> luis.a.cornejo@gmail.com> wrote:
>>>
>>>> Dave,
>>>>
>>>> Here is a run the way I think you wanted it.
>>>>
>>>> irtt running for 5 min to your dallas server, followed by a waveform
>>>> test, then a few seconds of inactivity, cloudflare test, a few more secs of
>>>> nothing, flent test to dallas. Packet capture is currently uploading (will
>>>> be done in 20 min or so), irtt JSON also in there (.zip file):
>>>>
>>>>
>>>> https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link
>>>>
>>>> -Luis
>>>>
>>>> On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
>>>> starlink@lists.bufferbloat.net> wrote:
>>>>
>>>>>
>>>>>
>>>>> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io>
>>>>> wrote:
>>>>>
>>>>>> Here's the data visualization for Johnathan's Data
>>>>>>
>>>>>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>>>>>
>>>>>> You can see the path change at :12, :27, :42, :57 after the minute.
>>>>>> Some paths are clearly busier than others with increased loss, latency, and
>>>>>> jitter.
>>>>>>
>>>>>
>>>>> I am so glad to see loss and bounded delay here. Also a bit of rigor
>>>>> regarding what traffic was active locally vs on the path would be nice,
>>>>> although it seems to line up with the known 15s starlink switchover thing
>>>>> (need a name for this), in this case, doing a few speedtests
>>>>> while that irtt is running would show the impact(s) of whatever else
>>>>> they are up to.
>>>>>
>>>>> It would also be my hope that the loss distribution in the middle
>>>>> portion of this data is good, not bursty, but we don't have a tool to take
>>>>> apart that. (I am so hopeless at json)
>>>>>
>>>>>
>>>>>
>>>>>>
>>>>>>
>>>>>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io>
>>>>>> wrote:
>>>>>>
>>>>>>> I’ll run my visualization code on this result this afternoon and
>>>>>>> report back!
>>>>>>>
>>>>>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>
>>>>>>>> The irtt command, run with normal, light usage:
>>>>>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>>>>>
>>>>>>>> Jonathan Bennett
>>>>>>>> Hackaday.com
>>>>>>>>
>>>>>>>>
>>>>>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> packet caps would be nice... all this is very exciting news.
>>>>>>>>>
>>>>>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>>>>>> results nowadays to duplicate and re-plot the original irtt tests
>>>>>>>>> we
>>>>>>>>> did:
>>>>>>>>>
>>>>>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net
>>>>>>>>> -o whatever.json
>>>>>>>>>
>>>>>>>>> They MUST have changed their scheduling to get such amazing uplink
>>>>>>>>> results, in addition to better queue management.
>>>>>>>>>
>>>>>>>>> (for the record, my servers are de, london, fremont, sydney,
>>>>>>>>> dallas,
>>>>>>>>> newark, atlanta, singapore, mumbai)
>>>>>>>>>
>>>>>>>>> There's an R and gnuplot script for plotting that output around
>>>>>>>>> here
>>>>>>>>> somewhere (I have largely personally put down the starlink project,
>>>>>>>>> loaning out mine) - that went by on this list... I should have
>>>>>>>>> written
>>>>>>>>> a blog entry so I can find that stuff again.
>>>>>>>>>
>>>>>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>>>>>> >
>>>>>>>>> >
>>>>>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>>> >>
>>>>>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>>>>>> >> >
>>>>>>>>> >> > From Auckland, New Zealand, using a roaming subscription, it
>>>>>>>>> puts me
>>>>>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>>>>>> >> >
>>>>>>>>> >> >
>>>>>>>>> >> > IP address: nix six.
>>>>>>>>> >> >
>>>>>>>>> >> > My thoughts shall follow later.
>>>>>>>>> >>
>>>>>>>>> >> OK, so here we go.
>>>>>>>>> >>
>>>>>>>>> >> I'm always a bit skeptical when it comes to speed tests -
>>>>>>>>> they're really
>>>>>>>>> >> laden with so many caveats that it's not funny. I took our new
>>>>>>>>> work
>>>>>>>>> >> Starlink kit home in December to give it a try and the other
>>>>>>>>> day finally
>>>>>>>>> >> got around to set it up. It's on a roaming subscription because
>>>>>>>>> our
>>>>>>>>> >> badly built-up campus really isn't ideal in terms of a clear
>>>>>>>>> view of the
>>>>>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>>>>>> adapter,
>>>>>>>>> >> not the WiFi?
>>>>>>>>> >>
>>>>>>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>>>>>>> Starlink
>>>>>>>>> >> promises is about 1/3 in terms of data rate you can actually
>>>>>>>>> get from
>>>>>>>>> >> fibre to the home at under half of Starlink's price. Read:
>>>>>>>>> There are few
>>>>>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>>>>>> >>
>>>>>>>>> >> Caveat 2: Auckland has three Starlink gateways close by:
>>>>>>>>> Clevedon (which
>>>>>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana
>>>>>>>>> and Puwera,
>>>>>>>>> >> the most distant of the three and about 130 km away from me as
>>>>>>>>> the crow
>>>>>>>>> >> flies. Read: My dishy can use any satellite that any of these
>>>>>>>>> three can
>>>>>>>>> >> see, and then depending on where I put it and how much of the
>>>>>>>>> southern
>>>>>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away,
>>>>>>>>> although that
>>>>>>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>>>>>>> options
>>>>>>>>> >> for my bits to travel without needing a lot of handovers. Why?
>>>>>>>>> Easy: If
>>>>>>>>> >> your nearest teleport is close by, then the set of satellites
>>>>>>>>> that the
>>>>>>>>> >> teleport can see and the set that you can see is almost the
>>>>>>>>> same, so you
>>>>>>>>> >> can essentially stick with the same satellite while it's in
>>>>>>>>> view for you
>>>>>>>>> >> because it'll also be in view for the teleport. Pretty much any
>>>>>>>>> bird
>>>>>>>>> >> above you will do.
>>>>>>>>> >>
>>>>>>>>> >> And because I don't get a lot of competition from other users
>>>>>>>>> in my area
>>>>>>>>> >> vying for one of the few available satellites that can see both
>>>>>>>>> us and
>>>>>>>>> >> the teleport, this is about as good as it gets at 37S latitude.
>>>>>>>>> If I'd
>>>>>>>>> >> want it any better, I'd have to move a lot further south.
>>>>>>>>> >>
>>>>>>>>> >> It'd be interesting to hear from Jonathan what the availability
>>>>>>>>> of home
>>>>>>>>> >> broadband is like in the Dallas area. I note that it's at a
>>>>>>>>> lower
>>>>>>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I
>>>>>>>>> notice
>>>>>>>>> >> two teleports each about 160 km away, which is also not too
>>>>>>>>> bad. I also
>>>>>>>>> >> note Starlink availability in the area is restricted at the
>>>>>>>>> moment -
>>>>>>>>> >> oversubscribed? But if Jonathan gets good data rates, then that
>>>>>>>>> means
>>>>>>>>> >> that competition for bird capacity can't be too bad - for
>>>>>>>>> whatever reason.
>>>>>>>>> >
>>>>>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>>>>>> broadband options. But there are also many people that live outside cities,
>>>>>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>>>>>> and Texas is probably ideal conditions for Starlink.
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>>>>>> whatever I
>>>>>>>>> >> talk to in terms of my communications. Traceroute shows about
>>>>>>>>> 10 hops
>>>>>>>>> >> between me and the University of Auckland via Starlink. That's
>>>>>>>>> 10
>>>>>>>>> >> queues, not one. Many of them will have cross traffic. So it's
>>>>>>>>> a bit
>>>>>>>>> >> hard to tell where our packets really get to wait or where they
>>>>>>>>> get
>>>>>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>>>>>> between 1
>>>>>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they
>>>>>>>>> can all
>>>>>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link of
>>>>>>>>> a few
>>>>>>>>> >> Mb/s where it's obvious where the dominant long latency
>>>>>>>>> bottleneck in
>>>>>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>>>>>> whether a
>>>>>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>>>>>> system or
>>>>>>>>> >> somewhere between Starlink's systems and the target system.
>>>>>>>>> >>
>>>>>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical
>>>>>>>>> latency
>>>>>>>>> >> should be under 15 ms. So there's clearly a bit of buffer here
>>>>>>>>> along the
>>>>>>>>> >> chain that occasionally fills up.
>>>>>>>>> >>
>>>>>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>>>>>> inevitably
>>>>>>>>> >> associated with a change in RTT and in most cases also available
>>>>>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>>>>>>> >> satellite while your TCP is still trying to respond to whatever
>>>>>>>>> it
>>>>>>>>> >> thought the queue on the previous bird was doing. Read:
>>>>>>>>> Whatever your
>>>>>>>>> >> cwnd is immediately after a handover, it's probably not what it
>>>>>>>>> should be.
>>>>>>>>> >>
>>>>>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four
>>>>>>>>> Ookla
>>>>>>>>> >> speedtest.net tests each to five local servers. Average upload
>>>>>>>>> rate was
>>>>>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the
>>>>>>>>> ISP that
>>>>>>>>> >> Starlink seems to be buying its local connectivity from (Vocus
>>>>>>>>> Group)
>>>>>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and
>>>>>>>>> 52.22
>>>>>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to
>>>>>>>>> rates
>>>>>>>>> >> observed. In fact, they were the ISP with consistently the
>>>>>>>>> worst rates.
>>>>>>>>> >>
>>>>>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up
>>>>>>>>> and
>>>>>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly
>>>>>>>>> correlating
>>>>>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>>>>>> >>
>>>>>>>>> >> Note the variation though: More or less a factor of two between
>>>>>>>>> highest
>>>>>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in
>>>>>>>>> my
>>>>>>>>> >> tests? Or is there something systematic behind this? Way too
>>>>>>>>> few tests
>>>>>>>>> >> to tell.
>>>>>>>>> >>
>>>>>>>>> >> What these tests do is establish a ballpark.
>>>>>>>>> >>
>>>>>>>>> >> I'm currently repeating tests with dish placed on a trestle
>>>>>>>>> closer to
>>>>>>>>> >> the heavens. This seems to have translated into fewer outages /
>>>>>>>>> ping
>>>>>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the
>>>>>>>>> ground on
>>>>>>>>> >> my deck). Still good enough for a lengthy video Skype call with
>>>>>>>>> my folks
>>>>>>>>> >> in Germany, although they did comment about reduced video
>>>>>>>>> quality. But
>>>>>>>>> >> maybe that was the lighting or the different background as I
>>>>>>>>> wasn't in
>>>>>>>>> >> my usual spot with my laptop when I called them.
>>>>>>>>> >
>>>>>>>>> > Clear view of the sky is king for Starlink reliability. I've got
>>>>>>>>> my dishy mounted on the back fence, looking up over an empty field, so it's
>>>>>>>>> pretty much best-case scenario here.
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >> --
>>>>>>>>> >>
>>>>>>>>> >> ****************************************************************
>>>>>>>>> >> Dr. Ulrich Speidel
>>>>>>>>> >>
>>>>>>>>> >> School of Computer Science
>>>>>>>>> >>
>>>>>>>>> >> Room 303S.594 (City Campus)
>>>>>>>>> >>
>>>>>>>>> >> The University of Auckland
>>>>>>>>> >> u.speidel@auckland.ac.nz
>>>>>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>>>>>> >> ****************************************************************
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >> _______________________________________________
>>>>>>>>> >> Starlink mailing list
>>>>>>>>> >> Starlink@lists.bufferbloat.net
>>>>>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>> >
>>>>>>>>> > _______________________________________________
>>>>>>>>> > Starlink mailing list
>>>>>>>>> > Starlink@lists.bufferbloat.net
>>>>>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>>>>
>>>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Starlink mailing list
>>>>>>>> Starlink@lists.bufferbloat.net
>>>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>
>>>>>>>
>>>>>
>>>>> --
>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>
>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>> Dave Täht CEO, TekLibre, LLC
>>>>> _______________________________________________
>>>>> Starlink mailing list
>>>>> Starlink@lists.bufferbloat.net
>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>
>>>> _______________________________________________
>>> Starlink mailing list
>>> Starlink@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/starlink
>>>
>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink
>>
>
>
> --
> This song goes out to all the folk that thought Stadia would work:
>
> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
> Dave Täht CEO, TekLibre, LLC
>

[-- Attachment #1.1.2: Type: text/html, Size: 21806 bytes --]

[-- Attachment #1.2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #1.3: Screenshot 2023-01-13 at 1.30.03 PM.png --]
[-- Type: image/png, Size: 855840 bytes --]

[-- Attachment #1.4: image.png --]
[-- Type: image/png, Size: 296989 bytes --]

[-- Attachment #1.5: image.png --]
[-- Type: image/png, Size: 222655 bytes --]

[-- Attachment #2: tcp_nup-2023-01-13T163813.572871.starlink_vs_irtt.flent.gz --]
[-- Type: application/gzip, Size: 336010 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 22:42                                                     ` Jonathan Bennett
@ 2023-01-13 22:49                                                       ` Dave Taht
  0 siblings, 0 replies; 44+ messages in thread
From: Dave Taht @ 2023-01-13 22:49 UTC (permalink / raw)
  To: Jonathan Bennett; +Cc: Nathan Owens, starlink


[-- Attachment #1.1.1: Type: text/plain, Size: 17821 bytes --]

selfishly if you could name the plot more after yourself, or your location,
it would help.

OK, in jonathan's case, I got a RTT sample out of flent. Sorry guys, this
is the same lousy TCP behavior as before, with a fixed size buffer, and
bandwidth that varies stepwise every 15 seconds, admittedly with much less
variance than we've seen before.

Waveform's test is broken, optimized for, or there's something crazy going
on. I so wanted to believe...

download of that cap almost done.

On Fri, Jan 13, 2023 at 2:42 PM Jonathan Bennett <
jonathanbennett@hackaday.com> wrote:

> Flent run using git, and ./run-flent -H atlanta.starlink.taht.net -t
> starlink_vs_irtt --step-size=.05 --socket-stats
> --test-parameter=upload_streams=4 tcp_nup
>
>
> Jonathan Bennett
> Hackaday.com
>
>
> On Fri, Jan 13, 2023 at 4:33 PM Dave Taht <dave.taht@gmail.com> wrote:
>
>> thank you all. in both the flent cases there appears to be no tcp_rtt
>> statistics, did you run with --socket-stats?
>>
>> (That seems to be a new bug, both with sampling correctly, and it's
>> either in newer linuxes or in flent itself. I hate to ask ya but could you
>> install the git version of flent?)
>>
>> Thank you for the packet capture!!!! I'm still downloading.
>>
>> Anyway, the "celabon" flent plot clearly shows the inverse relationship
>> between latency and throughput still in this starlink terminal, so there is
>> no AQM in play there, darn it. (in my fq_codeled world the latency stays
>> flat, only the throughput changes)
>>
>> So I am incredibly puzzled now at the ostensibly awesome waveform test
>> result (and need to look at that capture, and/or get tcp rtt stats)
>>
>> The other plot (luis's) shows incredibly consistent latency and bounded
>> throughput at about 6mbit.
>>
>> Patiently awaiting that download to complete.
>>
>>
>> On Fri, Jan 13, 2023 at 2:10 PM Jonathan Bennett via Starlink <
>> starlink@lists.bufferbloat.net> wrote:
>>
>>> The irtt run finished a few seconds before the flent run, but here are
>>> the results:
>>>
>>>
>>> https://drive.google.com/file/d/1FKve13ssUMW1LLWOXLM2931Yx6uMHw8K/view?usp=share_link
>>>
>>> https://drive.google.com/file/d/1ZXd64A0pfUedLr3FyhDNTHA7vxv8S2Gk/view?usp=share_link
>>>
>>> https://drive.google.com/file/d/1rx64UPQHHz3IMNiJtb1oFqtqw2DvhvEE/view?usp=share_link
>>>
>>>
>>> [image: image.png]
>>> [image: image.png]
>>>
>>>
>>> Jonathan Bennett
>>> Hackaday.com
>>>
>>>
>>> On Fri, Jan 13, 2023 at 3:30 PM Nathan Owens via Starlink <
>>> starlink@lists.bufferbloat.net> wrote:
>>>
>>>> Here's Luis's run -- the top line below the edge of the graph is 200ms
>>>> [image: Screenshot 2023-01-13 at 1.30.03 PM.png]
>>>>
>>>>
>>>> On Fri, Jan 13, 2023 at 1:25 PM Luis A. Cornejo <
>>>> luis.a.cornejo@gmail.com> wrote:
>>>>
>>>>> Dave,
>>>>>
>>>>> Here is a run the way I think you wanted it.
>>>>>
>>>>> irtt running for 5 min to your dallas server, followed by a waveform
>>>>> test, then a few seconds of inactivity, cloudflare test, a few more secs of
>>>>> nothing, flent test to dallas. Packet capture is currently uploading (will
>>>>> be done in 20 min or so), irtt JSON also in there (.zip file):
>>>>>
>>>>>
>>>>> https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link
>>>>>
>>>>> -Luis
>>>>>
>>>>> On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io>
>>>>>> wrote:
>>>>>>
>>>>>>> Here's the data visualization for Johnathan's Data
>>>>>>>
>>>>>>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>>>>>>
>>>>>>> You can see the path change at :12, :27, :42, :57 after the minute.
>>>>>>> Some paths are clearly busier than others with increased loss, latency, and
>>>>>>> jitter.
>>>>>>>
>>>>>>
>>>>>> I am so glad to see loss and bounded delay here. Also a bit of rigor
>>>>>> regarding what traffic was active locally vs on the path would be nice,
>>>>>> although it seems to line up with the known 15s starlink switchover thing
>>>>>> (need a name for this), in this case, doing a few speedtests
>>>>>> while that irtt is running would show the impact(s) of whatever else
>>>>>> they are up to.
>>>>>>
>>>>>> It would also be my hope that the loss distribution in the middle
>>>>>> portion of this data is good, not bursty, but we don't have a tool to take
>>>>>> apart that. (I am so hopeless at json)
>>>>>>
>>>>>>
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> I’ll run my visualization code on this result this afternoon and
>>>>>>>> report back!
>>>>>>>>
>>>>>>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>>
>>>>>>>>> The irtt command, run with normal, light usage:
>>>>>>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>>>>>>
>>>>>>>>> Jonathan Bennett
>>>>>>>>> Hackaday.com
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> packet caps would be nice... all this is very exciting news.
>>>>>>>>>>
>>>>>>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>>>>>>> results nowadays to duplicate and re-plot the original irtt tests
>>>>>>>>>> we
>>>>>>>>>> did:
>>>>>>>>>>
>>>>>>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net
>>>>>>>>>> -o whatever.json
>>>>>>>>>>
>>>>>>>>>> They MUST have changed their scheduling to get such amazing uplink
>>>>>>>>>> results, in addition to better queue management.
>>>>>>>>>>
>>>>>>>>>> (for the record, my servers are de, london, fremont, sydney,
>>>>>>>>>> dallas,
>>>>>>>>>> newark, atlanta, singapore, mumbai)
>>>>>>>>>>
>>>>>>>>>> There's an R and gnuplot script for plotting that output around
>>>>>>>>>> here
>>>>>>>>>> somewhere (I have largely personally put down the starlink
>>>>>>>>>> project,
>>>>>>>>>> loaning out mine) - that went by on this list... I should have
>>>>>>>>>> written
>>>>>>>>>> a blog entry so I can find that stuff again.
>>>>>>>>>>
>>>>>>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>>>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>>>>>>> >
>>>>>>>>>> >
>>>>>>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>>>> >>
>>>>>>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>>>>>>> >> >
>>>>>>>>>> >> > From Auckland, New Zealand, using a roaming subscription, it
>>>>>>>>>> puts me
>>>>>>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>>>>>>> >> >
>>>>>>>>>> >> >
>>>>>>>>>> >> > IP address: nix six.
>>>>>>>>>> >> >
>>>>>>>>>> >> > My thoughts shall follow later.
>>>>>>>>>> >>
>>>>>>>>>> >> OK, so here we go.
>>>>>>>>>> >>
>>>>>>>>>> >> I'm always a bit skeptical when it comes to speed tests -
>>>>>>>>>> they're really
>>>>>>>>>> >> laden with so many caveats that it's not funny. I took our new
>>>>>>>>>> work
>>>>>>>>>> >> Starlink kit home in December to give it a try and the other
>>>>>>>>>> day finally
>>>>>>>>>> >> got around to set it up. It's on a roaming subscription
>>>>>>>>>> because our
>>>>>>>>>> >> badly built-up campus really isn't ideal in terms of a clear
>>>>>>>>>> view of the
>>>>>>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>>>>>>> adapter,
>>>>>>>>>> >> not the WiFi?
>>>>>>>>>> >>
>>>>>>>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>>>>>>>> Starlink
>>>>>>>>>> >> promises is about 1/3 in terms of data rate you can actually
>>>>>>>>>> get from
>>>>>>>>>> >> fibre to the home at under half of Starlink's price. Read:
>>>>>>>>>> There are few
>>>>>>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>>>>>>> >>
>>>>>>>>>> >> Caveat 2: Auckland has three Starlink gateways close by:
>>>>>>>>>> Clevedon (which
>>>>>>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana
>>>>>>>>>> and Puwera,
>>>>>>>>>> >> the most distant of the three and about 130 km away from me as
>>>>>>>>>> the crow
>>>>>>>>>> >> flies. Read: My dishy can use any satellite that any of these
>>>>>>>>>> three can
>>>>>>>>>> >> see, and then depending on where I put it and how much of the
>>>>>>>>>> southern
>>>>>>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away,
>>>>>>>>>> although that
>>>>>>>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>>>>>>>> options
>>>>>>>>>> >> for my bits to travel without needing a lot of handovers. Why?
>>>>>>>>>> Easy: If
>>>>>>>>>> >> your nearest teleport is close by, then the set of satellites
>>>>>>>>>> that the
>>>>>>>>>> >> teleport can see and the set that you can see is almost the
>>>>>>>>>> same, so you
>>>>>>>>>> >> can essentially stick with the same satellite while it's in
>>>>>>>>>> view for you
>>>>>>>>>> >> because it'll also be in view for the teleport. Pretty much
>>>>>>>>>> any bird
>>>>>>>>>> >> above you will do.
>>>>>>>>>> >>
>>>>>>>>>> >> And because I don't get a lot of competition from other users
>>>>>>>>>> in my area
>>>>>>>>>> >> vying for one of the few available satellites that can see
>>>>>>>>>> both us and
>>>>>>>>>> >> the teleport, this is about as good as it gets at 37S
>>>>>>>>>> latitude. If I'd
>>>>>>>>>> >> want it any better, I'd have to move a lot further south.
>>>>>>>>>> >>
>>>>>>>>>> >> It'd be interesting to hear from Jonathan what the
>>>>>>>>>> availability of home
>>>>>>>>>> >> broadband is like in the Dallas area. I note that it's at a
>>>>>>>>>> lower
>>>>>>>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I
>>>>>>>>>> notice
>>>>>>>>>> >> two teleports each about 160 km away, which is also not too
>>>>>>>>>> bad. I also
>>>>>>>>>> >> note Starlink availability in the area is restricted at the
>>>>>>>>>> moment -
>>>>>>>>>> >> oversubscribed? But if Jonathan gets good data rates, then
>>>>>>>>>> that means
>>>>>>>>>> >> that competition for bird capacity can't be too bad - for
>>>>>>>>>> whatever reason.
>>>>>>>>>> >
>>>>>>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>>>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>>>>>>> broadband options. But there are also many people that live outside cities,
>>>>>>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>>>>>>> and Texas is probably ideal conditions for Starlink.
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>>>>>>> whatever I
>>>>>>>>>> >> talk to in terms of my communications. Traceroute shows about
>>>>>>>>>> 10 hops
>>>>>>>>>> >> between me and the University of Auckland via Starlink. That's
>>>>>>>>>> 10
>>>>>>>>>> >> queues, not one. Many of them will have cross traffic. So it's
>>>>>>>>>> a bit
>>>>>>>>>> >> hard to tell where our packets really get to wait or where
>>>>>>>>>> they get
>>>>>>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>>>>>>> between 1
>>>>>>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they
>>>>>>>>>> can all
>>>>>>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link
>>>>>>>>>> of a few
>>>>>>>>>> >> Mb/s where it's obvious where the dominant long latency
>>>>>>>>>> bottleneck in
>>>>>>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>>>>>>> whether a
>>>>>>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>>>>>>> system or
>>>>>>>>>> >> somewhere between Starlink's systems and the target system.
>>>>>>>>>> >>
>>>>>>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical
>>>>>>>>>> latency
>>>>>>>>>> >> should be under 15 ms. So there's clearly a bit of buffer here
>>>>>>>>>> along the
>>>>>>>>>> >> chain that occasionally fills up.
>>>>>>>>>> >>
>>>>>>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>>>>>>> inevitably
>>>>>>>>>> >> associated with a change in RTT and in most cases also
>>>>>>>>>> available
>>>>>>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>>>>>>>> >> satellite while your TCP is still trying to respond to
>>>>>>>>>> whatever it
>>>>>>>>>> >> thought the queue on the previous bird was doing. Read:
>>>>>>>>>> Whatever your
>>>>>>>>>> >> cwnd is immediately after a handover, it's probably not what
>>>>>>>>>> it should be.
>>>>>>>>>> >>
>>>>>>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four
>>>>>>>>>> Ookla
>>>>>>>>>> >> speedtest.net tests each to five local servers. Average
>>>>>>>>>> upload rate was
>>>>>>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the
>>>>>>>>>> ISP that
>>>>>>>>>> >> Starlink seems to be buying its local connectivity from (Vocus
>>>>>>>>>> Group)
>>>>>>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and
>>>>>>>>>> 52.22
>>>>>>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to
>>>>>>>>>> rates
>>>>>>>>>> >> observed. In fact, they were the ISP with consistently the
>>>>>>>>>> worst rates.
>>>>>>>>>> >>
>>>>>>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s
>>>>>>>>>> up and
>>>>>>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly
>>>>>>>>>> correlating
>>>>>>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>>>>>>> >>
>>>>>>>>>> >> Note the variation though: More or less a factor of two
>>>>>>>>>> between highest
>>>>>>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky
>>>>>>>>>> in my
>>>>>>>>>> >> tests? Or is there something systematic behind this? Way too
>>>>>>>>>> few tests
>>>>>>>>>> >> to tell.
>>>>>>>>>> >>
>>>>>>>>>> >> What these tests do is establish a ballpark.
>>>>>>>>>> >>
>>>>>>>>>> >> I'm currently repeating tests with dish placed on a trestle
>>>>>>>>>> closer to
>>>>>>>>>> >> the heavens. This seems to have translated into fewer outages
>>>>>>>>>> / ping
>>>>>>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the
>>>>>>>>>> ground on
>>>>>>>>>> >> my deck). Still good enough for a lengthy video Skype call
>>>>>>>>>> with my folks
>>>>>>>>>> >> in Germany, although they did comment about reduced video
>>>>>>>>>> quality. But
>>>>>>>>>> >> maybe that was the lighting or the different background as I
>>>>>>>>>> wasn't in
>>>>>>>>>> >> my usual spot with my laptop when I called them.
>>>>>>>>>> >
>>>>>>>>>> > Clear view of the sky is king for Starlink reliability. I've
>>>>>>>>>> got my dishy mounted on the back fence, looking up over an empty field, so
>>>>>>>>>> it's pretty much best-case scenario here.
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >> --
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> ****************************************************************
>>>>>>>>>> >> Dr. Ulrich Speidel
>>>>>>>>>> >>
>>>>>>>>>> >> School of Computer Science
>>>>>>>>>> >>
>>>>>>>>>> >> Room 303S.594 (City Campus)
>>>>>>>>>> >>
>>>>>>>>>> >> The University of Auckland
>>>>>>>>>> >> u.speidel@auckland.ac.nz
>>>>>>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>>>>>>> >>
>>>>>>>>>> ****************************************************************
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >> _______________________________________________
>>>>>>>>>> >> Starlink mailing list
>>>>>>>>>> >> Starlink@lists.bufferbloat.net
>>>>>>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>> >
>>>>>>>>>> > _______________________________________________
>>>>>>>>>> > Starlink mailing list
>>>>>>>>>> > Starlink@lists.bufferbloat.net
>>>>>>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>>>>>
>>>>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> Starlink mailing list
>>>>>>>>> Starlink@lists.bufferbloat.net
>>>>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>
>>>>>>>>
>>>>>>
>>>>>> --
>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>
>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>> _______________________________________________
>>>>>> Starlink mailing list
>>>>>> Starlink@lists.bufferbloat.net
>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>
>>>>> _______________________________________________
>>>> Starlink mailing list
>>>> Starlink@lists.bufferbloat.net
>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>
>>> _______________________________________________
>>> Starlink mailing list
>>> Starlink@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/starlink
>>>
>>
>>
>> --
>> This song goes out to all the folk that thought Stadia would work:
>>
>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>> Dave Täht CEO, TekLibre, LLC
>>
>

-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

[-- Attachment #1.1.2: Type: text/html, Size: 23183 bytes --]

[-- Attachment #1.2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #1.3: Screenshot 2023-01-13 at 1.30.03 PM.png --]
[-- Type: image/png, Size: 855840 bytes --]

[-- Attachment #1.4: image.png --]
[-- Type: image/png, Size: 296989 bytes --]

[-- Attachment #1.5: image.png --]
[-- Type: image/png, Size: 222655 bytes --]

[-- Attachment #2: bennet_tcp_rtt.png --]
[-- Type: image/png, Size: 202069 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 17:02                                 ` Jonathan Bennett
  2023-01-13 17:26                                   ` Dave Taht
@ 2023-01-13 22:51                                   ` Ulrich Speidel
  1 sibling, 0 replies; 44+ messages in thread
From: Ulrich Speidel @ 2023-01-13 22:51 UTC (permalink / raw)
  To: Jonathan Bennett; +Cc: starlink


[-- Attachment #1.1: Type: text/plain, Size: 2754 bytes --]

On 14/01/2023 6:02 am, Jonathan Bennett wrote:
>
>     It'd be interesting to hear from Jonathan what the availability of
>     home
>     broadband is like in the Dallas area. I note that it's at a lower
>     latitude (33N) than Auckland, but the difference isn't huge. I notice
>     two teleports each about 160 km away, which is also not too bad. I
>     also
>     note Starlink availability in the area is restricted at the moment -
>     oversubscribed? But if Jonathan gets good data rates, then that means
>     that competition for bird capacity can't be too bad - for whatever
>     reason.
>
> I'm in Southwest Oklahoma, but Dallas is the nearby Starlink gateway. 
> In cities, like Dallas, and Lawton where I live, there are good 
> broadband options. But there are also many people that live outside 
> cities, and the options are much worse. The low density userbase in 
> rural Oklahoma and Texas is probably ideal conditions for Starlink.
Ah. So what actually happens here is that you're relatively close to the 
Springer teleport (~120 km), and the IP address that Starlink assigns to 
you gets geolocated to Dallas. Given lack of current Starlink 
availability in the region that isn't correlating with population 
density like in other parts of the US, we're probably talking lots of 
satellite capacity for very few users here. Plus the extra fibre latency 
down to Dallas is peanuts. Which explains your data. You're at 
34-something north, which is not that different from Auckland.
> Clear view of the sky is king for Starlink reliability. I've got my 
> dishy mounted on the back fence, looking up over an empty field, so 
> it's pretty much best-case scenario here.

So-so. Your dishy probably orients itself north-facing, but you'll still 
be talking via your closest teleport or teleports, so Springer or Dumas.

If Starlink can talk to satellites that are straight overhead, it will - 
I've attached a photo of my current setup at home, dishy south-facing 
onto my roof. I've pinged for almost a day now in this configuration and 
have < 1% loss. Practically all loss events are one-offs, i.e., outages 
are under 2 s.

Before that, I had the dishy on the ground in the same position, and had 
around 2% ping loss. So being able to see more of the sky makes a bit of 
a difference obviously, but it's not huge. Internet was "usable" in both 
configurations.

-- 

****************************************************************
Dr. Ulrich Speidel

School of Computer Science

Room 303S.594 (City Campus)

The University of Auckland
u.speidel@auckland.ac.nz  
http://www.cs.auckland.ac.nz/~ulrich/
****************************************************************



[-- Attachment #1.2: Type: text/html, Size: 4137 bytes --]

[-- Attachment #2: 20230114_074033.jpg --]
[-- Type: image/*, Size: 4480431 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-13 22:09                                                 ` Jonathan Bennett
  2023-01-13 22:30                                                   ` Luis A. Cornejo
  2023-01-13 22:32                                                   ` Dave Taht
@ 2023-01-13 23:44                                                   ` Dave Taht
       [not found]                                                     ` <CALjsLJv5cbfHfkxqHnbjxoVHczspYvxc_jrshzs1CpHLEDWyew@mail.gmail.com>
  2 siblings, 1 reply; 44+ messages in thread
From: Dave Taht @ 2023-01-13 23:44 UTC (permalink / raw)
  To: Jonathan Bennett; +Cc: Nathan Owens, starlink, Sina Khanifar


[-- Attachment #1.1.1: Type: text/plain, Size: 15719 bytes --]

I am forced to conclude that waveform's upload test is broken in some cases
and has been for some time. All the joy many have been feeling about their
uplink speeds has to be cast away. Felt good, though, didn't it?

There is a *slight* possibility that there is some fq in the starlink
network on tcp port 433. The first showing normal rtt growth and loss for
cubic, the second, a low rate flow that is holding the line... but I didn't
check to see if these were sequential or parallel.

The last is a cwnd plot, clearly showing the cubic sawtooth on the upload.

It's weirdly nice to be able to follow a port 433 stream, see the tls
handshake put the website en'clair, and the rest go dark, and still trace
the behaviors.

we're so going to lose this analytical ability with quic. I'm enjoying it
while we still can.


On Fri, Jan 13, 2023 at 2:10 PM Jonathan Bennett via Starlink <
starlink@lists.bufferbloat.net> wrote:

> The irtt run finished a few seconds before the flent run, but here are the
> results:
>
>
> https://drive.google.com/file/d/1FKve13ssUMW1LLWOXLM2931Yx6uMHw8K/view?usp=share_link
>
> https://drive.google.com/file/d/1ZXd64A0pfUedLr3FyhDNTHA7vxv8S2Gk/view?usp=share_link
>
> https://drive.google.com/file/d/1rx64UPQHHz3IMNiJtb1oFqtqw2DvhvEE/view?usp=share_link
>
>
> [image: image.png]
> [image: image.png]
>
>
> Jonathan Bennett
> Hackaday.com
>
>
> On Fri, Jan 13, 2023 at 3:30 PM Nathan Owens via Starlink <
> starlink@lists.bufferbloat.net> wrote:
>
>> Here's Luis's run -- the top line below the edge of the graph is 200ms
>> [image: Screenshot 2023-01-13 at 1.30.03 PM.png]
>>
>>
>> On Fri, Jan 13, 2023 at 1:25 PM Luis A. Cornejo <luis.a.cornejo@gmail.com>
>> wrote:
>>
>>> Dave,
>>>
>>> Here is a run the way I think you wanted it.
>>>
>>> irtt running for 5 min to your dallas server, followed by a waveform
>>> test, then a few seconds of inactivity, cloudflare test, a few more secs of
>>> nothing, flent test to dallas. Packet capture is currently uploading (will
>>> be done in 20 min or so), irtt JSON also in there (.zip file):
>>>
>>>
>>> https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link
>>>
>>> -Luis
>>>
>>> On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
>>> starlink@lists.bufferbloat.net> wrote:
>>>
>>>>
>>>>
>>>> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io> wrote:
>>>>
>>>>> Here's the data visualization for Johnathan's Data
>>>>>
>>>>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>>>>
>>>>> You can see the path change at :12, :27, :42, :57 after the minute.
>>>>> Some paths are clearly busier than others with increased loss, latency, and
>>>>> jitter.
>>>>>
>>>>
>>>> I am so glad to see loss and bounded delay here. Also a bit of rigor
>>>> regarding what traffic was active locally vs on the path would be nice,
>>>> although it seems to line up with the known 15s starlink switchover thing
>>>> (need a name for this), in this case, doing a few speedtests
>>>> while that irtt is running would show the impact(s) of whatever else
>>>> they are up to.
>>>>
>>>> It would also be my hope that the loss distribution in the middle
>>>> portion of this data is good, not bursty, but we don't have a tool to take
>>>> apart that. (I am so hopeless at json)
>>>>
>>>>
>>>>
>>>>>
>>>>>
>>>>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io>
>>>>> wrote:
>>>>>
>>>>>> I’ll run my visualization code on this result this afternoon and
>>>>>> report back!
>>>>>>
>>>>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>
>>>>>>> The irtt command, run with normal, light usage:
>>>>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>>>>
>>>>>>> Jonathan Bennett
>>>>>>> Hackaday.com
>>>>>>>
>>>>>>>
>>>>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> packet caps would be nice... all this is very exciting news.
>>>>>>>>
>>>>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>>>>> results nowadays to duplicate and re-plot the original irtt tests we
>>>>>>>> did:
>>>>>>>>
>>>>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net -o
>>>>>>>> whatever.json
>>>>>>>>
>>>>>>>> They MUST have changed their scheduling to get such amazing uplink
>>>>>>>> results, in addition to better queue management.
>>>>>>>>
>>>>>>>> (for the record, my servers are de, london, fremont, sydney, dallas,
>>>>>>>> newark, atlanta, singapore, mumbai)
>>>>>>>>
>>>>>>>> There's an R and gnuplot script for plotting that output around here
>>>>>>>> somewhere (I have largely personally put down the starlink project,
>>>>>>>> loaning out mine) - that went by on this list... I should have
>>>>>>>> written
>>>>>>>> a blog entry so I can find that stuff again.
>>>>>>>>
>>>>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>>>>> >
>>>>>>>> >
>>>>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>> >>
>>>>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>>>>> >> >
>>>>>>>> >> > From Auckland, New Zealand, using a roaming subscription, it
>>>>>>>> puts me
>>>>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>>>>> >> >
>>>>>>>> >> >
>>>>>>>> >> > IP address: nix six.
>>>>>>>> >> >
>>>>>>>> >> > My thoughts shall follow later.
>>>>>>>> >>
>>>>>>>> >> OK, so here we go.
>>>>>>>> >>
>>>>>>>> >> I'm always a bit skeptical when it comes to speed tests -
>>>>>>>> they're really
>>>>>>>> >> laden with so many caveats that it's not funny. I took our new
>>>>>>>> work
>>>>>>>> >> Starlink kit home in December to give it a try and the other day
>>>>>>>> finally
>>>>>>>> >> got around to set it up. It's on a roaming subscription because
>>>>>>>> our
>>>>>>>> >> badly built-up campus really isn't ideal in terms of a clear
>>>>>>>> view of the
>>>>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>>>>> adapter,
>>>>>>>> >> not the WiFi?
>>>>>>>> >>
>>>>>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>>>>>> Starlink
>>>>>>>> >> promises is about 1/3 in terms of data rate you can actually get
>>>>>>>> from
>>>>>>>> >> fibre to the home at under half of Starlink's price. Read: There
>>>>>>>> are few
>>>>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>>>>> >>
>>>>>>>> >> Caveat 2: Auckland has three Starlink gateways close by:
>>>>>>>> Clevedon (which
>>>>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana and
>>>>>>>> Puwera,
>>>>>>>> >> the most distant of the three and about 130 km away from me as
>>>>>>>> the crow
>>>>>>>> >> flies. Read: My dishy can use any satellite that any of these
>>>>>>>> three can
>>>>>>>> >> see, and then depending on where I put it and how much of the
>>>>>>>> southern
>>>>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away,
>>>>>>>> although that
>>>>>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>>>>>> options
>>>>>>>> >> for my bits to travel without needing a lot of handovers. Why?
>>>>>>>> Easy: If
>>>>>>>> >> your nearest teleport is close by, then the set of satellites
>>>>>>>> that the
>>>>>>>> >> teleport can see and the set that you can see is almost the
>>>>>>>> same, so you
>>>>>>>> >> can essentially stick with the same satellite while it's in view
>>>>>>>> for you
>>>>>>>> >> because it'll also be in view for the teleport. Pretty much any
>>>>>>>> bird
>>>>>>>> >> above you will do.
>>>>>>>> >>
>>>>>>>> >> And because I don't get a lot of competition from other users in
>>>>>>>> my area
>>>>>>>> >> vying for one of the few available satellites that can see both
>>>>>>>> us and
>>>>>>>> >> the teleport, this is about as good as it gets at 37S latitude.
>>>>>>>> If I'd
>>>>>>>> >> want it any better, I'd have to move a lot further south.
>>>>>>>> >>
>>>>>>>> >> It'd be interesting to hear from Jonathan what the availability
>>>>>>>> of home
>>>>>>>> >> broadband is like in the Dallas area. I note that it's at a lower
>>>>>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I
>>>>>>>> notice
>>>>>>>> >> two teleports each about 160 km away, which is also not too bad.
>>>>>>>> I also
>>>>>>>> >> note Starlink availability in the area is restricted at the
>>>>>>>> moment -
>>>>>>>> >> oversubscribed? But if Jonathan gets good data rates, then that
>>>>>>>> means
>>>>>>>> >> that competition for bird capacity can't be too bad - for
>>>>>>>> whatever reason.
>>>>>>>> >
>>>>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>>>>> broadband options. But there are also many people that live outside cities,
>>>>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>>>>> and Texas is probably ideal conditions for Starlink.
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>>>>> whatever I
>>>>>>>> >> talk to in terms of my communications. Traceroute shows about 10
>>>>>>>> hops
>>>>>>>> >> between me and the University of Auckland via Starlink. That's 10
>>>>>>>> >> queues, not one. Many of them will have cross traffic. So it's a
>>>>>>>> bit
>>>>>>>> >> hard to tell where our packets really get to wait or where they
>>>>>>>> get
>>>>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>>>>> between 1
>>>>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they
>>>>>>>> can all
>>>>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link of
>>>>>>>> a few
>>>>>>>> >> Mb/s where it's obvious where the dominant long latency
>>>>>>>> bottleneck in
>>>>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>>>>> whether a
>>>>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>>>>> system or
>>>>>>>> >> somewhere between Starlink's systems and the target system.
>>>>>>>> >>
>>>>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical
>>>>>>>> latency
>>>>>>>> >> should be under 15 ms. So there's clearly a bit of buffer here
>>>>>>>> along the
>>>>>>>> >> chain that occasionally fills up.
>>>>>>>> >>
>>>>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>>>>> inevitably
>>>>>>>> >> associated with a change in RTT and in most cases also available
>>>>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>>>>>> >> satellite while your TCP is still trying to respond to whatever
>>>>>>>> it
>>>>>>>> >> thought the queue on the previous bird was doing. Read: Whatever
>>>>>>>> your
>>>>>>>> >> cwnd is immediately after a handover, it's probably not what it
>>>>>>>> should be.
>>>>>>>> >>
>>>>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four
>>>>>>>> Ookla
>>>>>>>> >> speedtest.net tests each to five local servers. Average upload
>>>>>>>> rate was
>>>>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the ISP
>>>>>>>> that
>>>>>>>> >> Starlink seems to be buying its local connectivity from (Vocus
>>>>>>>> Group)
>>>>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and
>>>>>>>> 52.22
>>>>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to
>>>>>>>> rates
>>>>>>>> >> observed. In fact, they were the ISP with consistently the worst
>>>>>>>> rates.
>>>>>>>> >>
>>>>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s up
>>>>>>>> and
>>>>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly
>>>>>>>> correlating
>>>>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>>>>> >>
>>>>>>>> >> Note the variation though: More or less a factor of two between
>>>>>>>> highest
>>>>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky in
>>>>>>>> my
>>>>>>>> >> tests? Or is there something systematic behind this? Way too few
>>>>>>>> tests
>>>>>>>> >> to tell.
>>>>>>>> >>
>>>>>>>> >> What these tests do is establish a ballpark.
>>>>>>>> >>
>>>>>>>> >> I'm currently repeating tests with dish placed on a trestle
>>>>>>>> closer to
>>>>>>>> >> the heavens. This seems to have translated into fewer outages /
>>>>>>>> ping
>>>>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the
>>>>>>>> ground on
>>>>>>>> >> my deck). Still good enough for a lengthy video Skype call with
>>>>>>>> my folks
>>>>>>>> >> in Germany, although they did comment about reduced video
>>>>>>>> quality. But
>>>>>>>> >> maybe that was the lighting or the different background as I
>>>>>>>> wasn't in
>>>>>>>> >> my usual spot with my laptop when I called them.
>>>>>>>> >
>>>>>>>> > Clear view of the sky is king for Starlink reliability. I've got
>>>>>>>> my dishy mounted on the back fence, looking up over an empty field, so it's
>>>>>>>> pretty much best-case scenario here.
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> --
>>>>>>>> >>
>>>>>>>> >> ****************************************************************
>>>>>>>> >> Dr. Ulrich Speidel
>>>>>>>> >>
>>>>>>>> >> School of Computer Science
>>>>>>>> >>
>>>>>>>> >> Room 303S.594 (City Campus)
>>>>>>>> >>
>>>>>>>> >> The University of Auckland
>>>>>>>> >> u.speidel@auckland.ac.nz
>>>>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>>>>> >> ****************************************************************
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> _______________________________________________
>>>>>>>> >> Starlink mailing list
>>>>>>>> >> Starlink@lists.bufferbloat.net
>>>>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>> >
>>>>>>>> > _______________________________________________
>>>>>>>> > Starlink mailing list
>>>>>>>> > Starlink@lists.bufferbloat.net
>>>>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>>>
>>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Starlink mailing list
>>>>>>> Starlink@lists.bufferbloat.net
>>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>
>>>>>>
>>>>
>>>> --
>>>> This song goes out to all the folk that thought Stadia would work:
>>>>
>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>> Dave Täht CEO, TekLibre, LLC
>>>> _______________________________________________
>>>> Starlink mailing list
>>>> Starlink@lists.bufferbloat.net
>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>
>>> _______________________________________________
>> Starlink mailing list
>> Starlink@lists.bufferbloat.net
>> https://lists.bufferbloat.net/listinfo/starlink
>>
> _______________________________________________
> Starlink mailing list
> Starlink@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/starlink
>


-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

[-- Attachment #1.1.2: Type: text/html, Size: 20796 bytes --]

[-- Attachment #1.2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #1.3: Screenshot 2023-01-13 at 1.30.03 PM.png --]
[-- Type: image/png, Size: 855840 bytes --]

[-- Attachment #1.4: image.png --]
[-- Type: image/png, Size: 296989 bytes --]

[-- Attachment #1.5: image.png --]
[-- Type: image/png, Size: 222655 bytes --]

[-- Attachment #2: 433-rtt-low.png --]
[-- Type: image/png, Size: 45065 bytes --]

[-- Attachment #3: 433-rtt.png --]
[-- Type: image/png, Size: 74416 bytes --]

[-- Attachment #4: waveformbroken_windowscale.png --]
[-- Type: image/png, Size: 64924 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
       [not found]                                                     ` <CALjsLJv5cbfHfkxqHnbjxoVHczspYvxc_jrshzs1CpHLEDWyew@mail.gmail.com>
@ 2023-01-14 14:20                                                       ` Nathan Owens
  2023-01-14 15:53                                                         ` Dave Taht
  2023-01-14 15:52                                                       ` Dave Taht
  1 sibling, 1 reply; 44+ messages in thread
From: Nathan Owens @ 2023-01-14 14:20 UTC (permalink / raw)
  To: Dave Taht; +Cc: Jonathan Bennett, starlink, Sina Khanifar


[-- Attachment #1.1: Type: text/plain, Size: 17272 bytes --]

Sorry, was rejected from the listserv, here's the google drive link for all
3 IRTT's visualized

https://drive.google.com/drive/folders/1SbUyvdlfdllgqrEcSSxvGdn3yN40vSx-?usp=share_link

On Sat, Jan 14, 2023 at 6:14 AM Nathan Owens <nathan@nathan.io> wrote:

> I realized I goofed up my visualizations, here's all of them again:
> I see a ton of loss on all of these!
>
> While Luis was downloading/uploading... oof.
> [image: Screenshot 2023-01-14 at 6.13.50 AM.png]
>
> [image: Screenshot 2023-01-14 at 6.12.53 AM.png]
>
>
>
>
> On Fri, Jan 13, 2023 at 3:44 PM Dave Taht <dave.taht@gmail.com> wrote:
>
>> I am forced to conclude that waveform's upload test is broken in some
>> cases and has been for some time. All the joy many have been feeling about
>> their uplink speeds has to be cast away. Felt good, though, didn't it?
>>
>> There is a *slight* possibility that there is some fq in the starlink
>> network on tcp port 433. The first showing normal rtt growth and loss for
>> cubic, the second, a low rate flow that is holding the line... but I didn't
>> check to see if these were sequential or parallel.
>>
>> The last is a cwnd plot, clearly showing the cubic sawtooth on the
>> upload.
>>
>> It's weirdly nice to be able to follow a port 433 stream, see the tls
>> handshake put the website en'clair, and the rest go dark, and still trace
>> the behaviors.
>>
>> we're so going to lose this analytical ability with quic. I'm enjoying it
>> while we still can.
>>
>>
>> On Fri, Jan 13, 2023 at 2:10 PM Jonathan Bennett via Starlink <
>> starlink@lists.bufferbloat.net> wrote:
>>
>>> The irtt run finished a few seconds before the flent run, but here are
>>> the results:
>>>
>>>
>>> https://drive.google.com/file/d/1FKve13ssUMW1LLWOXLM2931Yx6uMHw8K/view?usp=share_link
>>>
>>> https://drive.google.com/file/d/1ZXd64A0pfUedLr3FyhDNTHA7vxv8S2Gk/view?usp=share_link
>>>
>>> https://drive.google.com/file/d/1rx64UPQHHz3IMNiJtb1oFqtqw2DvhvEE/view?usp=share_link
>>>
>>>
>>> [image: image.png]
>>> [image: image.png]
>>>
>>>
>>> Jonathan Bennett
>>> Hackaday.com
>>>
>>>
>>> On Fri, Jan 13, 2023 at 3:30 PM Nathan Owens via Starlink <
>>> starlink@lists.bufferbloat.net> wrote:
>>>
>>>> Here's Luis's run -- the top line below the edge of the graph is 200ms
>>>> [image: Screenshot 2023-01-13 at 1.30.03 PM.png]
>>>>
>>>>
>>>> On Fri, Jan 13, 2023 at 1:25 PM Luis A. Cornejo <
>>>> luis.a.cornejo@gmail.com> wrote:
>>>>
>>>>> Dave,
>>>>>
>>>>> Here is a run the way I think you wanted it.
>>>>>
>>>>> irtt running for 5 min to your dallas server, followed by a waveform
>>>>> test, then a few seconds of inactivity, cloudflare test, a few more secs of
>>>>> nothing, flent test to dallas. Packet capture is currently uploading (will
>>>>> be done in 20 min or so), irtt JSON also in there (.zip file):
>>>>>
>>>>>
>>>>> https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link
>>>>>
>>>>> -Luis
>>>>>
>>>>> On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io>
>>>>>> wrote:
>>>>>>
>>>>>>> Here's the data visualization for Johnathan's Data
>>>>>>>
>>>>>>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>>>>>>
>>>>>>> You can see the path change at :12, :27, :42, :57 after the minute.
>>>>>>> Some paths are clearly busier than others with increased loss, latency, and
>>>>>>> jitter.
>>>>>>>
>>>>>>
>>>>>> I am so glad to see loss and bounded delay here. Also a bit of rigor
>>>>>> regarding what traffic was active locally vs on the path would be nice,
>>>>>> although it seems to line up with the known 15s starlink switchover thing
>>>>>> (need a name for this), in this case, doing a few speedtests
>>>>>> while that irtt is running would show the impact(s) of whatever else
>>>>>> they are up to.
>>>>>>
>>>>>> It would also be my hope that the loss distribution in the middle
>>>>>> portion of this data is good, not bursty, but we don't have a tool to take
>>>>>> apart that. (I am so hopeless at json)
>>>>>>
>>>>>>
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> I’ll run my visualization code on this result this afternoon and
>>>>>>>> report back!
>>>>>>>>
>>>>>>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>>
>>>>>>>>> The irtt command, run with normal, light usage:
>>>>>>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>>>>>>
>>>>>>>>> Jonathan Bennett
>>>>>>>>> Hackaday.com
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> packet caps would be nice... all this is very exciting news.
>>>>>>>>>>
>>>>>>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>>>>>>> results nowadays to duplicate and re-plot the original irtt tests
>>>>>>>>>> we
>>>>>>>>>> did:
>>>>>>>>>>
>>>>>>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net
>>>>>>>>>> -o whatever.json
>>>>>>>>>>
>>>>>>>>>> They MUST have changed their scheduling to get such amazing uplink
>>>>>>>>>> results, in addition to better queue management.
>>>>>>>>>>
>>>>>>>>>> (for the record, my servers are de, london, fremont, sydney,
>>>>>>>>>> dallas,
>>>>>>>>>> newark, atlanta, singapore, mumbai)
>>>>>>>>>>
>>>>>>>>>> There's an R and gnuplot script for plotting that output around
>>>>>>>>>> here
>>>>>>>>>> somewhere (I have largely personally put down the starlink
>>>>>>>>>> project,
>>>>>>>>>> loaning out mine) - that went by on this list... I should have
>>>>>>>>>> written
>>>>>>>>>> a blog entry so I can find that stuff again.
>>>>>>>>>>
>>>>>>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>>>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>>>>>>> >
>>>>>>>>>> >
>>>>>>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>>>> >>
>>>>>>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>>>>>>> >> >
>>>>>>>>>> >> > From Auckland, New Zealand, using a roaming subscription, it
>>>>>>>>>> puts me
>>>>>>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>>>>>>> >> >
>>>>>>>>>> >> >
>>>>>>>>>> >> > IP address: nix six.
>>>>>>>>>> >> >
>>>>>>>>>> >> > My thoughts shall follow later.
>>>>>>>>>> >>
>>>>>>>>>> >> OK, so here we go.
>>>>>>>>>> >>
>>>>>>>>>> >> I'm always a bit skeptical when it comes to speed tests -
>>>>>>>>>> they're really
>>>>>>>>>> >> laden with so many caveats that it's not funny. I took our new
>>>>>>>>>> work
>>>>>>>>>> >> Starlink kit home in December to give it a try and the other
>>>>>>>>>> day finally
>>>>>>>>>> >> got around to set it up. It's on a roaming subscription
>>>>>>>>>> because our
>>>>>>>>>> >> badly built-up campus really isn't ideal in terms of a clear
>>>>>>>>>> view of the
>>>>>>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>>>>>>> adapter,
>>>>>>>>>> >> not the WiFi?
>>>>>>>>>> >>
>>>>>>>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>>>>>>>> Starlink
>>>>>>>>>> >> promises is about 1/3 in terms of data rate you can actually
>>>>>>>>>> get from
>>>>>>>>>> >> fibre to the home at under half of Starlink's price. Read:
>>>>>>>>>> There are few
>>>>>>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>>>>>>> >>
>>>>>>>>>> >> Caveat 2: Auckland has three Starlink gateways close by:
>>>>>>>>>> Clevedon (which
>>>>>>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana
>>>>>>>>>> and Puwera,
>>>>>>>>>> >> the most distant of the three and about 130 km away from me as
>>>>>>>>>> the crow
>>>>>>>>>> >> flies. Read: My dishy can use any satellite that any of these
>>>>>>>>>> three can
>>>>>>>>>> >> see, and then depending on where I put it and how much of the
>>>>>>>>>> southern
>>>>>>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away,
>>>>>>>>>> although that
>>>>>>>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>>>>>>>> options
>>>>>>>>>> >> for my bits to travel without needing a lot of handovers. Why?
>>>>>>>>>> Easy: If
>>>>>>>>>> >> your nearest teleport is close by, then the set of satellites
>>>>>>>>>> that the
>>>>>>>>>> >> teleport can see and the set that you can see is almost the
>>>>>>>>>> same, so you
>>>>>>>>>> >> can essentially stick with the same satellite while it's in
>>>>>>>>>> view for you
>>>>>>>>>> >> because it'll also be in view for the teleport. Pretty much
>>>>>>>>>> any bird
>>>>>>>>>> >> above you will do.
>>>>>>>>>> >>
>>>>>>>>>> >> And because I don't get a lot of competition from other users
>>>>>>>>>> in my area
>>>>>>>>>> >> vying for one of the few available satellites that can see
>>>>>>>>>> both us and
>>>>>>>>>> >> the teleport, this is about as good as it gets at 37S
>>>>>>>>>> latitude. If I'd
>>>>>>>>>> >> want it any better, I'd have to move a lot further south.
>>>>>>>>>> >>
>>>>>>>>>> >> It'd be interesting to hear from Jonathan what the
>>>>>>>>>> availability of home
>>>>>>>>>> >> broadband is like in the Dallas area. I note that it's at a
>>>>>>>>>> lower
>>>>>>>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I
>>>>>>>>>> notice
>>>>>>>>>> >> two teleports each about 160 km away, which is also not too
>>>>>>>>>> bad. I also
>>>>>>>>>> >> note Starlink availability in the area is restricted at the
>>>>>>>>>> moment -
>>>>>>>>>> >> oversubscribed? But if Jonathan gets good data rates, then
>>>>>>>>>> that means
>>>>>>>>>> >> that competition for bird capacity can't be too bad - for
>>>>>>>>>> whatever reason.
>>>>>>>>>> >
>>>>>>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>>>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>>>>>>> broadband options. But there are also many people that live outside cities,
>>>>>>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>>>>>>> and Texas is probably ideal conditions for Starlink.
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>>>>>>> whatever I
>>>>>>>>>> >> talk to in terms of my communications. Traceroute shows about
>>>>>>>>>> 10 hops
>>>>>>>>>> >> between me and the University of Auckland via Starlink. That's
>>>>>>>>>> 10
>>>>>>>>>> >> queues, not one. Many of them will have cross traffic. So it's
>>>>>>>>>> a bit
>>>>>>>>>> >> hard to tell where our packets really get to wait or where
>>>>>>>>>> they get
>>>>>>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>>>>>>> between 1
>>>>>>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they
>>>>>>>>>> can all
>>>>>>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link
>>>>>>>>>> of a few
>>>>>>>>>> >> Mb/s where it's obvious where the dominant long latency
>>>>>>>>>> bottleneck in
>>>>>>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>>>>>>> whether a
>>>>>>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>>>>>>> system or
>>>>>>>>>> >> somewhere between Starlink's systems and the target system.
>>>>>>>>>> >>
>>>>>>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical
>>>>>>>>>> latency
>>>>>>>>>> >> should be under 15 ms. So there's clearly a bit of buffer here
>>>>>>>>>> along the
>>>>>>>>>> >> chain that occasionally fills up.
>>>>>>>>>> >>
>>>>>>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>>>>>>> inevitably
>>>>>>>>>> >> associated with a change in RTT and in most cases also
>>>>>>>>>> available
>>>>>>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>>>>>>>> >> satellite while your TCP is still trying to respond to
>>>>>>>>>> whatever it
>>>>>>>>>> >> thought the queue on the previous bird was doing. Read:
>>>>>>>>>> Whatever your
>>>>>>>>>> >> cwnd is immediately after a handover, it's probably not what
>>>>>>>>>> it should be.
>>>>>>>>>> >>
>>>>>>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four
>>>>>>>>>> Ookla
>>>>>>>>>> >> speedtest.net tests each to five local servers. Average
>>>>>>>>>> upload rate was
>>>>>>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the
>>>>>>>>>> ISP that
>>>>>>>>>> >> Starlink seems to be buying its local connectivity from (Vocus
>>>>>>>>>> Group)
>>>>>>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and
>>>>>>>>>> 52.22
>>>>>>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to
>>>>>>>>>> rates
>>>>>>>>>> >> observed. In fact, they were the ISP with consistently the
>>>>>>>>>> worst rates.
>>>>>>>>>> >>
>>>>>>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s
>>>>>>>>>> up and
>>>>>>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly
>>>>>>>>>> correlating
>>>>>>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>>>>>>> >>
>>>>>>>>>> >> Note the variation though: More or less a factor of two
>>>>>>>>>> between highest
>>>>>>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky
>>>>>>>>>> in my
>>>>>>>>>> >> tests? Or is there something systematic behind this? Way too
>>>>>>>>>> few tests
>>>>>>>>>> >> to tell.
>>>>>>>>>> >>
>>>>>>>>>> >> What these tests do is establish a ballpark.
>>>>>>>>>> >>
>>>>>>>>>> >> I'm currently repeating tests with dish placed on a trestle
>>>>>>>>>> closer to
>>>>>>>>>> >> the heavens. This seems to have translated into fewer outages
>>>>>>>>>> / ping
>>>>>>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the
>>>>>>>>>> ground on
>>>>>>>>>> >> my deck). Still good enough for a lengthy video Skype call
>>>>>>>>>> with my folks
>>>>>>>>>> >> in Germany, although they did comment about reduced video
>>>>>>>>>> quality. But
>>>>>>>>>> >> maybe that was the lighting or the different background as I
>>>>>>>>>> wasn't in
>>>>>>>>>> >> my usual spot with my laptop when I called them.
>>>>>>>>>> >
>>>>>>>>>> > Clear view of the sky is king for Starlink reliability. I've
>>>>>>>>>> got my dishy mounted on the back fence, looking up over an empty field, so
>>>>>>>>>> it's pretty much best-case scenario here.
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >> --
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> ****************************************************************
>>>>>>>>>> >> Dr. Ulrich Speidel
>>>>>>>>>> >>
>>>>>>>>>> >> School of Computer Science
>>>>>>>>>> >>
>>>>>>>>>> >> Room 303S.594 (City Campus)
>>>>>>>>>> >>
>>>>>>>>>> >> The University of Auckland
>>>>>>>>>> >> u.speidel@auckland.ac.nz
>>>>>>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>>>>>>> >>
>>>>>>>>>> ****************************************************************
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >> _______________________________________________
>>>>>>>>>> >> Starlink mailing list
>>>>>>>>>> >> Starlink@lists.bufferbloat.net
>>>>>>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>> >
>>>>>>>>>> > _______________________________________________
>>>>>>>>>> > Starlink mailing list
>>>>>>>>>> > Starlink@lists.bufferbloat.net
>>>>>>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>>>>>
>>>>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> Starlink mailing list
>>>>>>>>> Starlink@lists.bufferbloat.net
>>>>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>
>>>>>>>>
>>>>>>
>>>>>> --
>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>
>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>> _______________________________________________
>>>>>> Starlink mailing list
>>>>>> Starlink@lists.bufferbloat.net
>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>
>>>>> _______________________________________________
>>>> Starlink mailing list
>>>> Starlink@lists.bufferbloat.net
>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>
>>> _______________________________________________
>>> Starlink mailing list
>>> Starlink@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/starlink
>>>
>>
>>
>> --
>> This song goes out to all the folk that thought Stadia would work:
>>
>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>> Dave Täht CEO, TekLibre, LLC
>>
>

[-- Attachment #1.2: Type: text/html, Size: 22288 bytes --]

[-- Attachment #2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #3: Screenshot 2023-01-13 at 1.30.03 PM.png --]
[-- Type: image/png, Size: 855840 bytes --]

[-- Attachment #4: image.png --]
[-- Type: image/png, Size: 296989 bytes --]

[-- Attachment #5: image.png --]
[-- Type: image/png, Size: 222655 bytes --]

[-- Attachment #6: Screenshot 2023-01-14 at 6.12.53 AM.png --]
[-- Type: image/png, Size: 921914 bytes --]

[-- Attachment #7: Screenshot 2023-01-14 at 6.13.50 AM.png --]
[-- Type: image/png, Size: 337863 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
       [not found]                                                     ` <CALjsLJv5cbfHfkxqHnbjxoVHczspYvxc_jrshzs1CpHLEDWyew@mail.gmail.com>
  2023-01-14 14:20                                                       ` Nathan Owens
@ 2023-01-14 15:52                                                       ` Dave Taht
  1 sibling, 0 replies; 44+ messages in thread
From: Dave Taht @ 2023-01-14 15:52 UTC (permalink / raw)
  To: Nathan Owens; +Cc: Jonathan Bennett, starlink, Sina Khanifar


[-- Attachment #1.1: Type: text/plain, Size: 18546 bytes --]

On Sat, Jan 14, 2023 at 6:14 AM Nathan Owens <nathan@nathan.io> wrote:

> I realized I goofed up my visualizations, here's all of them again:
> I see a ton of loss on all of these!
>

Lovely pics, thank you, and the sawtooths are obvious, for a change, so in
my mind they've cut the buffering back quite a bit since we last checked.
The amount of TCP related loss remains low, at least according to the
packet captures (not enough loss for tcp! a BDP should settle in at 40ms
with a fifo, not 160+!).

Kind of related to my "glitches per minute" metric described here:
https://blog.cerowrt.org/post/speedtests/

What I'm interested in is the distribution pattern of the irtt udp loss. A
way to plot that might be along the -1 vertical axis, but incremented down
one further for every consecutive loss, and dots, colored,
blue->green->yellow->orange->red - voip can survive 3/5 losses in a row
fairly handily. Two ways

 + axis as it is now
----------------------------------------got a packet
---------------------------------------------
B  loss 1                                                       loss 1

G         2 loss in a row                                             loss 2
Y                                 3rd loss
         los 3
O
                        4
R
                           5
R
                              ....
R
R
R

and/or the same as above but plotting the rx vs tx loss



>
>
> While Luis was downloading/uploading... oof.
> [image: Screenshot 2023-01-14 at 6.13.50 AM.png]
>
> [image: Screenshot 2023-01-14 at 6.12.53 AM.png]
>
>
>
>
> On Fri, Jan 13, 2023 at 3:44 PM Dave Taht <dave.taht@gmail.com> wrote:
>
>> I am forced to conclude that waveform's upload test is broken in some
>> cases and has been for some time. All the joy many have been feeling about
>> their uplink speeds has to be cast away. Felt good, though, didn't it?
>>
>> There is a *slight* possibility that there is some fq in the starlink
>> network on tcp port 433. The first showing normal rtt growth and loss for
>> cubic, the second, a low rate flow that is holding the line... but I didn't
>> check to see if these were sequential or parallel.
>>
>> The last is a cwnd plot, clearly showing the cubic sawtooth on the
>> upload.
>>
>> It's weirdly nice to be able to follow a port 433 stream, see the tls
>> handshake put the website en'clair, and the rest go dark, and still trace
>> the behaviors.
>>
>> we're so going to lose this analytical ability with quic. I'm enjoying it
>> while we still can.
>>
>>
>> On Fri, Jan 13, 2023 at 2:10 PM Jonathan Bennett via Starlink <
>> starlink@lists.bufferbloat.net> wrote:
>>
>>> The irtt run finished a few seconds before the flent run, but here are
>>> the results:
>>>
>>>
>>> https://drive.google.com/file/d/1FKve13ssUMW1LLWOXLM2931Yx6uMHw8K/view?usp=share_link
>>>
>>> https://drive.google.com/file/d/1ZXd64A0pfUedLr3FyhDNTHA7vxv8S2Gk/view?usp=share_link
>>>
>>> https://drive.google.com/file/d/1rx64UPQHHz3IMNiJtb1oFqtqw2DvhvEE/view?usp=share_link
>>>
>>>
>>> [image: image.png]
>>> [image: image.png]
>>>
>>>
>>> Jonathan Bennett
>>> Hackaday.com
>>>
>>>
>>> On Fri, Jan 13, 2023 at 3:30 PM Nathan Owens via Starlink <
>>> starlink@lists.bufferbloat.net> wrote:
>>>
>>>> Here's Luis's run -- the top line below the edge of the graph is 200ms
>>>> [image: Screenshot 2023-01-13 at 1.30.03 PM.png]
>>>>
>>>>
>>>> On Fri, Jan 13, 2023 at 1:25 PM Luis A. Cornejo <
>>>> luis.a.cornejo@gmail.com> wrote:
>>>>
>>>>> Dave,
>>>>>
>>>>> Here is a run the way I think you wanted it.
>>>>>
>>>>> irtt running for 5 min to your dallas server, followed by a waveform
>>>>> test, then a few seconds of inactivity, cloudflare test, a few more secs of
>>>>> nothing, flent test to dallas. Packet capture is currently uploading (will
>>>>> be done in 20 min or so), irtt JSON also in there (.zip file):
>>>>>
>>>>>
>>>>> https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link
>>>>>
>>>>> -Luis
>>>>>
>>>>> On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io>
>>>>>> wrote:
>>>>>>
>>>>>>> Here's the data visualization for Johnathan's Data
>>>>>>>
>>>>>>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>>>>>>
>>>>>>> You can see the path change at :12, :27, :42, :57 after the minute.
>>>>>>> Some paths are clearly busier than others with increased loss, latency, and
>>>>>>> jitter.
>>>>>>>
>>>>>>
>>>>>> I am so glad to see loss and bounded delay here. Also a bit of rigor
>>>>>> regarding what traffic was active locally vs on the path would be nice,
>>>>>> although it seems to line up with the known 15s starlink switchover thing
>>>>>> (need a name for this), in this case, doing a few speedtests
>>>>>> while that irtt is running would show the impact(s) of whatever else
>>>>>> they are up to.
>>>>>>
>>>>>> It would also be my hope that the loss distribution in the middle
>>>>>> portion of this data is good, not bursty, but we don't have a tool to take
>>>>>> apart that. (I am so hopeless at json)
>>>>>>
>>>>>>
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> I’ll run my visualization code on this result this afternoon and
>>>>>>>> report back!
>>>>>>>>
>>>>>>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>>
>>>>>>>>> The irtt command, run with normal, light usage:
>>>>>>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>>>>>>
>>>>>>>>> Jonathan Bennett
>>>>>>>>> Hackaday.com
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> packet caps would be nice... all this is very exciting news.
>>>>>>>>>>
>>>>>>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>>>>>>> results nowadays to duplicate and re-plot the original irtt tests
>>>>>>>>>> we
>>>>>>>>>> did:
>>>>>>>>>>
>>>>>>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net
>>>>>>>>>> -o whatever.json
>>>>>>>>>>
>>>>>>>>>> They MUST have changed their scheduling to get such amazing uplink
>>>>>>>>>> results, in addition to better queue management.
>>>>>>>>>>
>>>>>>>>>> (for the record, my servers are de, london, fremont, sydney,
>>>>>>>>>> dallas,
>>>>>>>>>> newark, atlanta, singapore, mumbai)
>>>>>>>>>>
>>>>>>>>>> There's an R and gnuplot script for plotting that output around
>>>>>>>>>> here
>>>>>>>>>> somewhere (I have largely personally put down the starlink
>>>>>>>>>> project,
>>>>>>>>>> loaning out mine) - that went by on this list... I should have
>>>>>>>>>> written
>>>>>>>>>> a blog entry so I can find that stuff again.
>>>>>>>>>>
>>>>>>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>>>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>>>>>>> >
>>>>>>>>>> >
>>>>>>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>>>> >>
>>>>>>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>>>>>>> >> >
>>>>>>>>>> >> > From Auckland, New Zealand, using a roaming subscription, it
>>>>>>>>>> puts me
>>>>>>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>>>>>>> >> >
>>>>>>>>>> >> >
>>>>>>>>>> >> > IP address: nix six.
>>>>>>>>>> >> >
>>>>>>>>>> >> > My thoughts shall follow later.
>>>>>>>>>> >>
>>>>>>>>>> >> OK, so here we go.
>>>>>>>>>> >>
>>>>>>>>>> >> I'm always a bit skeptical when it comes to speed tests -
>>>>>>>>>> they're really
>>>>>>>>>> >> laden with so many caveats that it's not funny. I took our new
>>>>>>>>>> work
>>>>>>>>>> >> Starlink kit home in December to give it a try and the other
>>>>>>>>>> day finally
>>>>>>>>>> >> got around to set it up. It's on a roaming subscription
>>>>>>>>>> because our
>>>>>>>>>> >> badly built-up campus really isn't ideal in terms of a clear
>>>>>>>>>> view of the
>>>>>>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>>>>>>> adapter,
>>>>>>>>>> >> not the WiFi?
>>>>>>>>>> >>
>>>>>>>>>> >> Caveat 1: Location, location. I live in a place where the best
>>>>>>>>>> Starlink
>>>>>>>>>> >> promises is about 1/3 in terms of data rate you can actually
>>>>>>>>>> get from
>>>>>>>>>> >> fibre to the home at under half of Starlink's price. Read:
>>>>>>>>>> There are few
>>>>>>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>>>>>>> >>
>>>>>>>>>> >> Caveat 2: Auckland has three Starlink gateways close by:
>>>>>>>>>> Clevedon (which
>>>>>>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana
>>>>>>>>>> and Puwera,
>>>>>>>>>> >> the most distant of the three and about 130 km away from me as
>>>>>>>>>> the crow
>>>>>>>>>> >> flies. Read: My dishy can use any satellite that any of these
>>>>>>>>>> three can
>>>>>>>>>> >> see, and then depending on where I put it and how much of the
>>>>>>>>>> southern
>>>>>>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away,
>>>>>>>>>> although that
>>>>>>>>>> >> is obviously stretching it a bit. Either way, that's plenty of
>>>>>>>>>> options
>>>>>>>>>> >> for my bits to travel without needing a lot of handovers. Why?
>>>>>>>>>> Easy: If
>>>>>>>>>> >> your nearest teleport is close by, then the set of satellites
>>>>>>>>>> that the
>>>>>>>>>> >> teleport can see and the set that you can see is almost the
>>>>>>>>>> same, so you
>>>>>>>>>> >> can essentially stick with the same satellite while it's in
>>>>>>>>>> view for you
>>>>>>>>>> >> because it'll also be in view for the teleport. Pretty much
>>>>>>>>>> any bird
>>>>>>>>>> >> above you will do.
>>>>>>>>>> >>
>>>>>>>>>> >> And because I don't get a lot of competition from other users
>>>>>>>>>> in my area
>>>>>>>>>> >> vying for one of the few available satellites that can see
>>>>>>>>>> both us and
>>>>>>>>>> >> the teleport, this is about as good as it gets at 37S
>>>>>>>>>> latitude. If I'd
>>>>>>>>>> >> want it any better, I'd have to move a lot further south.
>>>>>>>>>> >>
>>>>>>>>>> >> It'd be interesting to hear from Jonathan what the
>>>>>>>>>> availability of home
>>>>>>>>>> >> broadband is like in the Dallas area. I note that it's at a
>>>>>>>>>> lower
>>>>>>>>>> >> latitude (33N) than Auckland, but the difference isn't huge. I
>>>>>>>>>> notice
>>>>>>>>>> >> two teleports each about 160 km away, which is also not too
>>>>>>>>>> bad. I also
>>>>>>>>>> >> note Starlink availability in the area is restricted at the
>>>>>>>>>> moment -
>>>>>>>>>> >> oversubscribed? But if Jonathan gets good data rates, then
>>>>>>>>>> that means
>>>>>>>>>> >> that competition for bird capacity can't be too bad - for
>>>>>>>>>> whatever reason.
>>>>>>>>>> >
>>>>>>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>>>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>>>>>>> broadband options. But there are also many people that live outside cities,
>>>>>>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>>>>>>> and Texas is probably ideal conditions for Starlink.
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>>>>>>> whatever I
>>>>>>>>>> >> talk to in terms of my communications. Traceroute shows about
>>>>>>>>>> 10 hops
>>>>>>>>>> >> between me and the University of Auckland via Starlink. That's
>>>>>>>>>> 10
>>>>>>>>>> >> queues, not one. Many of them will have cross traffic. So it's
>>>>>>>>>> a bit
>>>>>>>>>> >> hard to tell where our packets really get to wait or where
>>>>>>>>>> they get
>>>>>>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>>>>>>> between 1
>>>>>>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they
>>>>>>>>>> can all
>>>>>>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link
>>>>>>>>>> of a few
>>>>>>>>>> >> Mb/s where it's obvious where the dominant long latency
>>>>>>>>>> bottleneck in
>>>>>>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>>>>>>> whether a
>>>>>>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>>>>>>> system or
>>>>>>>>>> >> somewhere between Starlink's systems and the target system.
>>>>>>>>>> >>
>>>>>>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical
>>>>>>>>>> latency
>>>>>>>>>> >> should be under 15 ms. So there's clearly a bit of buffer here
>>>>>>>>>> along the
>>>>>>>>>> >> chain that occasionally fills up.
>>>>>>>>>> >>
>>>>>>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>>>>>>> inevitably
>>>>>>>>>> >> associated with a change in RTT and in most cases also
>>>>>>>>>> available
>>>>>>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a new
>>>>>>>>>> >> satellite while your TCP is still trying to respond to
>>>>>>>>>> whatever it
>>>>>>>>>> >> thought the queue on the previous bird was doing. Read:
>>>>>>>>>> Whatever your
>>>>>>>>>> >> cwnd is immediately after a handover, it's probably not what
>>>>>>>>>> it should be.
>>>>>>>>>> >>
>>>>>>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four
>>>>>>>>>> Ookla
>>>>>>>>>> >> speedtest.net tests each to five local servers. Average
>>>>>>>>>> upload rate was
>>>>>>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the
>>>>>>>>>> ISP that
>>>>>>>>>> >> Starlink seems to be buying its local connectivity from (Vocus
>>>>>>>>>> Group)
>>>>>>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33 and
>>>>>>>>>> 52.22
>>>>>>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to
>>>>>>>>>> rates
>>>>>>>>>> >> observed. In fact, they were the ISP with consistently the
>>>>>>>>>> worst rates.
>>>>>>>>>> >>
>>>>>>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s
>>>>>>>>>> up and
>>>>>>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly
>>>>>>>>>> correlating
>>>>>>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>>>>>>> >>
>>>>>>>>>> >> Note the variation though: More or less a factor of two
>>>>>>>>>> between highest
>>>>>>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky
>>>>>>>>>> in my
>>>>>>>>>> >> tests? Or is there something systematic behind this? Way too
>>>>>>>>>> few tests
>>>>>>>>>> >> to tell.
>>>>>>>>>> >>
>>>>>>>>>> >> What these tests do is establish a ballpark.
>>>>>>>>>> >>
>>>>>>>>>> >> I'm currently repeating tests with dish placed on a trestle
>>>>>>>>>> closer to
>>>>>>>>>> >> the heavens. This seems to have translated into fewer outages
>>>>>>>>>> / ping
>>>>>>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the
>>>>>>>>>> ground on
>>>>>>>>>> >> my deck). Still good enough for a lengthy video Skype call
>>>>>>>>>> with my folks
>>>>>>>>>> >> in Germany, although they did comment about reduced video
>>>>>>>>>> quality. But
>>>>>>>>>> >> maybe that was the lighting or the different background as I
>>>>>>>>>> wasn't in
>>>>>>>>>> >> my usual spot with my laptop when I called them.
>>>>>>>>>> >
>>>>>>>>>> > Clear view of the sky is king for Starlink reliability. I've
>>>>>>>>>> got my dishy mounted on the back fence, looking up over an empty field, so
>>>>>>>>>> it's pretty much best-case scenario here.
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >> --
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> ****************************************************************
>>>>>>>>>> >> Dr. Ulrich Speidel
>>>>>>>>>> >>
>>>>>>>>>> >> School of Computer Science
>>>>>>>>>> >>
>>>>>>>>>> >> Room 303S.594 (City Campus)
>>>>>>>>>> >>
>>>>>>>>>> >> The University of Auckland
>>>>>>>>>> >> u.speidel@auckland.ac.nz
>>>>>>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>>>>>>> >>
>>>>>>>>>> ****************************************************************
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >> _______________________________________________
>>>>>>>>>> >> Starlink mailing list
>>>>>>>>>> >> Starlink@lists.bufferbloat.net
>>>>>>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>> >
>>>>>>>>>> > _______________________________________________
>>>>>>>>>> > Starlink mailing list
>>>>>>>>>> > Starlink@lists.bufferbloat.net
>>>>>>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>>>>>
>>>>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> Starlink mailing list
>>>>>>>>> Starlink@lists.bufferbloat.net
>>>>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>
>>>>>>>>
>>>>>>
>>>>>> --
>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>
>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>> _______________________________________________
>>>>>> Starlink mailing list
>>>>>> Starlink@lists.bufferbloat.net
>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>
>>>>> _______________________________________________
>>>> Starlink mailing list
>>>> Starlink@lists.bufferbloat.net
>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>
>>> _______________________________________________
>>> Starlink mailing list
>>> Starlink@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/starlink
>>>
>>
>>
>> --
>> This song goes out to all the folk that thought Stadia would work:
>>
>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>> Dave Täht CEO, TekLibre, LLC
>>
>

-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

[-- Attachment #1.2: Type: text/html, Size: 24676 bytes --]

[-- Attachment #2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #3: Screenshot 2023-01-13 at 1.30.03 PM.png --]
[-- Type: image/png, Size: 855840 bytes --]

[-- Attachment #4: image.png --]
[-- Type: image/png, Size: 296989 bytes --]

[-- Attachment #5: image.png --]
[-- Type: image/png, Size: 222655 bytes --]

[-- Attachment #6: Screenshot 2023-01-14 at 6.12.53 AM.png --]
[-- Type: image/png, Size: 921914 bytes --]

[-- Attachment #7: Screenshot 2023-01-14 at 6.13.50 AM.png --]
[-- Type: image/png, Size: 337863 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-14 14:20                                                       ` Nathan Owens
@ 2023-01-14 15:53                                                         ` Dave Taht
  2023-01-14 16:33                                                           ` Dave Taht
  0 siblings, 1 reply; 44+ messages in thread
From: Dave Taht @ 2023-01-14 15:53 UTC (permalink / raw)
  To: Nathan Owens; +Cc: Jonathan Bennett, starlink, Sina Khanifar


[-- Attachment #1.1: Type: text/plain, Size: 19271 bytes --]

Lovely pics, thank you, and the sawtooths are obvious, for a change, so in
my mind they've cut the buffering back quite a bit since we last checked.
The amount of TCP related loss remains low, at least according to the
packet captures (not enough loss for tcp! a BDP should settle in at 40ms
with a fifo, not 160+!).

Kind of related to my "glitches per minute" metric described here:
https://blog.cerowrt.org/post/speedtests/

What I'm interested in is the distribution pattern of the irtt udp loss. A
way to plot that might be along the -1 vertical axis, but incremented down
one further for every consecutive loss, and dots, colored,
blue->green->yellow->orange->red - voip can survive 3/5 losses in a row
fairly handily. Two ways

 + axis as it is now
----------------------------------------got a packet
---------------------------------------------
B  loss 1                                                       loss 1

G         2 loss in a row                                             loss 2
Y                                 3rd loss
         los 3
O
                        4
R
                           5
R
                              ....
R
R
R

and/or the same as above but plotting the rx vs tx loss


On Sat, Jan 14, 2023 at 6:20 AM Nathan Owens <nathan@nathan.io> wrote:

> Sorry, was rejected from the listserv, here's the google drive link for
> all 3 IRTT's visualized
>
>
> https://drive.google.com/drive/folders/1SbUyvdlfdllgqrEcSSxvGdn3yN40vSx-?usp=share_link
>
> On Sat, Jan 14, 2023 at 6:14 AM Nathan Owens <nathan@nathan.io> wrote:
>
>> I realized I goofed up my visualizations, here's all of them again:
>> I see a ton of loss on all of these!
>>
>> While Luis was downloading/uploading... oof.
>> [image: Screenshot 2023-01-14 at 6.13.50 AM.png]
>>
>> [image: Screenshot 2023-01-14 at 6.12.53 AM.png]
>>
>>
>>
>>
>> On Fri, Jan 13, 2023 at 3:44 PM Dave Taht <dave.taht@gmail.com> wrote:
>>
>>> I am forced to conclude that waveform's upload test is broken in some
>>> cases and has been for some time. All the joy many have been feeling about
>>> their uplink speeds has to be cast away. Felt good, though, didn't it?
>>>
>>> There is a *slight* possibility that there is some fq in the starlink
>>> network on tcp port 433. The first showing normal rtt growth and loss for
>>> cubic, the second, a low rate flow that is holding the line... but I didn't
>>> check to see if these were sequential or parallel.
>>>
>>> The last is a cwnd plot, clearly showing the cubic sawtooth on the
>>> upload.
>>>
>>> It's weirdly nice to be able to follow a port 433 stream, see the tls
>>> handshake put the website en'clair, and the rest go dark, and still trace
>>> the behaviors.
>>>
>>> we're so going to lose this analytical ability with quic. I'm enjoying
>>> it while we still can.
>>>
>>>
>>> On Fri, Jan 13, 2023 at 2:10 PM Jonathan Bennett via Starlink <
>>> starlink@lists.bufferbloat.net> wrote:
>>>
>>>> The irtt run finished a few seconds before the flent run, but here are
>>>> the results:
>>>>
>>>>
>>>> https://drive.google.com/file/d/1FKve13ssUMW1LLWOXLM2931Yx6uMHw8K/view?usp=share_link
>>>>
>>>> https://drive.google.com/file/d/1ZXd64A0pfUedLr3FyhDNTHA7vxv8S2Gk/view?usp=share_link
>>>>
>>>> https://drive.google.com/file/d/1rx64UPQHHz3IMNiJtb1oFqtqw2DvhvEE/view?usp=share_link
>>>>
>>>>
>>>> [image: image.png]
>>>> [image: image.png]
>>>>
>>>>
>>>> Jonathan Bennett
>>>> Hackaday.com
>>>>
>>>>
>>>> On Fri, Jan 13, 2023 at 3:30 PM Nathan Owens via Starlink <
>>>> starlink@lists.bufferbloat.net> wrote:
>>>>
>>>>> Here's Luis's run -- the top line below the edge of the graph is 200ms
>>>>> [image: Screenshot 2023-01-13 at 1.30.03 PM.png]
>>>>>
>>>>>
>>>>> On Fri, Jan 13, 2023 at 1:25 PM Luis A. Cornejo <
>>>>> luis.a.cornejo@gmail.com> wrote:
>>>>>
>>>>>> Dave,
>>>>>>
>>>>>> Here is a run the way I think you wanted it.
>>>>>>
>>>>>> irtt running for 5 min to your dallas server, followed by a waveform
>>>>>> test, then a few seconds of inactivity, cloudflare test, a few more secs of
>>>>>> nothing, flent test to dallas. Packet capture is currently uploading (will
>>>>>> be done in 20 min or so), irtt JSON also in there (.zip file):
>>>>>>
>>>>>>
>>>>>> https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link
>>>>>>
>>>>>> -Luis
>>>>>>
>>>>>> On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Here's the data visualization for Johnathan's Data
>>>>>>>>
>>>>>>>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>>>>>>>
>>>>>>>> You can see the path change at :12, :27, :42, :57 after the minute.
>>>>>>>> Some paths are clearly busier than others with increased loss, latency, and
>>>>>>>> jitter.
>>>>>>>>
>>>>>>>
>>>>>>> I am so glad to see loss and bounded delay here. Also a bit of rigor
>>>>>>> regarding what traffic was active locally vs on the path would be nice,
>>>>>>> although it seems to line up with the known 15s starlink switchover thing
>>>>>>> (need a name for this), in this case, doing a few speedtests
>>>>>>> while that irtt is running would show the impact(s) of whatever else
>>>>>>> they are up to.
>>>>>>>
>>>>>>> It would also be my hope that the loss distribution in the middle
>>>>>>> portion of this data is good, not bursty, but we don't have a tool to take
>>>>>>> apart that. (I am so hopeless at json)
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> I’ll run my visualization code on this result this afternoon and
>>>>>>>>> report back!
>>>>>>>>>
>>>>>>>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>>>
>>>>>>>>>> The irtt command, run with normal, light usage:
>>>>>>>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>>>>>>>
>>>>>>>>>> Jonathan Bennett
>>>>>>>>>> Hackaday.com
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> packet caps would be nice... all this is very exciting news.
>>>>>>>>>>>
>>>>>>>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>>>>>>>> results nowadays to duplicate and re-plot the original irtt
>>>>>>>>>>> tests we
>>>>>>>>>>> did:
>>>>>>>>>>>
>>>>>>>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net
>>>>>>>>>>> -o whatever.json
>>>>>>>>>>>
>>>>>>>>>>> They MUST have changed their scheduling to get such amazing
>>>>>>>>>>> uplink
>>>>>>>>>>> results, in addition to better queue management.
>>>>>>>>>>>
>>>>>>>>>>> (for the record, my servers are de, london, fremont, sydney,
>>>>>>>>>>> dallas,
>>>>>>>>>>> newark, atlanta, singapore, mumbai)
>>>>>>>>>>>
>>>>>>>>>>> There's an R and gnuplot script for plotting that output around
>>>>>>>>>>> here
>>>>>>>>>>> somewhere (I have largely personally put down the starlink
>>>>>>>>>>> project,
>>>>>>>>>>> loaning out mine) - that went by on this list... I should have
>>>>>>>>>>> written
>>>>>>>>>>> a blog entry so I can find that stuff again.
>>>>>>>>>>>
>>>>>>>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>>>>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>>>>>>>> >
>>>>>>>>>>> >
>>>>>>>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>>>>> >>
>>>>>>>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>>>>>>>> >> >
>>>>>>>>>>> >> > From Auckland, New Zealand, using a roaming subscription,
>>>>>>>>>>> it puts me
>>>>>>>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>>>>>>>> >> >
>>>>>>>>>>> >> >
>>>>>>>>>>> >> > IP address: nix six.
>>>>>>>>>>> >> >
>>>>>>>>>>> >> > My thoughts shall follow later.
>>>>>>>>>>> >>
>>>>>>>>>>> >> OK, so here we go.
>>>>>>>>>>> >>
>>>>>>>>>>> >> I'm always a bit skeptical when it comes to speed tests -
>>>>>>>>>>> they're really
>>>>>>>>>>> >> laden with so many caveats that it's not funny. I took our
>>>>>>>>>>> new work
>>>>>>>>>>> >> Starlink kit home in December to give it a try and the other
>>>>>>>>>>> day finally
>>>>>>>>>>> >> got around to set it up. It's on a roaming subscription
>>>>>>>>>>> because our
>>>>>>>>>>> >> badly built-up campus really isn't ideal in terms of a clear
>>>>>>>>>>> view of the
>>>>>>>>>>> >> sky. Oh - and did I mention that I used the Starlink Ethernet
>>>>>>>>>>> adapter,
>>>>>>>>>>> >> not the WiFi?
>>>>>>>>>>> >>
>>>>>>>>>>> >> Caveat 1: Location, location. I live in a place where the
>>>>>>>>>>> best Starlink
>>>>>>>>>>> >> promises is about 1/3 in terms of data rate you can actually
>>>>>>>>>>> get from
>>>>>>>>>>> >> fibre to the home at under half of Starlink's price. Read:
>>>>>>>>>>> There are few
>>>>>>>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>>>>>>>> >>
>>>>>>>>>>> >> Caveat 2: Auckland has three Starlink gateways close by:
>>>>>>>>>>> Clevedon (which
>>>>>>>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana
>>>>>>>>>>> and Puwera,
>>>>>>>>>>> >> the most distant of the three and about 130 km away from me
>>>>>>>>>>> as the crow
>>>>>>>>>>> >> flies. Read: My dishy can use any satellite that any of these
>>>>>>>>>>> three can
>>>>>>>>>>> >> see, and then depending on where I put it and how much of the
>>>>>>>>>>> southern
>>>>>>>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away,
>>>>>>>>>>> although that
>>>>>>>>>>> >> is obviously stretching it a bit. Either way, that's plenty
>>>>>>>>>>> of options
>>>>>>>>>>> >> for my bits to travel without needing a lot of handovers.
>>>>>>>>>>> Why? Easy: If
>>>>>>>>>>> >> your nearest teleport is close by, then the set of satellites
>>>>>>>>>>> that the
>>>>>>>>>>> >> teleport can see and the set that you can see is almost the
>>>>>>>>>>> same, so you
>>>>>>>>>>> >> can essentially stick with the same satellite while it's in
>>>>>>>>>>> view for you
>>>>>>>>>>> >> because it'll also be in view for the teleport. Pretty much
>>>>>>>>>>> any bird
>>>>>>>>>>> >> above you will do.
>>>>>>>>>>> >>
>>>>>>>>>>> >> And because I don't get a lot of competition from other users
>>>>>>>>>>> in my area
>>>>>>>>>>> >> vying for one of the few available satellites that can see
>>>>>>>>>>> both us and
>>>>>>>>>>> >> the teleport, this is about as good as it gets at 37S
>>>>>>>>>>> latitude. If I'd
>>>>>>>>>>> >> want it any better, I'd have to move a lot further south.
>>>>>>>>>>> >>
>>>>>>>>>>> >> It'd be interesting to hear from Jonathan what the
>>>>>>>>>>> availability of home
>>>>>>>>>>> >> broadband is like in the Dallas area. I note that it's at a
>>>>>>>>>>> lower
>>>>>>>>>>> >> latitude (33N) than Auckland, but the difference isn't huge.
>>>>>>>>>>> I notice
>>>>>>>>>>> >> two teleports each about 160 km away, which is also not too
>>>>>>>>>>> bad. I also
>>>>>>>>>>> >> note Starlink availability in the area is restricted at the
>>>>>>>>>>> moment -
>>>>>>>>>>> >> oversubscribed? But if Jonathan gets good data rates, then
>>>>>>>>>>> that means
>>>>>>>>>>> >> that competition for bird capacity can't be too bad - for
>>>>>>>>>>> whatever reason.
>>>>>>>>>>> >
>>>>>>>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>>>>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>>>>>>>> broadband options. But there are also many people that live outside cities,
>>>>>>>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>>>>>>>> and Texas is probably ideal conditions for Starlink.
>>>>>>>>>>> >>
>>>>>>>>>>> >>
>>>>>>>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me and
>>>>>>>>>>> whatever I
>>>>>>>>>>> >> talk to in terms of my communications. Traceroute shows about
>>>>>>>>>>> 10 hops
>>>>>>>>>>> >> between me and the University of Auckland via Starlink.
>>>>>>>>>>> That's 10
>>>>>>>>>>> >> queues, not one. Many of them will have cross traffic. So
>>>>>>>>>>> it's a bit
>>>>>>>>>>> >> hard to tell where our packets really get to wait or where
>>>>>>>>>>> they get
>>>>>>>>>>> >> dropped. The insidious bit here is that a lot of them will be
>>>>>>>>>>> between 1
>>>>>>>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic, they
>>>>>>>>>>> can all
>>>>>>>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link
>>>>>>>>>>> of a few
>>>>>>>>>>> >> Mb/s where it's obvious where the dominant long latency
>>>>>>>>>>> bottleneck in
>>>>>>>>>>> >> your TCP connection's path is. Read: It's pretty hard to tell
>>>>>>>>>>> whether a
>>>>>>>>>>> >> drop in "speed" is due to a performance issue in the Starlink
>>>>>>>>>>> system or
>>>>>>>>>>> >> somewhere between Starlink's systems and the target system.
>>>>>>>>>>> >>
>>>>>>>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical
>>>>>>>>>>> latency
>>>>>>>>>>> >> should be under 15 ms. So there's clearly a bit of buffer
>>>>>>>>>>> here along the
>>>>>>>>>>> >> chain that occasionally fills up.
>>>>>>>>>>> >>
>>>>>>>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>>>>>>>> inevitably
>>>>>>>>>>> >> associated with a change in RTT and in most cases also
>>>>>>>>>>> available
>>>>>>>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a
>>>>>>>>>>> new
>>>>>>>>>>> >> satellite while your TCP is still trying to respond to
>>>>>>>>>>> whatever it
>>>>>>>>>>> >> thought the queue on the previous bird was doing. Read:
>>>>>>>>>>> Whatever your
>>>>>>>>>>> >> cwnd is immediately after a handover, it's probably not what
>>>>>>>>>>> it should be.
>>>>>>>>>>> >>
>>>>>>>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four
>>>>>>>>>>> Ookla
>>>>>>>>>>> >> speedtest.net tests each to five local servers. Average
>>>>>>>>>>> upload rate was
>>>>>>>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the
>>>>>>>>>>> ISP that
>>>>>>>>>>> >> Starlink seems to be buying its local connectivity from
>>>>>>>>>>> (Vocus Group)
>>>>>>>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33
>>>>>>>>>>> and 52.22
>>>>>>>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to
>>>>>>>>>>> rates
>>>>>>>>>>> >> observed. In fact, they were the ISP with consistently the
>>>>>>>>>>> worst rates.
>>>>>>>>>>> >>
>>>>>>>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s
>>>>>>>>>>> up and
>>>>>>>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly
>>>>>>>>>>> correlating
>>>>>>>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>>>>>>>> >>
>>>>>>>>>>> >> Note the variation though: More or less a factor of two
>>>>>>>>>>> between highest
>>>>>>>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky
>>>>>>>>>>> in my
>>>>>>>>>>> >> tests? Or is there something systematic behind this? Way too
>>>>>>>>>>> few tests
>>>>>>>>>>> >> to tell.
>>>>>>>>>>> >>
>>>>>>>>>>> >> What these tests do is establish a ballpark.
>>>>>>>>>>> >>
>>>>>>>>>>> >> I'm currently repeating tests with dish placed on a trestle
>>>>>>>>>>> closer to
>>>>>>>>>>> >> the heavens. This seems to have translated into fewer outages
>>>>>>>>>>> / ping
>>>>>>>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the
>>>>>>>>>>> ground on
>>>>>>>>>>> >> my deck). Still good enough for a lengthy video Skype call
>>>>>>>>>>> with my folks
>>>>>>>>>>> >> in Germany, although they did comment about reduced video
>>>>>>>>>>> quality. But
>>>>>>>>>>> >> maybe that was the lighting or the different background as I
>>>>>>>>>>> wasn't in
>>>>>>>>>>> >> my usual spot with my laptop when I called them.
>>>>>>>>>>> >
>>>>>>>>>>> > Clear view of the sky is king for Starlink reliability. I've
>>>>>>>>>>> got my dishy mounted on the back fence, looking up over an empty field, so
>>>>>>>>>>> it's pretty much best-case scenario here.
>>>>>>>>>>> >>
>>>>>>>>>>> >>
>>>>>>>>>>> >> --
>>>>>>>>>>> >>
>>>>>>>>>>> >>
>>>>>>>>>>> ****************************************************************
>>>>>>>>>>> >> Dr. Ulrich Speidel
>>>>>>>>>>> >>
>>>>>>>>>>> >> School of Computer Science
>>>>>>>>>>> >>
>>>>>>>>>>> >> Room 303S.594 (City Campus)
>>>>>>>>>>> >>
>>>>>>>>>>> >> The University of Auckland
>>>>>>>>>>> >> u.speidel@auckland.ac.nz
>>>>>>>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>>>>>>>> >>
>>>>>>>>>>> ****************************************************************
>>>>>>>>>>> >>
>>>>>>>>>>> >>
>>>>>>>>>>> >>
>>>>>>>>>>> >> _______________________________________________
>>>>>>>>>>> >> Starlink mailing list
>>>>>>>>>>> >> Starlink@lists.bufferbloat.net
>>>>>>>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>>> >
>>>>>>>>>>> > _______________________________________________
>>>>>>>>>>> > Starlink mailing list
>>>>>>>>>>> > Starlink@lists.bufferbloat.net
>>>>>>>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> This song goes out to all the folk that thought Stadia would
>>>>>>>>>>> work:
>>>>>>>>>>>
>>>>>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>>>>>>
>>>>>>>>>> _______________________________________________
>>>>>>>>>> Starlink mailing list
>>>>>>>>>> Starlink@lists.bufferbloat.net
>>>>>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>>
>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>> _______________________________________________
>>>>>>> Starlink mailing list
>>>>>>> Starlink@lists.bufferbloat.net
>>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>
>>>>>> _______________________________________________
>>>>> Starlink mailing list
>>>>> Starlink@lists.bufferbloat.net
>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>
>>>> _______________________________________________
>>>> Starlink mailing list
>>>> Starlink@lists.bufferbloat.net
>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>
>>>
>>>
>>> --
>>> This song goes out to all the folk that thought Stadia would work:
>>>
>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>> Dave Täht CEO, TekLibre, LLC
>>>
>>

-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

[-- Attachment #1.2: Type: text/html, Size: 25327 bytes --]

[-- Attachment #2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #3: Screenshot 2023-01-13 at 1.30.03 PM.png --]
[-- Type: image/png, Size: 855840 bytes --]

[-- Attachment #4: image.png --]
[-- Type: image/png, Size: 296989 bytes --]

[-- Attachment #5: image.png --]
[-- Type: image/png, Size: 222655 bytes --]

[-- Attachment #6: Screenshot 2023-01-14 at 6.12.53 AM.png --]
[-- Type: image/png, Size: 921914 bytes --]

[-- Attachment #7: Screenshot 2023-01-14 at 6.13.50 AM.png --]
[-- Type: image/png, Size: 337863 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

* Re: [Starlink] insanely great waveform result for starlink
  2023-01-14 15:53                                                         ` Dave Taht
@ 2023-01-14 16:33                                                           ` Dave Taht
  0 siblings, 0 replies; 44+ messages in thread
From: Dave Taht @ 2023-01-14 16:33 UTC (permalink / raw)
  To: Nathan Owens; +Cc: Jonathan Bennett, starlink, Sina Khanifar


[-- Attachment #1.1: Type: text/plain, Size: 20680 bytes --]

On Sat, Jan 14, 2023 at 7:53 AM Dave Taht <dave.taht@gmail.com> wrote:

> Lovely pics, thank you, and the sawtooths are obvious, for a change, so in
> my mind they've cut the buffering back quite a bit since we last checked.
> The amount of TCP related loss remains low, at least according to the
> packet captures (not enough loss for tcp! a BDP should settle in at 40ms
> with a fifo, not 160+!).
>
> Kind of related to my "glitches per minute" metric described here:
> https://blog.cerowrt.org/post/speedtests/
>
> What I'm interested in is the distribution pattern of the irtt udp loss. A
> way to plot that might be along the -1 vertical axis, but incremented down
> one further for every consecutive loss, and dots, colored,
> blue->green->yellow->orange->red - voip can survive 3/5 losses in a row
> fairly handily. Two ways
>
>  + axis as it is now
> ----------------------------------------got a packet
> ---------------------------------------------
> B  loss 1                                                       loss 1
>
> G         2 loss in a row                                             loss
> 2
> Y                                 3rd loss
>            los 3
> O
>                         4
> R
>                            5
> R
>                               ....
> R
> R
> R
>
>

> and/or the same as above but plotting the rx vs tx loss
>

Now that I've had a little coffee, a single packet recovered after the loss
of X in a row won't return the typical voip codec to "good", so  stepping
back a color, assuming we start at the 5th loss above, and then got 4 good
packets in a row would look like:

------------------------------------------------------------OYGB-------------------------
B     loss 1
G              loss 2
Y                       3rd loss
O                                     4th loss
R                                                    5th loss
R                                                  .
R
R
R



>
> On Sat, Jan 14, 2023 at 6:20 AM Nathan Owens <nathan@nathan.io> wrote:
>
>> Sorry, was rejected from the listserv, here's the google drive link for
>> all 3 IRTT's visualized
>>
>>
>> https://drive.google.com/drive/folders/1SbUyvdlfdllgqrEcSSxvGdn3yN40vSx-?usp=share_link
>>
>> On Sat, Jan 14, 2023 at 6:14 AM Nathan Owens <nathan@nathan.io> wrote:
>>
>>> I realized I goofed up my visualizations, here's all of them again:
>>> I see a ton of loss on all of these!
>>>
>>> While Luis was downloading/uploading... oof.
>>> [image: Screenshot 2023-01-14 at 6.13.50 AM.png]
>>>
>>> [image: Screenshot 2023-01-14 at 6.12.53 AM.png]
>>>
>>>
>>>
>>>
>>> On Fri, Jan 13, 2023 at 3:44 PM Dave Taht <dave.taht@gmail.com> wrote:
>>>
>>>> I am forced to conclude that waveform's upload test is broken in some
>>>> cases and has been for some time. All the joy many have been feeling about
>>>> their uplink speeds has to be cast away. Felt good, though, didn't it?
>>>>
>>>> There is a *slight* possibility that there is some fq in the starlink
>>>> network on tcp port 433. The first showing normal rtt growth and loss for
>>>> cubic, the second, a low rate flow that is holding the line... but I didn't
>>>> check to see if these were sequential or parallel.
>>>>
>>>> The last is a cwnd plot, clearly showing the cubic sawtooth on the
>>>> upload.
>>>>
>>>> It's weirdly nice to be able to follow a port 433 stream, see the tls
>>>> handshake put the website en'clair, and the rest go dark, and still trace
>>>> the behaviors.
>>>>
>>>> we're so going to lose this analytical ability with quic. I'm enjoying
>>>> it while we still can.
>>>>
>>>>
>>>> On Fri, Jan 13, 2023 at 2:10 PM Jonathan Bennett via Starlink <
>>>> starlink@lists.bufferbloat.net> wrote:
>>>>
>>>>> The irtt run finished a few seconds before the flent run, but here are
>>>>> the results:
>>>>>
>>>>>
>>>>> https://drive.google.com/file/d/1FKve13ssUMW1LLWOXLM2931Yx6uMHw8K/view?usp=share_link
>>>>>
>>>>> https://drive.google.com/file/d/1ZXd64A0pfUedLr3FyhDNTHA7vxv8S2Gk/view?usp=share_link
>>>>>
>>>>> https://drive.google.com/file/d/1rx64UPQHHz3IMNiJtb1oFqtqw2DvhvEE/view?usp=share_link
>>>>>
>>>>>
>>>>> [image: image.png]
>>>>> [image: image.png]
>>>>>
>>>>>
>>>>> Jonathan Bennett
>>>>> Hackaday.com
>>>>>
>>>>>
>>>>> On Fri, Jan 13, 2023 at 3:30 PM Nathan Owens via Starlink <
>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>
>>>>>> Here's Luis's run -- the top line below the edge of the graph is
>>>>>> 200ms
>>>>>> [image: Screenshot 2023-01-13 at 1.30.03 PM.png]
>>>>>>
>>>>>>
>>>>>> On Fri, Jan 13, 2023 at 1:25 PM Luis A. Cornejo <
>>>>>> luis.a.cornejo@gmail.com> wrote:
>>>>>>
>>>>>>> Dave,
>>>>>>>
>>>>>>> Here is a run the way I think you wanted it.
>>>>>>>
>>>>>>> irtt running for 5 min to your dallas server, followed by a waveform
>>>>>>> test, then a few seconds of inactivity, cloudflare test, a few more secs of
>>>>>>> nothing, flent test to dallas. Packet capture is currently uploading (will
>>>>>>> be done in 20 min or so), irtt JSON also in there (.zip file):
>>>>>>>
>>>>>>>
>>>>>>> https://drive.google.com/drive/folders/1FLWqrzNcM8aK-ZXQywNkZGFR81Fnzn-F?usp=share_link
>>>>>>>
>>>>>>> -Luis
>>>>>>>
>>>>>>> On Fri, Jan 13, 2023 at 2:50 PM Dave Taht via Starlink <
>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Fri, Jan 13, 2023 at 12:30 PM Nathan Owens <nathan@nathan.io>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Here's the data visualization for Johnathan's Data
>>>>>>>>>
>>>>>>>>> [image: Screenshot 2023-01-13 at 12.29.15 PM.png]
>>>>>>>>>
>>>>>>>>> You can see the path change at :12, :27, :42, :57 after the
>>>>>>>>> minute. Some paths are clearly busier than others with increased loss,
>>>>>>>>> latency, and jitter.
>>>>>>>>>
>>>>>>>>
>>>>>>>> I am so glad to see loss and bounded delay here. Also a bit of
>>>>>>>> rigor regarding what traffic was active locally vs on the path would be
>>>>>>>> nice, although it seems to line up with the known 15s starlink switchover
>>>>>>>> thing (need a name for this), in this case, doing a few speedtests
>>>>>>>> while that irtt is running would show the impact(s) of whatever
>>>>>>>> else they are up to.
>>>>>>>>
>>>>>>>> It would also be my hope that the loss distribution in the middle
>>>>>>>> portion of this data is good, not bursty, but we don't have a tool to take
>>>>>>>> apart that. (I am so hopeless at json)
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Fri, Jan 13, 2023 at 10:09 AM Nathan Owens <nathan@nathan.io>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> I’ll run my visualization code on this result this afternoon and
>>>>>>>>>> report back!
>>>>>>>>>>
>>>>>>>>>> On Fri, Jan 13, 2023 at 9:41 AM Jonathan Bennett via Starlink <
>>>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>>>>
>>>>>>>>>>> The irtt command, run with normal, light usage:
>>>>>>>>>>> https://drive.google.com/file/d/1SiVCiUYnx7nDTxIVOY5w-z20S2O059rA/view?usp=share_link
>>>>>>>>>>>
>>>>>>>>>>> Jonathan Bennett
>>>>>>>>>>> Hackaday.com
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Fri, Jan 13, 2023 at 11:26 AM Dave Taht <dave.taht@gmail.com>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> packet caps would be nice... all this is very exciting news.
>>>>>>>>>>>>
>>>>>>>>>>>> I'd so love for one or more of y'all reporting such great uplink
>>>>>>>>>>>> results nowadays to duplicate and re-plot the original irtt
>>>>>>>>>>>> tests we
>>>>>>>>>>>> did:
>>>>>>>>>>>>
>>>>>>>>>>>> irtt client -i3ms -d300s myclosestservertoyou.starlink.taht.net
>>>>>>>>>>>> -o whatever.json
>>>>>>>>>>>>
>>>>>>>>>>>> They MUST have changed their scheduling to get such amazing
>>>>>>>>>>>> uplink
>>>>>>>>>>>> results, in addition to better queue management.
>>>>>>>>>>>>
>>>>>>>>>>>> (for the record, my servers are de, london, fremont, sydney,
>>>>>>>>>>>> dallas,
>>>>>>>>>>>> newark, atlanta, singapore, mumbai)
>>>>>>>>>>>>
>>>>>>>>>>>> There's an R and gnuplot script for plotting that output around
>>>>>>>>>>>> here
>>>>>>>>>>>> somewhere (I have largely personally put down the starlink
>>>>>>>>>>>> project,
>>>>>>>>>>>> loaning out mine) - that went by on this list... I should have
>>>>>>>>>>>> written
>>>>>>>>>>>> a blog entry so I can find that stuff again.
>>>>>>>>>>>>
>>>>>>>>>>>> On Fri, Jan 13, 2023 at 9:02 AM Jonathan Bennett via Starlink
>>>>>>>>>>>> <starlink@lists.bufferbloat.net> wrote:
>>>>>>>>>>>> >
>>>>>>>>>>>> >
>>>>>>>>>>>> > On Fri, Jan 13, 2023 at 6:28 AM Ulrich Speidel via Starlink <
>>>>>>>>>>>> starlink@lists.bufferbloat.net> wrote:
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> On 13/01/2023 6:13 pm, Ulrich Speidel wrote:
>>>>>>>>>>>> >> >
>>>>>>>>>>>> >> > From Auckland, New Zealand, using a roaming subscription,
>>>>>>>>>>>> it puts me
>>>>>>>>>>>> >> > in touch with a server 2000 km away. OK then:
>>>>>>>>>>>> >> >
>>>>>>>>>>>> >> >
>>>>>>>>>>>> >> > IP address: nix six.
>>>>>>>>>>>> >> >
>>>>>>>>>>>> >> > My thoughts shall follow later.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> OK, so here we go.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> I'm always a bit skeptical when it comes to speed tests -
>>>>>>>>>>>> they're really
>>>>>>>>>>>> >> laden with so many caveats that it's not funny. I took our
>>>>>>>>>>>> new work
>>>>>>>>>>>> >> Starlink kit home in December to give it a try and the other
>>>>>>>>>>>> day finally
>>>>>>>>>>>> >> got around to set it up. It's on a roaming subscription
>>>>>>>>>>>> because our
>>>>>>>>>>>> >> badly built-up campus really isn't ideal in terms of a clear
>>>>>>>>>>>> view of the
>>>>>>>>>>>> >> sky. Oh - and did I mention that I used the Starlink
>>>>>>>>>>>> Ethernet adapter,
>>>>>>>>>>>> >> not the WiFi?
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> Caveat 1: Location, location. I live in a place where the
>>>>>>>>>>>> best Starlink
>>>>>>>>>>>> >> promises is about 1/3 in terms of data rate you can actually
>>>>>>>>>>>> get from
>>>>>>>>>>>> >> fibre to the home at under half of Starlink's price. Read:
>>>>>>>>>>>> There are few
>>>>>>>>>>>> >> Starlink users around. I might be the only one in my suburb.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> Caveat 2: Auckland has three Starlink gateways close by:
>>>>>>>>>>>> Clevedon (which
>>>>>>>>>>>> >> is at a stretch daytrip cycling distance from here), Te Hana
>>>>>>>>>>>> and Puwera,
>>>>>>>>>>>> >> the most distant of the three and about 130 km away from me
>>>>>>>>>>>> as the crow
>>>>>>>>>>>> >> flies. Read: My dishy can use any satellite that any of
>>>>>>>>>>>> these three can
>>>>>>>>>>>> >> see, and then depending on where I put it and how much of
>>>>>>>>>>>> the southern
>>>>>>>>>>>> >> sky it can see, maybe also the one in Hinds, 840 km away,
>>>>>>>>>>>> although that
>>>>>>>>>>>> >> is obviously stretching it a bit. Either way, that's plenty
>>>>>>>>>>>> of options
>>>>>>>>>>>> >> for my bits to travel without needing a lot of handovers.
>>>>>>>>>>>> Why? Easy: If
>>>>>>>>>>>> >> your nearest teleport is close by, then the set of
>>>>>>>>>>>> satellites that the
>>>>>>>>>>>> >> teleport can see and the set that you can see is almost the
>>>>>>>>>>>> same, so you
>>>>>>>>>>>> >> can essentially stick with the same satellite while it's in
>>>>>>>>>>>> view for you
>>>>>>>>>>>> >> because it'll also be in view for the teleport. Pretty much
>>>>>>>>>>>> any bird
>>>>>>>>>>>> >> above you will do.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> And because I don't get a lot of competition from other
>>>>>>>>>>>> users in my area
>>>>>>>>>>>> >> vying for one of the few available satellites that can see
>>>>>>>>>>>> both us and
>>>>>>>>>>>> >> the teleport, this is about as good as it gets at 37S
>>>>>>>>>>>> latitude. If I'd
>>>>>>>>>>>> >> want it any better, I'd have to move a lot further south.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> It'd be interesting to hear from Jonathan what the
>>>>>>>>>>>> availability of home
>>>>>>>>>>>> >> broadband is like in the Dallas area. I note that it's at a
>>>>>>>>>>>> lower
>>>>>>>>>>>> >> latitude (33N) than Auckland, but the difference isn't huge.
>>>>>>>>>>>> I notice
>>>>>>>>>>>> >> two teleports each about 160 km away, which is also not too
>>>>>>>>>>>> bad. I also
>>>>>>>>>>>> >> note Starlink availability in the area is restricted at the
>>>>>>>>>>>> moment -
>>>>>>>>>>>> >> oversubscribed? But if Jonathan gets good data rates, then
>>>>>>>>>>>> that means
>>>>>>>>>>>> >> that competition for bird capacity can't be too bad - for
>>>>>>>>>>>> whatever reason.
>>>>>>>>>>>> >
>>>>>>>>>>>> > I'm in Southwest Oklahoma, but Dallas is the nearby Starlink
>>>>>>>>>>>> gateway. In cities, like Dallas, and Lawton where I live, there are good
>>>>>>>>>>>> broadband options. But there are also many people that live outside cities,
>>>>>>>>>>>> and the options are much worse. The low density userbase in rural Oklahoma
>>>>>>>>>>>> and Texas is probably ideal conditions for Starlink.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> Caveat 3: Backhaul. There isn't just one queue between me
>>>>>>>>>>>> and whatever I
>>>>>>>>>>>> >> talk to in terms of my communications. Traceroute shows
>>>>>>>>>>>> about 10 hops
>>>>>>>>>>>> >> between me and the University of Auckland via Starlink.
>>>>>>>>>>>> That's 10
>>>>>>>>>>>> >> queues, not one. Many of them will have cross traffic. So
>>>>>>>>>>>> it's a bit
>>>>>>>>>>>> >> hard to tell where our packets really get to wait or where
>>>>>>>>>>>> they get
>>>>>>>>>>>> >> dropped. The insidious bit here is that a lot of them will
>>>>>>>>>>>> be between 1
>>>>>>>>>>>> >> Gb/s and 10 Gb/s links, and with a bit of cross traffic,
>>>>>>>>>>>> they can all
>>>>>>>>>>>> >> turn into bottlenecks. This isn't like a narrowband GEO link
>>>>>>>>>>>> of a few
>>>>>>>>>>>> >> Mb/s where it's obvious where the dominant long latency
>>>>>>>>>>>> bottleneck in
>>>>>>>>>>>> >> your TCP connection's path is. Read: It's pretty hard to
>>>>>>>>>>>> tell whether a
>>>>>>>>>>>> >> drop in "speed" is due to a performance issue in the
>>>>>>>>>>>> Starlink system or
>>>>>>>>>>>> >> somewhere between Starlink's systems and the target system.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> I see RTTs here between 20 ms and 250 ms, where the physical
>>>>>>>>>>>> latency
>>>>>>>>>>>> >> should be under 15 ms. So there's clearly a bit of buffer
>>>>>>>>>>>> here along the
>>>>>>>>>>>> >> chain that occasionally fills up.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> Caveat 4: Handovers. Handover between birds and teleports is
>>>>>>>>>>>> inevitably
>>>>>>>>>>>> >> associated with a change in RTT and in most cases also
>>>>>>>>>>>> available
>>>>>>>>>>>> >> bandwidth. Plus your packets now arrive at a new queue on a
>>>>>>>>>>>> new
>>>>>>>>>>>> >> satellite while your TCP is still trying to respond to
>>>>>>>>>>>> whatever it
>>>>>>>>>>>> >> thought the queue on the previous bird was doing. Read:
>>>>>>>>>>>> Whatever your
>>>>>>>>>>>> >> cwnd is immediately after a handover, it's probably not what
>>>>>>>>>>>> it should be.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> I ran a somewhat hamstrung (sky view restricted) set of four
>>>>>>>>>>>> Ookla
>>>>>>>>>>>> >> speedtest.net tests each to five local servers. Average
>>>>>>>>>>>> upload rate was
>>>>>>>>>>>> >> 13 Mb/s, average down 75.5 Mb/s. Upload to the server of the
>>>>>>>>>>>> ISP that
>>>>>>>>>>>> >> Starlink seems to be buying its local connectivity from
>>>>>>>>>>>> (Vocus Group)
>>>>>>>>>>>> >> varied between 3.04 and 14.38 Mb/s, download between 23.33
>>>>>>>>>>>> and 52.22
>>>>>>>>>>>> >> Mb/s, with RTTs between 37 and 56 ms not correlating well to
>>>>>>>>>>>> rates
>>>>>>>>>>>> >> observed. In fact, they were the ISP with consistently the
>>>>>>>>>>>> worst rates.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> Another ISP (MyRepublic) scored between 11.81 and 21.81 Mb/s
>>>>>>>>>>>> up and
>>>>>>>>>>>> >> between 106.5 and 183.8 Mb/s down, again with RTTs badly
>>>>>>>>>>>> correlating
>>>>>>>>>>>> >> with rates. Average RTT was the same as for Vocus.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> Note the variation though: More or less a factor of two
>>>>>>>>>>>> between highest
>>>>>>>>>>>> >> and lowest rates for each ISP. Did MyRepublic just get lucky
>>>>>>>>>>>> in my
>>>>>>>>>>>> >> tests? Or is there something systematic behind this? Way too
>>>>>>>>>>>> few tests
>>>>>>>>>>>> >> to tell.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> What these tests do is establish a ballpark.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> I'm currently repeating tests with dish placed on a trestle
>>>>>>>>>>>> closer to
>>>>>>>>>>>> >> the heavens. This seems to have translated into fewer
>>>>>>>>>>>> outages / ping
>>>>>>>>>>>> >> losses (around 1/4 of what I had yesterday with dishy on the
>>>>>>>>>>>> ground on
>>>>>>>>>>>> >> my deck). Still good enough for a lengthy video Skype call
>>>>>>>>>>>> with my folks
>>>>>>>>>>>> >> in Germany, although they did comment about reduced video
>>>>>>>>>>>> quality. But
>>>>>>>>>>>> >> maybe that was the lighting or the different background as I
>>>>>>>>>>>> wasn't in
>>>>>>>>>>>> >> my usual spot with my laptop when I called them.
>>>>>>>>>>>> >
>>>>>>>>>>>> > Clear view of the sky is king for Starlink reliability. I've
>>>>>>>>>>>> got my dishy mounted on the back fence, looking up over an empty field, so
>>>>>>>>>>>> it's pretty much best-case scenario here.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> --
>>>>>>>>>>>> >>
>>>>>>>>>>>> >>
>>>>>>>>>>>> ****************************************************************
>>>>>>>>>>>> >> Dr. Ulrich Speidel
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> School of Computer Science
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> Room 303S.594 (City Campus)
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> The University of Auckland
>>>>>>>>>>>> >> u.speidel@auckland.ac.nz
>>>>>>>>>>>> >> http://www.cs.auckland.ac.nz/~ulrich/
>>>>>>>>>>>> >>
>>>>>>>>>>>> ****************************************************************
>>>>>>>>>>>> >>
>>>>>>>>>>>> >>
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> _______________________________________________
>>>>>>>>>>>> >> Starlink mailing list
>>>>>>>>>>>> >> Starlink@lists.bufferbloat.net
>>>>>>>>>>>> >> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>>>> >
>>>>>>>>>>>> > _______________________________________________
>>>>>>>>>>>> > Starlink mailing list
>>>>>>>>>>>> > Starlink@lists.bufferbloat.net
>>>>>>>>>>>> > https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>> This song goes out to all the folk that thought Stadia would
>>>>>>>>>>>> work:
>>>>>>>>>>>>
>>>>>>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>>>>>>>
>>>>>>>>>>> _______________________________________________
>>>>>>>>>>> Starlink mailing list
>>>>>>>>>>> Starlink@lists.bufferbloat.net
>>>>>>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> This song goes out to all the folk that thought Stadia would work:
>>>>>>>>
>>>>>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>>>>>> Dave Täht CEO, TekLibre, LLC
>>>>>>>> _______________________________________________
>>>>>>>> Starlink mailing list
>>>>>>>> Starlink@lists.bufferbloat.net
>>>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>>>
>>>>>>> _______________________________________________
>>>>>> Starlink mailing list
>>>>>> Starlink@lists.bufferbloat.net
>>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>>
>>>>> _______________________________________________
>>>>> Starlink mailing list
>>>>> Starlink@lists.bufferbloat.net
>>>>> https://lists.bufferbloat.net/listinfo/starlink
>>>>>
>>>>
>>>>
>>>> --
>>>> This song goes out to all the folk that thought Stadia would work:
>>>>
>>>> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
>>>> Dave Täht CEO, TekLibre, LLC
>>>>
>>>
>
> --
> This song goes out to all the folk that thought Stadia would work:
>
> https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
> Dave Täht CEO, TekLibre, LLC
>


-- 
This song goes out to all the folk that thought Stadia would work:
https://www.linkedin.com/posts/dtaht_the-mushroom-song-activity-6981366665607352320-FXtz
Dave Täht CEO, TekLibre, LLC

[-- Attachment #1.2: Type: text/html, Size: 27311 bytes --]

[-- Attachment #2: Screenshot 2023-01-13 at 12.29.15 PM.png --]
[-- Type: image/png, Size: 1380606 bytes --]

[-- Attachment #3: Screenshot 2023-01-13 at 1.30.03 PM.png --]
[-- Type: image/png, Size: 855840 bytes --]

[-- Attachment #4: image.png --]
[-- Type: image/png, Size: 296989 bytes --]

[-- Attachment #5: image.png --]
[-- Type: image/png, Size: 222655 bytes --]

[-- Attachment #6: Screenshot 2023-01-14 at 6.12.53 AM.png --]
[-- Type: image/png, Size: 921914 bytes --]

[-- Attachment #7: Screenshot 2023-01-14 at 6.13.50 AM.png --]
[-- Type: image/png, Size: 337863 bytes --]

^ permalink raw reply	[flat|nested] 44+ messages in thread

end of thread, other threads:[~2023-01-14 16:33 UTC | newest]

Thread overview: 44+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2023-01-04 17:26 [Starlink] the grinch meets cloudflare's christmas present Dave Taht
2023-01-04 19:20 ` [Starlink] [Rpm] " jf
2023-01-04 20:02   ` rjmcmahon
2023-01-04 21:16     ` Ulrich Speidel
2023-01-04 23:54       ` Bruce Perens
2023-01-05  2:48         ` Dave Collier-Brown
2023-01-05  3:11           ` Dick Roy
2023-01-05 11:25             ` Sebastian Moeller
2023-01-06  0:01               ` Dick Roy
2023-01-06  9:43                 ` Sebastian Moeller
2023-01-05  6:11         ` rjmcmahon
2023-01-05 11:11     ` [Starlink] [Bloat] " Sebastian Moeller
2023-01-06 16:38 ` [Starlink] [LibreQoS] " MORTON JR., AL
2023-01-06 20:38   ` [Starlink] [Rpm] " rjmcmahon
2023-01-06 20:47     ` rjmcmahon
     [not found]     ` <89D796E75967416B9723211C183A8396@SRA6>
     [not found]       ` <a082b2436e6ba7892d2de8e0dfcc5acd@rjmcmahon.com>
     [not found]         ` <3696AEA5409D4303ABCBC439727A5E40@SRA6>
     [not found]           ` <CAKJdXWBb0VxSSoGAQTe3BXFLXCHd6NSspRnXd1frK2f66SLiUw@mail.gmail.com>
     [not found]             ` <CAA93jw6B_9-WE9EEFuac+FAH-2dcULk=_3i_HfhCSVSOxyM7Eg@mail.gmail.com>
     [not found]               ` <CA+Ld8r8hR8KF35Yv7A3hb1QvC9v9ka2Nh2J=HEm0XhPfvAAcag@mail.gmail.com>
     [not found]                 ` <CAKJdXWC+aEy1b3vB-FFd+tnfT+Ni5O9bZ+p4kkhj-FzMPVGGcQ@mail.gmail.com>
     [not found]                   ` <CAA93jw4DcBhA8CevRQoMbzjO-3Jt+emr+xvnJ-hUGkT+n0KJzg@mail.gmail.com>
     [not found]                     ` <CH0PR02MB79800FF2E40CE037D6802D71D3FD9@CH0PR02MB7980.namprd02.prod.outlook.com>
     [not found]                       ` <CAKJdXWDOFbzsam2C_24e9DLkc18ed4uhV51hOKVjDipk1Uhc2g@mail.gmail.com>
2023-01-13  4:08                         ` [Starlink] insanely great waveform result for starlink Dave Taht
2023-01-13  4:26                           ` Jonathan Bennett
2023-01-13  5:13                             ` Ulrich Speidel
2023-01-13  5:25                               ` Dave Taht
2023-01-13 12:27                               ` Ulrich Speidel
2023-01-13 17:02                                 ` Jonathan Bennett
2023-01-13 17:26                                   ` Dave Taht
2023-01-13 17:41                                     ` Jonathan Bennett
2023-01-13 18:09                                       ` Nathan Owens
2023-01-13 20:30                                         ` Nathan Owens
2023-01-13 20:37                                           ` Dave Taht
2023-01-13 21:24                                             ` Nathan Owens
2023-01-13 20:49                                           ` Dave Taht
2023-01-13 21:25                                             ` Luis A. Cornejo
2023-01-13 21:30                                               ` Nathan Owens
2023-01-13 22:09                                                 ` Jonathan Bennett
2023-01-13 22:30                                                   ` Luis A. Cornejo
2023-01-13 22:32                                                   ` Dave Taht
2023-01-13 22:36                                                     ` Luis A. Cornejo
2023-01-13 22:42                                                     ` Jonathan Bennett
2023-01-13 22:49                                                       ` Dave Taht
2023-01-13 23:44                                                   ` Dave Taht
     [not found]                                                     ` <CALjsLJv5cbfHfkxqHnbjxoVHczspYvxc_jrshzs1CpHLEDWyew@mail.gmail.com>
2023-01-14 14:20                                                       ` Nathan Owens
2023-01-14 15:53                                                         ` Dave Taht
2023-01-14 16:33                                                           ` Dave Taht
2023-01-14 15:52                                                       ` Dave Taht
2023-01-13 20:25                                     ` Pat Jensen
2023-01-13 20:40                                       ` Dave Taht
2023-01-13 22:51                                   ` Ulrich Speidel

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox