From: David Lang <david@lang.hm>
To: Ulrich Speidel <u.speidel@auckland.ac.nz>
Cc: Sebastian Moeller <moeller0@gmx.de>,
Ulrich Speidel via Starlink <starlink@lists.bufferbloat.net>
Subject: Re: [Starlink] Starlink hidden buffers
Date: Sat, 13 May 2023 16:00:43 -0700 (PDT) [thread overview]
Message-ID: <pr8q701r-4o82-q138-921o-q250p05qs568@ynat.uz> (raw)
In-Reply-To: <a47520af-37c0-4fc3-9472-ce20d8203724@auckland.ac.nz>
[-- Attachment #1: Type: text/plain, Size: 5538 bytes --]
On Sun, 14 May 2023, Ulrich Speidel via Starlink wrote:
> I'd also imaging total user numbers to be lower and the
> bandwidth demand per user to be less (hands up who takes their 50" TV onto
> trains to watch Netflix in HD?).
most phones are >HD resolution, and the higher end are >4k resolution. the
network bandwidth doesn't care if the resulting screen is 5" or 50", the
resolution is all that matters.
> The other is that most places have 3+ networks serving the train line, which
> brings down user numbers, or you have in-train cells, which communicate with
> off-train POPs that have no extra users.
but the density of people in a train car is MUCH higher than in an office, even
if split a couple of ways.
David Lang
> But yes, good question IMHO!
>
> Cheers,
>
> Ulrich
>
> On 13/05/2023 11:20 pm, Sebastian Moeller wrote:
>> Hi Ulrich,
>>
>> This situation is not completely different from say a train full of LTE/5G
>> users moving through a set of cells with already established 'static'
>> users, no?
>>
>>
>> On 13 May 2023 12:10:17 CEST, Ulrich Speidel via Starlink
>> <starlink@lists.bufferbloat.net> wrote:
>>
>> Here's a bit of a question to you all. See what you make of it.
>> I've been thinking a bit about the latencies we see in the
>> Starlink network. This is why this list exist (right, Dave?). So
>> what do we know? 1) We know that RTTs can be in the 100's of ms
>> even in what appear to be bent-pipe scenarios where the physical
>> one-way path should be well under 3000 km, with physical RTT under
>> 20 ms. 2) We know from plenty of traceroutes that these RTTs
>> accrue in the Starlink network, not between the Starlink handover
>> point (POP) to the Internet. 3) We know that they aren't an
>> artifact of the Starlink WiFi router (our traceroutes were done
>> through their Ethernet adaptor, which bypasses the router), so
>> they must be delays on the satellites or the teleports. 4) We know
>> that processing delay isn't a huge factor because we also see RTTs
>> well under 30 ms. 5) That leaves queuing delays. This issue has
>> been known for a while now. Starlink have been innovating their
>> heart out around pretty much everything here - and yet, this
>> bufferbloat issue hasn't changed, despite Dave proposing what
>> appears to be an easy fix compared to a lot of other things they
>> have done. So what are we possibly missing here? Going back to
>> first principles: The purpose of a buffer on a network device is
>> to act as a shock absorber against sudden traffic bursts. If I
>> want to size that buffer correctly, I need to know at the very
>> least (paraphrasing queueing theory here) something about my
>> packet arrival process. If I look at conventional routers, then
>> that arrival process involves traffic generated by a user
>> population that changes relatively slowly: WiFi users come and go.
>> One at a time. Computers in a company get turned on and off and
>> rebooted, but there are no instantaneous jumps in load - you don't
>> suddenly have a hundred users in the middle of watching Netflix
>> turning up that weren't there a second ago. Most of what we know
>> about Internet traffic behaviour is based on this sort of network,
>> and this is what we've designed our queuing systems around, right?
>> Observation: Starlink potentially breaks that paradigm. Why?
>> Imagine a satellite X handling N users that are located closely
>> together in a fibre-less rural town watching a range of movies.
>> Assume that N is relatively large. Say these users are currently
>> handled through ground station teleport A some distance away to
>> the west (bent pipe with switching or basic routing on the
>> satellite). X is in view of both A and the N users, but with X
>> being a LEO satellite, that bliss doesn't last. Say X is moving to
>> the (south- or north-)east and out of A's range. Before connection
>> is lost, the N users migrate simultaneously to a new satellite Y
>> that has moved into view of both A and themselves. Y is doing so
>> from the west and is also catering to whatever users it can see
>> there, and let's suppose has been using A for a while already. The
>> point is that the user load on X and Y from users other than our N
>> friends could be quite different. E.g., one of them could be over
>> the ocean with few users, the other over countryside with a lot of
>> customers. The TCP stacks of our N friends are (hopefully)
>> somewhat adapted to the congestion situation on X with their cwnds
>> open to reasonable sizes, but they are now thrown onto a
>> completely different congestion scenario on Y. Similarly, say that
>> Y had less than N users before the handover. For existing users on
>> Y, there is now a huge surge of competing traffic that wasn't
>> there a second ago - surging far faster than we would expect this
>> to happen in a conventional network because there is no slow start
>> involved. This seems to explain the huge jumps you see on Starlink
>> in TCP goodput over time. But could this be throwing a few
>> spanners into the works in terms of queuing? Does it invalidate
>> what we know about queues and queue management? Would surges like
>> these justify larger buffers?
>>
>> --
>> Sent from my Android device with K-9 Mail. Please excuse my brevity.
>
>
[-- Attachment #2: Type: text/plain, Size: 149 bytes --]
_______________________________________________
Starlink mailing list
Starlink@lists.bufferbloat.net
https://lists.bufferbloat.net/listinfo/starlink
next prev parent reply other threads:[~2023-05-13 23:00 UTC|newest]
Thread overview: 34+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-05-13 10:10 Ulrich Speidel
2023-05-13 11:20 ` Sebastian Moeller
2023-05-13 12:16 ` Ulrich Speidel
2023-05-13 23:00 ` David Lang [this message]
2023-05-13 22:57 ` David Lang
2023-05-14 6:06 ` Ulrich Speidel
2023-05-14 6:55 ` David Lang
2023-05-14 8:43 ` Ulrich Speidel
2023-05-14 9:00 ` David Lang
2023-05-15 2:41 ` Ulrich Speidel
2023-05-15 3:33 ` David Lang
2023-05-15 6:36 ` Sebastian Moeller
2023-05-15 11:07 ` David Lang
2023-05-24 12:55 ` Ulrich Speidel
2023-05-24 13:44 ` Dave Taht
2023-05-24 14:05 ` David Lang
2023-05-24 14:49 ` Michael Richardson
2023-05-24 15:09 ` Dave Collier-Brown
2023-05-24 15:31 ` Dave Taht
2023-05-24 18:30 ` Michael Richardson
2023-05-24 18:45 ` Sebastian Moeller
2023-05-24 13:59 ` David Lang
2023-05-24 22:39 ` Ulrich Speidel
2023-05-25 0:06 ` David Lang
2023-07-27 20:37 ` Ulrich Speidel
2023-05-24 15:18 ` Mark Handley
2023-05-24 21:50 ` Ulrich Speidel
2023-05-25 0:17 ` David Lang
2023-05-14 9:06 ` Sebastian Moeller
2023-05-14 9:13 ` David Lang
2023-05-14 9:57 ` Oleg Kutkov
2023-05-14 9:59 ` Oleg Kutkov
2023-05-24 15:26 ` Bjørn Ivar Teigen
2023-05-24 21:53 ` Ulrich Speidel
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
List information: https://lists.bufferbloat.net/postorius/lists/starlink.lists.bufferbloat.net/
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=pr8q701r-4o82-q138-921o-q250p05qs568@ynat.uz \
--to=david@lang.hm \
--cc=moeller0@gmx.de \
--cc=starlink@lists.bufferbloat.net \
--cc=u.speidel@auckland.ac.nz \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox