[Cerowrt-devel] [Make-wifi-fast] [tsvwg] Comments on draft-szigeti-tsvwg-ieee-802-11e
David Lang
david at lang.hm
Mon Aug 3 12:14:16 EDT 2015
On Mon, 3 Aug 2015, dpreed at reed.com wrote:
> It's not infeasible to make queues shorter. In any case, the throughput of a
> link does not increase above the point where there is always one packet ready
> to go by the time the currently outgoing packet is completed. It physically
> cannot do better than that.
change 'one packet' to 'one transmissions worth of packets' and I'll agree
> If hardware designers can't create an interface that achieves that bound I'd
> be suspicious that they understand how to design hardware. In the case of
> WiFi, this also includes the MAC protocol being designed so that when the
> current packet on the air terminates, the next packet can be immediately begun
> - that's a little more subtle.
on a shared medium (like radio) things are a bit messier.
There are two issues
1. You shouldn't just transmit to a new station once you finish sending to the
first. Fairness requires that you pause and give other stations a chance to
transmit as well.
1. There is per-transmission overhead (including the pause mentioned above) that
can be very significant for small packets, so there is considerable value in
sending multiple packets at once. It's a lighter version of what you run into
inserting into reliable databases. You can insert 1000 records in about the same
time you can insert 2 records sperately.
The "stock" answer to this is for hardware and software folks to hold off on
sending anything in case there is more to send later that it can be batched
with. This maximizes throughput at the cost of latency.
What should be done instead is to send what you have immediatly, and while it's
sending, queue whatever continues to arrive and the next chance you have to
send, you will have more to send. This scales the batch size with congestion,
minimizing latency at the cost of keeping the channel continually busy, but
inefficiently busy if you aren't at capacity.
> But my point here is that one needs to look at the efficiency of the system as
> a whole (in context), and paradoxically to the hardware designer mindset, the
> proper way to think about that efficiency is NOT about link throughput
> maximization - instead it is an end-to-end property. One has very little to
> do with the other. Queueing doesn't affect link throughput beyond the "double
> buffering" effect noted above: at most one packet queued behind the currently
> transmitting packet.
>
> Regarding TXOP overhead - rather than complicated queueing, just allow packets
> to be placed in line *while the currently transmitting packet is going out*,
> and changed up to the point in time when they begin transmitting. This is
> trivial in hardware.
This is a key thing that a lot of hardware and software folks get wrong. All the
complexity and bugs that you see around 'blocking/plugging' flows are the result
of this mistake. As you say, send something as soon as you have it to send. If
there's more arriving, let it accumulate while the first bit is being sent and
the next chance you get to send, send everything that's accumulated. This
minimizes latency, greatly simplifies the code (no need for timers to
unblock/release the data if more doesn't arrive), and results in the exact same
throughput under load.
It does have some interesting changes to the utilization curve at part load.
These could be a problem with wifi under some conditions, but I think the
trade-off is worth it since the wifi is going to end up running up to it's limit
sometime anyway, and the part load problems are just previews of what you would
run into at full load.
David Lang
>
> On Friday, July 31, 2015 1:04pm, "Jonathan Morton" <chromatix99 at gmail.com> said:
>
>
>
>> I think that is achievable, *even if there is a WiFi network in the middle*, by thinking about the fact that the shared airwaves in a WiFi network behaves like a single link, so all the queues on individual stations are really *one queue*, and that the optimal behavior of that link will be achieved if there is at most one packet queued at a time.
> I agree that queues should be kept short in general. However I don't think single packet queues are achievable in the general case.
> The general case includes Wi-Fi networks, whose TXOP overhead is so ruinously heavy that sending single MTU sized packets is inefficient. Aggregating multiple packets into one TXOP requires those several packets to be present in the buffer at that moment.
> The general case includes links which vary in throughput frequently, perhaps on shorter timescales than an RTT, so either packets must be buffered or capacity is left unused. This also happens to include Wi-Fi, but could easily include a standard wired link whose competing load varies.
> The endpoints do not have and do not receive sufficient information in sufficient time to reliably make packets arrive at nodes just in time to be transmitted. Not even with ECN, not even with the wet dreams of the DCTCP folks, and not even with ELR (though ELR should be able to make it happen under steady conditions, there are still transient conditions in the general case).
> - Jonathan Morton
-------------- next part --------------
_______________________________________________
Cerowrt-devel mailing list
Cerowrt-devel at lists.bufferbloat.net
https://lists.bufferbloat.net/listinfo/cerowrt-devel
More information about the Cerowrt-devel
mailing list