kbps = quantum = time20000 = 3000 = 1.2ms30000 = 6000 = 1.6ms40000 = 12000 = 2.4ms50000 = 24000 = 3.84ms60000 = 48000 = 6.4ms80000 = 96000 = 9.6msSo it appears that the goal of these values was to keep increases the quantum as rates went up to provide more bytes per operation, but that's going to risk adding latency as the time-per-quantum crosses the delay target in fq_codel (if I'm understanding this correctly).So one thing that I can do is play around with this, and see if I can keep that quantum time at a linear level (ie, 10ms, which seems _awfully_ long), or continue increasing it (which seems like a bad idea). I'd love to hear from whoever put this in as to what it's goal was (or was it just empirically tuned?)Empirical and tested only to about 60Mbits. I got back about 15% cpu to do it this way at the time I did it on the wndr3800.
and WOW, thx for the analysis! I did not think much about this crossover point at the time - because we'd maxed on cpu long beforehand.
I can certainly see this batching interacting with the codel target.
On the other hand, you gotta not be running out of cpu in the first place. I am liking where cake is going.
One of my daydreams is that once we have writable custom ethernet hardware that we can easily do hardware outbound rate limiting/shaping merely by programming a register to return a completion interrupt at the set rate rather than the actual rate.