[Bloat] Measuring latency-under-load consistently

Jonathan Morton chromatix99 at gmail.com
Fri Mar 11 20:44:05 EST 2011


On 12 Mar, 2011, at 3:09 am, Rick Jones wrote:

>>> You may be able to get most of what you want with a top-of-trunk netperf
>>> "burst mode" TCP_RR test. It isn't quite an exact match though.
>> 
>> I don't really see how that would get the measurement I want.
> 
> Then one TCP_STREAM (or TCP_MAERTS) and one TCP_RR with all the RTT
> stats and histogram enabled :)

Closer, but that doesn't seem to be self-tuning - and often one bulk flow doesn't fill the pipe any more.

I want to scale up to FTTH and GigE, and down to V.34bis and GPRS - all of which are (still) relevant network technologies today - without any error-prone parameters to enter except the identity of the server.

Ideally it should be robust enough for use by random ISP technicians, regulatory officials and end-users, and the output should be accordingly simple to interpret.

So yes, code reuse is great, but only if it does what I need it to.  I need the practice anyway.  :-)

 - Jonathan




More information about the Bloat mailing list