* Re: [Ecn-sane] ect(1) queue selector question @ 2021-03-09 0:36 Pete Heist 2021-03-09 1:18 ` Dave Taht 0 siblings, 1 reply; 15+ messages in thread From: Pete Heist @ 2021-03-09 0:36 UTC (permalink / raw) To: ECN-Sane Sorry for reviving an old thread as I haven't been on this list in a while: > > SCE proposes to use ect(1) as an indicator of some congestion and does > > not explictly > > require a dscp codepoint in a FQ'd implementation. > Pretty much. I do think that a demonstration using an > additional DSCP to create a similar HOV lane for SCE would have gone > miles in convincing people in the WG that L4S might really not be as > swell as its proponents argue, IMHO it won the day more with its > attractive promise of low latency for all instead of what it delivers. On that, I don't think any of us knows how things will end up or how long it will take to get there... I do agree that the interim meeting leading up to the codepoint decision could have gone better. Everything went great until it came to how to deploy SCE in a small number of queues. We had dismissed the idea of using DSCP, because we thought it would be panned for its poor traversal over the Internet. That may still have been the case, but it also may have worked if sold right. We thought that AF alone might be enough to get past that part, but it wasn't. We already implemented a two-queue design that uses DSCP, but either there wasn't much interest, or we didn't sell it enough. Plus, for those who demand a two queue solution that requires no flow awareness at all, DSCP alone may not get you there, because you still need some reasonably fair way to schedule the two queues. So that might have been the next line of criticism. Scheduling in proportion to the number of flows each queue contains is one effective way to do that, but that requires at least some concept of a flow. Perhaps there's another way that doesn't introduce too much RTT unfairness, but I'm not sure. In our defense, there was already a lot of support built up for L4S, and stepping in front of that was like stepping in front of a freight train no matter what we did. I think we've made a decent argument in the most recent version of the SCE draft that ECN is a "network feature" which carries higher risks than drop-based signaling, and warrants the requirement for unresponsive flow mitigation, for starters. That of course requires some level of flow awareness, which then makes various queueing designs possible. And, there may still be deployment possibilities with DSCP, as Rodney mentioned. Anyway, there's progress being made on SCE, with some new ideas and improvements to testing tools coming along. Pete ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: [Ecn-sane] ect(1) queue selector question 2021-03-09 0:36 [Ecn-sane] ect(1) queue selector question Pete Heist @ 2021-03-09 1:18 ` Dave Taht 2021-03-09 9:11 ` Pete Heist 2021-03-09 15:19 ` Rodney W. Grimes 0 siblings, 2 replies; 15+ messages in thread From: Dave Taht @ 2021-03-09 1:18 UTC (permalink / raw) To: Pete Heist; +Cc: ECN-Sane I am of course much happier interacting here than over on tsvwg. I would like very much A large scale bit twiddling test from the larger providers of fq_codel and cake based hw and middleboxes. extensive testing on lte and wifi transports. even emulated. the sce patches polished up and submitted for review to netdev as what we call there, an RFC the ability to test both SCE and L4S on openwrt (backport thereof) I know there's been a lot of work on other qdiscs than cake, but haven't paid much attention. netdev review would be nice. A simple internet-wide test of say bittorrent, or using an adserver method like what apnic uses. Most of all I'd really like someone to take a stab at RFC3168 support for BBR. And by that, I don't mean a reduction by half, but to institute an RTT probe phase. A fixed rate reduction per RTT is simply not going to work IMHO, period, on lte and wifi. I'm sure dave miller would accept such a patch, and this would also lead towards better ecn support across the internet in general... at least from the perspective of the one provider I have an in with, dropbox. SCE support for BBRv2 would be nice also. And I know, a pony, all sparkly and stuff. I find it very difficult to summon the cojones to do a drop of work in this area, and I keep cheering you on - especially on the bug fixing and wonderful scripts and data you've provided so far, and I do wish I could find some way to help that didn't cause so much ptsd in me. I wish I merely knew more about how dctp was configured in the data center. So far as *I* know its on dedicated switches mostly. I would vastly prefer a dscp codepoint for this, also. On Mon, Mar 8, 2021 at 4:36 PM Pete Heist <pete@heistp.net> wrote: > > Sorry for reviving an old thread as I haven't been on this list in a > while: > > > > SCE proposes to use ect(1) as an indicator of some congestion and > does > > > not explictly > > > require a dscp codepoint in a FQ'd implementation. > > Pretty much. I do think that a demonstration using an > > additional DSCP to create a similar HOV lane for SCE would have gone > > miles in convincing people in the WG that L4S might really not be as > > swell as its proponents argue, IMHO it won the day more with its > > attractive promise of low latency for all instead of what it > delivers. > > On that, I don't think any of us knows how things will end up or how > long it will take to get there... > > I do agree that the interim meeting leading up to the codepoint > decision could have gone better. Everything went great until it came to > how to deploy SCE in a small number of queues. We had dismissed the > idea of using DSCP, because we thought it would be panned for its poor > traversal over the Internet. That may still have been the case, but it > also may have worked if sold right. We thought that AF alone might be > enough to get past that part, but it wasn't. > > We already implemented a two-queue design that uses DSCP, but either > there wasn't much interest, or we didn't sell it enough. Plus, for > those who demand a two queue solution that requires no flow awareness > at all, DSCP alone may not get you there, because you still need some > reasonably fair way to schedule the two queues. So that might have been > the next line of criticism. Scheduling in proportion to the number of > flows each queue contains is one effective way to do that, but that > requires at least some concept of a flow. Perhaps there's another way > that doesn't introduce too much RTT unfairness, but I'm not sure. > > In our defense, there was already a lot of support built up for L4S, > and stepping in front of that was like stepping in front of a freight > train no matter what we did. I think we've made a decent argument in > the most recent version of the SCE draft that ECN is a "network > feature" which carries higher risks than drop-based signaling, and > warrants the requirement for unresponsive flow mitigation, for > starters. That of course requires some level of flow awareness, which > then makes various queueing designs possible. And, there may still be > deployment possibilities with DSCP, as Rodney mentioned. > > Anyway, there's progress being made on SCE, with some new ideas and > improvements to testing tools coming along. > > Pete > > > _______________________________________________ > Ecn-sane mailing list > Ecn-sane@lists.bufferbloat.net > https://lists.bufferbloat.net/listinfo/ecn-sane -- "For a successful technology, reality must take precedence over public relations, for Mother Nature cannot be fooled" - Richard Feynman dave@taht.net <Dave Täht> CTO, TekLibre, LLC Tel: 1-831-435-0729 ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: [Ecn-sane] ect(1) queue selector question 2021-03-09 1:18 ` Dave Taht @ 2021-03-09 9:11 ` Pete Heist 2021-03-09 15:19 ` Rodney W. Grimes 1 sibling, 0 replies; 15+ messages in thread From: Pete Heist @ 2021-03-09 9:11 UTC (permalink / raw) To: Dave Taht; +Cc: ECN-Sane Fwiw I captured a "Dave's wish list" file so we have it around. We do come back to these lists sometimes. There's not much here currently in our project plan except for more testing on bursty mac layers, in the lab at first, as part of work that should make it much easier to run more types of tests, and concurrently so they get done faster. If HFCC has a deployment problem with bursty traffic that can't be solved, it seems like something that needs to be worked out asap. We made quite a utilization improvement by using Codel to do SCE signaling, but what are the tradeoffs and how does that work in different conditions? Worth trying to answer with testing. On Mon, 2021-03-08 at 17:18 -0800, Dave Taht wrote: > I am of course much happier interacting here than over on tsvwg. > > I would like very much > > A large scale bit twiddling test from the larger providers of > fq_codel > and cake based hw and middleboxes. > > extensive testing on lte and wifi transports. even emulated. > the sce patches polished up and submitted for review to netdev as > what > we call there, an RFC > the ability to test both SCE and L4S on openwrt (backport thereof) > > I know there's been a lot of work on other qdiscs than cake, but > haven't paid much attention. netdev review would be nice. > > A simple internet-wide test of say bittorrent, or using an adserver > method like what apnic uses. > > Most of all I'd really like someone to take a stab at RFC3168 support > for BBR. And by that, I don't mean a reduction by half, but to > institute an RTT probe phase. A fixed rate reduction per RTT is > simply > not > going to work IMHO, period, on lte and wifi. I'm sure dave miller > would accept such a patch, and this would also lead towards better > ecn > support across the internet in general... at least from the > perspective of the one provider I have an in with, dropbox. > > SCE support for BBRv2 would be nice also. > > And I know, a pony, all sparkly and stuff. I find it very difficult > to > summon the cojones to do a drop of work in this area, and I keep > cheering you on - especially on the bug fixing and wonderful scripts > and data you've provided so far, and I do wish I could find some way > to help that didn't cause so much ptsd in me. > > I wish I merely knew more about how dctp was configured in the data > center. So far as *I* know its on dedicated switches mostly. I would > vastly prefer a dscp codepoint for this, also. > > On Mon, Mar 8, 2021 at 4:36 PM Pete Heist <pete@heistp.net> wrote: > > > > Sorry for reviving an old thread as I haven't been on this list in > > a > > while: > > > > > > SCE proposes to use ect(1) as an indicator of some congestion > > > > and > > does > > > > not explictly > > > > require a dscp codepoint in a FQ'd implementation. > > > Pretty much. I do think that a demonstration using an > > > additional DSCP to create a similar HOV lane for SCE would have > > > gone > > > miles in convincing people in the WG that L4S might really not be > > > as > > > swell as its proponents argue, IMHO it won the day more with its > > > attractive promise of low latency for all instead of what it > > delivers. > > > > On that, I don't think any of us knows how things will end up or > > how > > long it will take to get there... > > > > I do agree that the interim meeting leading up to the codepoint > > decision could have gone better. Everything went great until it > > came to > > how to deploy SCE in a small number of queues. We had dismissed the > > idea of using DSCP, because we thought it would be panned for its > > poor > > traversal over the Internet. That may still have been the case, but > > it > > also may have worked if sold right. We thought that AF alone might > > be > > enough to get past that part, but it wasn't. > > > > We already implemented a two-queue design that uses DSCP, but > > either > > there wasn't much interest, or we didn't sell it enough. Plus, for > > those who demand a two queue solution that requires no flow > > awareness > > at all, DSCP alone may not get you there, because you still need > > some > > reasonably fair way to schedule the two queues. So that might have > > been > > the next line of criticism. Scheduling in proportion to the number > > of > > flows each queue contains is one effective way to do that, but that > > requires at least some concept of a flow. Perhaps there's another > > way > > that doesn't introduce too much RTT unfairness, but I'm not sure. > > > > In our defense, there was already a lot of support built up for > > L4S, > > and stepping in front of that was like stepping in front of a > > freight > > train no matter what we did. I think we've made a decent argument > > in > > the most recent version of the SCE draft that ECN is a "network > > feature" which carries higher risks than drop-based signaling, and > > warrants the requirement for unresponsive flow mitigation, for > > starters. That of course requires some level of flow awareness, > > which > > then makes various queueing designs possible. And, there may still > > be > > deployment possibilities with DSCP, as Rodney mentioned. > > > > Anyway, there's progress being made on SCE, with some new ideas and > > improvements to testing tools coming along. > > > > Pete > > > > > > _______________________________________________ > > Ecn-sane mailing list > > Ecn-sane@lists.bufferbloat.net > > https://lists.bufferbloat.net/listinfo/ecn-sane > > > ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: [Ecn-sane] ect(1) queue selector question 2021-03-09 1:18 ` Dave Taht 2021-03-09 9:11 ` Pete Heist @ 2021-03-09 15:19 ` Rodney W. Grimes 2021-03-09 15:28 ` Dave Taht 1 sibling, 1 reply; 15+ messages in thread From: Rodney W. Grimes @ 2021-03-09 15:19 UTC (permalink / raw) To: Dave Taht; +Cc: Pete Heist, ECN-Sane > I am of course much happier interacting here than over on tsvwg. > > I would like very much > > A large scale bit twiddling test from the larger providers of fq_codel > and cake based hw and middleboxes. > > extensive testing on lte and wifi transports. even emulated. > the sce patches polished up and submitted for review to netdev as what > we call there, an RFC > the ability to test both SCE and L4S on openwrt (backport thereof) > > I know there's been a lot of work on other qdiscs than cake, but > haven't paid much attention. netdev review would be nice. > > A simple internet-wide test of say bittorrent, or using an adserver > method like what apnic uses. > > Most of all I'd really like someone to take a stab at RFC3168 support > for BBR. And by that, I don't mean a reduction by half, but to > institute an RTT probe phase. A fixed rate reduction per RTT is simply > not > going to work IMHO, period, on lte and wifi. I'm sure dave miller > would accept such a patch, and this would also lead towards better ecn > support across the internet in general... at least from the > perspective of the one provider I have an in with, dropbox. > > SCE support for BBRv2 would be nice also. > > And I know, a pony, all sparkly and stuff. I find it very difficult to > summon the cojones to do a drop of work in this area, and I keep > cheering you on - especially on the bug fixing and wonderful scripts > and data you've provided so far, and I do wish I could find some way > to help that didn't cause so much ptsd in me. > > I wish I merely knew more about how dctp was configured in the data > center. So far as *I* know its on dedicated switches mostly. I would > vastly prefer a dscp codepoint for this, also. If you want I can take a discussion on how DCTCP is configured in data center switches in a side channel. Its basically using a drop function that has a starting queue depth, and slope that as you go higher on the slope your propability of marking increases. The 40 gbit experiments we did with HPE modified that part of the switch asic code to do SCE marks starting at 1% mark probability for 1% Queue depth, up to 100% marking at 100% Queue depth. Sadly that was a slight mistake on my part, and I thought we would get a chance to iterate and retune these, I just pulled those values out as a WOG and gave them to HPE to let them get on with setting it up. For most practical purposes an SCE mark and a DCTCP CE mark convey very similiar, if not the same information, as does a CE mark in L4S. Usually tunning of these queue depth and slope values involes data collection and several iterations of experiments. TOR vs SPINE switches are usually configured with different values. I believe there is even some work by someone that does data collection across the whole datacenter and tries to adjust these automagically and dynamically. > > On Mon, Mar 8, 2021 at 4:36 PM Pete Heist <pete@heistp.net> wrote: > > > > Sorry for reviving an old thread as I haven't been on this list in a > > while: > > > > > > SCE proposes to use ect(1) as an indicator of some congestion and > > does > > > > not explictly > > > > require a dscp codepoint in a FQ'd implementation. > > > Pretty much. I do think that a demonstration using an > > > additional DSCP to create a similar HOV lane for SCE would have gone > > > miles in convincing people in the WG that L4S might really not be as > > > swell as its proponents argue, IMHO it won the day more with its > > > attractive promise of low latency for all instead of what it > > delivers. > > > > On that, I don't think any of us knows how things will end up or how > > long it will take to get there... > > > > I do agree that the interim meeting leading up to the codepoint > > decision could have gone better. Everything went great until it came to > > how to deploy SCE in a small number of queues. We had dismissed the > > idea of using DSCP, because we thought it would be panned for its poor > > traversal over the Internet. That may still have been the case, but it > > also may have worked if sold right. We thought that AF alone might be > > enough to get past that part, but it wasn't. > > > > We already implemented a two-queue design that uses DSCP, but either > > there wasn't much interest, or we didn't sell it enough. Plus, for > > those who demand a two queue solution that requires no flow awareness > > at all, DSCP alone may not get you there, because you still need some > > reasonably fair way to schedule the two queues. So that might have been > > the next line of criticism. Scheduling in proportion to the number of > > flows each queue contains is one effective way to do that, but that > > requires at least some concept of a flow. Perhaps there's another way > > that doesn't introduce too much RTT unfairness, but I'm not sure. > > > > In our defense, there was already a lot of support built up for L4S, > > and stepping in front of that was like stepping in front of a freight > > train no matter what we did. I think we've made a decent argument in > > the most recent version of the SCE draft that ECN is a "network > > feature" which carries higher risks than drop-based signaling, and > > warrants the requirement for unresponsive flow mitigation, for > > starters. That of course requires some level of flow awareness, which > > then makes various queueing designs possible. And, there may still be > > deployment possibilities with DSCP, as Rodney mentioned. > > > > Anyway, there's progress being made on SCE, with some new ideas and > > improvements to testing tools coming along. > > > > Pete > > > > > > Ecn-sane@lists.bufferbloat.net > dave@taht.net <Dave T?ht> CTO, TekLibre, LLC Tel: 1-831-435-0729 -- Rod Grimes rgrimes@freebsd.org ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: [Ecn-sane] ect(1) queue selector question 2021-03-09 15:19 ` Rodney W. Grimes @ 2021-03-09 15:28 ` Dave Taht 2021-03-09 15:38 ` Rodney W. Grimes 2021-03-09 18:09 ` [Ecn-sane] A brick wall threshold for SCE_threshold Dave Taht 0 siblings, 2 replies; 15+ messages in thread From: Dave Taht @ 2021-03-09 15:28 UTC (permalink / raw) To: Rodney W. Grimes; +Cc: Pete Heist, ECN-Sane I have basically hoped to find some sysadmin out there with battle experience with a real dctcp deployment (such as the one at facebook) that could share his or her insights as to this debate. Most of the public testing with BBRv2 has instead been with a brick wall setting to CE_threshold, the published data on it was a 260us, which was about as low as raw iron linux can get. I tend to favor the brick wall approach over anything more complex for the AQM component of a SCE architecture, and to modify the transports to suit this assumption. On Tue, Mar 9, 2021 at 7:19 AM Rodney W. Grimes <4bone@gndrsh.dnsmgr.net> wrote: > > > I am of course much happier interacting here than over on tsvwg. > > > > I would like very much > > > > A large scale bit twiddling test from the larger providers of fq_codel > > and cake based hw and middleboxes. > > > > extensive testing on lte and wifi transports. even emulated. > > the sce patches polished up and submitted for review to netdev as what > > we call there, an RFC > > the ability to test both SCE and L4S on openwrt (backport thereof) > > > > I know there's been a lot of work on other qdiscs than cake, but > > haven't paid much attention. netdev review would be nice. > > > > A simple internet-wide test of say bittorrent, or using an adserver > > method like what apnic uses. > > > > Most of all I'd really like someone to take a stab at RFC3168 support > > for BBR. And by that, I don't mean a reduction by half, but to > > institute an RTT probe phase. A fixed rate reduction per RTT is simply > > not > > going to work IMHO, period, on lte and wifi. I'm sure dave miller > > would accept such a patch, and this would also lead towards better ecn > > support across the internet in general... at least from the > > perspective of the one provider I have an in with, dropbox. > > > > SCE support for BBRv2 would be nice also. > > > > And I know, a pony, all sparkly and stuff. I find it very difficult to > > summon the cojones to do a drop of work in this area, and I keep > > cheering you on - especially on the bug fixing and wonderful scripts > > and data you've provided so far, and I do wish I could find some way > > to help that didn't cause so much ptsd in me. > > > > I wish I merely knew more about how dctp was configured in the data > > center. So far as *I* know its on dedicated switches mostly. I would > > vastly prefer a dscp codepoint for this, also. > > If you want I can take a discussion on how DCTCP is configured in > data center switches in a side channel. Its basically using a > drop function that has a starting queue depth, and slope that > as you go higher on the slope your propability of marking > increases. The 40 gbit experiments we did with HPE modified > that part of the switch asic code to do SCE marks starting > at 1% mark probability for 1% Queue depth, up to 100% marking > at 100% Queue depth. Sadly that was a slight mistake on > my part, and I thought we would get a chance to iterate > and retune these, I just pulled those values out as a WOG > and gave them to HPE to let them get on with setting it up. > > For most practical purposes an SCE mark and a DCTCP CE > mark convey very similiar, if not the same information, > as does a CE mark in L4S. > > Usually tunning of these queue depth and slope values involes > data collection and several iterations of experiments. > > TOR vs SPINE switches are usually configured with different values. > > I believe there is even some work by someone that does data > collection across the whole datacenter and tries to adjust > these automagically and dynamically. > > > > > On Mon, Mar 8, 2021 at 4:36 PM Pete Heist <pete@heistp.net> wrote: > > > > > > Sorry for reviving an old thread as I haven't been on this list in a > > > while: > > > > > > > > SCE proposes to use ect(1) as an indicator of some congestion and > > > does > > > > > not explictly > > > > > require a dscp codepoint in a FQ'd implementation. > > > > Pretty much. I do think that a demonstration using an > > > > additional DSCP to create a similar HOV lane for SCE would have gone > > > > miles in convincing people in the WG that L4S might really not be as > > > > swell as its proponents argue, IMHO it won the day more with its > > > > attractive promise of low latency for all instead of what it > > > delivers. > > > > > > On that, I don't think any of us knows how things will end up or how > > > long it will take to get there... > > > > > > I do agree that the interim meeting leading up to the codepoint > > > decision could have gone better. Everything went great until it came to > > > how to deploy SCE in a small number of queues. We had dismissed the > > > idea of using DSCP, because we thought it would be panned for its poor > > > traversal over the Internet. That may still have been the case, but it > > > also may have worked if sold right. We thought that AF alone might be > > > enough to get past that part, but it wasn't. > > > > > > We already implemented a two-queue design that uses DSCP, but either > > > there wasn't much interest, or we didn't sell it enough. Plus, for > > > those who demand a two queue solution that requires no flow awareness > > > at all, DSCP alone may not get you there, because you still need some > > > reasonably fair way to schedule the two queues. So that might have been > > > the next line of criticism. Scheduling in proportion to the number of > > > flows each queue contains is one effective way to do that, but that > > > requires at least some concept of a flow. Perhaps there's another way > > > that doesn't introduce too much RTT unfairness, but I'm not sure. > > > > > > In our defense, there was already a lot of support built up for L4S, > > > and stepping in front of that was like stepping in front of a freight > > > train no matter what we did. I think we've made a decent argument in > > > the most recent version of the SCE draft that ECN is a "network > > > feature" which carries higher risks than drop-based signaling, and > > > warrants the requirement for unresponsive flow mitigation, for > > > starters. That of course requires some level of flow awareness, which > > > then makes various queueing designs possible. And, there may still be > > > deployment possibilities with DSCP, as Rodney mentioned. > > > > > > Anyway, there's progress being made on SCE, with some new ideas and > > > improvements to testing tools coming along. > > > > > > Pete > > > > > > > > > Ecn-sane@lists.bufferbloat.net > > dave@taht.net <Dave T?ht> CTO, TekLibre, LLC Tel: 1-831-435-0729 > -- > Rod Grimes rgrimes@freebsd.org -- "For a successful technology, reality must take precedence over public relations, for Mother Nature cannot be fooled" - Richard Feynman dave@taht.net <Dave Täht> CTO, TekLibre, LLC Tel: 1-831-435-0729 ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: [Ecn-sane] ect(1) queue selector question 2021-03-09 15:28 ` Dave Taht @ 2021-03-09 15:38 ` Rodney W. Grimes 2021-03-09 18:09 ` [Ecn-sane] A brick wall threshold for SCE_threshold Dave Taht 1 sibling, 0 replies; 15+ messages in thread From: Rodney W. Grimes @ 2021-03-09 15:38 UTC (permalink / raw) To: Dave Taht; +Cc: Rodney W. Grimes, Pete Heist, ECN-Sane > I have basically hoped to find some sysadmin out there with battle > experience with a real dctcp deployment (such as the one at facebook) > that could share his or her > insights as to this debate. My "experience" comes from working with HPE who are very experienced at building data centers running RDMA using DCTCP. I do not know who the end customer is of the products though. This is now mostly 100Gb/s and above Aruba switch gear. > > Most of the public testing with BBRv2 has instead been with a brick > wall setting to CE_threshold, the published data on it was a 260us, > which was about as low as raw iron linux can get. I tend to favor the > brick wall approach over anything more complex for the AQM component > of a SCE architecture, and to modify the transports to suit this > assumption. The current information I have from HPE on this is that infact they do use a brick wall or cliff setting, ie the slope is infinity, as basically with all the offload going on in NIC cards you simply can not react in time to much queue depth at all. I told them to turn the port speed down, to which they pretty much ran me out.... I call it poor mans pacing :-) > > On Tue, Mar 9, 2021 at 7:19 AM Rodney W. Grimes <4bone@gndrsh.dnsmgr.net> wrote: > > > > > I am of course much happier interacting here than over on tsvwg. > > > > > > I would like very much > > > > > > A large scale bit twiddling test from the larger providers of fq_codel > > > and cake based hw and middleboxes. > > > > > > extensive testing on lte and wifi transports. even emulated. > > > the sce patches polished up and submitted for review to netdev as what > > > we call there, an RFC > > > the ability to test both SCE and L4S on openwrt (backport thereof) > > > > > > I know there's been a lot of work on other qdiscs than cake, but > > > haven't paid much attention. netdev review would be nice. > > > > > > A simple internet-wide test of say bittorrent, or using an adserver > > > method like what apnic uses. > > > > > > Most of all I'd really like someone to take a stab at RFC3168 support > > > for BBR. And by that, I don't mean a reduction by half, but to > > > institute an RTT probe phase. A fixed rate reduction per RTT is simply > > > not > > > going to work IMHO, period, on lte and wifi. I'm sure dave miller > > > would accept such a patch, and this would also lead towards better ecn > > > support across the internet in general... at least from the > > > perspective of the one provider I have an in with, dropbox. > > > > > > SCE support for BBRv2 would be nice also. > > > > > > And I know, a pony, all sparkly and stuff. I find it very difficult to > > > summon the cojones to do a drop of work in this area, and I keep > > > cheering you on - especially on the bug fixing and wonderful scripts > > > and data you've provided so far, and I do wish I could find some way > > > to help that didn't cause so much ptsd in me. > > > > > > I wish I merely knew more about how dctp was configured in the data > > > center. So far as *I* know its on dedicated switches mostly. I would > > > vastly prefer a dscp codepoint for this, also. > > > > If you want I can take a discussion on how DCTCP is configured in > > data center switches in a side channel. Its basically using a > > drop function that has a starting queue depth, and slope that > > as you go higher on the slope your propability of marking > > increases. The 40 gbit experiments we did with HPE modified > > that part of the switch asic code to do SCE marks starting > > at 1% mark probability for 1% Queue depth, up to 100% marking > > at 100% Queue depth. Sadly that was a slight mistake on > > my part, and I thought we would get a chance to iterate > > and retune these, I just pulled those values out as a WOG > > and gave them to HPE to let them get on with setting it up. > > > > For most practical purposes an SCE mark and a DCTCP CE > > mark convey very similiar, if not the same information, > > as does a CE mark in L4S. > > > > Usually tunning of these queue depth and slope values involes > > data collection and several iterations of experiments. > > > > TOR vs SPINE switches are usually configured with different values. > > > > I believe there is even some work by someone that does data > > collection across the whole datacenter and tries to adjust > > these automagically and dynamically. > > > > > > > > On Mon, Mar 8, 2021 at 4:36 PM Pete Heist <pete@heistp.net> wrote: > > > > > > > > Sorry for reviving an old thread as I haven't been on this list in a > > > > while: > > > > > > > > > > SCE proposes to use ect(1) as an indicator of some congestion and > > > > does > > > > > > not explictly > > > > > > require a dscp codepoint in a FQ'd implementation. > > > > > Pretty much. I do think that a demonstration using an > > > > > additional DSCP to create a similar HOV lane for SCE would have gone > > > > > miles in convincing people in the WG that L4S might really not be as > > > > > swell as its proponents argue, IMHO it won the day more with its > > > > > attractive promise of low latency for all instead of what it > > > > delivers. > > > > > > > > On that, I don't think any of us knows how things will end up or how > > > > long it will take to get there... > > > > > > > > I do agree that the interim meeting leading up to the codepoint > > > > decision could have gone better. Everything went great until it came to > > > > how to deploy SCE in a small number of queues. We had dismissed the > > > > idea of using DSCP, because we thought it would be panned for its poor > > > > traversal over the Internet. That may still have been the case, but it > > > > also may have worked if sold right. We thought that AF alone might be > > > > enough to get past that part, but it wasn't. > > > > > > > > We already implemented a two-queue design that uses DSCP, but either > > > > there wasn't much interest, or we didn't sell it enough. Plus, for > > > > those who demand a two queue solution that requires no flow awareness > > > > at all, DSCP alone may not get you there, because you still need some > > > > reasonably fair way to schedule the two queues. So that might have been > > > > the next line of criticism. Scheduling in proportion to the number of > > > > flows each queue contains is one effective way to do that, but that > > > > requires at least some concept of a flow. Perhaps there's another way > > > > that doesn't introduce too much RTT unfairness, but I'm not sure. > > > > > > > > In our defense, there was already a lot of support built up for L4S, > > > > and stepping in front of that was like stepping in front of a freight > > > > train no matter what we did. I think we've made a decent argument in > > > > the most recent version of the SCE draft that ECN is a "network > > > > feature" which carries higher risks than drop-based signaling, and > > > > warrants the requirement for unresponsive flow mitigation, for > > > > starters. That of course requires some level of flow awareness, which > > > > then makes various queueing designs possible. And, there may still be > > > > deployment possibilities with DSCP, as Rodney mentioned. > > > > > > > > Anyway, there's progress being made on SCE, with some new ideas and > > > > improvements to testing tools coming along. > > > > > > > > Pete > > > > > > > > > > > > Ecn-sane@lists.bufferbloat.net > > > dave@taht.net <Dave T?ht> CTO, TekLibre, LLC Tel: 1-831-435-0729 > > -- > > Rod Grimes rgrimes@freebsd.org > > > > -- > "For a successful technology, reality must take precedence over public > relations, for Mother Nature cannot be fooled" - Richard Feynman > > dave@taht.net <Dave T?ht> CTO, TekLibre, LLC Tel: 1-831-435-0729 > > -- Rod Grimes rgrimes@freebsd.org ^ permalink raw reply [flat|nested] 15+ messages in thread
* [Ecn-sane] A brick wall threshold for SCE_threshold 2021-03-09 15:28 ` Dave Taht 2021-03-09 15:38 ` Rodney W. Grimes @ 2021-03-09 18:09 ` Dave Taht 2021-03-09 18:42 ` Jonathan Morton 1 sibling, 1 reply; 15+ messages in thread From: Dave Taht @ 2021-03-09 18:09 UTC (permalink / raw) To: Rodney W. Grimes; +Cc: Pete Heist, ECN-Sane while I appreciated the experiments with making the assertion of SCE_threshold spread out, and not a brick wall, to me it seems best to coalesce on a brick wall for it at a fixed period of time the operator considers "queuing". Does the SCE team agree with this approach? I note, in trying to wrap my head around this as to how to make it work at all, for wifi, that I was kind of stuck on it happening after 1-2 txops, as that was the minimum queue depth that can be achieved today with today's hw. However, that is a minimum !250us, and a maximum of over 10ms, and that doesn't count at all the arbitratraton delay potential from many stations which can stretch out to hundreds of ms. Nowadays I think about it as merely being a fixed amount of time, not txops, in the 2.5ms range. I do not like jitter and interference inducing long txops in the first place, and present day 802.11ac can pack so much data into a single aggregate, that finding ways to get rid of longer txops in general has long been on my mind so as wifi could multiplex better over many more stations. as for LTE, gawd knows. Still, that multiplexing over stations takes a LONG time, and perhaps it would be better to apply a delta from the last station service time to the sce_threshold time, before considering it too much queuing. On Tue, Mar 9, 2021 at 7:28 AM Dave Taht <dave.taht@gmail.com> wrote: > > I have basically hoped to find some sysadmin out there with battle > experience with a real dctcp deployment (such as the one at facebook) > that could share his or her > insights as to this debate. > > Most of the public testing with BBRv2 has instead been with a brick > wall setting to CE_threshold, the published data on it was a 260us, > which was about as low as raw iron linux can get. I tend to favor the > brick wall approach over anything more complex for the AQM component > of a SCE architecture, and to modify the transports to suit this > assumption. > > On Tue, Mar 9, 2021 at 7:19 AM Rodney W. Grimes <4bone@gndrsh.dnsmgr.net> wrote: > > > > > I am of course much happier interacting here than over on tsvwg. > > > > > > I would like very much > > > > > > A large scale bit twiddling test from the larger providers of fq_codel > > > and cake based hw and middleboxes. > > > > > > extensive testing on lte and wifi transports. even emulated. > > > the sce patches polished up and submitted for review to netdev as what > > > we call there, an RFC > > > the ability to test both SCE and L4S on openwrt (backport thereof) > > > > > > I know there's been a lot of work on other qdiscs than cake, but > > > haven't paid much attention. netdev review would be nice. > > > > > > A simple internet-wide test of say bittorrent, or using an adserver > > > method like what apnic uses. > > > > > > Most of all I'd really like someone to take a stab at RFC3168 support > > > for BBR. And by that, I don't mean a reduction by half, but to > > > institute an RTT probe phase. A fixed rate reduction per RTT is simply > > > not > > > going to work IMHO, period, on lte and wifi. I'm sure dave miller > > > would accept such a patch, and this would also lead towards better ecn > > > support across the internet in general... at least from the > > > perspective of the one provider I have an in with, dropbox. > > > > > > SCE support for BBRv2 would be nice also. > > > > > > And I know, a pony, all sparkly and stuff. I find it very difficult to > > > summon the cojones to do a drop of work in this area, and I keep > > > cheering you on - especially on the bug fixing and wonderful scripts > > > and data you've provided so far, and I do wish I could find some way > > > to help that didn't cause so much ptsd in me. > > > > > > I wish I merely knew more about how dctp was configured in the data > > > center. So far as *I* know its on dedicated switches mostly. I would > > > vastly prefer a dscp codepoint for this, also. > > > > If you want I can take a discussion on how DCTCP is configured in > > data center switches in a side channel. Its basically using a > > drop function that has a starting queue depth, and slope that > > as you go higher on the slope your propability of marking > > increases. The 40 gbit experiments we did with HPE modified > > that part of the switch asic code to do SCE marks starting > > at 1% mark probability for 1% Queue depth, up to 100% marking > > at 100% Queue depth. Sadly that was a slight mistake on > > my part, and I thought we would get a chance to iterate > > and retune these, I just pulled those values out as a WOG > > and gave them to HPE to let them get on with setting it up. > > > > For most practical purposes an SCE mark and a DCTCP CE > > mark convey very similiar, if not the same information, > > as does a CE mark in L4S. > > > > Usually tunning of these queue depth and slope values involes > > data collection and several iterations of experiments. > > > > TOR vs SPINE switches are usually configured with different values. > > > > I believe there is even some work by someone that does data > > collection across the whole datacenter and tries to adjust > > these automagically and dynamically. > > > > > > > > On Mon, Mar 8, 2021 at 4:36 PM Pete Heist <pete@heistp.net> wrote: > > > > > > > > Sorry for reviving an old thread as I haven't been on this list in a > > > > while: > > > > > > > > > > SCE proposes to use ect(1) as an indicator of some congestion and > > > > does > > > > > > not explictly > > > > > > require a dscp codepoint in a FQ'd implementation. > > > > > Pretty much. I do think that a demonstration using an > > > > > additional DSCP to create a similar HOV lane for SCE would have gone > > > > > miles in convincing people in the WG that L4S might really not be as > > > > > swell as its proponents argue, IMHO it won the day more with its > > > > > attractive promise of low latency for all instead of what it > > > > delivers. > > > > > > > > On that, I don't think any of us knows how things will end up or how > > > > long it will take to get there... > > > > > > > > I do agree that the interim meeting leading up to the codepoint > > > > decision could have gone better. Everything went great until it came to > > > > how to deploy SCE in a small number of queues. We had dismissed the > > > > idea of using DSCP, because we thought it would be panned for its poor > > > > traversal over the Internet. That may still have been the case, but it > > > > also may have worked if sold right. We thought that AF alone might be > > > > enough to get past that part, but it wasn't. > > > > > > > > We already implemented a two-queue design that uses DSCP, but either > > > > there wasn't much interest, or we didn't sell it enough. Plus, for > > > > those who demand a two queue solution that requires no flow awareness > > > > at all, DSCP alone may not get you there, because you still need some > > > > reasonably fair way to schedule the two queues. So that might have been > > > > the next line of criticism. Scheduling in proportion to the number of > > > > flows each queue contains is one effective way to do that, but that > > > > requires at least some concept of a flow. Perhaps there's another way > > > > that doesn't introduce too much RTT unfairness, but I'm not sure. > > > > > > > > In our defense, there was already a lot of support built up for L4S, > > > > and stepping in front of that was like stepping in front of a freight > > > > train no matter what we did. I think we've made a decent argument in > > > > the most recent version of the SCE draft that ECN is a "network > > > > feature" which carries higher risks than drop-based signaling, and > > > > warrants the requirement for unresponsive flow mitigation, for > > > > starters. That of course requires some level of flow awareness, which > > > > then makes various queueing designs possible. And, there may still be > > > > deployment possibilities with DSCP, as Rodney mentioned. > > > > > > > > Anyway, there's progress being made on SCE, with some new ideas and > > > > improvements to testing tools coming along. > > > > > > > > Pete > > > > > > > > > > > > Ecn-sane@lists.bufferbloat.net > > > dave@taht.net <Dave T?ht> CTO, TekLibre, LLC Tel: 1-831-435-0729 > > -- > > Rod Grimes rgrimes@freebsd.org > > > > -- > "For a successful technology, reality must take precedence over public > relations, for Mother Nature cannot be fooled" - Richard Feynman > > dave@taht.net <Dave Täht> CTO, TekLibre, LLC Tel: 1-831-435-0729 -- "For a successful technology, reality must take precedence over public relations, for Mother Nature cannot be fooled" - Richard Feynman dave@taht.net <Dave Täht> CTO, TekLibre, LLC Tel: 1-831-435-0729 ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: [Ecn-sane] A brick wall threshold for SCE_threshold 2021-03-09 18:09 ` [Ecn-sane] A brick wall threshold for SCE_threshold Dave Taht @ 2021-03-09 18:42 ` Jonathan Morton 2021-03-09 18:48 ` Dave Taht 0 siblings, 1 reply; 15+ messages in thread From: Jonathan Morton @ 2021-03-09 18:42 UTC (permalink / raw) To: Dave Taht; +Cc: Rodney W. Grimes, ECN-Sane > On 9 Mar, 2021, at 8:09 pm, Dave Taht <dave.taht@gmail.com> wrote: > > while I appreciated the experiments with making the assertion of SCE_threshold spread out, and not a brick wall, to me it seems best to coalesce on a brick wall for it at a fixed period of time the operator considers "queuing". > > Does the SCE team agree with this approach? In a word, no. While it might be an acceptable compromise versus implementation cost in the datacentre, in the general case you're not going to get good performance that way. Formally, what a brick-wall control law gives you is control of *peak* delay in that queue. It only takes a small number of packets in each RTT going over the threshold to make the flow back off. If there's more variation of delay than the threshold is set at, then the trough will be with the queue empty and the link idle - and the link introducing the variation might be remote from the one serving as the bottleneck and doing the marking, so you can't merely set the threshold higher at those particular links. Kathy Nichols designed Codel specifically to control the *trough* of a delay with variance. This strategy has worked well and applies nicely to paths including a wifi link, so long as the target sojourn time is greater than the time between txops. Since the limit on wifi aggregation is 4ms, the default setting of 5ms dovetails nicely with that. The SCE marking target is 2.5ms is actually a little low, but using this with Codel still works markedly better than with a fixed ramp based at the same level, which would tend to control the peak. Some of my current effort is going into a new AQM algorithm called DelTiC (Delay Time Control). You can think of it as taking the best ideas from Codel and PIE and putting them together. Like Codel, it measures queue sojourn time and manages a marking frequency (not probability), making it a fully time-domain algorithm. The control law is PID, not merely PI, so it can detect a rapidly rising queue depth (due to slow-start exponential growth, say) and act sooner to control it than Codel does. I think that was one of the big shortcomings of Codel you were keen to see solved. Unlike Codel, DelTiC has a true steady-state condition where the marking frequency can remain constant. This goes well with the SCE response function, which permits a constant cwnd when an appropriate marking frequency is received. This in turn is completely unlike traditional AIMD congestion control, which exhibits a large sawtooth. Implementation of the first qdisc incorporating DelTiC is not yet complete, and then we have to see how well it works in practice. I'll talk about it more when it's ready. - Jonathan Morton ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: [Ecn-sane] A brick wall threshold for SCE_threshold 2021-03-09 18:42 ` Jonathan Morton @ 2021-03-09 18:48 ` Dave Taht 2021-03-09 18:58 ` Jonathan Morton 0 siblings, 1 reply; 15+ messages in thread From: Dave Taht @ 2021-03-09 18:48 UTC (permalink / raw) To: Jonathan Morton; +Cc: Rodney W. Grimes, ECN-Sane I do not think anything other than a brick wall can be implemented in high speed hardware. I really wish you'd try a brick wall, and fiddle on bbr. It is also heavily influenced by kathies work on codel. On Tue, Mar 9, 2021 at 10:43 AM Jonathan Morton <chromatix99@gmail.com> wrote: > > > On 9 Mar, 2021, at 8:09 pm, Dave Taht <dave.taht@gmail.com> wrote: > > > > while I appreciated the experiments with making the assertion of SCE_threshold spread out, and not a brick wall, to me it seems best to coalesce on a brick wall for it at a fixed period of time the operator considers "queuing". > > > > Does the SCE team agree with this approach? > > In a word, no. While it might be an acceptable compromise versus implementation cost in the datacentre, in the general case you're not going to get good performance that way. > > Formally, what a brick-wall control law gives you is control of *peak* delay in that queue. It only takes a small number of packets in each RTT going over the threshold to make the flow back off. If there's more variation of delay than the threshold is set at, then the trough will be with the queue empty and the link idle - and the link introducing the variation might be remote from the one serving as the bottleneck and doing the marking, so you can't merely set the threshold higher at those particular links. > > Kathy Nichols designed Codel specifically to control the *trough* of a delay with variance. This strategy has worked well and applies nicely to paths including a wifi link, so long as the target sojourn time is greater than the time between txops. Since the limit on wifi aggregation is 4ms, the default setting of 5ms dovetails nicely with that. The SCE marking target is 2.5ms is actually a little low, but using this with Codel still works markedly better than with a fixed ramp based at the same level, which would tend to control the peak. > > Some of my current effort is going into a new AQM algorithm called DelTiC (Delay Time Control). You can think of it as taking the best ideas from Codel and PIE and putting them together. Like Codel, it measures queue sojourn time and manages a marking frequency (not probability), making it a fully time-domain algorithm. The control law is PID, not merely PI, so it can detect a rapidly rising queue depth (due to slow-start exponential growth, say) and act sooner to control it than Codel does. I think that was one of the big shortcomings of Codel you were keen to see solved. > > Unlike Codel, DelTiC has a true steady-state condition where the marking frequency can remain constant. This goes well with the SCE response function, which permits a constant cwnd when an appropriate marking frequency is received. This in turn is completely unlike traditional AIMD congestion control, which exhibits a large sawtooth. > > Implementation of the first qdisc incorporating DelTiC is not yet complete, and then we have to see how well it works in practice. I'll talk about it more when it's ready. > > - Jonathan Morton -- "For a successful technology, reality must take precedence over public relations, for Mother Nature cannot be fooled" - Richard Feynman dave@taht.net <Dave Täht> CTO, TekLibre, LLC Tel: 1-831-435-0729 ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: [Ecn-sane] A brick wall threshold for SCE_threshold 2021-03-09 18:48 ` Dave Taht @ 2021-03-09 18:58 ` Jonathan Morton 2021-03-09 19:20 ` Dave Taht 0 siblings, 1 reply; 15+ messages in thread From: Jonathan Morton @ 2021-03-09 18:58 UTC (permalink / raw) To: Dave Taht; +Cc: Rodney W. Grimes, ECN-Sane > On 9 Mar, 2021, at 8:48 pm, Dave Taht <dave.taht@gmail.com> wrote: > > I do not think anything other than a brick wall can be implemented in high speed hardware. We've spoken to a Cisco guy who says they can already do Approximate Fairness at 100Gbps per port. That's our benchmark. > I really wish you'd try a brick wall We did. We didn't like the results. - Jonathan Morton ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: [Ecn-sane] A brick wall threshold for SCE_threshold 2021-03-09 18:58 ` Jonathan Morton @ 2021-03-09 19:20 ` Dave Taht 0 siblings, 0 replies; 15+ messages in thread From: Dave Taht @ 2021-03-09 19:20 UTC (permalink / raw) To: Jonathan Morton; +Cc: Rodney W. Grimes, ECN-Sane On Tue, Mar 9, 2021 at 10:58 AM Jonathan Morton <chromatix99@gmail.com> wrote: > > > On 9 Mar, 2021, at 8:48 pm, Dave Taht <dave.taht@gmail.com> wrote: > > > > I do not think anything other than a brick wall can be implemented in high speed hardware. > > We've spoken to a Cisco guy who says they can already do Approximate Fairness at 100Gbps per port. That's our benchmark. I have long wondered how much of AFD they were selling. > > > I really wish you'd try a brick wall > > We did. We didn't like the results. But you did not try a bbr-like approach. Observe a mark, and the rtt inflation. bbrv2 works pretty decently with CE_threshold already. making it work with a brick wall SCE made some sense. > > - Jonathan Morton -- "For a successful technology, reality must take precedence over public relations, for Mother Nature cannot be fooled" - Richard Feynman dave@taht.net <Dave Täht> CTO, TekLibre, LLC Tel: 1-831-435-0729 ^ permalink raw reply [flat|nested] 15+ messages in thread
* [Ecn-sane] ect(1) queue selector question @ 2021-02-20 19:27 Dave Taht 2021-02-21 1:51 ` Sebastian Moeller 0 siblings, 1 reply; 15+ messages in thread From: Dave Taht @ 2021-02-20 19:27 UTC (permalink / raw) To: ECN-Sane I note that I have done my best to stay out of this for a while, so long that I (thankfully) mis-remember various aspects of the debate. Today I have a question about l4s vs SCE as it's come up again. l4s uses both a dscp codepoint AND ect(1) to indicate dctcp style congestion control is in use, and also can dump other protocols into that queue lacking any ecn markings. SCE proposes to use ect(1) as an indicator of some congestion and does not explictly require a dscp codepoint in a FQ'd implementation. Do I have that right? Now, my question was, simply, in MPLS or X-G are they out of bits, and that's why they want to use up this one in L4S? -- "For a successful technology, reality must take precedence over public relations, for Mother Nature cannot be fooled" - Richard Feynman dave@taht.net <Dave Täht> CTO, TekLibre, LLC Tel: 1-831-435-0729 ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: [Ecn-sane] ect(1) queue selector question 2021-02-20 19:27 [Ecn-sane] ect(1) queue selector question Dave Taht @ 2021-02-21 1:51 ` Sebastian Moeller 2021-02-21 17:14 ` Rodney W. Grimes 0 siblings, 1 reply; 15+ messages in thread From: Sebastian Moeller @ 2021-02-21 1:51 UTC (permalink / raw) To: Dave Täht; +Cc: ECN-Sane Hi Dave, > On Feb 20, 2021, at 20:27, Dave Taht <dave.taht@gmail.com> wrote: > > I note that I have done my best to stay out of this for a while, so > long that I (thankfully) mis-remember various aspects of the debate. > Today I have a question about l4s vs SCE as it's come up again. > > l4s uses both a dscp codepoint AND ect(1) to indicate dctcp style > congestion control > is in use, and also can dump other protocols into that queue lacking > any ecn markings. Mmmh, according to the L4S internet drafts, L4S does not want to use a DSCP at all. Interestingly enough, Greg White is proposing a related internet draft about the NQB PHB and DSCP that sounds awfully like the missing DSCP in the L4S drafts. IMHO if the whole thing would be guarded behind a DSCP I would be less annoyed by the design and process of L4S.... > > SCE proposes to use ect(1) as an indicator of some congestion and does > not explictly > require a dscp codepoint in a FQ'd implementation. Pretty much. I do think that a demonstration using an additional DSCP to create a similar HOV lane for SCE would have gone miles in convincing people in the WG that L4S might really not be as swell as its proponents argue, IMHO it won the day more with its attractive promise of low latency for all instead of what it delivers. > > Do I have that right? Now, my question was, simply, in MPLS or X-G are > they out of bits, and > that's why they want to use up this one in L4S? I do not think that MPLS folks are a driver in all of this, no? No idea what X-G is. > > -- > "For a successful technology, reality must take precedence over public > relations, for Mother Nature cannot be fooled" - Richard Feynman > > dave@taht.net <Dave Täht> CTO, TekLibre, LLC Tel: 1-831-435-0729 > _______________________________________________ > Ecn-sane mailing list > Ecn-sane@lists.bufferbloat.net > https://lists.bufferbloat.net/listinfo/ecn-sane ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: [Ecn-sane] ect(1) queue selector question 2021-02-21 1:51 ` Sebastian Moeller @ 2021-02-21 17:14 ` Rodney W. Grimes 2021-02-21 20:26 ` Dave Taht 0 siblings, 1 reply; 15+ messages in thread From: Rodney W. Grimes @ 2021-02-21 17:14 UTC (permalink / raw) To: Sebastian Moeller; +Cc: Dave Täht, ECN-Sane Hello Dave, and Sebastian, > Hi Dave, > > > > On Feb 20, 2021, at 20:27, Dave Taht <dave.taht@gmail.com> wrote: > > > > I note that I have done my best to stay out of this for a while, so > > long that I (thankfully) mis-remember various aspects of the debate. > > Today I have a question about l4s vs SCE as it's come up again. > > > > l4s uses both a dscp codepoint AND ect(1) to indicate dctcp style > > congestion control > > is in use, and also can dump other protocols into that queue lacking > > any ecn markings. > > Mmmh, according to the L4S internet drafts, L4S does not want to use a DSCP at all. Interestingly enough, Greg White is proposing a related internet draft about the NQB PHB and DSCP that sounds awfully like the missing DSCP in the L4S drafts. IMHO if the whole thing would be guarded behind a DSCP I would be less annoyed by the design and process of L4S.... That is correct, the L4S proponents absolutely do not want anything to do with a DSCP and L4S. If they would of simply agreed that this was actually a good idea they could propbably be at deployment state now.... but they insist it is pointless to take this route due to the traveral problem. > > > > > SCE proposes to use ect(1) as an indicator of some congestion and does > > not explictly > > require a dscp codepoint in a FQ'd implementation. > > Pretty much. I do think that a demonstration using an additional DSCP to create a similar HOV lane for SCE would have gone miles in convincing people in the WG that L4S might really not be as swell as its proponents argue, IMHO it won the day more with its attractive promise of low latency for all instead of what it delivers. The original SCE design was without any mention of DSCP. Since that pretty much stahled us in the battle with L4S there actually *IS* an SCE design that uses one or more DSCP. This technically allows SCE to proceed forward on its own without any blessing from IETF so that we may gain some additional experience and data. This *IS NOT* the ideal design, but I feel by taking this road we might gain ground. > > Do I have that right? Now, my question was, simply, in MPLS or X-G are > > they out of bits, and > > that's why they want to use up this one in L4S? > > I do not think that MPLS folks are a driver in all of this, no? No idea what X-G is. We do not know fully all of there motivations for being so very insistant on doing it this way, only this way, and with no changes of any kind. But they do seem very stuck on there design and unflexiable in any and all suggestions of change. > > -- > > "For a successful technology, reality must take precedence over public > > relations, for Mother Nature cannot be fooled" - Richard Feynman > > > > dave@taht.net <Dave T?ht> CTO, TekLibre, LLC Tel: 1-831-435-0729 Regards, -- Rod Grimes rgrimes@freebsd.org ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: [Ecn-sane] ect(1) queue selector question 2021-02-21 17:14 ` Rodney W. Grimes @ 2021-02-21 20:26 ` Dave Taht 0 siblings, 0 replies; 15+ messages in thread From: Dave Taht @ 2021-02-21 20:26 UTC (permalink / raw) To: Rodney W. Grimes; +Cc: Sebastian Moeller, ECN-Sane Thx for the clarifications. What I meant by X-G was 3G, 4G, 5G. On Sun, Feb 21, 2021 at 9:14 AM Rodney W. Grimes <4bone@gndrsh.dnsmgr.net> wrote: > > Hello Dave, and Sebastian, > > > Hi Dave, > > > > > > > On Feb 20, 2021, at 20:27, Dave Taht <dave.taht@gmail.com> wrote: > > > > > > I note that I have done my best to stay out of this for a while, so > > > long that I (thankfully) mis-remember various aspects of the debate. > > > Today I have a question about l4s vs SCE as it's come up again. > > > > > > l4s uses both a dscp codepoint AND ect(1) to indicate dctcp style > > > congestion control > > > is in use, and also can dump other protocols into that queue lacking > > > any ecn markings. > > > > Mmmh, according to the L4S internet drafts, L4S does not want to use a DSCP at all. Interestingly enough, Greg White is proposing a related internet draft about the NQB PHB and DSCP that sounds awfully like the missing DSCP in the L4S drafts. IMHO if the whole thing would be guarded behind a DSCP I would be less annoyed by the design and process of L4S.... > > That is correct, the L4S proponents absolutely do not want anything to do with a DSCP and L4S. > If they would of simply agreed that this was actually a good idea they could propbably be at deployment state now.... but they insist it is pointless to take this route due to the traveral problem. > > > > > > > > > SCE proposes to use ect(1) as an indicator of some congestion and does > > > not explictly > > > require a dscp codepoint in a FQ'd implementation. > > > > Pretty much. I do think that a demonstration using an additional DSCP to create a similar HOV lane for SCE would have gone miles in convincing people in the WG that L4S might really not be as swell as its proponents argue, IMHO it won the day more with its attractive promise of low latency for all instead of what it delivers. > > The original SCE design was without any mention of DSCP. Since that pretty much stahled us in the battle with L4S there actually *IS* an SCE design that uses one or more DSCP. This technically allows SCE to proceed forward on its own without any blessing from IETF so that we may gain some additional experience and data. This *IS NOT* the ideal design, but I feel by taking this road we might gain ground. > > > > Do I have that right? Now, my question was, simply, in MPLS or X-G are > > > they out of bits, and > > > that's why they want to use up this one in L4S? > > > > I do not think that MPLS folks are a driver in all of this, no? No idea what X-G is. > > We do not know fully all of there motivations for being so very insistant on doing it this way, only this way, and with no changes of any kind. But they do seem very stuck on there design and unflexiable in any and all suggestions of change. > > > > -- > > > "For a successful technology, reality must take precedence over public > > > relations, for Mother Nature cannot be fooled" - Richard Feynman > > > > > > dave@taht.net <Dave T?ht> CTO, TekLibre, LLC Tel: 1-831-435-0729 > > Regards, > -- > Rod Grimes rgrimes@freebsd.org -- "For a successful technology, reality must take precedence over public relations, for Mother Nature cannot be fooled" - Richard Feynman dave@taht.net <Dave Täht> CTO, TekLibre, LLC Tel: 1-831-435-0729 ^ permalink raw reply [flat|nested] 15+ messages in thread
end of thread, other threads:[~2021-03-09 19:20 UTC | newest] Thread overview: 15+ messages (download: mbox.gz / follow: Atom feed) -- links below jump to the message on this page -- 2021-03-09 0:36 [Ecn-sane] ect(1) queue selector question Pete Heist 2021-03-09 1:18 ` Dave Taht 2021-03-09 9:11 ` Pete Heist 2021-03-09 15:19 ` Rodney W. Grimes 2021-03-09 15:28 ` Dave Taht 2021-03-09 15:38 ` Rodney W. Grimes 2021-03-09 18:09 ` [Ecn-sane] A brick wall threshold for SCE_threshold Dave Taht 2021-03-09 18:42 ` Jonathan Morton 2021-03-09 18:48 ` Dave Taht 2021-03-09 18:58 ` Jonathan Morton 2021-03-09 19:20 ` Dave Taht -- strict thread matches above, loose matches on Subject: below -- 2021-02-20 19:27 [Ecn-sane] ect(1) queue selector question Dave Taht 2021-02-21 1:51 ` Sebastian Moeller 2021-02-21 17:14 ` Rodney W. Grimes 2021-02-21 20:26 ` Dave Taht
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox