TSVWG B. Briscoe Internet Draft P. Eardley draft-briscoe-tsvwg-cl-architecture-04.txt D. Songhurst Expires: April 2007 BT F. Le Faucheur A. Charny Cisco Systems, Inc J. Babiarz K. Chan S. Dudley Nortel G. Karagiannis University of Twente / Ericsson A. Bader L. Westberg Ericsson 25 October, 2006 An edge-to-edge Deployment Model for Pre-Congestion Notification: Admission Control over a DiffServ Region draft-briscoe-tsvwg-cl-architecture-04.txt Status of this Memo By submitting this Internet-Draft, each author represents that any applicable patent or other IPR claims of which he or she is aware have been or will be disclosed, and any of which he or she becomes aware will be disclosed, in accordance with Section 6 of BCP 79. Internet-Drafts are working documents of the Internet Engineering Task Force (IETF), its areas, and its working groups. Note that other groups may also distribute working documents as Internet- Drafts. Internet-Drafts are draft documents valid for a maximum of six months and may be updated, replaced, or obsoleted by other documents at any time. It is inappropriate to use Internet-Drafts as reference material or to cite them other than as "work in progress". The list of current Internet-Drafts can be accessed at http://www.ietf.org/1id-abstracts.html Briscoe Expires April 25, 2007 [Page 1]
Internet-Draft Deployment Model using PCN October 2006 The list of Internet-Draft Shadow Directories can be accessed at http://www.ietf.org/shadow.html This Internet-Draft will expire on September 6, 2006. Copyright Notice Copyright (C) The Internet Society (2006). All Rights Reserved. Abstract This document describes a deployment model for pre-congestion notification (PCN) operating in a large DiffServ-based region of the Internet. PCN-based admission control protects the quality of service of existing flows in normal circumstances, whilst if necessary (eg after a large failure) pre-emption of some flows preserves the quality of service of the remaining flows. Each link has a configured- admission-rate and a configured-pre-emption-rate, and a router marks packets that exceed these rates. Hence routers give an early warning of their own potential congestion, before packets need to be dropped. Gateways around the edges of the PCN-region convert measurements of packet rates and their markings into decisions about whether to admit new flows, and (if necessary) into the rate of excess traffic that should be pre-empted. Per-flow admission states are kept at the gateways only, while the PCN markers that are required for all routers operate on the aggregate traffic - hence there is no scalability impact on interior routers. Authors' Note (TO BE DELETED BY THE RFC EDITOR UPON PUBLICATION) This document is posted as an Internet-Draft with the intention of eventually becoming an INFORMATIONAL RFC. Briscoe Expires April 25, 2007 [Page 2]
Internet-Draft Deployment Model using PCN October 2006 Table of Contents 1. Introduction................................................5 1.1. Summary................................................5 1.2. Key benefits...........................................8 1.3. Terminology............................................9 1.4. Existing terminology...................................11 1.5. Standardisation requirements...........................11 1.6. Structure of rest of the document......................12 2. Key aspects of the deployment model.........................13 2.1. Key goals.............................................13 2.2. Key assumptions........................................14 3. Deployment model...........................................17 3.1. Admission control......................................17 3.1.1. Pre-Congestion Notification for Admission Marking..17 3.1.2. Measurements to support admission control..........17 3.1.3. How edge-to-edge admission control supports end-to-end QoS signalling..........................................18 3.1.4. Use case.........................................18 3.2. Flow pre-emption.......................................20 3.2.1. Alerting an ingress gateway that flow pre-emption may be needed..................................................20 3.2.2. Determining the right amount of CL traffic to drop.23 3.2.3. Use case for flow pre-emption.....................24 3.3. Both admission control and pre-emption.................25 4. Summary of Functionality....................................27 4.1. Ingress gateways.......................................27 4.2. Interior routers.......................................28 4.3. Egress gateways........................................28 4.4. Failures..............................................29 5. Limitations and some potential solutions....................31 5.1. ECMP..................................................31 5.2. Beat down effect.......................................33 5.3. Bi-directional sessions................................35 5.4. Global fairness........................................37 5.5. Flash crowds..........................................39 5.6. Pre-empting too fast...................................41 5.7. Other potential extensions.............................42 5.7.1. Tunnelling........................................42 5.7.2. Multi-domain and multi-operator usage.............43 5.7.3. Preferential dropping of pre-emption marked packets44 5.7.4. Adaptive bandwidth for the Controlled Load service.44 5.7.5. Controlled Load service with end-to-end Pre-Congestion Notification............................................45 5.7.6. MPLS-TE..........................................45 6. Relationship to other QoS mechanisms........................46 Briscoe Expires April 25, 2007 [Page 3]
Internet-Draft Deployment Model using PCN October 2006 6.1. IntServ Controlled Load................................46 6.2. Integrated services operation over DiffServ............46 6.3. Differentiated Services................................46 6.4. ECN...................................................47 6.5. RTECN.................................................47 6.6. RMD...................................................48 6.7. RSVP Aggregation over MPLS-TE..........................48 6.8. Other Network Admission Control Approaches.............48 7. Security Considerations.....................................49 8. Acknowledgements...........................................49 9. Comments solicited.........................................50 10. Changes from earlier versions of the draft.................50 11. Appendices................................................52 11.1. Appendix A: Explicit Congestion Notification..........52 11.2. Appendix B: What is distributed measurement-based admission control?...................................................53 11.3. Appendix C: Calculating the Exponentially weighted moving average (EWMA).............................................54 12. References................................................56 Authors' Addresses............................................61 Intellectual Property Statement................................63 Disclaimer of Validity........................................63 Copyright Statement...........................................63 Briscoe Expires April 25, 2007 [Page 4]
Internet-Draft Deployment Model using PCN October 2006 1. Introduction 1.1. Summary This document describes a deployment model to achieve an end-to-end Controlled Load service by using (within a large region of the Internet) DiffServ and edge-to-edge distributed measurement-based admission control and flow pre-emption. Controlled load service is a quality of service (QoS) closely approximating the QoS that the same flow would receive from a lightly loaded network element [RFC2211]. Controlled Load (CL) is useful for inelastic flows such as those for real-time media. In line with the "IntServ over DiffServ" framework defined in [RFC2998], the CL service is supported end-to-end and RSVP signalling [RFC2205] is used end-to-end, over an edge-to-edge DiffServ region. We call the DiffServ region the "CL-region". ___ ___ _______________________________________ ____ ___ | | | | |Ingress Interior Egress| | | | | | | | | |gateway routers gateway| | | | | | | | | |-------+ +-------+ +-------+ +------| | | | | | | | | | PCN- | | PCN- | | PCN- | | | | | | | | |..| |..|marking|..|marking|..|marking|..| Meter|..| |..| | | | | | |-------+ +-------+ +-------+ +------| | | | | | | | | | \ / | | | | | | | | | | \ / | | | | | | | | | | \ Congestion-Level-Estimate / | | | | | | | | | | \ (for admission control) / | | | | | | | | | | --<-----<----<----<-----<-- | | | | | | | | | | Sustainable-Aggregate-Rate | | | | | | | | | | (for flow pre-emption) | | | | | |___| |___| |_______________________________________| |____| |___| Sx Access CL-region Access Rx End Network Network End Host Host <------ edge-to-edge signalling -----> (for admission control & flow pre-emption) <-------------------end-to-end QoS signalling protocol---------------> Figure 1: Overall QoS architecture (NB terminology explained later) Briscoe Expires April 25, 2007 [Page 5]
Internet-Draft Deployment Model using PCN October 2006 Figure 1 shows an example of an overall QoS architecture, where the two access networks are connected by a CL-region. Another possibility is that there are several CL-regions between the access networks - each would operate the Pre-Congestion Notification mechanisms separately. The document assumes RSVP as the end-to-end QoS signalling protocol. However, the RSVP signalling may itself be originated or terminated by proxies still closer to the edge of the network, such as home hubs or the like, triggered in turn by application layer signalling. [RFC2998] and our approach are compared further in Section 6.2. Flows must enter and leave the CL-region through its ingress and egress gateways, and they need traffic descriptors that are policed by the ingress gateway (NB the policing function is out of this document's scope). The overall CL-traffic between two border routers is called a "CL-region-aggregate". The document introduces a mechanism for flow admission control: should a new flow be admitted into a specific CL-region-aggregate? Admission control protects the QoS of existing CL-flows in normal circumstances. In abnormal circumstances, for instance a disaster affecting multiple interior routers, then the QoS on existing CL microflows may degrade even if care was exercised when admitting those microflows before those circumstances. Therefore we also propose a mechanism for flow pre-emption: how much traffic, in a specific CL-region-aggregate, should be pre-empted in order to preserve the QoS of the remaining CL-flows? Flow pre-emption also restores QoS to lower priority traffic. As a fundamental building block to enable these two mechanisms, each link of the CL-region is associated with a configured-admission-rate and configured-pre-emption-rate; the former is usually significantly larger than the latter. If traffic in a specific DiffServ class ("CL- traffic") on the link exceeds these rates then packets are marked with "Admission Marking" or "Pre-emption Marking". The algorithms that determine the number of packets marked are outlined in Section 3 and detailed in [PCN]. PCN marking (Pre-Congestion Notification) builds on the concepts of RFC 3168, "The addition of Explicit Congestion Notification to IP" (which is briefly summarised in Appendix A). Briscoe Expires April 25, 2007 [Page 6]
Internet-Draft Deployment Model using PCN October 2006 Traffic rate on link ^ | | Drop packets link bandwidth -|--------------------------- | | Pre-emption Mark packets configured-pre-emption-rate -|--------------------------- | | Admission Mark packets configured-admission-rate -|--------------------------- | | No marking of packets | +--------------------------- Figure 2: Packet Marking by Routers Gateways of the CL-region make measurements of packet rates and their PCN markings and convert them into decisions about whether to admit new flows, and (if necessary) into the rate of excess traffic that should be pre-empted. These mechanisms are detailed in Section 3 and briefly outlined in the next few paragraphs. The admission control mechanism for a new flow entering the network at ingress gateway G0 and leaving it at egress gateway G1 relies on feedback from the egress gateway G1 about the existing CL-region- aggregate between G0 and G1. This feedback is generated as follows. All routers meter the rate of the CL-traffic on their outgoing links and mark the packets with the Admission Mark if the configured- admission-rate is exceeded. Egress gateway G1 measures the Admission Marks for each of its CL-region-aggregates separately. If the fraction of traffic on a CL-region-aggregate that is Admission Marked exceeds some threshold, no further flows should be admitted into this CL-region-aggregate. Because sources vary their data rates (amongst other reasons) the rate of the CL-traffic on a link may fluctuate above and below the configured-admission-rate. Hence to get more stable information, the egress gateway measures the fraction as a moving average, called the Congestion-Level-Estimate. This is signalled from the egress G1 to the ingress G0, to enable the ingress to block new flows. Admission control seems most useful for DiffServ's Controlled load service. In order to support CL traffic we would expect PCN to supplement the existing scheduling behaviour Expedited Forwarding Briscoe Expires April 25, 2007 [Page 7]
Internet-Draft Deployment Model using PCN October 2006 (EF). Since PCN gives an "early warning" of potential congestion (hence "pre-congestion notification"), admission control can kick in before there is any significant build up of packets in routers - which is exactly the performance required for CL. However, PCN is not only intended to supplement EF. PCN is specified (in [PCN]) as a building block which can supplement the scheduling behaviour of other PHBs. The function to pre-empt flows (or allow the potential to pre-empt them) relies on feedback from the egress gateway about the CL-region- aggregates. This feedback is generated as follows. All routers meter the rate of the CL-traffic on their outgoing links, and if the rate is in excess of the configured-pre-emption-rate then packets amounting to the excess rate are Pre-emption Marked. If the egress gateway G1 sees a Pre-emption Marked packet then it measures, for this CL-region-aggregate, the rate of all received packets that aren't Pre-emption Marked. This is the rate of CL-traffic that the network can actually support from G0 to G1, and we thus call it the Sustainable-Aggregate-Rate. The ingress gateway G0 compares the Sustainable-Aggregate-Rate with the rate that it is sending towards G1, and hence determines the required traffic rate reduction. The document assumes flow pre-emption as the way of reacting to this information, ie stopping sufficient flows to reduce the rate to the Sustainable-Aggregate-Rate. However, this isn't mandated, for instance policy or regulation may prevent pre-emption of some flows - such considerations are out of scope of this document. 1.2. Key benefits We believe that the mechanisms described in this document are simple, scalable, and robust because: o Per flow state is only required at the ingress gateways to prevent non-admitted CL traffic from entering the PCN-region. Other network entities are not aware of individual flows. o For each of its links a router has Admission Marking and Pre- emption Marking behaviours. These markers operate on the overall CL traffic of the respective link. Therefore, there are no scalability concerns. o The information of these measurements is implicitly signalled to the egress gateways by the marks in the packet headers. No protocol actions (explicit messages) are required. Briscoe Expires April 25, 2007 [Page 8]
Internet-Draft Deployment Model using PCN October 2006 o The egress gateways make separate measurements for each ingress gateway of packets. Each meter operates on the overall CL traffic of a particular CL-region-aggregate. Therefore, there are no scalability concerns as long as the number of ingress gateways is not overwhelmingly large. o Feedback signalling is required between all pairs of ingress and egress gateways and the signalled information is on the basis of the corresponding CL-region-aggregate, i.e. it is also unaware of individual flows. o The configured-admission-rates can be chosen small enough that admitted traffic can still be carried after a rerouting in most failure cases. This is an important feature as QoS violations in core networks due to link failures are more likely than QoS violations due to increased traffic volume. o The admitted load is controlled dynamically. Therefore it adapts as the traffic matrix changes, and also if the network topology changes (eg after a link failure). Hence an operator can be less conservative when deploying network capacity, and less accurate in their prediction of the traffic matrix. Also, controlling the load using statically provisioned capacity per ingress (regardless of the egress of a flow), as is typical in the DiffServ architecture [RFC2475], can lead to focussed overload: many flows happen to focus on a particular link and then all flows through the congested link fail catastrophically (Section 6.2). o The pre-emption function complements admission control. It allows the network to recover from sudden unexpected surges of CL-traffic on some links, thus restoring QoS to the remaining flows. Such scenarios are very unlikely but not impossible. They can be caused by large network failures that redirect lots of admitted CL- traffic to other links, or by malfunction of the measurement-based admission control in the presence of admitted flows that send for a while with an atypically low rate and increase their rates in a correlated way. 1.3. Terminology EDITOR'S NOTE: Terminology in this document is (hopefully) consistent with that in [PCN]. However, it may not be consistent with the Briscoe Expires April 25, 2007 [Page 9]
Internet-Draft Deployment Model using PCN October 2006 terminology in other PCN-related documents. The PCN Working Group (if formed) will need to agree a single set of terminology. This terminology is copied from the pre-congestion notification marking draft [PCN]: o Pre-Congestion Notification (PCN): two new algorithms that determine when a PCN-enabled router Admission Marks and Pre- emption Marks a packet, depending on the traffic level. o Admission Marking condition: the traffic level is such that the router Admission Marks packets. The router provides an "early warning" that the load is nearing the engineered admission control capacity, before there is any significant build-up of CL packets in the queue. o Pre-emption Marking condition: the traffic level is such that the router Pre-emption Marks packets. The router warns explicitly that pre-emption may be needed. o Configured-admission-rate: the reference rate used by the admission marking algorithm in a PCN-enabled router. o Configured-pre-emption-rate - the reference rate used by the pre- emption marking algorithm in a PCN-enabled router. The following terms are defined here: o Ingress gateway: router at an ingress to the CL-region. A CL- region may have several ingress gateways. o Egress gateway: router at an egress from the CL-region. A CL- region may have several egress gateways. o Interior router: a router which is part of the CL-region, but isn't an ingress or egress gateway. o CL-region: A region of the Internet in which all traffic enters/leaves through an ingress/egress gateway and all routers run Pre-Congestion Notification marking. A CL-region is a DiffServ region (a DiffServ region is either a single DiffServ domain or set of contiguous DiffServ domains), but note that the CL-region does not use the traffic conditioning agreements (TCAs) of the (informational) DiffServ architecture. Briscoe Expires April 25, 2007 [Page 10]
Internet-Draft Deployment Model using PCN October 2006 o CL-region-aggregate: all the microflows between a specific pair of ingress and egress gateways. Note there is no field in the flow packet headers that uniquely identifies the aggregate. o Congestion-Level-Estimate: the number of bits in CL packets that are admission marked (or pre-emption marked), divided by the number of bits in all CL packets. It is calculated as an exponentially weighted moving average. It is calculated by an egress gateway for the CL packets from a particular ingress gateway, i.e. there is a Congestion-Level-Estimate for each CL- region-aggregate. o Sustainable-Aggregate-Rate: the rate of traffic that the network can actually support for a specific CL-region-aggregate. So it is measured by an egress gateway for the CL packets from a particular ingress gateway. o Ingress-Aggregate-Rate: the rate of traffic that is being sent on a specific CL-region-aggregate. So it is measured by an ingress gateway for the CL packets sent towards a particular egress gateway. 1.4. Existing terminology This is a placeholder for useful terminology that is defined elsewhere. 1.5. Standardisation requirements The framework described in this document has two new standardisation requirements: o new Pre-Congestion Notification for Admission Marking and Pre- emption Marking are required, as detailed in [PCN]. o the end-to-end signalling protocol needs to be modified to carry the Congestion-Level-Estimate report (for admission control) and the Sustainable-Aggregate-Rate (for flow pre-emption). With our assumption of RSVP (Section 2.2) as the end-to-end signalling protocol, it means that extensions to RSVP are required, as detailed in [RSVP-PCN], for example to carry the Congestion-Level- Estimate and Sustainable-Aggregate-Rate information from egress gateway to ingress gateway. Briscoe Expires April 25, 2007 [Page 11]
Internet-Draft Deployment Model using PCN October 2006 o We are discussing what to standardise about the gateway's behaviour. Other than these things, the arrangement uses existing IETF protocols throughout, although not in their usual architecture. 1.6. Structure of rest of the document Section 2 describes some key aspects of the deployment model: our goals and assumptions. Section 3 describes the deployment model, whilst Section 4 summarises the required changes to the various routers in the CL-region. Section 5 outlines some limitations of PCN that we've identified in this deployment model; it also discusses some potential solutions, and other possible extensions. Section 6 provides some comparison with existing QoS mechanisms. Briscoe Expires April 25, 2007 [Page 12]
Internet-Draft Deployment Model using PCN October 2006 2. Key aspects of the deployment model EDITOR'S NOTE: The material in Section 2 will eventually disappear, as it will be covered by the problem statement of the PCN Working Group (if formed). In this section we discuss the key aspects of the deployment model: o At a high level, our key goals, i.e. the functionality that we want to achieve o The assumptions that we're prepared to make 2.1. Key goals The deployment model achieves an end-to-end controlled load (CL) service where a segment of the end-to-end path is an edge-to-edge Pre-Congestion Notification region. CL is a quality of service (QoS) closely approximating the QoS that the same flow would receive from a lightly loaded network element [RFC2211]. It is useful for inelastic flows such as those for real-time media. o The CL service should be achieved despite varying load levels of other sorts of traffic, which may or may not be rate adaptive (i.e. responsive to packet drops or ECN marks). o The CL service should be supported for a variety of possible CL sources: Constant Bit Rate (CBR), Variable Bit Rate (VBR) and voice with silence suppression. VBR is the most challenging to support. o After a localised failure in the interior of the CL-region causing heavy congestion, the CL service should recover gracefully by pre- empting (dropping) some of the admitted CL microflows, whilst preserving as many of them as possible with their full CL QoS. Briscoe Expires April 25, 2007 [Page 13]
Internet-Draft Deployment Model using PCN October 2006 o It needs to be possible to complete flow pre-emption within 1-2 seconds. Operators will have varying requirements but, at least for voice, it has been estimated that after a few seconds then many affected users will start to hang up, making the flow pre- emption mechanism redundant and possibly even counter-productive. Until flow pre-emption kicks in, other applications using CL (e.g. video) and lower priority traffic (e.g. Assured Forwarding (AF)) could be receiving reduced service. Therefore an even faster flow pre-emption mechanism would be desirable (even if, in practice, operators have to add a deliberate pause to ride out a transient while the natural rate of call tear down or lower layer protection mechanisms kick in). o The CL service should support emergency services ([EMERG-RQTS], [EMERG-TEL]) as well as the Assured Service which is the IP implementation of the existing ITU-T/NATO/DoD telephone system architecture known as Multi-Level Pre-emption and Precedence [ITU.MLPP.1990] [ANSI.MLPP.Spec][ANSI.MLPP.Supplement], or MLPP. In particular, this involves admitting new flows that are part of high priority sessions even when admission control would reject new routine flows. Similarly, when having to choose which flows to pre-empt, this involves taking into account the priorities and properties of the sessions that flows are part of. 2.2. Key assumptions The framework does not try to deliver the above functionality in all scenarios. We make the following assumptions about the type of scenario to be solved. o Edge-to-edge: all the routers in the CL-region are upgraded with Pre-Congestion Notification, and all the ingress and egress gateways are upgraded to perform the measurement-based admission control and flow pre-emption. Note that although the upgrades required are edge-to-edge, the CL service is provided end-to-end. o Additional load: we assume that any additional load offered within the reaction time of the admission control mechanism doesn't move the CL-region directly from no congestion to overload. So it assumes there will always be an intermediate stage where some CL packets are Admission Marked, but they are still delivered without significant QoS degradation. We believe this is valid for core and backbone networks with typical call arrival patterns (given the reaction time is little more than one round trip time across the CL-region), but is unlikely to be valid in access networks where the granularity of an individual call becomes significant. Briscoe Expires April 25, 2007 [Page 14]
Internet-Draft Deployment Model using PCN October 2006 o Aggregation: we assume that in normal operations, there are many CL microflows within the CL-region, typically at least hundreds between any pair of ingress and egress gateways. The implication is that the solution is targeted at core and backbone networks and possibly parts of large access networks. o Trust: we assume that there is trust between all the routers in the CL-region. For example, this trust model is satisfied if one operator runs the whole of the CL-region. But we make no such assumptions about the end hosts, i.e. depending on the scenario they may be trusted or untrusted by the CL-region. o Signalling: we assume that the end-to-end signalling protocol is RSVP. Section 3 describes how the CL-region fits into such an end- to-end QoS scenario, whilst [RSVP-PCN] describes the extensions to RSVP that are required. o Separation: we assume that all routers within the CL-region are upgraded with the CL mechanism, so the requirements of [Floyd] are met because the CL-region is an enclosed environment. Also, an operator separates CL-traffic in the CL-region from outside traffic by administrative configuration of the ring of gateways around the region. Within the CL-region we assume that the CL- traffic is separated from non-CL traffic. o Routing: we assume that all packets between a pair of ingress and egress gateways follow the same path, or that they follow different paths but that the load balancing scheme is tuned in the CL-region to distribute load such that the different paths always receive comparable relative load. This ensures that the Congestion-Level-Estimate used in the admission control procedure (and which is computed taking into account packets travelling on all the paths) approximately reflects the status of the actual path that will be followed by the new microflow's packets. We are investigating ways of loosening the restrictions set by some of these assumptions, for instance: o Trust: to allow the CL-region to span multiple, non-trusting operators, using the technique of [Re-PCN] as mentioned in Section 5.7.2. Briscoe Expires April 25, 2007 [Page 15]
Internet-Draft Deployment Model using PCN October 2006 o Signalling: we believe that the solution could operate with another signalling protocol, such as the one produced by the NSIS working group. It could also work with application level signalling as suggested in [RT-ECN]. o Additional load: we believe that the assumption is valid for core and backbone networks, with an appropriate margin between the configured-admission-rate and the capacity for CL traffic. However, in principle a burst of admission requests can occur in a short time. We expect this to be a rare event under normal conditions, but it could happen e.g. due to a 'flash crowd'. If it does, then more flows may be admitted than should be, triggering the pre-emption mechanism. There are various ways an operator might try to alleviate this issue, which are discussed in the 'Flash crowds' section 5.5 later. o Separation: the assumption that CL traffic is separated from non- CL traffic implies that the CL traffic has its own PHB, not shared with other traffic. We are looking at whether it could share Expedited Forwarding's PHB, but supplemented with Pre-Congestion Notification. If this is possible, other PHBs (like Assured Forwarding) could be supplemented with the same new behaviours. This is similar to how RFC3168 ECN was defined to supplement any PHB. o Routing: we are looking in greater detail at the solution in the presence of Equal Cost Multi-Path routing and at suitable enhancements. See also the 'ECMP' section 5.1 later. Briscoe Expires April 25, 2007 [Page 16]
Internet-Draft Deployment Model using PCN October 2006 3. Deployment model 3.1. Admission control In this section we describe the admission control mechanism. We discuss the three pieces of the solution and then give an example of how they fit together in a use case: o the new Pre-Congestion Notification for Admission Marking used by all routers in the CL-region o how the measurements made support our admission control mechanism o how the edge to edge mechanism fits into the end to end RSVP signalling 3.1.1. Pre-Congestion Notification for Admission Marking This is discussed in [PCN]. Here we only give a brief outline. To support our admission control mechanism, each router in the CL- region runs an algorithm to determine whether to Admission Mark the packet. The algorithm measures the aggregate CL traffic on the link and ensures that packets are admission marked before the actual queue builds up, but when it is in danger of doing so soon; the probability of admission marking increases with the danger. The algorithm's main parameter is the configured-admission-rate, which is set lower than the link speed, perhaps considerably so. Admission marked packets indicate that the CL traffic rate is reaching the configured- admission-rate and so act as an "early warning" that the engineered capacity is nearly reached. Therefore they indicate that requests to admit prospective new CL flows may need to be refused. 3.1.2. Measurements to support admission control To support our admission control mechanism the egress measures the Congestion-Level-Estimate for traffic from each remote ingress gateway, i.e. per CL-region-aggregate. The Congestion-Level-Estimate is the number of bits in CL packets that are admission marked or pre- emption marked, divided by the number of bits in all CL packets. It is calculated as an exponentially weighted moving average. It is calculated by an egress gateway separately for the CL packets from each particular ingress gateway. Briscoe Expires April 25, 2007 [Page 17]
Internet-Draft Deployment Model using PCN October 2006 Why are pre-emption marked packets included in the Congestion-Level- Estimate? Pre-emption marking over-writes admission marking, i.e. a packet cannot be both admission and pre-emption marked. So if pre- emption marked packets weren't counted we would have the anomaly that as the traffic rate grew above the configured-pre-emption-rate, the Congestion-Level-Estimate would fall. If a particular encoding scheme is chosen where a packet can be both admission and pre-emption marked (such as Alternative 4 in Appendix C of [PCN]), then this is not necessary. This Congestion-Level-Estimate provides an estimate of how near the links on the path inside the CL-region are getting to the configured- admission-rate. Note that the metering is done separately per ingress gateway, because there may be sufficient capacity on all the routers on the path between one ingress gateway and a particular egress, but not from a second ingress to that same egress gateway. 3.1.3. How edge-to-edge admission control supports end-to-end QoS signalling Consider a scenario that consists of two end hosts, each connected to their own access networks, which are linked by the CL-region. A source tries to set up a new CL microflow by sending an RSVP PATH message, and the receiving end host replies with an RSVP RESV message. Outside the CL-region some other method, for instance IntServ, is used to provide QoS. From the perspective of RSVP the CL- region is a single hop, so the RSVP PATH and RESV messages are processed by the ingress and egress gateways but are carried transparently across all the interior routers; hence, the ingress and egress gateways hold per microflow state, whilst no per microflow state is kept by the interior routers. So far this is as in IntServ over DiffServ [RFC2998]. However, in order to support our admission control mechanism, the egress gateway adds to the RESV message an opaque object which states the current Congestion-Level-Estimate for the relevant CL-region-aggregate. Details of the corresponding RSVP extensions are described in [RSVP-PCN]. 3.1.4. Use case To see how the three pieces of the solution fit together, we imagine a scenario where some microflows are already in place between a given pair of ingress and egress gateways, but the traffic load is such that no packets from these flows are admission marked as they travel across the CL-region. A source wanting to start a new CL microflow sends an RSVP PATH message. The egress gateway adds an object to the RESV message with the Congestion-Level-Estimate, which is zero. The ingress gateway sees this and consequently admits the new flow. It Briscoe Expires April 25, 2007 [Page 18]
Internet-Draft Deployment Model using PCN October 2006 then forwards the RSVP RESV message upstream towards the source end host. Hence, assuming there's sufficient capacity in the access networks, the new microflow is admitted end-to-end. The source now sends CL packets, which arrive at the ingress gateway. The ingress uses a five-tuple filter to identify that the packets are part of a previously admitted CL microflow, and it also polices the microflow to ensure it remains within its traffic profile. (The ingress has learnt the required information from the RSVP messages.) When forwarding a packet belonging to an admitted microflow, the ingress sets the packet's DSCP and ECN fields to the appropriate values configured for the CL region. The CL packet now travels across the CL-region, getting admission marked if necessary. Next, we imagine the same scenario but at a later time when load is higher at one (or more) of the interior routers, which start to Admission Mark CL packets, because their load on the outgoing link is nearing the configured-admission-rate. The next time a source tries to set up a CL microflow, the ingress gateway learns (from the egress) the relevant Congestion-Level-Estimate. If it is greater than some CLE-threshold value then the ingress refuses the request, otherwise it is accepted. The ingress gateway could also take into account attributes of the RSVP reservation (such as for example the RSVP pre-emption priority of [RSVP-PREEMPTION] or the RSVP admission priority of [RSVP-EMERGENCY]) as well as information provided by a policy decision point in order to make a more sophisticated admission decision. This way, flow admission can help emergency/military calls by taking into account the corresponding priorities (as conveyed in RSVP policy elements) when deciding to admit or reject a new reservation. Use of RSVP for the support of emergency/military applications is discussed in further detail in [RFC4542] and [RSVP- EMERGENCY]. It is also possible for an egress gateway to get a RSVP RESV message and not know what the Congestion-Level-Estimate is. For example, if there are no CL microflows at present between the relevant ingress and egress gateways. In this case the egress requests the ingress to send probe packets, from which it can initialise its meter. RSVP Extensions for such a request to send probe data can be found in [RSVP-PCN]. Briscoe Expires April 25, 2007 [Page 19]
Internet-Draft Deployment Model using PCN October 2006 3.2. Flow pre-emption In this section we describe the flow pre-emption mechanism. We discuss the two parts of the solution and then give an example of how they fit together in a use case: o How an ingress gateway is triggered to test whether flow pre- emption may be needed o How an ingress gateway determines the right amount of CL traffic to drop The mechanism is defined in [PCN] and [RSVP-PCN]. Two subsequent steps could be: o Choose which flows to shed, influenced by their priority and other policy information o Tear down the reservations for the chosen flows We provide some hints about these latter two steps in Section 3.2.3, but don't try to provide full guidance as it greatly depends on the particular detailed operational situation. An essential QoS issue in core and backbone networks is being able to cope with failures of routers and links. The consequent re-routing can cause severe congestion on some links and hence degrade the QoS experienced by on-going microflows and other, lower priority traffic. Even when the network is engineered to sustain a single link failure, multiple link failures (e.g. due to a fibre cut, router failure or a natural disaster) can cause violation of capacity constraints and resulting QoS failures. Our solution uses rate-based flow pre- emption, so that sufficient of the previously admitted CL microflows are dropped to ensure that the remaining ones again receive QoS commensurate with the CL service and at least some QoS is quickly restored to other traffic classes. 3.2.1. Alerting an ingress gateway that flow pre-emption may be needed Alerting an ingress gateway that flow pre-emption may be needed is a two stage process: a router in the CL-region alerts an egress gateway that flow pre-emption may be needed; in turn the egress gateway alerts the relevant ingress gateway. Every router in the CL-region has the ability to alert egress gateways, which may be done either explicitly or implicitly: Briscoe Expires April 25, 2007 [Page 20]
Internet-Draft Deployment Model using PCN October 2006 o Explicit - the router per-hop behaviour is supplemented with a new Pre-emption Marking behaviour, which is outlined below. Reception of such a packet by the egress gateway alerts it that pre-emption may be needed. o Implicit - the router behaviour is unchanged from the Admission Marking behaviour described earlier. The egress gateway treats a Congestion-Level-Estimate of (almost) 100% as an implicit alert that pre-emption may be required. ('Almost' because the Congestion-Level-Estimate is a moving average, so can never reach exactly 100%.) To support explicit pre-emption alerting, each router in the CL- region runs an algorithm to determine whether to Pre-emption Mark the packet. The algorithm measures the aggregate CL traffic and ensures that packets are pre-emption marked before the actual queue builds up. The algorithm's main parameter is the configured-pre-emption- rate, which is set lower than the link speed (but higher than the configured-admission-rate). Thus pre-emption marked packets indicate that the CL traffic rate is reaching the configured-pre-emption-rate and so act as an "early warning" that the engineered capacity is nearly reached. Therefore they indicate that it may be advisable to pre-empt some of the existing CL flows in order to preserve the QoS of the others. Note that the pre-emption marking algorithm doesn't measure the packets that are already Pre-emption Marked. This ensures that in a scenario with several links that are above their configured-pre- emption-rate, then at the egress gateway the rate of packets excluding Pre-emption Marked ones truly does represent the Sustainable-Aggregate-Rate(see below for explanation). Note that the explicit mechanism only makes sense if all the routers in the CL-region have the functionality so that the egress gateways can rely on the explicit mechanism. Otherwise there is the danger that the traffic happens to focus on a router without it, and egress gateways then have also to watch for implicit pre-emption alerts. When one or more packets in a CL-region-aggregate alert the egress gateway of the need for flow pre-emption, whether explicitly or implicitly, the egress puts that CL-region-aggregate into the Pre- emption Alert state. For each CL-region-aggregate in alert state it measures the rate of traffic at the egress gateway (i.e. the traffic rate of the appropriate CL-region-aggregate) and reports this to the relevant ingress gateway. The steps are: Briscoe Expires April 25, 2007 [Page 21]
Internet-Draft Deployment Model using PCN October 2006 o Determine the relevant ingress gateway - for the explicit case the egress gateway examines the pre-emption marked packet and uses the state installed at the time of admission to determine which ingress gateway the packet came from. For the implicit case the egress gateway has already determined this information, because the Congestion-Level-Estimate is calculated per ingress gateway. o Measure the traffic rate of CL packets - as soon as the egress gateway is alerted (whether explicitly or implicitly) it measures the rate of CL traffic from this ingress gateway (i.e. for this CL-region-aggregate). Note that pre-emption marked packets are excluded from that measurement. It should make its measurement quickly and accurately, but exactly how is up to the implementation. o Alert the ingress gateway - the egress gateway then immediately alerts the relevant ingress gateway about the fact that flow pre- emption may be required. This Alert message also includes the measured Sustainable-Aggregate-Rate, i.e. the rate of CL-traffic received from this ingress gateway. The Alert message is sent using reliable delivery. Procedures for the support of such an Alert using RSVP are defined in [RSVP-PCN]. -------------- _ _ ----------------- CL packet |Update | / Is it a \ Y | Measure CL rate | arrives --->|Congestion- |--->/pre-emption\-----> | from ingress and| |Level-Estimate| \ marked / | alert ingress | -------------- \ packet? / ----------------- \_ _/ Figure 2: Egress gateway action for explicit Pre-emption Alert _ _ -------------- / \ ----------------- CL packet |Update | / Is \ Y | Measure CL rate | arrives --->|Congestion- |--->/ C.L.E. \-----> | from ingress and| |Level-Estimate| \ (nearly) / | alert ingress | -------------- \ 100%? / ----------------- \_ _/ Figure 3: Egress gateway action for implicit Pre-emption Alert Briscoe Expires April 25, 2007 [Page 22]
Internet-Draft Deployment Model using PCN October 2006 3.2.2. Determining the right amount of CL traffic to drop The method relies on the insight that the amount of CL traffic that can be supported between a particular pair of ingress and egress gateways, is the amount of CL traffic that is actually getting across the CL-region to the egress gateway without being Pre-emption Marked. Hence we term it the Sustainable-Aggregate-Rate. So when the ingress gateway gets the Alert message from an egress gateway, it compares: o The traffic rate that it is sending to this particular egress gateway (which we term Ingress-Aggregate-Rate) o The traffic rate that the egress gateway reports (in the Alert message) that it is receiving from this ingress gateway (which is the Sustainable-Aggregate-Rate) If the difference is significant, then the ingress gateway pre-empts some microflows. It only pre-empts if: Ingress-Aggregate-Rate > Sustainable-Aggregate-Rate + error The "error" term is partly to allow for inaccuracies in the measurements of the rates. It is also needed because the Ingress- Aggregate-Rate is measured at a slightly later moment than the Sustainable-Aggregate-Rate, and it is quite possible that the Ingress-Aggregate-Rate has increased in the interim due to natural variation of the bit rate of the CL sources. So the "error" term allows for some variation in the ingress rate without triggering pre- emption. The ingress gateway should pre-empt enough microflows to ensure that: New Ingress-Aggregate-Rate < Sustainable-Aggregate-Rate - error The "error" term here is used for similar reasons but in the other direction, to ensure slightly more load is shed than seems necessary, in case the two measurements were taken during a short-term fall in load. When the routers in the CL-region are using explicit pre-emption alerting, the ingress gateway would normally pre-empt microflows whenever it gets an alert (it always would if it were possible to set "error" equal to zero). For the implicit case however this is not so. It receives an Alert message when the Congestion-Level-Estimate reaches (almost) 100%, which is roughly when traffic exceeds the Briscoe Expires April 25, 2007 [Page 23]
Internet-Draft Deployment Model using PCN October 2006 configured-admission-rate. However, it is only when packets are indeed dropped en route that the Sustainable-Aggregate-Rate becomes less than the Ingress-Aggregate-Rate so only then will pre-emption actually occur on the ingress gateway. Hence with the implicit scheme, pre-emption can only be triggered once the system starts dropping packets and thus the QoS of flows starts being significantly degraded. This is in contrast with the explicit scheme which allows flow pre-emption to be triggered before any packet drop, simply when the traffic reaches the configured-pre- emption-rate. Therefore we believe that the explicit mechanism is superior. However it does require new functionality on all the routers (although this is little more than a bulk token bucket - see [PCN] for details). 3.2.3. Use case for flow pre-emption To see how the pieces of the solution fit together in a use case, we imagine a scenario where many microflows have already been admitted. We confine our description to the explicit pre-emption mechanism. Now an interior router in the CL-region fails. The network layer routing protocol re-routes round the problem, but as a consequence traffic on other links increases. In fact let's assume the traffic on one link now exceeds its configured-pre-emption-rate and so the router pre- emption marks CL packets. When the egress sees the first one of the pre-emption marked packets it immediately determines which microflow this packet is part of (by using a five-tuple filter and comparing it with state installed at admission) and hence which ingress gateway the packet came from. It sets up a meter to measure the traffic rate from this ingress gateway, and as soon as possible sends a message to the ingress gateway. This message alerts the ingress gateway that pre-emption may be needed and contains the traffic rate measured by the egress gateway. Then the ingress gateway determines the traffic rate that it is sending towards this egress gateway and hence it can calculate the amount of traffic that needs to be pre-empted. The solution operates within a little over one round trip time - the time required for microflow packets that have experienced Pre-emption Marking to travel downstream through the CL-region and arrive at the egress gateway, plus some additional time for the egress gateway to measure the rate seen after it has been alerted that pre-emption may be needed, and the time for the egress gateway to report this information to the ingress gateway. Briscoe Expires April 25, 2007 [Page 24]
Internet-Draft Deployment Model using PCN October 2006 The ingress gateway could now just shed random microflows, but it is better if the least important ones are dropped. The ingress gateway could use information stored locally in each reservation's state (such as for example the RSVP pre-emption priority of [RSVP- PREEMPTION] or the RSVP admission priority of [RSVP-EMERGENCY]) as well as information provided by a policy decision point in order to decide which of the flows to shed (or perhaps which ones not to shed). This way, flow pre-emption can also helps emergency/military calls by taking into account the corresponding priorities (as conveyed in RSVP policy elements) when selecting calls to be pre- empted, which is likely to be particularly important in a disaster scenario. Use of RSVP for support of emergency/military applications is discussed in further details in [RFC4542] and [RSVP-EMERGENCY]. The ingress gateway then initiates RSVP signalling to instruct the relevant destinations that their reservation has been terminated, and to tell (RSVP) nodes along the path to tear down associated RSVP state. To guard against recalcitrant sources, normal IntServ policing may be used to block any future traffic from the dropped flows from entering the CL-region. Note that - with the explicit Pre-emption Alert mechanism - since the configured-pre-emption-rate may be significantly less than the physical line capacity, flow pre-emption may be triggered before any congestion has actually occurred and before any packet is dropped. We extend the scenario further by imagining that (due to a disaster of some kind) further routers in the CL-region fail during the time taken by the pre-emption process described above. This is handled naturally, as packets will continue to be pre-emption marked and so the pre-emption process will happen for a second time. 3.3. Both admission control and pre-emption This document describes both the admission control and pre-emption mechanisms, and we suggest that an operator uses both. However, we do not require this and some operators may want to implement only one. For example, an operator could use just admission control, solving heavy congestion (caused by re-routing) by 'just waiting' - as sessions end, existing microflows naturally depart from the system over time, and the admission control mechanism will prevent admission of new microflows that use the affected links. So the CL-region will naturally return to normal controlled load service, but with reduced capacity. The drawback of this approach would be that until flows naturally depart to relieve the congestion, all flows and lower Briscoe Expires April 25, 2007 [Page 25]
Internet-Draft Deployment Model using PCN October 2006 priority services will be adversely affected. As another example, an operator could use just admission control, avoiding heavy congestion (caused by re-routing) by 'capacity planning' - by configuring admission control thresholds to lower levels than the network could accept in normal situations such that the load after failure is expected to stay below acceptable levels even with reduced network resources. On the other hand, an operator could just rely for admission control on the traffic conditioning agreements of the DiffServ architecture [RFC2475]. The pre-emption mechanism described in this document would be used to counteract the problem described at the end of Section 1.1.1. Briscoe Expires April 25, 2007 [Page 26]
Internet-Draft Deployment Model using PCN October 2006 4. Summary of Functionality This section is intended to provide a systematic summary of the new functionality required by the routers in the CL-region. A network operator upgrades normal IP routers by: o Adding functionality related to admission control and flow pre- emption to all its ingress and egress gateways o Adding Pre-Congestion Notification for Admission Marking and Pre- emption Marking to all the routers in the CL-region. We consider the detailed actions required for each of the types of router in turn. 4.1. Ingress gateways Ingress gateways perform the following tasks: o Classify incoming packets - decide whether they are CL or non-CL packets. This is done using an IntServ filter spec (source and destination addresses and port numbers), whose details have been gathered from the RSVP messaging. o Police - check that the microflow conforms with what has been agreed (i.e. it keeps to its agreed data rate). If necessary, packets which do not correspond to any reservations, packets which are in excess of the rate agreed for their reservation, and packets for a reservation that has earlier been pre-empted may be policed. Policing may be achieved via dropping or via re-marking of the packet's DSCP to a value different from the CL behaviour aggregate. o ECN colouring packets - for CL microflows, set the ECN field of packets appropriately (see [PCN] for some discussion of encoding). o Perform 'interior router' functions (see next sub-section). o Admission Control - on new session establishment, consider the Congestion-Level-Estimate received from the corresponding egress gateway and most likely based on a simple configured CLE-threshold decide if a new call is to be admitted or rejected (taking into account local policy information as well as optionally information provided by a policy decision point). Briscoe Expires April 25, 2007 [Page 27]
Internet-Draft Deployment Model using PCN October 2006 o Probe - if requested by the egress gateway to do so, the ingress gateway generates probe traffic so that the egress gateway can compute the Congestion-Level-Estimate from this ingress gateway. Probe packets may be simple data addressed to the egress gateway and require no protocol standardisation, although there will be best practice for their number, size and rate. o Measure - when it receives a Pre-emption Alert message from an egress gateway, it determines the rate at which it is sending packets to that egress gateway o Pre-empt - calculate how much CL traffic needs to be pre-empted; decide which microflows should be dropped, perhaps in consultation with a Policy Decision Point; and do the necessary signalling to drop them. 4.2. Interior routers Interior routers do the following tasks: o Classify packets - examine the DSCP and ECN field to see if it's a CL packet o Non-CL packets are handled as usual, with respect to dropping them or setting their CE codepoint. o Pre-Congestion Notification - CL packets are Admission Marked and Pre-emption Marked according to the algorithm detailed in [PCN] and outlined in Section 3. 4.3. Egress gateways Egress gateways do the following tasks: o Classify packets - determine which ingress gateway a CL packet has come from. This is the previous RSVP hop, hence the necessary details are obtained just as with IntServ from the state associated with the packet five-tuple, which has been built using information from the RSVP messages. Briscoe Expires April 25, 2007 [Page 28]
Internet-Draft Deployment Model using PCN October 2006 o Meter - for CL packets, calculate the fraction of the total number of bits which are in Admission marked packets or in Pre-emption Marked packets. The calculation is done as an exponentially weighted moving average (see Appendix C). A separate calculation is made for CL packets from each ingress gateway. The meter works on an aggregate basis and not per microflow. o Signal the Congestion-Level-Estimate - this is piggy-backed on the reservation reply. An egress gateway's interface is configured to know it is an egress gateway, so it always appends this to the RESV message. If the Congestion-Level-Estimate is unknown or is too stale, then the egress gateway can request the ingress gateway to send probes. o Packet colouring - for CL packets, set the DSCP and the ECN field to whatever has been agreed as appropriate for the next domain. By default the ECN field is set to the Not-ECT codepoint. See also the discussion in the Tunnelling section later. o Measure the rate - measure the rate of CL traffic from a particular ingress gateway, excluding packets that are Pre-emption Marked (i.e. the Sustainable-Aggregate-Rate for the CL-region- aggregate), when alerted (either explicitly or implicitly) that pre-emption may be required. The measured rate is reported back to the appropriate ingress gateway [RSVP-PCN]. 4.4. Failures If an interior router fails, then the regular IP routing protocol will re-route round it. If the new route can carry all the admitted traffic, flows will gracefully continue. If instead this causes early warning of pre-congestion on the new route, then admission control based on pre-congestion notification will ensure new flows will not be admitted until enough existing flows have departed. Finally re- routing may result in heavy congestion, when the flow pre-emption mechanism will kick in. If a gateway fails then we would like regular RSVP procedures [RFC2205] to take care of things. With the local repair mechanism of [RFC2205], when a route changes the next RSVP PATH refresh message will establish path state along the new route, and thus attempt to re-establish reservations through the new ingress gateway. Essentially the same procedure is used as described earlier in this document, with the re-routed session treated as a new session request. Briscoe Expires April 25, 2007 [Page 29]
Internet-Draft Deployment Model using PCN October 2006 In more detail, consider what happens if an ingress gateway of the CL-region fails. Then RSVP routers upstream of it do IP re-routing to a new ingress gateway. The next time the upstream RSVP router sends a PATH refresh message it reaches the new ingress gateway which therefore installs the associated RSVP state. The next RSVP RESV refresh will pick up the Congestion-Level-Estimate from the egress gateway, and the ingress compares this with its threshold to decide whether to admit the new session. This could result in some of the flows being rejected, but those accepted will receive the full QoS. An issue with this is that we have to wait until a PATH and RESV refresh messages are sent - which may not be very often - the default value is 30 seconds. [RFC2205] discusses how to speed up the local repair mechanism. First, the RSVP module is notified by the local routing protocol module of a route change to particular destinations, which triggers it to rapidly send out PATH refresh messages. Further, when a PATH refresh arrives with a previous hop address different from the one stored, then RESV refreshes are immediately sent to that previous hop. Where RSVP is operating hop-by-hop, i.e. on every router, then triggering the PATH refresh is easy as the router can simply monitor its local link. Thus, this fast local repair mechanism can be used to deal with failures upstream of the ingress gateway, with failures of the ingress gateway and with failures downstream of the egress gateway. But where RSVP is not operating hop-by-hop (as is the case within the CL-region), it is not so easy to trigger the PATH refresh. Unfortunately, this problem applies if an egress gateway fails, since it's very likely that an egress gateway is several IP hops from the ingress gateway. (If the ingress is several IP hops from its previous RSVP node, then there is the same issue.) The options appear to be: o the ingress gateway has a link state database for the CL-region, so it can detect that an egress gateway has failed or became unreachable o there is an inter-gateway protocol, so the ingress can continuously check that the egress gateways are still alive o (default) do nothing and wait for the regular PATH/RESV refreshes (and, if needed, the pre-emption mechanism) to sort things out. Briscoe Expires April 25, 2007 [Page 30]
Internet-Draft Deployment Model using PCN October 2006 5. Limitations and some potential solutions In this section we describe various limitations of the deployment model, and some suggestions about potential ways of alleviating them. The limitations fall into three broad categories: o ECMP (Section 5.1): the assumption about routing (Section 2.2) is that all packets between a pair of ingress and egress gateways follow the same path; ECMP breaks this assumption. A study regarding the accuracy of load balancing schemes can be found in [LoadBalancing-a] and [LoadBalancing-b]. o The lack of global coordination (Sections 5.2, 5.3 and 5.4): a decision about admission control or flow pre-emption is made for one aggregate independently of other aggregates o Timing and accuracy of measurements (Sections 5.5 and 5.6): the assumption (Section 2.2) that additional load, offered within the reaction time of the measurement-based admission control mechanism, doesn't move the system directly from no congestion to overload (dropping packets). A 'flash crowd' may break this assumption (Section 5.5). There are a variety of more general issues associated with marking measurements, which may mean it's a good idea to do pre-emption 'slower' (Section 5.6). Each section describes a limitation and some possible solutions to alleviate the limitation. These are intended as options for an operator to consider, based on their particular requirements. We would welcome feedback, for example suggestions as to which potential solutions are worth working out in more detail, and ideas on new potential solutions. Finally Section 5.7 considers some other potential extensions. 5.1. ECMP If the CL-region uses Equal Cost Multipath Routing (ECMP), then traffic between a particular pair of ingress and egress gateways may follow several different paths. Why? An ECMP-enabled router runs an algorithm to choose between potential outgoing links, based on a hash of fields such as the packet's source and destination addresses - exactly what depends on the proprietary algorithm. Packets are addressed to the CL flow's Briscoe Expires April 25, 2007 [Page 31]
Internet-Draft Deployment Model using PCN October 2006 end-point, and therefore different flows may follow different paths through the CL-region. (All packets of an individual flow follow the same ECMP path.) The problem is that if one of the paths is congested such that packets are being admission marked, then the Congestion-Level- Estimate measured by the egress gateway will be diluted by unmarked packets from other non-congested paths. Similarly, the measurement of the Sustainable-Aggregate-Rate will also be diluted. Possible solution approaches are: o tunnel: traffic is tunnelled across the CL-region. Then the destination address (and so on) seen by the ECMP algorithm is that of the egress gateway, so all flows follow the same path. Effectively ECMP is turned off. As a compromise, to try to retain some of the benefits of ECMP, there could be several tunnels, each following a different ECMP path, with flows randomly assigned to different tunnels. o assume worst case: the operator sets the configured-admission-rate (and configured-pre-emption-rate) to below the optimum level to compensate for the fact that the effect on the Congestion-Level- Estimate (and Sustainable-Aggregate-Rate) of the congestion experienced over one of the paths may be diluted by traffic received over non-congested paths. Hence lower thresholds need to be used to ensure early admission control rejection and pre- emption over the congested path. This approach will waste capacity (e.g. flows following a non-congested ECMP path are not admitted or are pre-empted), and there is still the danger that for some traffic mixes the operator hasn't been cautious enough. o for admission control, probe to obtain a flow-specific congestion- level-estimate. Earlier this document suggests continuously monitoring the congestion-level-estimate. Instead, probe packets could be sent for each prospective new flow. The probe packets have the same IP address etc as the data packets would have, and hence follow the same ECMP path. However, probing is an extra overhead, depending on how many probe packets need to be sent to get a sufficiently accurate congestion-level-estimate. Probes also cause a processing overhead, either for the machine at the destination address or for the egress gateway to identify and remove the probe packets. Briscoe Expires April 25, 2007 [Page 32]
Internet-Draft Deployment Model using PCN October 2006 o for flow pre-emption, only select flows for pre-emption from amongst those that have actually received a Pre-emption Marked packet. Because these flows must have followed an ECMP path that goes through an overloaded router. However, it needs some extra work by the egress gateway, to record this information and report it to the ingress gateway. o for flow pre-emption, a variant of this idea involves introducing a new marking behaviour, 'Router Marking'. A router that is pre- emption marking packets on an outgoing link, also 'Router Marks' all other packets. When selecting flows for pre-emption, the selection is made from amongst those that have actually received a Router Marked or Pre-emption Marked packet. Hence compared with the previous bullet, it may extend the range of flows from which the pre-emption selection is made (i.e. it includes those which, by chance, haven't had any pre-emption marked packets). However, it also requires that the 'Router Marking' state is somehow encoded into a packet, i.e. it makes harder the encoding challenge discussed in Appendix C of [PCN]. The extra work required by the egress gateway would also be somewhat higher than for the previous bullet. 5.2. Beat down effect This limitation concerns the pre-emption mechanism in the case where more than one router is pre-emption marking packets. The result (explained in the next paragraph) is that the measurement of sustainable-aggregate-rate is lower than its true value, so more traffic is pre-empted than necessary. Imagine the scenario: +-------+ +-------+ +-------+ IAR-b=3 >@@@@@| CPR=2 |@@@@@| CPR>2 |@@@@@| CPR=1 |@@> SAR-b=1 IAR-a=1 >#####| R1 |#####| R2 | | R3 | +-------+ +-------+ +-------+ # # # v SAR-a=0.5 Figure 4: Scenario to illustrate 'beat down effect' limitation Aggregate-a (ingress-aggregate-rate, IAR, 1 unit) takes a 'short' route through two routers, one of which (R1) is above its configured- Briscoe Expires April 25, 2007 [Page 33]
Internet-Draft Deployment Model using PCN October 2006 pre-emption-rate (CPR, 2 units). Aggregate-b takes a 'long' route, going through a second congested router (R3, with a CPR of 1 unit). R1's input traffic is 4 units, twice its configured-pre-emption-rate, so 50% of packets are pre-emption marked. Hence the measured sustainable-aggregate-rate (SAR) for aggregate-a is 0.5, and half of its traffic will be pre-empted. R3's input of non-pre-emption-marked traffic is 1.5 units, and therefore it has to do further marking. But this means that aggregate-a has taken a bigger hit than it needed to; the router R1 could have let through all of aggregate-a's traffic unmarked if it had known that the second router R2 was going to "beat down" aggregate-b's traffic further. Generalising, the result is that in a scenario where more than one router is pre-emption marking packets, only the final router is sure to be fully loaded after flow pre-emption. The fundamental reason is that a router makes a local decision about which packets to pre- emption mark, i.e. independently of how other routers are pre-emption marking. A very similar effect has been noted in XCP [Low]. Potential solutions: o a full solution would involve routers learning about other routers that are pre-emption marking, and being able to differentially mark flows (e.g. in the example above, aggregate-a's packets wouldn't be marked by R1). This seems hard and complex. o do nothing about this limitation. It causes over-pre-emption, which is safe. At the moment this is our suggested option. Briscoe Expires April 25, 2007 [Page 34]
Internet-Draft Deployment Model using PCN October 2006 o do pre-emption 'slowly'. The description earlier in this document assumes that after the measurements of ingress-aggregate-rate and sustainable-aggregate-rate, then sufficient flows are pre-empted in 'one shot' to eliminate the excess traffic. An alternative is to spread pre-emption over several rounds: initially, only pre- empt enough to eliminate some of the excess traffic, then re- measure the sustainable-aggregate-rate, and then pre-empt some more, etc. In the scenario above, the re-measurement would be lower than expected, due to the beat down effect, and hence in the second round of pre-emption less of aggregate-a's traffic would be pre-empted (perhaps none). Overall, therefore the impact of the 'beat down' effect would be lessened, i.e. there would be a smaller degree of over pre-emption. The downside is that the overall pre-emption is slower, and therefore routers will be congested longer. 5.3. Bi-directional sessions The document earlier describes how to decide whether or not to admit (or pre-empt) a particular flow. However, from a user/application perspective, the session is the relevant unit of granularity. A session can consist of several flows which may not all be part of the same aggregate. The most obvious example is a bi-directional session, where the two flows should ideally be admitted or pre-empted as a pair - for instance a voice call only makes sense if A can send to B as well as B to A! But the admission and pre-emption mechanisms described earlier in this document operate on a per-aggregate basis, independently of what's happening with other aggregates. For admission control the problem isn't serious: e.g. the SIP server for the voice call can easily detect that the A-to-B flow has been admitted but the B-to-A flow blocked, and inform the user perhaps via a busy tone. For flow pre-emption, the problem is similar but more serious. If both the aggregate-1-to-2 (i.e. from gateway 1 to gateway 2) and the aggregate-2-to-1 have to pre-empt flows, then it would be good if either all of the flows of a particular session were pre- empted or none of them. Therefore if the two aggregates pre-empt flows independently of each other, more sessions will end up being torn down than is really necessary. For instance, pre-empting one direction of a voice call will result in the SIP server tearing down the other direction anyway. Briscoe Expires April 25, 2007 [Page 35]
Internet-Draft Deployment Model using PCN October 2006 Potential solutions: o if it's known that all session are bi-directional, simply pre- empting roughly half as many flows as suggested by the measurements of {ingress-aggregate-rate - sustainable-aggregate- rate}. But this makes a big assumption about the nature of sessions, and also that the aggregate-1-to-2 and aggregate-2-to-1 are equally overloaded. o ignore the limitation. The penalty will be quite small if most sessions consist of one flow or of flows part of the same aggregate. o introduce a gateway controller. It would receive reports for all aggregates where the ingress-aggregate-rate exceeds the sustainable-aggregate-rate. It then would make a global decision about which flows to pre-empt. However it requires quite some complexity, for example the controller needs to understand which flows map to which sessions. This may be an option in some scenarios, for example where gateways aren't handling too many flows (but note that this breaks the aggregation assumption of Section 2.2). A variant of this idea would be to introduce a gateway controller per pair of gateways, in order to handle bi- directional sessions but not try to deal with more complex sessions that include flows from an arbitrary number of aggregates. o do pre-emption 'slowly'. As in the "beat down" solution 4, this would reduce the impact of this limitation. The downside is that the overall pre-emption is slower, and therefore router(s) will be congested longer. Briscoe Expires April 25, 2007 [Page 36]
Internet-Draft Deployment Model using PCN October 2006 o each ingress gateway 'loosely coordinates' with other gateways its decision about which specific flows to pre-empt. Each gateway numbers flows in the order they arrive (note that this number has no meaning outside the gateway), and when pre-empting flows, the most recent (or most recent low priority flow) is selected for pre-emption; the gateway then works backwards selecting as many flows as needed. Gateways will therefore tend to pre-empt flows that are part of the same session (as they were admitted at the same time). Of course this isn't guaranteed for several reasons, for instance gateway A's most recent bi-directional sessions may be with gateway C, whereas gateway B's are with gateway A (so gateway A will pre-empt A-to-C flows and gateway B will pre-empt B-to-A flows). Rather than pre-empting the most recent (low priority) flow, an alternative algorithm (for further study) may be to select flows based on a hash of particular fields in the packet, such that both gateways produce the same hash for flows of the same bi-directional session. We believe that this approach should be investigated further. 5.4. Global fairness The limitation here is that 'high priority' traffic may be pre-empted (or not admitted) when a global decision would instead pre-empt (or not admit) 'lower priority' traffic on a different aggregate. Imagine the following scenario (extreme to illustrate the point clearly). Aggregate_a is all Assured Services (MLPP) traffic, whilst aggregate_b is all ordinary traffic (i.e. comparatively low priority). Together the two aggregates cause a router to be at twice its configured-pre-emption-rate. Ideally we'd like all of aggregate_b to be pre-empted, as then all of aggregate_a could be carried. However, the approach described earlier in this document leads to half of each aggregate being pre-empted. Briscoe Expires April 25, 2007 [Page 37]
Internet-Draft Deployment Model using PCN October 2006 IAR_b=1 v v +-------+ IAR_a=1 ---->-----| CPR=1 |-----> SAR_a=0.5 | | +-------+ v v SAR_a=0.5 Figure 5: Scenario to illustrate 'global fairness' limitation Similarly, for admission control - Section 4.1 describes how if the Congestion-Level-Estimate is greater than the CLE-threshold all new sessions are refused. But it is unsatisfactory to block emergency calls, for instance. Potential solutions: o in the admission control case, it is recommended that an 'emergency / Assured Services' call is admitted immediately even if the CLE-threshold is exceeded. Usually the network can actually handle the additional microflow, because there is a safety margin between the configured-admission-rate and the configured-pre- emption-rate. Normal call termination behaviour will soon bring the traffic level down below the configured-admission-rate. However, in exceptional circumstances the 'emergency / higher precedence' call may cause the traffic level to exceed the configured-pre-emption-rate; then the usual pre-emption mechanism will pre-empt enough (non 'emergency / higher precedence') microflows to bring the total traffic back under the configured- pre-emption-rate. o all egress gateways report to a global coordinator that makes decisions about what flows to pre-empt. However this solution adds complexity and probably isn't scalable, but it may be an option in some scenarios, for example where gateways aren't handling too many flows (but note that this breaks the aggregation assumption of Section 2.2). Briscoe Expires April 25, 2007 [Page 38]
Internet-Draft Deployment Model using PCN October 2006 o introduce a heuristic rule: before pre-empting a 'high priority' flow the egress gateway should wait to see if sufficient (lower priority) traffic is pre-empted on other aggregates. This is a reasonable option. o enhance the functionality of all the interior routers, so they can detect the priority of a packet, and then differentially mark them. As well as adding complexity, in general this would be an unacceptable security risk for MLPP traffic, since only controlled nodes (like gateways) should know which packets are high priority, as this information can be abused by an attacker. o do nothing, i.e. accept the limitation. Whilst it's unlikely that high priority calls will be quite so unbalanced as in the scenario above, just accepting this limitation may be risky. The sorts of situations that cause routers to start pre-emption marking are also likely to cause a surge of emergency / MLPP calls. 5.5. Flash crowds This limitation concerns admission control and arises because there is a time lag between the admission control decision (which depends on the Congestion-Level-Estimate during RSVP signalling during call set-up) and when the data is actually sent (after the called party has answered). In PSTN terms this is the time the phone rings. Normally the time lag doesn't matter much because (1) in the CL- region there are many flows and they terminate and are answered at roughly the same rate, and (2) the network can still operate safely when the traffic level is some margin above the configured-admission- rate. A 'flash crowd' occurs when something causes many calls to be initiated in a short period of time - for instance a 'tele-vote'. So there is a danger that a 'flash' of calls is accepted, but when the calls are answered and data flows the traffic overloads the network. Therefore potentially the 'additional load' assumption of Section 2.2 doesn't hold. Potential solutions: Briscoe Expires April 25, 2007 [Page 39]
Internet-Draft Deployment Model using PCN October 2006 o The simplest option is to do nothing; an operator relies on the pre-emption mechanism if there is a problem. This doesn't seem a good choice, as 'flash crowds' are reasonably common on the PSTN, unless the operator can ensure that nearly all 'flash crowd' events are blocked in the access network and so do not impact on the CL-region. o A second option is to send 'dummy data' as soon as the call is admitted, thus effectively reserving the bandwidth whilst waiting for the called party to answer. Reserving bandwidth in advance means that the network cannot admit as many calls. For example, suppose sessions last 100 seconds and ringing for 10 seconds, the cost is a 10% loss of capacity. It may be possible to offset this somewhat by increasing the configured-admission-rate in the routers, but it would need further investigation. A concern with this 'dummy data' option is that it may allow an attacker to initiate many calls that are never answered (by a cooperating attacker), so eventually the network would only be carrying 'dummy data'. The attack exploits that charging only starts when the call is answered and not when it is dialled. It may be possible to alleviate the attack at the session layer - for example, when the ingress gateway gets an RSVP PATH message it checks that the source has been well-behaved recently; and limiting the maximum time that ringing can last. We believe that if this attack can be dealt with then this is a good option. o A third option is that the egress gateway limits the rate at which it sends out the Congestion-Level-Estimate, or limits the rate at which calls are accepted by replying with a Congestion-Level- Estimate of 100% (this is the equivalent of 'call gapping' in the PSTN). There is a trade-off, which would need to be investigated further, between the degree of protection and possible adverse side-effects like slowing down call set-up. o A final option is to re-perform admission control before the call is answered. The ingress gateway monitors Congestion-Level- Estimate updates received from each egress. If it notices that a Congestion-Level-Estimate has risen above the CLE-threshold, then it terminates all unanswered calls through that egress (e.g. by instructing the session protocol to stop the 'ringing tone'). For extra safety the Congestion-Level-Estimate could be re-checked when the call is answered. A potential drawback for an operator that wants to emulate the PSTN is that the PSTN network never drops a 'ringing' PSTN call. Briscoe Expires April 25, 2007 [Page 40]
Internet-Draft Deployment Model using PCN October 2006 5.6. Pre-empting too fast As a general idea it seems good to pre-empt excess flows rapidly, so that the full QoS is restored to the remaining CL users as soon as possible, and partial service is restored to lower priority traffic classes on shared links. Therefore the pre-emption mechanism described earlier in this document works in 'one shot', i.e. one measurement is made of the sustainable-aggregate-rate and the ingress-aggregate-rate, and the excess is pre-empted immediately. However, there are some reasons why an operator may potentially want to pre-empt 'more slowly': o To allow time to modify the ingress gateway's policer, as the ingress wants to be able to drop any packets that arrive from a pre-empted flow. There will be a limit on how many new filters an ingress gateway can install in a certain time period. Otherwise the source may cheat and ignore the instruction to drop its flow. o The operator may decide to slow down pre-emption in order to ameliorate the 'beat down' and/or 'bi-directional sessions' limitations (see above) o To help combat inaccuracies in measurements of the sustainable- aggregate-rate and ingress-aggregate-rate. For a CL-region where it's assumed there are many flows in an aggregate these measurements can be obtained in a short period of time, but where there are fewer flows it will take longer. o To help combat over pre-emption because, during the time it takes to pre-empt flows, others may be ending anyway (either the call has naturally ended, or the user hangs up due to poor QoS). Slowing pre-emption may seem counter-intuitive here, as it makes it more likely that calls will terminate anyway - however it also gives time to adjust the amount pre-empted to take account of this. o Earlier in this document we said that an egress starts measuring the sustainable-aggregate-rate immediately it sees a single pre- emption marked packet. However, when a link or router fails the network's underlying recovery mechanism will kick in (e.g. switching to a back up path), which may result in the network again being able to support all the traffic. Briscoe Expires April 25, 2007 [Page 41]
Internet-Draft Deployment Model using PCN October 2006 Potential solutions o To combat the final issue, the egress could measure the sustainable-aggregate-rate over a longer time period than the network recovery time (say 100ms vs. 50ms). If it detects no pre- emption marked packets towards the end of its measurement period (say in the last 30 ms) then it doesn't send a pre-emption alert message to the ingress. o We suggest that optionally (the choice of the operator) pre- emption is slowed by pre-empting traffic in several rounds rather than in one shot. One possible algorithm is to pre-empt most of the traffic in the first round and the rest in the second round; the amount pre-empted in the second round is influenced by both the first and second round measurements: * Round 1: pre-empt h * S_1 where 0.5 <= h <= 1 where S_1 is the amount the normal mechanism calculates that it should shed, i.e. {ingress-aggregate-rate - sustainable-aggregate- rate} * Round 2: pre-empt Predicted-S_2 - h * (Predicted- S_2 - Measured-S_2) where Predicted-S_2 = (1-h)*S_1 Note that the second measurement should be made when sufficient time has elapsed for the first round of pre-emption to have happened. One idea to achieve this is for the egress gateway to continuously measure and report its sustainable-aggregate-rate, in (say) 100ms windows. Therefore the ingress gateway knows when the egress gateway made its measurement (assuming the round trip time is known). Therefore the ingress gateway knows when measurements should reflect that it has pre-empted flows. 5.7. Other potential extensions In this section we discuss some other potential extensions not already covered above. 5.7.1. Tunnelling It is possible to tunnel all CL packets across the CL-region. Although there is a cost of tunnelling (additional header on each packet, additional processing at tunnel ingress and egress), there are three reasons it may be interesting. Briscoe Expires April 25, 2007 [Page 42]
Internet-Draft Deployment Model using PCN October 2006 ECMP: Tunnelling is one of the possible solutions given earlier in Section 5.1 on Equal Cost Multipath Routing (ECMP). Ingress gateway determination: If packets are tunnelled from ingress gateway to egress gateway, the egress gateway can very easily determine in the data path which ingress gateway a packet comes from (by simply looking at the source address of the tunnel header). This can facilitate operations such as computing the Congestion-Level-Estimate on a per ingress gateway basis. End-to-end ECN: The ECN field is used for PCN marking (see [PCN] for details), and so it needs to be re-set by the egress gateway to whatever has been agreed as appropriate for the next domain. Therefore if a packet arrives at the ingress gateway with its ECN field already set (i.e. not '00'), it may leave the egress gateway with a different value. Hence the end-to-end meaning of the ECN field is lost. It is open to debate whether end-to-end congestion control is ever necessary within an end-to-end reservation. But if a genuine need is identified for end-to-end ECN semantics within a reservation, then one solution is to tunnel CL packets across the CL-region. When the egress gateway decapsulates them the original ECN field is recovered. 5.7.2. Multi-domain and multi-operator usage This potential extension would eliminate the trust assumption (Section 2.2), so that the CL-region could consist of multiple domains run by different operators that did not trust each other. Then only the ingress and egress gateways of the CL-region would take part in the admission control procedure, i.e. at the ingress to the first domain and the egress from the final domain. The border routers between operators within the CL-region would only have to do bulk accounting - they wouldn't do per microflow metering and policing, and they wouldn't take part in signal processing or hold per flow state [Briscoe]. [Re-feedback] explains how a downstream domain can police that its upstream domain does not 'cheat' by admitting traffic when the downstream path is congested. [Re-PCN] proposes how to achieve this with the help of another recently proposed extension to ECN, involving re-echoing ECN feedback [Re-ECN]. Briscoe Expires April 25, 2007 [Page 43]
Internet-Draft Deployment Model using PCN October 2006 5.7.3. Preferential dropping of pre-emption marked packets When the rate of real-time traffic in the specified class exceeds the maximum configured rate, then a router has to drop some packet(s) instead of forwarding them on the out-going link. Now when the egress gateway measures the Sustainable-Aggregate-Rate, neither dropped packets nor pre-emption marked packets contribute to it. Dropping non-pre-emption-marked packets therefore reduces the measured Sustainable-Aggregate-Rate below its true value. Thus a router should preferentially drop pre-emption marked packets. Note that it is important that the operator doesn't set the configured-pre-emption-rate equal to the rate at which packets start being dropped (for the specified real-time service class). Otherwise the egress gateway may never see a pre-emption marked packet and so won't be triggered into the Pre-emption Alert state. This optimisation is optional. When considering whether to use it an operator will consider issues such as whether the over-pre-emption is serious, and whether the particular routers can easily do this sort of selective drop. 5.7.4. Adaptive bandwidth for the Controlled Load service The admission control mechanism described in this document assumes that each router has a fixed bandwidth allocated to CL flows. A possible extension is that the bandwidth is flexible, depending on the level of non-CL traffic. If a large share of the current load on a path is CL, then more CL traffic can be admitted. And if the greater share of the load is non-CL, then the admission threshold can be proportionately lower. The approach re-arranges sharing between classes to aim for economic efficiency, whatever the traffic load matrix. It also deals with unforeseen changes to capacity during failures better than configuring fixed engineered rates. Adaptive bandwidth allocation can be achieved by changing the admission marking behaviour, so that the probability of admission marking a packet would now depend on the number of queued non-CL packets as well as the size of the virtual queue. The adaptive bandwidth approach would be supplemented by placing limits on the adaptation to prevent starvation of the CL by other traffic classes and of other classes by CL traffic. [Songhurst] has more details of the adaptive bandwidth approach. Briscoe Expires April 25, 2007 [Page 44]
Internet-Draft Deployment Model using PCN October 2006 5.7.5. Controlled Load service with end-to-end Pre-Congestion Notification It may be possible to extend the framework to parts of the network where there are only a low number of CL microflows, i.e. the aggregation assumption (Section 2.2) doesn't hold. In the extreme it may be possible to operate the framework end-to-end, i.e. between end hosts. One potential method is to send probe packets to test whether the network can support a prospective new CL microflow. The probe packets would be sent at the same traffic rate as expected for the actual microflow, but in order not to disturb existing CL traffic a router would always schedule probe packets behind CL ones (compare [Breslau00]); this implies they have a new DSCP. Otherwise the routers would treat probe packets identically to CL packets. In order to perform admission control quickly, in parts of the network where there are only a few CL microflows, the algorithm for Admission Marking described in [PCN] would need to "switch on" very rapidly, ie go from marking no packets to marking them all for only a minimal increase in the size of the virtual queue. 5.7.6. MPLS-TE [ECN-MPLS] discusses how to extend the deployment model to MPLS, i.e. for admission control of microflows into a set of MPLS-TE aggregates (Multi-protocol label switching traffic engineering). It would require that the MPLS header could include the ECN field, which is not precluded by RFC3270. See [ECN-MPLS]. Briscoe Expires April 25, 2007 [Page 45]
Internet-Draft Deployment Model using PCN October 2006 6. Relationship to other QoS mechanisms 6.1. IntServ Controlled Load The CL mechanism delivers QoS similar to Integrated Services controlled load, but rather better. The reason the QoS is better is that the CL mechanism keeps the real queues empty, by driving admission control from a bulk virtual queue on each interface. The virtual queue [AVQ, vq] can detect a rise in load before the real queue builds. It is also more robust to route changes. 6.2. Integrated services operation over DiffServ Our approach to end-to-end QoS is similar to that described in [RFC2998] for Integrated services operation over DiffServ networks. Like [RFC2998], an IntServ class (CL in our case) is achieved end-to- end, with a CL-region viewed as a single reservation hop in the total end-to-end path. Interior routers of the CL-region do not process flow signalling nor do they hold per flow state. Unlike [RFC2998] we do not require the end-to-end signalling mechanism to be RSVP, although it can be. Bearing in mind these differences, we can describe our architecture in the terms of the options in [RFC2998]. The DiffServ network region is RSVP-aware, but awareness is confined to (what [RFC2998] calls) the "border routers" of the DiffServ region. We use explicit admission control into this region, with static provisioning within it. The ingress "border router" does per microflow policing and sets the DSCP and ECN fields to indicate the packets are CL ones (i.e. we use router marking rather than host marking). 6.3. Differentiated Services The DiffServ architecture does not specify any way for devices outside the domain to dynamically reserve resources or receive indications of network resource availability. In practice, service providers rely on subscription-time Service Level Agreements (SLAs) that statically define the parameters of the traffic that will be accepted from a customer. The CL mechanism allows dynamic reservation of resources through the DiffServ domain and, with the potential extension mentioned in Section 5.7.2, it can span multiple domains without active policing mechanisms at the borders (unlike DiffServ). Therefore we do not use the traffic conditioning agreements (TCAs) of the (informational) DiffServ architecture [RFC2475]. An important benefit arises from the fact that the load is controlled dynamically rather than with traffic conditioning agreements (TCAs). Briscoe Expires April 25, 2007 [Page 46]
Internet-Draft Deployment Model using PCN October 2006 TCAs were originally introduced in the (informational) DiffServ architecture [RFC2475] as an alternative to reservation processing in the interior region in order to reduce the burden on interior routers. With TCAs, in practice service providers rely on subscription-time Service Level Agreements that statically define the parameters of the traffic that will be accepted from a customer. The problem arises because the TCA at the ingress must allow any destination address, if it is to remain scalable. But for longer topologies, the chances increase that traffic will focus on an interior resource, even though it is within contract at the ingress [Reid], e.g. all flows converge on the same egress gateway. Even though networks can be engineered to make such failures rare, when they occur all inelastic flows through the congested resource fail catastrophically. [Johnson] compares admission control with a 'generously dimensioned' DiffServ network as ways to achieve QoS. The former is recommended. 6.4. ECN The marking behaviour described in this document complies with the ECN aspects of the IP wire protocol RFC3168, but provides its own edge-to-edge feedback instead of the TCP aspects of RFC3168. All routers within the CL-region are upgraded with the admission marking and pre-emption marking of Pre-Congestion Notification, so the requirements of [Floyd] are met because the CL-region is an enclosed environment. The operator prevents traffic arriving at a router that doesn't understand CL by administrative configuration of the ring of gateways around the CL-region. 6.5. RTECN Real-time ECN (RTECN) [RTECN, RTECN-usage] has a similar aim to this document (to achieve a low delay, jitter and loss service suitable for RT traffic) and a similar approach (per microflow admission control combined with an "early warning" of potential congestion through setting the CE codepoint). But it explores a different architecture without the aggregation assumption: host-to-host rather than edge-to-edge. We plan to document such a host-to-host framework in a parallel draft to this one, and to describe if and how [PCN] can work in this framework. Briscoe Expires April 25, 2007 [Page 47]
Internet-Draft Deployment Model using PCN October 2006 6.6. RMD Resource Management in DiffServ (RMD) [RMD] is similar to this work, in that it pushes complex classification, traffic conditioning and admission control functions to the edge of a DiffServ domain and simplifies the operation of the interior routers. One of the RMD modes ("Congestion notification function based on probing") uses measurement-based admission control in a similar way to this document. The main difference is that in RMD probing plays a significant role in the admission control process. Other differences are that the admission control decision is taken on the egress gateway (rather than the ingress); 'admission marking' is encoded in a packet as a new DSCP (rather than in the ECN field), and that the NSIS protocols are used for signalling (rather than RSVP). RMD also includes the concept of Severe Congestion handling. The pre- emption mechanism described in the CL architecture has similar objectives but relies on different mechanisms. The main difference is that the interior routers measure the data rate that causes an overload and mark packets according to this rate. 6.7. RSVP Aggregation over MPLS-TE Multi-protocol label switching traffic engineering (MPLS-TE) allows scalable reservation of resources in the core for an aggregate of many microflows. To achieve end-to-end reservations, admission control and policing of microflows into the aggregate can be achieved using techniques such as RSVP Aggregation over MPLS TE Tunnels as per [AGGRE-TE]. However, in the case of inter-provider environments, these techniques require that admission control and policing be repeated at each trust boundary or that MPLS TE tunnels span multiple domains. 6.8. Other Network Admission Control Approaches Link admission control (LAC) describes how admission control (AC) can be done on a single link and comprises, e.g., the calculation of effective bandwidths which may be the base for a parameter-based AC. In contrast, network AC (NAC) describes how AC can be done for a network and focuses on the locations from which data is gathered for the admission decision. Most approaches implement a link budget based NAC (LB NAC) where each link has a certain AC-budget. RSVP works according to that principle, but also the new concept admits additional flows as long as each link on the new flow's path still has resources available. The border-to-border budget based NAC (BBB NAC) pre-configures an AC budget for all border-to-border relationships (= CL-region-aggregates) and if this capacity budget is Briscoe Expires April 25, 2007 [Page 48]
Internet-Draft Deployment Model using PCN October 2006 exhausted, new flows are rejected. The TCA-based admission control which is associated with the DiffServ architecture implements an ingress budget based NAC (IB NAC). These basically different concepts have different flexibility and efficiency with regard to the use of link bandwidths [NAC-a,NAC-b]. They can be made resilient by choosing the budgets in such a way that the network will not be congested after rerouting due to a failure. The efficiency of the approaches is different with and without such resilient requirements. 7. Security Considerations To protect against denial of service attacks, the ingress gateway of the CL-region needs to police all CL packets and drop packets in excess of the reservation. This is similar to operations with existing IntServ behaviour. For pre-emption, it is considered acceptable from a security perspective that the ingress gateway can treat "emergency/military" CL flows preferentially compared with "ordinary" CL flows. However, in the rest of the CL-region they are not distinguished (nonetheless, our proposed technique does not preclude the use of different DSCPs at the packet level as well as different priorities at the flow level.). Keeping emergency traffic indistinguishable at the packet level minimises the opportunity for new security attacks. For example, if instead a mechanism used different DSCPs for "emergency/military" and "ordinary" packets, then an attacker could specifically target the former in the data plane (perhaps for DoS or for eavesdropping). Further security aspects to be considered later. 8. Acknowledgements The admission control mechanism evolved from the work led by Martin Karsten on the Guaranteed Stream Provider developed in the M3I project [GSPa, GSP-TR], which in turn was based on the theoretical work of Gibbens and Kelly [DCAC]. Kennedy Cheng, Gabriele Corliano, Carla Di Cairano-Gilfedder, Kashaf Khan, Peter Hovell, Arnaud Jacquet and June Tay (BT) helped develop and evaluate this approach. Many thanks to those who have commented on this work at Transport Area Working Group meetings and on the mailing list, including: Ken Briscoe Expires April 25, 2007 [Page 49]
Internet-Draft Deployment Model using PCN October 2006 Carlberg, Ruediger Geib, Lars Westberg, David Black, Robert Hancock, Cornelia Kappler, Michael Menth. 9. Comments solicited Comments and questions are encouraged and very welcome. They can be sent to the Transport Area Working Group's mailing list, tsvwg@ietf.org, and/or to the authors. 10. Changes from earlier versions of the draft The main changes are: From -00 to -01 The whole of the Pre-emption mechanism is added. There are several modifications to the admission control mechanism. From -01 to -02 The pre-congestion notification algorithms for admission marking and pre-emption marking are now described in [PCN]. There are new sub-sections in Section 4 on Failures, Admission of 'emergency / higher precedence' session, and Tunnelling; and a new sub-section in Section 5 on Mechanisms to deal with 'Flash crowds'. From -02 to -03 Section 5 has been updated and expanded. It is now about the 'limitations' of the PCN mechanism, as described in the earlier sections, plus discussion of 'possible solutions' to those limitations. The measurement of the Congestion-Level-Estimate now includes pre- emption marked packets as well as admission marked ones. Section 3.1.2 explains. Briscoe Expires April 25, 2007 [Page 50]
Internet-Draft Deployment Model using PCN October 2006 From -03 to -04 Detailed review by Michael Menth. In response, Abstract, Summary and Key benefits sections re-written. Numerous detailed comments on Sections 5 and following sections. Briscoe Expires April 25, 2007 [Page 51]
Internet-Draft Deployment Model using PCN October 2006 11. Appendices 11.1. Appendix A: Explicit Congestion Notification This Appendix provides a brief summary of Explicit Congestion Notification (ECN). [RFC3168] specifies the incorporation of ECN to TCP and IP, including ECN's use of two bits in the IP header. It specifies a method for indicating incipient congestion to end-hosts (e.g. as in RED, Random Early Detection), where the notification is through ECN marking packets rather than dropping them. ECN uses two bits in the IP header of both IPv4 and IPv6 packets: 0 1 2 3 4 5 6 7 +-----+-----+-----+-----+-----+-----+-----+-----+ | DS FIELD, DSCP | ECN FIELD | +-----+-----+-----+-----+-----+-----+-----+-----+ DSCP: differentiated services codepoint ECN: Explicit Congestion Notification Figure A.1: The Differentiated Services and ECN Fields in IP. The two bits of the ECN field have four ECN codepoints, '00' to '11': +-----+-----+ | ECN FIELD | +-----+-----+ ECT CE 0 0 Not-ECT 0 1 ECT(1) 1 0 ECT(0) 1 1 CE Figure A.2: The ECN Field in IP. The not-ECT codepoint '00' indicates a packet that is not using ECN. The CE codepoint '11' is set by a router to indicate congestion to the end hosts. The term 'CE packet' denotes a packet that has the CE codepoint set. The ECN-Capable Transport (ECT) codepoints '10' and '01' (ECT(0) and ECT(1) respectively) are set by the data sender to indicate that the Briscoe Expires April 25, 2007 [Page 52]
Internet-Draft Deployment Model using PCN October 2006 end-points of the transport protocol are ECN-capable. Routers treat the ECT(0) and ECT(1) codepoints as equivalent. Senders are free to use either the ECT(0) or the ECT(1) codepoint to indicate ECT, on a packet-by-packet basis. The motivation for having two codepoints (the 'ECN nonce') is the desire to check two things: for the data sender to verify that network elements are not erasing the CE codepoint; and for the data sender to verify that data receivers are properly reporting to the sender the receipt of packets with the CE codepoint set. ECN requires support from the transport protocol, in addition to the functionality given by the ECN field in the IP packet header. [RFC3168] addresses the addition of ECN Capability to TCP, specifying three new pieces of functionality: negotiation between the endpoints during connection setup to determine if they are both ECN-capable; an ECN-Echo (ECE) flag in the TCP header so that the data receiver can inform the data sender when a CE packet has been received; and a Congestion Window Reduced (CWR) flag in the TCP header so that the data sender can inform the data receiver that the congestion window has been reduced. The transport layer (e.g.. TCP) must respond, in terms of congestion control, to a *single* CE packet as it would to a packet drop. The advantage of setting the CE codepoint as an indication of congestion, instead of relying on packet drops, is that it allows the receiver(s) to receive the packet, thus avoiding the potential for excessive delays due to retransmissions after packet losses. 11.2. Appendix B: What is distributed measurement-based admission control? This Appendix briefly explains what distributed measurement-based admission control is [Breslau99]. Traditional admission control algorithms for 'hard' real-time services (those providing a firm delay bound for example) guarantee QoS by using 'worst case analysis'. Each time a flow is admitted its traffic parameters are examined and the network re-calculates the remaining resources. When the network gets a new request it therefore knows for certain whether the prospective flow, with its particular parameters, should be admitted. However, parameter-based admission control algorithms result in under-utilisation when the traffic is bursty. Therefore 'soft' real time services - like Controlled Load - can use a more relaxed admission control algorithm. Briscoe Expires April 25, 2007 [Page 53]
Internet-Draft Deployment Model using PCN October 2006 This insight suggests measurement-based admission control (MBAC). The aim of MBAC is to provide a statistical service guarantee. The classic scenario for MBAC is where each router participates in hop- by-hop admission control, characterising existing traffic locally through measurements (instead of keeping an accurate track of traffic as it is admitted), in order to determine the current value of some parameter e.g. load. Note that for scalability the measurement is of the aggregate of the flows in the local system. The measured parameter(s) is then compared to the requirements of the prospective flow to see whether it should be admitted. MBAC may also be performed centrally for a network, it which case it uses centralised measurements by a bandwidth broker. We use distributed MBAC. "Distributed" means that the measurement is accumulated for the 'whole-path' using in-band signalling. In our case, this means that the measurement of existing traffic is for the same pair of ingress and egress gateways as the prospective microflow. In fact our mechanism can be said to be distributed in three ways: all routers on the ingress-egress path affect the Congestion-Level- Estimate; the admission control decision is made just once on behalf of all the routers on the path across the CL-region; and the ingress and egress gateways cooperate to perform MBAC. 11.3. Appendix C: Calculating the Exponentially weighted moving average (EWMA) At the egress gateway, for every CL packet arrival: [EWMA-total-bits]n+1 = (w * bits-in-packet) + ((1-w) * [EWMA- total-bits]n ) [EWMA-M-bits]n+1 = (B * w * bits-in-packet) + ((1-w) * [EWMA-M- bits]n ) Then, per new flow arrival: [Congestion-Level-Estimate]n+1 = [EWMA-M-bits]n+1 / [EWMA-total- bits]n+1 where Briscoe Expires April 25, 2007 [Page 54]
Internet-Draft Deployment Model using PCN October 2006 EWMA-total-bits is the total number of bits in CL packets, calculated as an exponentially weighted moving average (EWMA) EWMA-M-bits is the total number of bits in CL packets that are Admission Marked or Pre-emption Marked, again calculated as an EWMA. B is either 0 or 1: B = 0 if the CL packet is not admission marked B = 1 if the CL packet is admission marked w is the exponential weighting factor. Varying the value of the weight trades off between the smoothness and responsiveness of the Congestion-Level-Estimate. However, in general both can be achieved, given our original assumption of many CL microflows and remembering that the EWMA is calculated on the basis of aggregate traffic between the ingress and egress gateways. There will be a threshold inter-arrival time between packets of the same aggregate below which the egress will consider the estimate of the Congestion-Level-Estimate as too stale, and it will then trigger generation of probes by the ingress. The first two per-packet algorithms can be simplified, if their only use will be where the result of one is divided by the result of the other in the third, per-flow algorithm. [EWMA-total-bits]'n+1 = bits-in-packet + (w' * [EWMA- total- bits]n ) [EWMA-AM-bits]'n+1 = (B * bits-in-packet) + (w' * [EWMA-AM-bits]n ) where w' = (1-w)/w. If w' is arranged to be a power of 2, these per packet algorithms can be implemented solely with a shift and an add. Briscoe Expires April 25, 2007 [Page 55]
Internet-Draft Deployment Model using PCN October 2006 There are alternative possibilities for smoothing out the congestion- level-estimate. For example [TEWMA] deals better with the issue of stale information when the traffic rate for 12. References A later version will distinguish normative and informative references. [AGGRE-TE] Francois Le Faucheur, Michael Dibiasio, Bruce Davie, Michael Davenport, Chris Christou, Jerry Ash, Bur Goode, 'Aggregation of RSVP Reservations over MPLS TE/DS-TE Tunnels', draft-ietf-tsvwg-rsvp-dste-03 (work in progress), June 2006 [ANSI.MLPP.Spec] American National Standards Institute, "Telecommunications- Integrated Services Digital Network (ISDN) - Multi-Level Precedence and Pre- emption (MLPP) Service Capability", ANSI T1.619-1992 (R1999), 1992. [ANSI.MLPP.Supplement] American National Standards Institute, "MLPP Service Domain Cause Value Changes", ANSI ANSI T1.619a-1994 (R1999), 1990. [AVQ] S. Kunniyur and R. Srikant "Analysis and Design of an Adaptive Virtual Queue (AVQ) Algorithm for Active Queue Management", In: Proc. ACM SIGCOMM'01, Computer Communication Review 31 (4) (October, 2001). [Breslau99] L. Breslau, S. Jamin, S. Shenker "Measurement-based admission control: what is the research agenda?", In: Proc. Int'l Workshop on Quality of Service 1999. [Breslau00] L. Breslau, E. Knightly, S. Shenker, I. Stoica, H. Zhang "Endpoint Admission Control: Architectural Issues and Performance", In: ACM SIGCOMM 2000 [Briscoe] Bob Briscoe and Steve Rudkin, "Commercial Models for IP Quality of Service Interconnect", BT Technology Journal, Vol 23 No 2, April 2005. Briscoe Expires April 25, 2007 [Page 56]
Internet-Draft Deployment Model using PCN October 2006 [DCAC] Richard J. Gibbens and Frank P. Kelly "Distributed connection acceptance control for a connectionless network", In: Proc. International Teletraffic Congress (ITC16), Edinburgh, pp. 941952 (1999). [ECN-MPLS] Bruce Davie, Bob Briscoe, June Tay, "Explicit Congestion Marking in MPLS", draft- davie-ecn-mpls-00.txt (work in progress), June 2006 [EMERG-RQTS] Carlberg, K. and R. Atkinson, "General Requirements for Emergency Telecommunication Service (ETS)", RFC 3689, February 2004. [EMERG-TEL] Carlberg, K. and R. Atkinson, "IP Telephony Requirements for Emergency Telecommunication Service (ETS)", RFC 3690, February 2004. [Floyd] S. Floyd, 'Specifying Alternate Semantics for the Explicit Congestion Notification (ECN) Field', draft- floyd-ecn-alternates-02.txt (work in progress), August 2005 [GSPa] Karsten (Ed.), Martin "GSP/ECN Technology & Experiments", Deliverable: 15.3 PtIII, M3I Eu Vth Framework Project IST-1999-11429, URL: http://www.m3i.org/ (February, 2002) (superseded by [GSP-TR]) [GSP-TR] Martin Karsten and Jens Schmitt, "Admission Control Based on Packet Marking and Feedback Signalling -- Mechanisms, Implementation and Experiments", TU- Darmstadt Technical Report TR-KOM-2002-03, URL: http://www.kom.e-technik.tu- darmstadt.de/publications/abstracts/KS02-5.html (May, 2002) [ITU.MLPP.1990] International Telecommunications Union, "Multilevel Precedence and Pre-emption Service (MLPP)", ITU-T Recommendation I.255.3, 1990. [Johnson] DM Johnson, 'QoS control versus generous dimensioning', BT Technology Journal, Vol 23 No 2, April 2005 Briscoe Expires April 25, 2007 [Page 57]
Internet-Draft Deployment Model using PCN October 2006 [LoadBalancing-a] Ruediger Martin, Michael Menth, and Michael Hemmkeppler: "Accuracy and Dynamics of Hash-Based Load Balancing Algorithms for Multipath Internet Routing", IEEE Broadnets, San Jose, CA, USA, October 2006 http://www3.informatik.uni- wuerzburg.de/~menth/Publications/Menth06p.pdf [LoadBalancing-b] Ruediger Martin, Michael Menth, and Michael Hemmkeppler: "Accuracy and Dynamics of Multi-Stage Load Balancing for Multipath Internet Routing", currently under submission http://www3.informatik.uni- wuerzburg.de/~menth/Publications/Menth07-Sub-6.pdf [Low] S. Low, L. Andrew, B. Wydrowski, 'Understanding XCP: equilibrium and fairness', IEEE InfoCom 2005 [NAC-a] Michael Menth: "Efficient Admission Control and Routing in Resilient Communication Networks", PhD thesis, July 2004, http://opus.bibliothek.uni- wuerzburg.de/opus/volltexte/2004/994/pdf/Menth04.pdf [NAC-b] Michael Menth, Stefan Kopf, Joachim Charzinski, and Karl Schrodi: "Resilient Network Admission Control", currently under submission. http://www3.informatik.uni- wuerzburg.de/~menth/Publications/Menth07-Sub-3.pdf [PCN] B. Briscoe, P. Eardley, D. Songhurst, F. Le Faucheur, A. Charny, V. Liatsos, S. Dudley, J. Babiarz, K. Chan, G. Karagiannis, A. Bader, L. Westberg. 'Pre-Congestion Notification marking', draft-briscoe-tsvwg-cl-phb-02 (work in progress), June 2006. [Re-ECN] Bob Briscoe, Arnaud Jacquet, Alessandro Salvatori, 'Re-ECN: Adding Accountability for Causing Congestion to TCP/IP', draft-briscoe-tsvwg-re-ecn-tcp-01 (work in progress), March 2006. [Re-feedback] Bob Briscoe, Arnaud Jacquet, Carla Di Cairano- Gilfedder, Andrea Soppera, 'Re-feedback for Policing Congestion Response in an Inter-network', ACM SIGCOMM 2005, August 2005. [Re-PCN] B. Briscoe, 'Emulating Border Flow Policing using Re- ECN on Bulk Data', draft-briscoe-tsvwg-re-ecn-border- cheat-00 (work in progress), February 2006. Briscoe Expires April 25, 2007 [Page 58]
Internet-Draft Deployment Model using PCN October 2006 [Reid] ABD Reid, 'Economics and scalability of QoS solutions', BT Technology Journal, Vol 23 No 2, April 2005 [RFC2211] J. Wroclawski, Specification of the Controlled-Load Network Element Service, September 1997 [RFC2309] Braden, B., et al., "Recommendations on Queue Management and Congestion Avoidance in the Internet", RFC 2309, April 1998. [RFC2474] Nichols, K., Blake, S., Baker, F. and D. Black, "Definition of the Differentiated Services Field (DS Field) in the IPv4 and IPv6 Headers", RFC 2474, December 1998 [RFC2475] Blake, S., Black, D., Carlson, M., Davies, E., Wang, Z. and W. Weiss, 'A framework for Differentiated Services', RFC 2475, December 1998. [RFC2597] Heinanen, J., Baker, F., Weiss, W. and J. Wrocklawski, "Assured Forwarding PHB Group", RFC 2597, June 1999. [RFC2998] Bernet, Y., Yavatkar, R., Ford, P., Baker, F., Zhang, L., Speer, M., Braden, R., Davie, B., Wroclawski, J. and E. Felstaine, "A Framework for Integrated Services Operation Over DiffServ Networks", RFC 2998, November 2000. [RFC3168] Ramakrishnan, K., Floyd, S. and D. Black "The Addition of Explicit Congestion Notification (ECN) to IP", RFC 3168, September 2001. [RFC3246] B. Davie, A. Charny, J.C.R. Bennet, K. Benson, J.Y. Le Boudec, W. Courtney, S. Davari, V. Firoiu, D. Stiliadis, 'An Expedited Forwarding PHB (Per-Hop Behavior)', RFC 3246, March 2002. [RFC3270] Le Faucheur, F., Wu, L., Davie, B., Davari, S., Vaananen, P., Krishnan, R., Cheval, P., and J. Heinanen, "Multi- Protocol Label Switching (MPLS) Support of Differentiated Services", RFC 3270, May 2002. [RFC4542] F. Baker & J. Polk, "Implementing an Emergency Telecommunications Service for Real Time Services in the Internet Protocol Suite", RFC 4542, May 2006. Briscoe Expires April 25, 2007 [Page 59]
Internet-Draft Deployment Model using PCN October 2006 [RMD] Attila Bader, Lars Westberg, Georgios Karagiannis, Cornelia Kappler, Tom Phelan, 'RMD-QOSM - The Resource Management in DiffServ QoS model', draft-ietf-nsis- rmd-03 Work in Progress, June 2005. [RSVP-PCN] Francois Le Faucheur, Anna Charny, Bob Briscoe, Philip Eardley, Joe Barbiaz, Kwok-Ho Chan, 'RSVP Extensions for Admission Control over DiffServ using Pre- Congestion Notification (PCN)', draft-lefaucheur-rsvp- ecn-01 (work in progress), June 2006. [RSVP-PREEMPTION] Herzog, S., "Signaled Preemption Priority Policy Element", RFC 3181, October 2001. [RSVP-EMERGENCY] Le Faucheur et al., RSVP Extensions for Emergency Services, draft-lefaucheur-emergency-rsvp-02.txt [RTECN] Babiarz, J., Chan, K. and V. Firoiu, 'Congestion Notification Process for Real-Time Traffic', draft- babiarz-tsvwg-rtecn-04 Work in Progress, July 2005. [RTECN-usage] Alexander, C., Ed., Babiarz, J. and J. Matthews, 'Admission Control Use Case for Real-time ECN', draft- alexander-rtecn-admission-control-use-case-00, Work in Progress, February 2005. [Songhurst] David J. Songhurst, Philip Eardley, Bob Briscoe, Carla Di Cairano Gilfedder and June Tay, 'Guaranteed QoS Synthesis for Admission Control with Shared Capacity', BT Technical Report TR-CXR9-2006-001, Feb 2006, http://www.cs.ucl.ac.uk/staff/B.Briscoe/projects/ipe2e qos/gqs/papers/GQS_shared_tr.pdf [vq] Costas Courcoubetis and Richard Weber "Buffer Overflow Asymptotics for a Switch Handling Many Traffic Sources" In: Journal Applied Probability 33 pp. 886-- 903 (1996). Briscoe Expires April 25, 2007 [Page 60]
Internet-Draft Deployment Model using PCN October 2006 Authors' Addresses Bob Briscoe BT Research B54/77, Sirius House Adastral Park Martlesham Heath Ipswich, Suffolk IP5 3RE United Kingdom Email: bob.briscoe@bt.com Dave Songhurst BT Research B54/69, Sirius House Adastral Park Martlesham Heath Ipswich, Suffolk IP5 3RE United Kingdom Email: dsonghurst@jungle.bt.co.uk Philip Eardley BT Research B54/77, Sirius House Adastral Park Martlesham Heath Ipswich, Suffolk IP5 3RE United Kingdom Email: philip.eardley@bt.com Francois Le Faucheur Cisco Systems, Inc. Village d'Entreprise Green Side - Batiment T3 400, Avenue de Roumanille 06410 Biot Sophia-Antipolis France Email: flefauch@cisco.com Anna Charny Cisco Systems 300 Apollo Drive Chelmsford, MA 01824 USA Email: acharny@cisco.com Briscoe Expires April 25, 2007 [Page 61]
Internet-Draft Deployment Model using PCN October 2006 Kwok Ho Chan Nortel Networks 600 Technology Park Drive Billerica, MA 01821 USA Email: khchan@nortel.com Jozef Z. Babiarz Nortel Networks 3500 Carling Avenue Ottawa, Ont K2H 8E9 Canada Email: babiarz@nortel.com Stephen Dudley Nortel Networks 4001 E. Chapel Hill Nelson Highway P.O. Box 13010, ms 570-01-0V8 Research Triangle Park, NC 27709 USA Email: smdudley@nortel.com Georgios Karagiannis University of Twente P.O. BOX 217 7500 AE Enschede, The Netherlands EMail: g.karagiannis@ewi.utwente.nl Attila Báder attila.bader@ericsson.com Lars Westberg Ericsson AB SE-164 80 Stockholm Sweden EMail: Lars.Westberg@ericsson.com Briscoe Expires April 25, 2007 [Page 62]
Internet-Draft Deployment Model using PCN October 2006 Intellectual Property Statement The IETF takes no position regarding the validity or scope of any Intellectual Property Rights or other rights that might be claimed to pertain to the implementation or use of the technology described in this document or the extent to which any license under such rights might or might not be available; nor does it represent that it has made any independent effort to identify any such rights. Information on the procedures with respect to rights in RFC documents can be found in BCP 78 and BCP 79. Copies of IPR disclosures made to the IETF Secretariat and any assurances of licenses to be made available, or the result of an attempt made to obtain a general license or permission for the use of such proprietary rights by implementers or users of this specification can be obtained from the IETF on-line IPR repository at http://www.ietf.org/ipr. The IETF invites any interested party to bring to its attention any copyrights, patents or patent applications, or other proprietary rights that may cover technology that may be required to implement this standard. Please address the information to the IETF at ietf-ipr@ietf.org Disclaimer of Validity This document and the information contained herein are provided on an "AS IS" basis and THE CONTRIBUTOR, THE ORGANIZATION HE/SHE REPRESENTS OR IS SPONSORED BY (IF ANY), THE INTERNET SOCIETY AND THE INTERNET ENGINEERING TASK FORCE DISCLAIM ALL WARRANTIES, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO ANY WARRANTY THAT THE USE OF THE INFORMATION HEREIN WILL NOT INFRINGE ANY RIGHTS OR ANY IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Copyright Statement Copyright (C) The Internet Society (2006). This document is subject to the rights, licenses and restrictions contained in BCP 78, and except as set forth therein, the authors retain all their rights. Briscoe Expires April 25, 2007 [Page 63]