Internet-Draft Multicast Scaling October 2022
Zhang, et al. Expires 27 April 2023 [Page]
Workgroup:
pim
Internet-Draft:
draft-zzhang-pim-multicast-scaling-considerations-00
Published:
Intended Status:
Informational
Expires:
Authors:
Z. Zhang
Juniper Networks
R. Parekh
Cisco
H. Bidgoli
Nokia
Z. Zhang
ZTE

Multicast Scaling Considerations

Abstract

This informational document discusses various multicast scaling aspects, compares different multicast technologies with respect to scaling, and suggests a general approach of combined solutions to scale multicast. This discussion is independent of IPv4/IPv6 or MPLS/SRv6 data planes.

Status of This Memo

This Internet-Draft is submitted in full conformance with the provisions of BCP 78 and BCP 79.

Internet-Drafts are working documents of the Internet Engineering Task Force (IETF). Note that other groups may also distribute working documents as Internet-Drafts. The list of current Internet-Drafts is at https://datatracker.ietf.org/drafts/current/.

Internet-Drafts are draft documents valid for a maximum of six months and may be updated, replaced, or obsoleted by other documents at any time. It is inappropriate to use Internet-Drafts as reference material or to cite them other than as "work in progress."

This Internet-Draft will expire on 27 April 2023.

1. Different Scaling Aspects

IP Multicast [RFC1112] architecture involves an IP multicast group/destination address that, together with the source address, identifies a multicast tree rooted at the First Hop Router (FHR) that the source is attached to. Multicast traffic is forwarded along the tree, which is typically set up by Protocol Independent Multicast (PIM) [RFC7761].

In this document, each (source, group) pair is referred to a flow or a tree. Typically, each flow is for a "piece of content", e.g., an audio or video content.

While a bidirectional tree [RFC5015] can be used for multicast among a set of sources and receivers using only group/destination address, it is typically used only in a small domain and is not considered in this document.

1.1. Scaling in the number of receivers

With the above-mentioned multicast tree, each tree node only need to replicate to a minimum number of downstream nodes, which could be transit or leaf nodes. Except in the case of Broadband Node Gateway (BNG) connecting to a huge number of home subscribers or in the case of a spine switch connecting to a large number of leaf switches in a Data Center, the number of replication is small, yet the number of total receivers can be unlimited as the tree grows lengthwise and width-wise.

1.2. Scaling in the number of trees

The number of (source, group) pairs could be huge. Each (source, group) pair would need a tree, with each replication point being a tree node. Each tree node needs to have state specifically for each tree both in forwarding plane and control plane (that is used to set up the forwarding plane state).

The number of multicast trees that a network can support may have to be limited because of the tree state, yet the number of multicast trees transiting a network may be huge (simply because of the huge number of multicast flows).

Chances are, many of the flows have a common set of ingress and egress points in the transit network. In that case, instead of setting up individual multicast (source, group) trees across this network, tunnels can be to transport those individual trees. For example, a single mLDP [RFC6388] tunnel can be used to transport thousands of IP multicast flows, greatly reducing the number of trees in the network.

2. Multicast Tunnels

A multicast tunnel also corresponds to a multicast tree and requires corresponding state on tree nodes, though it is not identified by a (source, group) pair.

In case of mLDP [RFC6388] or RSVP-TE P2MP [RFC4875] tunnel, the tree is identified by an mLDP FEC or RSVP P2MP Session Object in the control plane, and a label in the forwarding plane. The tree will likely have transit (non-root/leaf) tree nodes.

An MPLS SR-P2MP [I-D.ietf-pim-sr-p2mp-policy] tunnel is identical to mLDP/RSVP-TE in the forwarding plane. It's only different from the latter in the control plane - different control plane identifier and different setup procedures, but still requires per-tree state on all tree nodes.

An SRv6-P2MP tunnel is also similar to mLDP/RSVP-TE/SR MPLS P2MP tunnel even in data plane - it's just that the data plane identifier is now part of IPv6 destination address.

An Ingress Replication (IR) [RFC7988] tunnel is a special/degraded multicast tree in that it does not have transit tree nodes. The tree root (ingress) replicates traffic and tunnel to tree leaves directly using unicast, without the need for tree state on transit routers. Obviously, it does not have efficient replication - the ingress may need to send multiple copies through common downstream nodes - but it may still be desired in certain situation.

We refer the tunnels to as underlay tunnels and multicast traffic (e.g. IP Multicast) that the undelay tunnels carry to as overlay flows. Notice that, an underlay tunnel could be carried by another underlay tunnel (e.g. a part of an mLDP tunnel is transported by RSVP-TE P2MP tunnel).

3. New Multicast Technologies

The IP multicast and non-IR tunnels described above require per-tree/tunnel state on transit tree nodes. While use of tunnels removes individual IP Multicast trees in the tunnels' region, the number of tunnels may still be large. New multicast technologies have been developed to remove the per-tree/tunnel state while still allow efficient replication, as summarized below.

3.1. BIER

Bit Index Explicit Replication (BIER) [RFC8279] is a multicast architecture in which forwarding is based on a BitString in the BIER header preceding the payload. Each Bit Position (BP) that is set in the BitString represents a BIER Forwarding Egress Router (BFER) that needs to receive the packet. A BIER Forwarding Router (BFR) checks the set BPs to determine the set of next hop BFR to replicate the packet to, by repeatedly using a BP as lookup key in a BIER Index Forwarding Table (BIFT). A clever way ensures that the number of lookup is bound by the number of replications, no matter how many and which BPs are set.

The length of the BitString is limited by maximum transmission unit (MTU) of the BIER domain, the reserved space in a packet for payloads, and more importantly the implementation tradeoff in the forwarding plane. A typical value is 256. While larger BitString lengths could be used, it may be suboptimal if the BFERs for a packet are sparsely distributed.

If the number of BFERs is larger than the BitString length, there are two ways to make it work:

  • Send multiple copies - each copy is for a different set of BFERs. The same BP in different copies corresponds to different BFERs. This not only requires multiple copies but also multiple BIFTs - one BIFT for each set.
  • Divide the BIER domain into per-region sub-domains to use tunnel segmentation (Section 4). A segmentation point decapsulates the BIER packets received in an upstream region and re-encapsulate BIER packets for downstream regions.

BIER can be considered as a tunneling technology - overlay flows (e.g. IP multicast) are tunneled over BIER, even though there are not per-tunnel state in a BIER domain.

The BIER architecture includes a flow overlay on top of BIER layer that does BIER signaling and forwarding, which is in turn over a routing underlay that forwards BIER traffic from BIER Forwarding Ingress Routers (BFIRs) to BFERs. The purpose of flow overlay is for BFERs to signal BFIRs that they are interested in certain overlay flows so that the BFIRs can derive what BitString to use for an overlay flow.

3.2. BIER-TE

BIER does not provide per-flow traffic engineering capability. BIER Traffic Engineering (BIER-TE) [I-D.ietf-bier-te-arch] does so without requiring per-flow state inside the BIER domain.

With BIER-TE, a BP may indicate a replication branch - not just a BFER anymore. Because of that, with the same BitString length, fewer BFERs will be covered by a packet - whether we send multiple copies or use tunnel segmentation.

3.3. CGM2

With BIER and BIER-TE, the BPs in the BitString have "global" ("subdomain-wide" to be strict) significance and all BFRs in a subdomain must have entries for all of them in the BIFTs. However, the BIER-TE concept can be augmented to work with BPs of local significance. This is referred to as Carrier Grade Minimalist Multicast (CGM2) [I-D.eckert-bier-cgm2-rbs].

The CGM2 concept could be more easily understood through [I-D.chen-pim-srv6-p2mp-path], which uses SRv6 SIDs to encode replication branches. Replacing the SIDs with BPs you get CGM2.

Obviously CGM2 scales better than [I-D.chen-pim-srv6-p2mp-path] in that many fewer bits are needed in the packet header so this document focuses on CGM2.

While CGM2 does use fewer BIFT entries, it does not save on the number of bits needed to encode all replication branches. In fact, it uses more bits to turn the flat-structured BitString in BIER-TE to recursive/hierarchical constructs, and that also leads to more complicated forwarding behavior.

Given this limitation, while CGM2 can be used as underlay tunnels, it is not really a good fit for Carrier Grade multicast when overlay IP multicast is not used (i.e. the end receivers are directly encoded as BPs).

4. Multicast Tunnel Segmentation

An end-to-end (E2E) multicast flow may need to go through a vast network consisting of many ASes and/or IGP areas (referred to regions). When tunneling overlay multicast flows, different types/instances of tunnels may need to be used in different regions. This is referred to as tunnel segmentation [RFC6514] [RFC7524], and it is because of the following reasons:

  • Due to technological or administrative reasons, different tunnel technologies may have to be used in different regions. For example, one region may use mLDP while another may use IR.
  • For the same reasons, different tunnel instances of the same type may have to be used in different regions.
  • Even if the same tunnel could be established across multiple regions, different tunnel instances in different regions may be desired for better optimization. For example, if a set of flows have very different sets of egress points, it is not desired to carry them in a single across-region tunnel but they may be carried by a single intra-region tunnel in one region and different intra-region tunnels in other regions.
  • In case of IR, instead of directly replicating to all egress points, it is more efficient to ingress replicate to segmentation points who in turn ingress replicate to the next set of segmentation and/or egress points.
  • In case of BIER, when the number of BFERs is larger than the BitString length, segmentation may be used together with smaller subdomains.

At the segmentation points, overlay state is needed to stitch the upstream segments and downstream segments together.

5. Signaling for Tunneling and Segmentation

To carry multicast traffic over a tunnel, the tunnel ingress needs to know what flows to be put onto which tunnel. Depending on the tunnel type, it may also need to know the tunnel egresses. The tunnel egresses may also need to know that they need to join the tunnel. This requires signaling - note that this is not for setting up the underlay tunnel, but for "flow overlay".

5.1. MVPN as Flow Overlay Signaling for Tunneling

Consider some E2E IP multicast trees of a customer with part of them go over a VPN provider network. The PE routers tunnel the traffic across the provider network based on PIM-MVPN or BGP-MVPN signaling specified in [RFC6037] [RFC6513] and [RFC6514]. In particular, Data MDT or I/S-PMSI A-D routes are used to announce the binding of overlay flows to underlay tunnels. The binding could be inclusive (one tunnel for all flows in a VPN and to all egress PEs - referred to as an inclusive tunnel) or selective (one tunnel for one or more flows and only to the egress PEs that need to receive the traffic - referred to as a selective tunnel). While selective binding prevents multicast traffic from being sent to PEs that do not need to receive traffic, inclusive binding can significantly reduce the number of tunnels needed (only one tunnel is used for each VPN but traffic is sent to all the PEs of the VPN).

Two other features of BGP-MVPN can further reduce the number of underlay tunnels. One is to use the same tunnel for flows in different VPNs (referred to as tunnel aggregation in this document). This can be done for all flows or just for some selective flows in the VPNs, achieved by advertising a de-multiplex VPN label in the PMSI Tunnel Attribute (PTA) attached to the PMSI A-D routes and imposing the de-multiplex label before imposing the tunnel label to the traffic. The other is inter-as aggregation where per-PE inclusive tunnels are confined to the local AS while per-AS inclusive tunnels are used outside the local AS. This is achieved with Inter-AS I-PMSI A-D routes [RFC6514] and the concept is further explained and extended to per-Region Aggregation (Section 6.2 of [I-D.ietf-bess-evpn-bum-procedure-updates]).

The same BGP-MVPN procedures can be applied to Global Table Multicast (GTM) for multicast in the default/master routing instance over an underlay network, as specified in [RFC7716].

Note that this section is about MVPN as Flow Overlay Signaling for any kind of tunneling technology including BIER [RFC8556], and the next section is about two flow overlay signaling methods specifically for BIER.

5.2. PIM/mLDP as Flow Overlay Signaling over BIER

Consider an E2E IP multicast tree with part of it tunneled over BIER. The flow overlay signaling can be PIM, which is both the signaling protocol for the E2E tree and the overlay signaling protocol for tunneling over BIER [I-D.ietf-bier-pim-signaling].

The above applies to IP multicast in both the default/global routing instance and in VRFs in case of MVPN, and it replaces PIM-MVPN [RFC6037] [RFC6513] in this context.

Similarly, an E2E mLDP tree can have part of it tunneled over BIER. In this case, the flow overlay signaling is mLDP as specified in [I-D.ietf-bier-mldp-signaling-over-bier].

Note that, "tunnel segmentation" is originally documented in BGP-MVPN (Section 4) and it refers to that a PE-PE multicast provider tunnel can be instantiated with different types/instances of tunnels in different ASes/regions. Strictly speaking, an IP multicast tree going over different tunnels in different regions is different from the MVPN "tunnel segmentation". However, in the rest of the document we use "tunnel segmentation" for both situations.

5.3. Flow Overlay Signaling and Tunnel Segmentation

In case of BGP-MVPN as overlay signaling (for any kind of tunnels - see Section 5.1) with tunnel segmentation, the segmentation points maintain overlay state in the form of I-PMSI A-D routes that are for all flows or S-PMSI A-D routes for specific flows, instead of overlay (e.g., IP) multicast tree state.

In case of PIM/mLDP as BIER flow overlay signaling (Section 5.2), when the BIER domain is divided into multiple sub-domains for segmentation purpose, the overlay state that the segmentation points maintain is the overlay multicast tree state itself (i.e., IP multicast tree state or mLDP tree state).

6. Overall Considerations for Multicast Scaling

With all the background laid out above, we have the following observations and considerations.

6.1. Observations

For a massive number of flows to reach a massive number of receivers, the best solution is IP multicast (Section 1.1) carried over tunnels (Section 1.2).

With massive number of receivers and a large span of an E2E network, an E2E multicast overlay flow may need to be carried by multiple tunnels/segments of different types or instances in different parts of the E2E network (Section 1.2, Section 4).

Even if BIER/BIER-TE/CGM2 is supported on all devices E2E, it may be impractical to encode all BFERs or BIER-TE/CGM2 replication branches in a single BitString (flat or recursive), and sending multiple copies is just not efficient hence undesired or impractical. Tunnel segmentation needs to be used so that different sub-domains can be used in different regions of the E2E network, and that requires the overlay state/signaling at the segmentation points.

However, the segmentation points may become the scaling bottleneck due to the overlay state/signaling. That may be mitigated by solutions described below.

6.2. Considerations

As observed above, to massively scale multicast in both dimensions (number of receivers and number of flows) and over a vast network, IP multicast over underlay tunnel segments is needed. This section focuses on how to reduce the number of underlay tunnels and how to scale segmentation points.

6.2.1. Reduce Number of Underlay Tunnels or Amount of Tunnel Binding Signaling

As discussed in Section 5.1, underlay tunnels can be reduced by use of following:

  • Inclusive tunnels
  • Tunnel aggregation
  • Per-region aggregation

Notice that while they all reduce the number of underlay tunnels (which is useful, otherwise the underlay network need to maintain more tunnel state unless BIER/BIER-TE/CGM2 is used), tunnel aggregation does not reduce the overlay signaling of binding between tunnel and overlay flows.

Therefore, if segmented selective tunnels are used, or if there are just too many VPNs to support (such that even the number of I-PMSI A-D routes is too large), segmentation point scaling is still needed as discussed below.

On the other hand, the trend of multicast application seems to be that it is mainly for large scale delivery of real-time high rate data that most likely need to be sent everywhere, e.g., broadcasting of World Cup Soccer or Chinese Spring Festival Gala. In this case inclusive tunnels in a few VPNs may very well suffice.

Note that, while [RFC6514] and [RFC6625] specifies that the source/group length fields of S-PMSI A-D routes to be either 0 (wildcard) or 32/128 (IPv4/IPv6 host address), it could be extended to have lengths between 0 and 32/128. With that, a set of overlay flows not only can share the same underlay tunnel but can also share the same S-PMSI A-D route.

Part of an underlay tunnel itself can be stacked on top of another multicast tunnel. When multiple upper tunnels stack on top of the same lower tunnel, the lower tunnel's domain does not need to maintain state for the upper tunnels. Examples include mLDP over RSVP-TE/mLDP/other P2MP tunnels ([RFC7060], mLDP over MVPN [RFC6514] where mLDP is the C-multicast protocol, and mLDP over BIER [I-D.ietf-bier-mldp-signaling-over-bier].

6.2.2. Scale Up Segmentation Points

The scaling burden on a segmentation point falls on both control plane (for overlay signaling) and forwarding plane. While a typical router implementation supports much lower number of multicast flows than unicast ones, it does not mean the multicast scale cannot be increased.

6.2.2.1. Forwarding Plane Scaling

A modern edge router can handle hundreds of thousands if not millions of (unicast) routes in the forwarding plane. For multicast forwarding, the scaling property is actually similar to the unicase case - the only difference is that the lookup key in case of IP multicast is (source, group/destination) pair instead of just a destination address. The forwarding action is based on the forwarding instructions (e.g. a set of replication branches), referred to as forwarding nexthop associated with the route that is looked up.

On the other hand, while many multicast routes can have the same forwarding nexthop just like in unicast case, on segmentation points the multicast forwarding nexthop sharing is limited as explained in Section 2.1 of [I-D.zzhang-bess-mvpn-evpn-segmented-forwarding].

Therefore, a modern router designed with multicast scaling in consideration (e.g. has enough memory for multicast state especially more unshared forwarding nexthop) should be able to handle hundreds of thousands if not millions of multicast flows.

Of course, scaling must always be considered together with convergence. For example, when something happens, how fast can the FIB state be updated. This is further discussed below.

6.2.2.2. Control Plane Scaling

While it might be challenging for some PIM [RFC7761] implementations to handle hundreds of thousands of multicast flows due to the following reasons:

  • Periodic refreshes due to soft state nature
  • Potentially huge number of impacted flows when an interface goes up/down

Overlay multicast scaling does not have to rely on PIM (so no soft state refresh problem) and is less prone to topology changes.

Taking the example of BGP-MVPN [RFC6514] as the overlay protocol, overlay multicast interest is signaled with BGP-MVPN type-6/7 (C-Multicast Auto-Discovery or A-D) routes, augmented with type-4 (Leaf A-D) routes in case of selective IR/BIER tunnels. None of the above-mentioned scaling concerns are applicable here.

It is possible that an underlay topology change could impact some underlay tunnels. A good implementation should be able to minimize the impact to the overlay forwarding state. In the BIER case, no overlay forwarding state needs to be changed at all.

6.2.3. Scale out Segmentation Points

Consider that between two segmentation regions there are 2N segmentation points (Regional Border Routers, or RBRs) in parallel. They are divided into N pairs where each pair is responsible for 1/N of overlay flows (a pair is used for redundancy purposes). By increasing the number N, we can scale out the segmentation points (whether scale up is done at the same time or not).

7. Summary

As discussed above, existing and in-development multicast solutions, when implemented properly and deployed with appropriate combinations, can scale very well to massive number of multicast flows with massive number of receivers over a vast network:

  • Use IP multicast to scale for massive number of receivers (Section 1.1)
  • Use tunneling to scale for massive number of flows (Section 1.2)
  • Use inclusive tunnels and/or aggregation to reduce the number of underlay tunnels needed (Section 5.1)
  • Use BIER to achieve efficient replication without requiring per-tree state (Section 3.1)
  • Use BIER-TE/CGM2 to achieve per-flow traffic steering without requiring per-tree state (Section 3.2, Section 3.3)
  • Use tunnel segmentation to scale for a vast network that may deploy different types/instances of tunnels in different regions (Section 4)
  • Scale up (Section 6.2.2) and scale out (Section 6.2.3) tunnel segmentation points

It's worth pointing out that the above is independent of IPv4/IPv6 or MPLS/SRv6 data planes.

8. Informative References

[I-D.chen-pim-srv6-p2mp-path]
Chen, H., McBride, M., Fan, Y., Li, Z., Geng, X., Toy, M., Gyan Mishra, S., Wang, A., Liu, L., and X. Liu, "Stateless SRv6 Point-to-Multipoint Path", Work in Progress, Internet-Draft, draft-chen-pim-srv6-p2mp-path-06, , <https://www.ietf.org/archive/id/draft-chen-pim-srv6-p2mp-path-06.txt>.
[I-D.eckert-bier-cgm2-rbs]
Eckert, T. and B. (. Xu, "Carrier Grade Minimalist Multicast (CGM2) using Bit Index Explicit Replication (BIER) with Recursive BitString Structure (RBS) Addresses", Work in Progress, Internet-Draft, draft-eckert-bier-cgm2-rbs-01, , <https://www.ietf.org/archive/id/draft-eckert-bier-cgm2-rbs-01.txt>.
[I-D.ietf-bess-evpn-bum-procedure-updates]
Zhang, Z., Lin, W., Rabadan, J., Patel, K., and A. Sajassi, "Updates on EVPN BUM Procedures", Work in Progress, Internet-Draft, draft-ietf-bess-evpn-bum-procedure-updates-14, , <https://www.ietf.org/archive/id/draft-ietf-bess-evpn-bum-procedure-updates-14.txt>.
[I-D.ietf-bier-mldp-signaling-over-bier]
Bidgoli, H., Kotalwar, J., Wijnands, I., Mishra, M., Zhang, Z., and E. Leyton, "M-LDP Signaling Through BIER Core", Work in Progress, Internet-Draft, draft-ietf-bier-mldp-signaling-over-bier-01, , <https://www.ietf.org/archive/id/draft-ietf-bier-mldp-signaling-over-bier-01.txt>.
[I-D.ietf-bier-pim-signaling]
Bidgoli, H., Xu, F., Kotalwar, J., Wijnands, I., Mishra, M., and Z. Zhang, "PIM Signaling Through BIER Core", Work in Progress, Internet-Draft, draft-ietf-bier-pim-signaling-12, , <https://www.ietf.org/archive/id/draft-ietf-bier-pim-signaling-12.txt>.
[I-D.ietf-bier-te-arch]
Eckert, T., Menth, M., and G. Cauchie, "Tree Engineering for Bit Index Explicit Replication (BIER-TE)", Work in Progress, Internet-Draft, draft-ietf-bier-te-arch-13, , <https://www.ietf.org/archive/id/draft-ietf-bier-te-arch-13.txt>.
[I-D.ietf-pim-sr-p2mp-policy]
(editor), D. V., Filsfils, C., Parekh, R., Bidgoli, H., and Z. Zhang, "Segment Routing Point-to-Multipoint Policy", Work in Progress, Internet-Draft, draft-ietf-pim-sr-p2mp-policy-05, , <https://www.ietf.org/archive/id/draft-ietf-pim-sr-p2mp-policy-05.txt>.
[I-D.zzhang-bess-mvpn-evpn-segmented-forwarding]
Zhang, Z. and J. Xie, "MVPN/EVPN Segmentated Forwarding Options", Work in Progress, Internet-Draft, draft-zzhang-bess-mvpn-evpn-segmented-forwarding-00, , <https://www.ietf.org/archive/id/draft-zzhang-bess-mvpn-evpn-segmented-forwarding-00.txt>.
[RFC1112]
Deering, S., "Host extensions for IP multicasting", STD 5, RFC 1112, DOI 10.17487/RFC1112, , <https://www.rfc-editor.org/info/rfc1112>.
[RFC4875]
Aggarwal, R., Ed., Papadimitriou, D., Ed., and S. Yasukawa, Ed., "Extensions to Resource Reservation Protocol - Traffic Engineering (RSVP-TE) for Point-to-Multipoint TE Label Switched Paths (LSPs)", RFC 4875, DOI 10.17487/RFC4875, , <https://www.rfc-editor.org/info/rfc4875>.
[RFC5015]
Handley, M., Kouvelas, I., Speakman, T., and L. Vicisano, "Bidirectional Protocol Independent Multicast (BIDIR-PIM)", RFC 5015, DOI 10.17487/RFC5015, , <https://www.rfc-editor.org/info/rfc5015>.
[RFC6037]
Rosen, E., Ed., Cai, Y., Ed., and IJ. Wijnands, "Cisco Systems' Solution for Multicast in BGP/MPLS IP VPNs", RFC 6037, DOI 10.17487/RFC6037, , <https://www.rfc-editor.org/info/rfc6037>.
[RFC6388]
Wijnands, IJ., Ed., Minei, I., Ed., Kompella, K., and B. Thomas, "Label Distribution Protocol Extensions for Point-to-Multipoint and Multipoint-to-Multipoint Label Switched Paths", RFC 6388, DOI 10.17487/RFC6388, , <https://www.rfc-editor.org/info/rfc6388>.
[RFC6513]
Rosen, E., Ed. and R. Aggarwal, Ed., "Multicast in MPLS/BGP IP VPNs", RFC 6513, DOI 10.17487/RFC6513, , <https://www.rfc-editor.org/info/rfc6513>.
[RFC6514]
Aggarwal, R., Rosen, E., Morin, T., and Y. Rekhter, "BGP Encodings and Procedures for Multicast in MPLS/BGP IP VPNs", RFC 6514, DOI 10.17487/RFC6514, , <https://www.rfc-editor.org/info/rfc6514>.
[RFC6625]
Rosen, E., Ed., Rekhter, Y., Ed., Hendrickx, W., and R. Qiu, "Wildcards in Multicast VPN Auto-Discovery Routes", RFC 6625, DOI 10.17487/RFC6625, , <https://www.rfc-editor.org/info/rfc6625>.
[RFC7060]
Napierala, M., Rosen, E., and IJ. Wijnands, "Using LDP Multipoint Extensions on Targeted LDP Sessions", RFC 7060, DOI 10.17487/RFC7060, , <https://www.rfc-editor.org/info/rfc7060>.
[RFC7524]
Rekhter, Y., Rosen, E., Aggarwal, R., Morin, T., Grosclaude, I., Leymann, N., and S. Saad, "Inter-Area Point-to-Multipoint (P2MP) Segmented Label Switched Paths (LSPs)", RFC 7524, DOI 10.17487/RFC7524, , <https://www.rfc-editor.org/info/rfc7524>.
[RFC7716]
Zhang, J., Giuliano, L., Rosen, E., Ed., Subramanian, K., and D. Pacella, "Global Table Multicast with BGP Multicast VPN (BGP-MVPN) Procedures", RFC 7716, DOI 10.17487/RFC7716, , <https://www.rfc-editor.org/info/rfc7716>.
[RFC7761]
Fenner, B., Handley, M., Holbrook, H., Kouvelas, I., Parekh, R., Zhang, Z., and L. Zheng, "Protocol Independent Multicast - Sparse Mode (PIM-SM): Protocol Specification (Revised)", STD 83, RFC 7761, DOI 10.17487/RFC7761, , <https://www.rfc-editor.org/info/rfc7761>.
[RFC7988]
Rosen, E., Ed., Subramanian, K., and Z. Zhang, "Ingress Replication Tunnels in Multicast VPN", RFC 7988, DOI 10.17487/RFC7988, , <https://www.rfc-editor.org/info/rfc7988>.
[RFC8279]
Wijnands, IJ., Ed., Rosen, E., Ed., Dolganow, A., Przygienda, T., and S. Aldrin, "Multicast Using Bit Index Explicit Replication (BIER)", RFC 8279, DOI 10.17487/RFC8279, , <https://www.rfc-editor.org/info/rfc8279>.
[RFC8556]
Rosen, E., Ed., Sivakumar, M., Przygienda, T., Aldrin, S., and A. Dolganow, "Multicast VPN Using Bit Index Explicit Replication (BIER)", RFC 8556, DOI 10.17487/RFC8556, , <https://www.rfc-editor.org/info/rfc8556>.

Authors' Addresses

Zhaohui Zhang
Juniper Networks
Rishabh Parekh
Cisco
Hooman Bidgoli
Nokia
Zheng Zhang
ZTE