Triple Play multicast
Introduction to multicast
IP multicast provides an effective method of many-to-many communication. Delivering unicast datagrams is simple. Normally, IP packets are sent from a single source to a single recipient. The source inserts the address of the target host in the IP header destination field of an IP datagram, intermediate routers (if present) simply forward the datagram toward the target in accordance with their respective routing tables.
Sometimes distribution needs individual IP packets be delivered to multiple destinations (like audio or video streaming broadcasts). Multicast is a method of distributing datagrams sourced from one (or possibly more) hosts to a set of receivers that may be distributed over different (sub) networks. This makes delivery of multicast datagrams significantly more complex.
Multicast sources can send a single copy of data using a single address for the entire group of recipients. The routers between the source and recipients route the data using the group address route. Multicast packets are delivered to a multicast group. A multicast group specifies a set of recipients who are interested in a data stream and is represented by an IP address from a specified range. Data addressed to the IP address is forwarded to the members of the group. A source host sends data to a multicast group by specifying the multicast group address in datagram’s destination IP address. A source does not have to register to send data to a group nor do they need to be a member of the group.
Routers and Layer 3 switches use the Internet Group Management Protocol (IGMP) and Multicast Listener Discovery (MLD) to manage membership for a multicast session. When a host wants to receive one or more multicast sessions, it sends a join message for each multicast group it wants to join. When a host wants to leave a multicast group, it sends a leave message.
Multicast in the broadband service router
This section describes the multicast protocols employed when a Nokia router is used as a Broadband Service Router (BSR) in a Triple Play aggregation network.
The protocols used are:
Internet Group Management Protocol (Internet Group Management Protocol)
Multicast Listener Discovery (Multicast Listener Discovery)
Source Specific Multicast Groups (Internet Group Management Protocol)
Protocol Independent Multicast (Sparse Mode) (Protocol Independent Multicast Sparse Mode)
Internet Group Management Protocol
Internet Group Management Protocol (IGMP) is used by IPv4 hosts and routers to report their IP multicast group memberships to neighboring multicast routers. A multicast router keeps a list of multicast group memberships for each attached network, and a timer for each membership.
Multicast group memberships include at least one member of a multicast group on a given attached network, not a list of all the members. With respect to each of its attached networks, a multicast router can assume one of two roles, querier or non-querier. There is normally only one querier per physical network.
A querier issues two types of queries, a general query and a group-specific query. General queries are issued to solicit membership information about any multicast group. Group-specific queries are issued when a router receives a leave message from the node it perceives as the last group member remaining on that network segment.
Hosts wanting to receive a multicast session issue a multicast group membership report. These reports must be sent to all multicast enabled routers.
IGMP versions and interoperability requirements
If routers run different versions of IGMP, they negotiates the lowest common version of IGMP that is supported on their subnet and operate in that version.
Version 1
Specified in RFC 1112, Host extensions for IP Multicasting, was the first widely deployed version and the first version to become an Internet standard.
Version 2
Specified in RFC 2236, Internet Group Management Protocol, added support for low leave latency, that is, a reduction in the time it takes for a multicast router to learn that there are no longer any members of a group present on an attached network.
Version 3
Specified in RFC 3376, Internet Group Management Protocol, adds support for source filtering, that is, the ability for a system to report interest in receiving packets only from specific source addresses, as required to support Source-Specific Multicast (See Source Specific Multicast (SSM)), or from all but specific source addresses, sent to a multicast address.
IGMPv3 must keep state per group per attached network. This group state consists of a filter-mode, a list of sources, and various timers. For each attached network running IGMP, a multicast router records the necessary reception state for that network.
IGMP version transition
Nokia’s SRs are capable of interoperating with routers and hosts running IGMPv1, IGMPv2, or IGMPv3. Draft-ietf-magma-igmpv3-and-routing-0x.txt explores some of the interoperability issues and how they affect the various routing protocols.
IGMP version 3 specifies that if at any point a router receives an older version query message on an interface that it must immediately switch into a compatibility mode with that earlier version. Because none of the previous versions of IGMP are source aware, should this occur and the interface switch to Version 1 or 2 compatibility mode, any previously learned group memberships with specific sources (learned from the IGMPv3 specific INCLUDE or EXCLUDE mechanisms) must be converted to non-source specific group memberships. The routing protocol then treats this as if there is no EXCLUDE definition present.
Multicast Listener Discovery
Multicast Listener Discovery (MLD) is used by IPv6 hosts and routers to report their IP multicast group memberships to neighboring multicast routers. A multicast router keeps a list of multicast group memberships for each attached network, and a timer for each membership. Multicast group memberships include at least one member of a multicast group on a specific attached network, not a list of all the members. With respect to each of its attached networks, a multicast router can assume one of two roles, querier or non-querier. There is normally only one querier per physical network.
A querier issues two types of queries, a general query and a group-specific query. General queries are issued to solicit membership information about any multicast group. Group-specific queries are issued when a router receives a leave message from the node it perceives as the last group member remaining on that network segment.
Hosts wanting to receive a multicast session issue a multicast group membership report. These reports must be sent to all multicast enabled routers.
MLD versions and interoperability requirements
If routers run different versions of MLD, they negotiates the lowest common version of MLD that is supported on their subnet and operate in that version.
Version 1
Specified in RFC 2710, Multicast Listener Discovery (MLD) for IPv6, was the first deployed version and included low leave latency, that is, a reduction in the time it takes for a multicast router to learn that there are no longer any members of a group present on an attached network.
Version 2
Specified in RFC 3810, Multicast Listener Discovery Version 2 (MLDv2) for IPv6, adds support for source filtering, that is, the ability for a system to report interest in receiving packets only from specific source addresses, as required to support Source-Specific Multicast.
Multicast (SSM)), or from all but specific source addresses, sent to a multicast address. MLDv2 must keep state per group per attached network. This group state consists of a filter mode, a list of sources, and various timers. For each attached network running MLD, a multicast router records the wanted reception state for that network.
Source-specific multicast groups
IGMPv3and MLDv2 allows a receiver to join a group and specify that it only wants to receive traffic for a group if that traffic comes from a specific source. If a receiver does this, and no other receiver on the LAN requires all the traffic for the group, then the Designated Router (DR) can omit performing a (*,G) join to set up the shared tree, and instead issue a source-specific (S,G) join only.
For IPv4, the range of multicast addresses from 232.0.0.0 to 232.255.255.255 is currently set aside for source-specific multicast. For groups in this range, receivers should only issue source specific IGMPv3 joins. If a PIM router receives a non-source-specific join for a group in this range, it should ignore it.
For IPv6, the multicast prefix FF3x::/32 is currently set aside for source-specific multicast. For groups in this range, receivers should only issue source specific MLDv2 joins. If a PIM router receives a non-source-specific join for a group in this range, it should ignore it.
A Nokia PIM router must silently ignore a received (*, G) PIM join message where G is a multicast group address from the multicast address group range that has been explicitly configured for SSM. This occurrence should generate an event. If configured, the IGMPv2 (MLDv1for IPv6) request can be translated into IGMPv3 (MLDv2 for IPv6). The SR allows for the conversion of an IGMPv2 (*,G) request into a IGMPv3 (S,G) request based on manual entries. A maximum of 32 SSM ranges is supported.
IGMPv3 and MLDv2 also allows a receiver to join a group and specify that it only wants to receive traffic for a group if that traffic does not come from a specific source or sources. In this case, the DR performs a (*,G) join as normal, but can combine this with a prune for each of the sources the receiver does not want to receive.
Protocol Independent Multicast Sparse Mode
Protocol Independent Multicast Sparse Mode (PIM-SM) leverages the unicast routing protocols that are used to create the unicast routing table: OSPF, IS-IS, BGP, and static routes. Because PIM uses this unicast routing information to perform the multicast forwarding function it is effectively IP protocol independent. Unlike DVMRP, PIM does not send multicast routing tables updates to its neighbors.
PIM-SM uses the unicast routing table to perform the Reverse Path Forwarding (RPF) check function instead of building up a completely independent multicast routing table.
PIM-SM only forwards data to network segments with active receivers that have explicitly requested the multicast group. PIM-SM in the ASM model initially uses a shared tree to distribute information about active sources. Depending on the configuration options, the traffic can remain on the shared tree or switch over to an optimized source distribution tree. As multicast traffic starts to flow down the shared tree, routers along the path determine if there is a better path to the source. If a more direct path exists, then the router closest to the receiver sends a join message toward the source and then reroutes the traffic along this path.
As stated above, PIM-SM relies on an underlying topology-gathering protocol to populate a routing table with routes. This routing table is called the Multicast Routing Information Base (MRIB). The routes in this table can be taken directly from the unicast routing table, or it can be different and provided by a separate routing protocol such as MBGP. Regardless of how it is created, the primary role of the MRIB in the PIM-SM protocol is to provide the next hop router along a multicast-capable path to each destination subnet. The MRIB is used to determine the next hop neighbor to whom any PIM join/prune message is sent. Data flows along the reverse path of the join messages. Thus, in contrast to the unicast RIB that specifies the next hop that a data packet would take to get to some subnet, the MRIB gives reverse-path information, and indicates the path that a multicast data packet would take from its origin subnet to the router that has the MRIB.
Ingress multicast Path Management (IMPM) enhancements
See the Advanced Configuration Guide for more information about IMPM as well as detailed configuration examples.
IMPM allows the system to manage Layer 2 and Layer 3 IP multicast flows by sorting them into the available multicast paths through the switch fabric. The ingress multicast manager tracks the amount of available multicast bandwidth per path and the amount of bandwidth used per IP multicast stream. The following traffic is managed by IMPM when enabled:
IPv4 and IPv6 routed multicast traffic
VPLS IGMP snooping traffic
VPLS PIM snooping for IPv4 traffic
VPLS PIM snooping for IPv6 traffic (only when sg-based forwarding is configured)
Two policies define how each path should be managed, the bandwidth policy, and how multicast channels compete for the available bandwidth, the multicast information policy.
Chassis multicast planes should not be confused with IOM/IMM multicast paths. The IOM/IMM uses multicast paths to reach multicast planes on the switch fabric. An IOM/IMM may have less or more multicast paths than the number of multicast planes available in the chassis.
Each IOM/IMM multicast path is either a primary or secondary path type. The path type indicates the multicast scheduling priority within the switch fabric. Multicast flows sent on primary paths are scheduled at multicast high priority while secondary paths are associated with multicast low priority.
The system determines the number of primary and secondary paths from each IOM/IMM forwarding plane and distributes them as equally as possible between the available switch fabric multicast planes. Each multicast plane may terminate multiple paths of both the primary and secondary types.
The system ingress multicast management module evaluates the ingress multicast flows from each ingress forwarding plane and determines the best multicast path for the flow. A specific path can be used until the terminating multicast plane is ‟maxed” out (based on the rate limit defined in the per-mcast-plane-capacity commands) at which time either flows are moved to other paths or potentially blackholed (flows with the lowest preference are dropped first). In this way, the system makes the best use of the available multicast capacity without congesting individual multicast planes.
The switch fabric is simultaneously handling both unicast and multicast flows. The switch fabric uses a weighted scheduling scheme between multicast high, unicast high, multicast low and unicast low when deciding which cell to forward to the egress forwarding plane next. The weighted mechanism allows some amount of unicast and lower priority multicast (secondary) to drain on the egress switch fabric links used by each multicast plane. The amount is variable based on the number of switch fabric planes available on the amount of traffic attempting to use the fabric planes. The per-mcast-plane-capacity commands allows the amount of managed multicast traffic to be tuned to compensate for the expected available egress multicast bandwidth per multicast plane. In conditions where it is highly desirable to prevent multicast plane congestion, the per-mcast-plane-capacity commands should be used to compensate for the non-multicast or secondary multicast switch fabric traffic.
Multicast in the BSA
IP Multicast is normally not a function of the Broadband Service Aggregator (BSA) in a Triple Play aggregation network being a Layer 2 device. However, the BSA does use IGMP snooping to optimize bandwidth utilization.
IGMP snooping
For most Layer 2 switches, multicast traffic is treated like an unknown MAC address or broadcast frame, which causes the incoming frame to be flooded out (broadcast) on every port within a VLAN. While this is acceptable behavior for unknowns and broadcasts, as IP Multicast hosts may join and be interested in only specific multicast groups, all this flooded traffic results in wasted bandwidth on network segments and end stations.
IGMP snooping entails using information in layer 3 protocol headers of multicast control messages to determine the processing at layer 2. By doing so, an IGMP snooping switch provides the benefit of conserving bandwidth on those segments of the network where no node has expressed interest in receiving packets addressed to the group address.
On the 7450 ESS and 7750 SR, IGMP snooping can be enabled in the context of VPLS services. The IGMP snooping feature allows for optimization of the multicast data flow for a group within a service to only those Service Access Points (SAPs) and Service Distribution Points (SDPs) that are members of the group. In fact, the 7450 ESS and 7750 SR implementation performs more than pure snooping of IGMP data, because it also summarizes upstream IGMP reports and responds to downstream queries.
The 7450 ESS and 7750 SR maintain several multicast databases:
A port database on each SAP and SDP lists the multicast groups that are active on this SAP or SDP.
All port databases are compiled into a central proxy database. Towards the multicast routers, summarized group membership reports are sent based on the information in the proxy database.
The information in the different port databases is also used to compile the multicast forwarding information base (MFIB). This contains the active SAPs and SDPs for every combination of source router and group address (S,G), and is used for the actual multicast replication and forwarding.
When the router receives a join report from a host for a multicast group, it adds the group to the port database and (if it is a new group) to the proxy database. It also adds the SAP or SDP to existing (S,G) in the MFIB, or builds a new MFIB entry.
When the router receives a leave report from a host, it first checks if other devices on the SAP or SDP still want to receive the group (unless fast leave is enabled). Then it removes the group from the port database, and from the proxy database if it was the only receiver of the group. The router also deletes entries if it does not receive periodic membership confirmations from the hosts.
The fast leave feature finds its use in multicast TV delivery systems, for example. Fast Leave speeds up the membership leave process by terminating the multicast session immediately, instead of the standard procedure of issuing a group specific query to check if other group members are present on the SAP or SDP.
IGMP/MLD message processing
IGMP/MLD message processing illustrates the basic IGMP message processing in several situations.
IGMP message processing
Scenario A: A host joins a multicast group (TV channel) which is not yet being received by other hosts on the router, and therefore, is not yet present in the proxy database. The 7450 ESS or 7750 SR adds the group to the proxy database and sends a new IGMP Join group-specific membership report upstream to the multicast router.
Scenario B: A host joins a channel which is already being received by one or more hosts on the router, and is already present in the proxy database. No upstream IGMP report is generated by the router.
Scenario C: The multicast router periodically sends IGMP queries to the router, requesting it to respond with generic membership reports. Upon receiving such a query, the router compiles a report from its proxy database and send it back to the multicast router.
In addition, the router floods the received IGMP query to all hosts (on SAPs and spoke SDPs), and updates its proxy database based on the membership reports received back.
Scenario D: A host leaves a channel by sending an IGMP leave message. If fast-leave is not enabled, the router first checks whether there are other hosts on the same SAP or spoke SDP by sending a query. If no other host responds, the 7450 ESS or 7750 SR removes the channel from the SAP. In addition, if there are no other SAPs or spoke SDPs with hosts subscribing to the same channel, the channel is removed from the proxy database and an IGMP leave report is sent to the upstream Multicast Router.
Scenario E: A host leaves a channel by sending an IGMP leave message. If fast-leave is not enabled, the router checks whether there are other hosts on the same SAP or spoke SDP by sending a query. Another device on the same SAP or spoke SDP still wants to receive the channel and responds with a membership report. The router does not remove the channel from the SAP.
Scenario F: A host leaves a channel by sending an IGMP leave report. Fast-leave is enabled, so the router does not check whether there are other hosts on the same SAP or spoke SDP but immediately removes the group from the SAP. In addition, if there are no other SAPs or spoke SDPs with hosts subscribing to the same group, the group is removed from the proxy database and an IGMP leave report is sent to the upstream multicast router.
MLD message processing
MLD message processing differs from IGMP. An IPv6 host can have two WAN IPv6 addresses and an IPv6 prefix. MLD messages source address are link local addresses. This makes it difficult to know if the originating host is a WAN host or a PD host. By default, all requested IPv6 (s,g) are first associated with a WAN host. If this WAN host disconnects or ends its IPv6 session, the (s,g) is then associated with the remaining WAN host. If there are no more WAN hosts, the (s,g) is then associated with the remaining PD host. The (s,g) is always transferred to the remaining IPv6 host until there are no more report replies to corresponding to the queries. Scenarios A — F do not differ for IPv6 hosts.
IGMP/MLD filtering
A provider may want to block receive or transmit permission to individual hosts or a range of hosts. To this end, the 7450 ESS and 7750 SR support IGMP/MLD filtering. Two types of filter can be defined:
Filter IGMP/MLD membership reports from a host or range of hosts. This is performed by importing an appropriately defined routing policy into the SAP or spoke SDP.
Filter to prevent a host from transmitting multicast streams into the network. The operator can define a data-plane filter (ACL) which drops all multicast traffic, and apply this filter to a SAP or spoke SDP.
Multicast VPLS Registration (MVR)
Multicast VPLS Registration (MVR) is a bandwidth optimization method for multicast in a broadband services network. MVR allows a subscriber on a port to subscribe and unsubscribe to a multicast stream on one or more network-wide multicast VPLS instances.
MVR assumes that subscribers join and leave multicast streams by sending IGMP join and leave messages. The IGMP leave and join message are sent inside the VPLS to which the subscriber port is assigned. The multicast VPLS is shared in the network while the subscribers remain in separate VPLS services. Using MVR, users on different VPLS cannot exchange any information between them, but still multicast services are provided.
On the MVR VPLS, IGMP snooping must be enabled. On the user VPLS, IGMP snooping and MVR work independently. If IGMP snooping and MVR are both enabled, MVR reacts only to join and leave messages from multicast groups configured under MVR. Join and leave messages from all other multicast groups are managed by IGMP snooping in the local VPLS. This way, potentially several MVR VPLS instances could be configured, each with its own set of multicast channels.
MVR by proxy — In some situations, the multicast traffic should not be copied from the MVR VPLS to the SAP on which the IGMP message was received (standard MVR behavior) but to another SAP. This is called MVR by proxy.
MVR and MVR by proxy shows a MVR and MVR by proxy configuration.
Layer 3 multicast load balancing
Layer 3 multicast load balancing establishes a more efficient distribution of Layer 3 multicast data over ECMP and LAG links. Operators have the option to redistribute multicast groups over ECMP or LAG links if the number of links changes either up or down.
When implementing this feature, there are several considerations. When multicast load balancing is not configured, the distribution remains as is. Multicast load balancing is based on the number of ‟s,g” groups. This means that bandwidth considerations are not considered. The multicast groups are distributed over the available links as joins are processed. When link failure occurs, the load is distributed on the failed channel to the remaining channels so multicast groups are evenly distributed over the remaining links. When a link is added (or failed link returned) all multicast joins on the added links are allocated until a balance is achieved.
When multicast load balancing is configured, but the channels are not found in the multicast-info-policy, then multicast load balancing is based on the number of ‟s,g” groups. This means that bandwidth considerations are not considered. The multicast groups are distributed over the available links as joins are processed. The multicast groups are evenly distributed over the remaining links. When link failure occurs, the load is distributed on the failed channel to the remaining channels. When a link is added (or failed link returned) all multicast joins on the added links are allocated until a balance is achieved. A manual redistribute command enables the operator to re-evaluate the current balance and, if required, move channels to different links to achieve a balance. A timed redistribute parameter allows the system to automatically, at regular intervals, redistribute multicast groups over available links. If no links have been added or removed from the ECMP/LAG interface, then no redistribution is attempted.
When multicast load balancing is configured, multicast groups are distributed over the available links as joins are processed based on bandwidth configured for the specified group address. If the bandwidth is not configured for the multicast stream then the configured default value is used.
If link failure occurs, the load is distributed on the failed channel to the remaining channels. The bandwidth required over each individual link is evenly distributed over the remaining links.
When an additional link is available for a specific multicast stream, then it is considered in all multicast stream additions applied to the interface. This multicast stream is included in the next scheduled automatic rebalance run. A rebalance run re-evaluates the current balance with regard to the bandwidth utilization and if required, move multicast streams to different links to achieve a balance.
A rebalance, either timed or executing the mc-ecmp-rebalance command, should be administered gradually to minimize the effect of the rebalancing process on the different multicast streams. If multicast re-balancing is disabled and subsequently enabled, keeping with the rebalance process, the gradual and least invasive method is used to minimize the effect of the changes to the customer.
By default multicast load balancing over ECMP links is enabled and set at 30 minutes.
The rebalance process can be executed as a low priority background task while control of the console is returned to the operator. When multicast load rebalancing is not enabled, then ECMP changes are not be optimized, however, when a link is added occurs an attempt is made to balance the number of multicast streams on the available ECMP links. This however may not result in balanced utilization of ECMP links.
Only a single mc-ecmp-rebalance command can be executed any specific time, if a rebalance is in progress and the command is entered, it is rejected with the message saying that a rebalance is already in progress. A low priority event is generated when an actual change for a specific multicast stream occurs as a result of the rebalance process.
IGMP state reporter
The target application for this feature is linear TV delivery. In some countries, wholesale Service Providers are obligated by the government regulation to provide information about channel viewership per subscriber to retailers.
A service provider (wholesaler or retailer) my use this information for:
billing purposes
market research/data mining to gain view into the most frequently watched channels, duration of the channel viewing, frequency of channel zapping by the time of the day, and so on
The information about channel viewership is based on IGMP states maintained per each subscriber host. Each event related to the IGMP state creation is recorded and formatted by the IGMP process. The formatted event is then sent to another task in the system (Exporter), which allocates a TX buffer and start a timer.
The event is then be written by the Exporter into the buffer. The buffer in essence corresponds to the packet that contains a single event or a set of events. Those events are transported as data records over UDP transport to an external collector node. The packet itself has a header followed by a set of TLV type data structures, each describing a unique filed within the IGMP event.
The packet is transmitted when it reaches a preconfigured size (1400 bytes), or when the timer expires, whichever comes first.
The receiving end (collector node) accepts the data on the destination UDP port. It must be aware of the data format so that it can interpret incoming data accordingly. The implementation details of the receiving node are outside of the scope of this description and are left to the network operator.
The IGMP state recording per subscriber host must be supported for hosts which are replicating multicast traffic directly as well as for those host that are only keeping track of IGMP states for the HQoS Adjustment purpose. The latter is implemented by redirection and not the Host Tracking (HT) feature as originally proposed. The IGMP reporting must differentiate events between direct replication and redirection.
It further distinguish events that are related to denial of IGMP state creation (because of filters, MCAC failure, and so on) and the ones that are related to removal of an already existing IGMP state in the system.
IGMP data records
Each IGMP state change generates a data record that is formatted by the IGMP task and written into the buffer. IGMP state transitions configured statically through CLI are not reported.
To minimize the size of the records when transported over the network, most fields in the data record are HEX coded (as opposed to ASCII descriptive strings).
Each data record has a common header as shown in Common IGMP data record header:
Application:
0x01 - IGMP
0x02 - IGMP Host Tracking Event:
Event:
Related to denial of a new state creation:
0x01
Join
0x02 (Join_Deny_Filter)
Join denied because of filtering by an import policy
0x03 (Join_Deny_CAC)
Join denied because of MCAC
0x04 (Join_Deny_MaxGrps)
Join denied because of maximum groups per host limit reached
0x05 (Join_Deny_MaxSrcs)
Join denied because of maximum sources limit reached
-
0x06 Join (Join_Deny_SysErr)
Join denied because of an internal error (for example: out of memory)
-
Related to removal of an existing IGMP state:
- 0x07
(Drop_Leave_Rx) IGMP state is removed because of the Leave message
-
0x08
(Drop_Expiry) IGMP state is removed because of the time out (by default 2*query_interval + query_response_interval = 260sec)
-
0x09
(Drop_Filter) IGMP state is removed because of the filter (import policy) change
-
0x0A
(Drop_CfgChange) IGMP state is removed because of a configuration change (clear grp, intf shutdown, PPPoE session goes unexpectedly down)
-
0x0B
(Drop_CAC) an existing stream is stopped because of a configuration change in MCAC
- 0x07
Length: The length of the entire data record (including the header and TLVs) in octets.
16 bit Sequence Number
Because IGMP Reporting is based on connectionless transport (UDP), a 16 bit sequence numbers are used in each data record so that data loss in the network can be tracked.
The 16 bit sequence number is located after the time stamp field. The sequence numbers increases sequentially from 0 to 65535 and then rolls over back to 0.
Timestamp: Timestamp is in UNIX format (32 bit integer in seconds because 01/01/1970) plus an extra 8 bits for 10msec resolution.
TLVs describing the IGMP state record has the following structure shown in Data record field TLV structure and the data record field descriptions in Data record field description:
Type | Description | Encoding/length | Mandatory/optional |
---|---|---|---|
0x02 |
Subscriber ID |
ASCII |
M |
0x03 |
Sub Host IP |
4 Bytes IPv4 |
M |
0x04 |
Mcast Group IP |
4 Bytes IPv4 |
M |
0x05 |
Mcast Source IP |
4 Bytes IPv4 |
M |
0x06 |
Host MAC |
6 Bytes |
M |
0x07 |
PPPoE Session-ID |
2 Bytes |
M |
0x08 |
Service ID |
4 Bytes |
M |
0x09 |
SAP ID |
ASCII |
M |
0x0A |
Redirection vRtrId |
4 Bytes |
M |
0x0B |
Redirection ifIndex |
4 Bytes |
M |
The redirection destination TLV is a mandatory TLV that is sent only in cases where redirection is enabled. It contains two 32 bit integer numbers. The first number identifies the VRF where IGMPs are redirected; the second number identifies the interface index.
Optional fields can be included in the data records according to the configuration.
In IGMPv3, if an IGMP message (Join or Leave) contains multiple multicast groups or a multicast group contains multiple IP sources, only a single event is generated per group-source combination. In other words, data records are transmitted with a single source IP address and multiple mcast group addresses or a single multicast group address with multiple source IP addresses, depending on the content of the IGMP message. (*,G)
Transport mechanism
Data is transported by a UDP socket. The destination IP address, the destination port and the source IP address are configurable. The default UDP source and destination port number is 1037.
Upon the arrival of an IGMP event, the Exporter allocates a buffer for the packet (if not already allocated) and starts writing the events into the buffer (packet). Along with the initial buffer creation, a timer is started. The trigger for the transmission of the packet is either the TX buffer being filled up to 1400B (hard coded value), or the timer expiry, whichever comes first.
The source IP address is configurable within GRT (by default system IP), and the destination IP address can be reachable only by GRT. The source IP address is modified in the system>security>source-address>application CLI hierarchy.
The receiving end (the collector node) collects the data and process them according to the formatting rules defined in this document. The capturing and processing of the data on the collector node is outside of the context of this description.
The processing node should have sufficient resources to accept and process packets that contain information about every IGMP state change for every host from a set of network BRASes that are transporting data to this collector node.
Multicast Reporter traffic is marked as BE (all 6 DSCP bits are set to 0) exiting our system.
HA compliance
IGMP events are synchronized between two CPMs before they are transported out of the system.
QoS awareness
IGMP Reporter is a client of sgt-qos so that DSCP or dot1p bits can be appropriately be marked when egressing the system.
IGMP reporting restrictions
The following are not supported:
Regular (non-subscriber) interfaces
SAM support as the collector device
Multicast support over subscriber interfaces in Routed CO model
Applications for multicast over subscriber interfaces in Routed CO ESM model can be divided in two main categories:
Residential customers where the driver applications are:
IPTV in an environment with legacy non-multicasting DSLAMs
Internet multicast where users connect to a multicast stream sourced from the Internet.
For the business customers, the main drivers are enterprise multicast and Internet multicast applications.
On multicast-capable ANs, a single copy of each multicast stream is delivered over a separate regular IP interface. AN would then perform the replication. This is how multicast would be deployed in Routed CO environment with the 7750 SR and 7450 ESS.
On legacy, non-multicast ANs, or in environments with low volume multicast traffic where it is not worth setting up a separate multicast topology (from BNG to AN), multicast replication is performed with subscriber-interfaces in the 7750 SR and 7450 ESS. There are differences in replicating multicast traffic on IPoE vs PPPoX which are described in subsequent sessions.
An example of a business connectivity model is shown in Typical business connectivity model.
In this example, HSI is terminated in a Global Routing Table (GRT) whereas VPRN services are terminated in Wholesale/Retail VPRN method, with each customer using a separate VPRN.
The actual connectivity model that is deployed depends on many operational aspects that are present in the customer environment.
Multicast over subscriber-interfaces in a Routed CO model is supported for both types of hosts, IPoE and PPPoE which can be simultaneously enabled on a shared SAP.
There are some fundamental differences in multicast behavior between two host types (IPoE and PPPoX). The differences are discussed further in the next sections.
Multicast over IPoE
There are several deployment scenarios for delivering multicast directly over subscriber hosts:
1:1 model (subscriber per VLAN/SAP) with the Access Node (AN) that is not IGMP/MLD aware.
N:1 model (service per VLAN/SAP) with the AN in the Snooping mode.
N:1 model with the AN in the Proxy mode.
N:1 model with the AN that is not IGMP/MLD aware.
There are two modes of operation for subscriber multicast that can be chosen to address the above mentioned deployment scenarios:
Per SAP replication — A single multicast stream per group is forwarded on any given SAP. Even if the SAP has a multicast group (channel) that is registered to multiple hosts, only a single copy of the multicast stream is forwarded over this SAP. The multicast stream has a multicast destination MAC address (as opposed to unicast). IGMP/MLD states are maintained per host. This is the default mode of operation.
Per subscriber host replication in this mode of operation, multicast is replicated per subscriber host even if this means that multiple copies of the same stream that are forwarded over the same SAP. For example, if two hosts on the same SAP are registered to receive the same multicast group (channel), then this multicast channel is replicated twice on the same SAP. The streams have a unique unicast destination MAC address (otherwise it would not make sense to replicate the streams twice).
In all deployment scenarios and modes of operation the IGMP/MLD states per source IP address of the incoming IGMP/MLD message is maintained. This source IP address might represent a subscriber hosts or the AN (proxy mode).
For MLD, the source IP address is the host link local address. Therefore, the MLD message is associated with all IP address/prefix of the host.
Per SAP replication mode
In the per SAP replication mode a single copy of the multicast channel is forwarded per SAP. In other words, if a subscriber (in 1:1 mode) or a group of subscribers (in N:1 mode) have multiple hosts and all of them are subscribed to the same multicast group (watching the same channel), then only a single copy of the multicast stream for that group is sent. The destination MAC address is always a multicast MAC (there are no conversion to unicast MAC address).
IGMP/MLD states are maintained per subscriber host and per SAP.
Per SAP queue
Multicast traffic over subscribers in a per-SAP replication mode flows through a SAP queue which is outside of the subscriber queues context. Sending the multicast traffic over the default SAP queue is characterized by:
the inability to classify multicast traffic into separate subscriber queues and therefore include it natively in the subscriber HQoS hierarchy; however, multicast traffic can be classified into specific (M-)SAP queues, assuming that such queues are enabled by (M-)SAP based QoS policy
redirection of multicast traffic by internal queues in case the SAP queue in subscriber environment is disabled (sub-sla-mgmt>single-sub-parameters>profiled-traffic-only); this is applicable only to 1:1 subscriber model
a possible necessity for HQoS Adjustment as multicast traffic is flowing outside of the subscriber queues
de-coupling of the multicast forwarding statistics from the overall subscriber forwarding statistics obtained by subscriber specific show commands
IPoE 1:1 model (subscriber per VLAN/SAP) — no IGMP/MLD in AN
The AN is not IGMP/MLD aware, all replications are performed in the BNG. From the BNG perspective this deployment model has the following characteristics:
IGMP/MLD states are kept per hosts and SAPs. Each host can be registered to more than one group.
MLD uses link-local as source address, which makes it difficult to associate with the originating host. MLD (S,G) are always first associated with a IPv6 WAN host if available. If this WAN host terminate its IPv6 session (by lease expiry, session terminate, and so on), the (S,G) is then associated with any remaining WAN host. Only when there are no more WAN hosts available is the (S,G) associated with the PD host, if any.
IGMP/MLD Joins are accepted only from the active subscriber hosts as dictated by antispoofing.
IGMP/MLD statistics can be displayed per host or per group.
Multicast traffic for the subscriber is forwarded through the egress SAP queue. In case that the SAP queue is disabled (profiled-traffic-only command), multicast traffic flows through internal queues outside of the subscriber context.
A single copy of any multicast stream is generated per SAP. This can be viewed as replication per unique multicast group per SAP, instead of the replication per host. In other words, the number of multicast streams on this SAP is equal to the number of unique groups across all hosts on this SAP (subscriber).
Traffic statistics are kept per the SAP queue. Consequently multicast traffic stats are shown outside of the subscriber context.
HQoS adjustment may be necessary.
Traffic cannot be explicitly classified (forwarding classes and queue mappings) inside of the subscriber queues.
Redirection to the common multicast VLAN (or Layer 3 interface) is supported.
Multicast streams have multicast destination MAC.
MLD uses link-local as source address, which makes it difficult to associate with the originating host. MLD (s,g) is associated with the IPv6 host. Therefore, if any WAN host or PD host end their IPv6 session (by lease expire, and so on), the (s,g) is associated with the remaining host address/prefix. The (s,g) is be delivered to the subscriber as long as a IPv6 address or prefix remains.
This model is shown in 1:1 model.
IPoE N:1 model (service per VLAN/SAP) — IGMP/MLD snooping in the AN
The AN is IGMP/MLD aware and is participating in multicast replication. From the BNG perspective this deployment model has the following characteristics:
IGMP/MLD states are kept per hosts and SAPs. Each host can be registered to more than one group.
IGMP/MLD Joins are accepted only from the active subscriber hosts as dictated by antispoofing.
IGMP/MLD statistics are displayed per host, per group or per subscriber.
Multicast traffic for all subscribers on this SAP is forwarded through the egress SAP queues.
A single copy of any multicast stream is generated per SAP. This can be viewed as the replication per unique multicast group per SAP, instead of the replication per host or subscriber. In other words, the number of multicast streams on this SAP is equal to the number of unique groups across all hosts and subscribers on this SAP.
The AN receives a single multicast stream and based on its own (AN) IGMP/MLD snooping information, it replicates the mcast stream to the appropriate subscribers.
Traffic statistics are kept per the SAP queue. Consequently multicast traffic stats are shown on a per SAP basis (aggregate of all subscribers on this SAP).
Traffic cannot be explicitly classified (forwarding classes and queue mappings) inside of the subscriber queues.
Redirection to the common multicast VLAN is supported.
Multicast streams have multicast destination MAC.
IGMP Joins are accepted (src IP address) only for the sub hosts that are already created in the system. IGMP Joins coming from the hosts that are nonexistent in the system are rejected, unless this functionality is explicitly enabled by the sub-hosts-only command under the IGMP group-int CLI hierarchy level.
MLD uses link-local as source address, which makes it difficult to associate with the originating host. MLD (S,G) are always first associated with a IPv6 WAN host, if available. If this WAN host terminate its IPv6 session (by lease expiry, session terminate, and so on), the (S,G) is then associated with any remaining WAN host. Only when there are no more WAN hosts available, is the (S,G) associated with the PD host, if any.
MLD join are only accepted if it matches the subscriber host link local address. MLD Joins coming from the hosts that are nonexistent in the system is rejected, unless this functionality is explicitly enabled by the sub-hosts-only command under the MLD group-int CLI hierarchy level.
This model is shown in N:1 model - AN in IGMP snooping mode.
IPoE N:1 model (service per VLAN/SAP) — IGMP/MLD proxy in the AN
The AN is configured as IGMP/MLD Proxy node and is participating in downstream multicast replication.
For IPv4, IGMP messages from multiple sources (subscribers hosts) for the same multicast group are consolidated in the AN into a single IGMP messages. This single IGMP message has the source IP address of the AN.
For IPv6, MLD messages from multiple sources (subscribers hosts) for the same multicast group are consolidated in the AN into a single MLD messages. This single MLD message has the link-local IP address of the AN.
From the BNG perspective this deployment model has the following characteristics:
Subscriber IGMP/MLD states are maintained in the AN.
IGMP Joins are accepted from the source IP address that is different from any of the subscriber’s IP addresses already existing in the BNG. This is controlled by an IGMP filter on a per group-interface level assuming that the IGMP processing for subscriber hosts is disabled with the no sub-hosts-only command under the router/service vprn>igmp>group-interface CLI hierarchy. In this case all IGMP messages that cannot be related to existing hosts are treated in the context of the SAP while IGMP messages from the existing hosts are treated in the context of the subscriber hosts.
MLD Joins are only accepted if the link-local address matches the subscriber’ link local address. To allow processing of foreign link-local address such as the AN link local address, the MLD processing for subscriber hosts should be disabled with the no sub-hosts-only command under the router/service vprn>mld>group-interface CLI hierarchy. In this case all MLD messages that cannot be related to existing hosts are treated in the context of the SAP while MLD messages from the existing hosts are treated in the context of the subscriber hosts.
IGMP/MLD statistics can be displayed per group-interface.
Multicast traffic for all subscribers on this SAP is forwarded through the egress SAP queue.
A single copy of any multicast stream is generated per SAP.
The AN receives a single multicast stream. Based on the IGMP/MLD proxy information, the AN replicates the mcast stream to the appropriate subscribers.
Traffic statistics are maintained per SAP queue.
HQoS Adjustment is not useful because the per host/subscriber IGMP/MLD granularity is lost. IGMP/MLD states are aggregated per AN.
Traffic can be explicitly classified into a specific SAP queues by a QoS policy applied under the SAP.
Multicast streams have multicast destination MAC.
In the following example, IGMPs from the source IP address <ip> is accepted even though there is no subscriber-host with that IP addresses present in the system. An IGMP state is created under the SAP context (service per vlan, or N:1 model) for the group pref-definition. All other IGMP messages originated from non-subscriber hosts is rejected. IGMP messages for subscriber hosts are processed according to the IGMP policy applied to each subscriber host.
configure
service vprn <id>
igmp
group-interface <ip-int-name>
import <policy-name>
configure
router
policy-options
begin
prefix-list <pref-name>
prefix <pref-definition>
policy-statement proxy-policy
entry 1
from
group-address <pref-name>
source-address <ip>
protocol igmp
exit
action accept
exit
exit
default-action reject
This functionality (accepting IGMP from non-subscriber hosts) can be disabled with the following flag.
configure
service vprn <id>
igmp
group-interface <ip-int-name>
sub-host-only
In this case, only per host IGMP processing is allowed.
In the following example, MLDs with foreign link-local-address is accepted even though there is no subscriber-host with that link local addresses present in the system. An MLD state is created under the SAP context (service per vlan, or N:1 model) for the group pref-definition. All other MLD messages originated from non-subscriber hosts are rejected. MLD messages for subscriber hosts are processed according to the IGMP policy applied to each subscriber host.
configure
service vprn <id>
mld
group-interface <ip-int-name>
import <policy-name>
configure
router
policy-options
begin
prefix-list <pref-name>
prefix <pref-definition>
policy-statement proxy-policy
entry 1
from
group-address <pref-name>
source-address <ip>
protocol igmp
exit
action accept
exit
exit
default-action reject
This functionality (accepting MLD from non-subscriber hosts) can be disabled with the following flag.
configure
service vprn <id>
mld
group-interface <ip-int-name>
sub-host-only
In this case, only per host MLD processing is allowed.
This model is shown in N:1 model - AN in proxy mode.
Per subscriber host replication mode
In this mode a multicast stream is transmitted per subscriber hosts for each registered multicast group (channel). As a result, multiple copies of the same multicast stream destined for different destinations can be transmitted over the same SAP. In this case, traffic flows within the subscriber queues and consequently it is accounted in HQoS. As a result, HQoS Adjustment is not needed. Each copy of the same multicast stream have a unique unicast destination MAC addresses. The per host unicast MAC destination addresses are necessary to differentiate multiple copies between different receivers on the same SAP.
Per host replication mode can be enabled on a subscriber basis with the per-host-replication command in the config>subscr-mgmt>igmp-policy context.
For IPv6, the command is in the config>subscr-mgmt>mld-policy context.
IPoE 1:1 model (subscriber per VLAN/SAP) — no IGMP/MLD in AN
The AN is not IGMP/MLD aware and multicast replication is performed in the BNG. Multicast streams are sent directly to the hosts using their unicast MAC addresses. HQoS adjustment is not needed as multicast traffic is flowing through subscriber queues. From the BNG perspective this deployment model has the following characteristics:
IGMP/MLD states are kept per hosts. Each host can be registered to multiple IGMP/MLD groups.
MLD uses link-local as source address, which makes it difficult to associate with the originating host. MLD (S,G) are always first associated with a IPv6 WAN host if available. If this WAN host terminate its IPv6 session (with lease expiry, session terminate, and so on), the (S,G) is then associated with any remaining WAN host. Only when there are no more WAN hosts available, are the (S,G) associated with the PD host, if any.
IGMP/MLD Joins are accepted only from the active subscriber hosts. In other words antispoofing is in effect for IGMP/MLD messages.
IGMP/MLD statistics can be displayed per host, per group or per subscriber.
Multicast traffic is forwarded through subscriber queues using unicast destination MAC address of the destination host.
Multiple copies of the same multicast stream can be generated per SAP. The number of copies depends on the number of hosts on the SAP that are registered to the same multicast group (channel). In other words, the number of multicast streams on the SAP is equal to the number of groups registered across all hosts on this SAP.
Traffic statistics are kept per the host queue. In case that multicast statistics need to be separated from unicast, the multicast traffic should be classified in a subscriber separate queue.
HQoS Adjustment is not needed as traffic is flowing within the subscriber queues and is automatically accounted in HQoS.
Multicast traffic can be explicitly classified into forwarding classes and consequently directed into needed queues.
MCAC is supported.
profiled-traffic-only mode defined under sub-sla-mgmt is supported. This mode (profiled-traffic-only) is used to save the number of queues in 1:1 model (sub-sla-mgmt-> no multi-sub-SAP) by preventing the creation of the SAP queues. Because multicast traffic is not using the SAP queue, enabling this feature does not have any effect on the multicast operation.
This model is shown in 1:1 model.
IPoE N:1 model (service per VLAN/SAP) — no IGMP/MLD in the AN
The AN is not IGMP/MLD aware and is not participating in multicast replication. From the BNG perspective this deployment model has the following characteristics:
IGMP/MLD states are kept per hosts. Each host can be registered to multiple multicast groups.
MLD uses link-local as source address, which makes it difficult to associate with the originating host. MLD (S,G) are always first associated with a IPv6 WAN host if available. If this WAN host terminate its IPv6 session (by lease expiry, session terminate, and so on), the (S,G) is then associated with any remaining WAN host. Only when there are no more WAN hosts available, are the (S,G) associated with the PD host, if any.
IGMP/MLD Joins are accepted only from the active subscriber hosts, subject to antispoofing.
IGMP/MLD statistics can be displayed per host, per group or per subscriber.
Multicast traffic is forwarded through subscriber queues using unicast destination MAC address of the destination host.
Multiple copies of the same multicast stream can be generated per SAP. The number of copies depends on the number of hosts on the SAP that are registered to the same multicast group (channel). In other words, the number of multicast streams on the SAP is equal to the number of groups registered across all hosts on this SAP.
Traffic statistics are kept per the host queue. In case that multicast statistics need to be separated from unicast, the multicast traffic should be classified in a separate subscriber queue.
HQoS Adjustment is not needed as traffic is flowing within the subscriber queues and is automatically accounted in HQoS.
Multicast traffic can be explicitly classified into forwarding classes and consequently directed into needed queues.
MCAC is supported.
This model is shown in N:1 model — no IGMP/MLD in the AN.
Per-SLA profile instance replication mode
In the per-SLA Profile Instance (SPI) replication mode, a multicast stream is transmitted for the subscriber host SLA profile for each registered multicast group (channel). As a result, multiple copies of the same multicast stream are transmitted over the same SAP. The multicast packet replication depends on the number of SLA profile instances configured for the subscriber. For example, if the subscriber hosts use three SLA profiles, the (S,G) is replicated three times. But if the hosts use the same SLA profile, the (S,G) is replicated only once.
MCAC is supported for per-SPI replication.
Multicast traffic flows on the SAP queue and each copy of the multicast stream use the same multicast destination MAC address.
The per-SPI replication mode can be enabled per subscriber:
in classic CLI, using the per-spi-replication command in the following contexts:
config>subscriber-mgmt>igmp-policy
config>subscriber-mgmt>mld-policy (IPv6)
in MD-CLI, using the replication command in the following contexts:
configure subscriber-mgmt igmp-policy
configure subscriber-mgmt mld-policy
Multicast over PPPoE
In a PPPoE environment, multicast replication is performed per session (host) regardless of whether those sessions are shared per SAP or they reside on individual SAPs. This is because of the point-to-point nature of PPPoE sessions. HQoS adjustment is not needed as multicast is part of the PPPoE session traffic that is flowing through subscriber queues. Multicast packets are sent with unicast MAC address to each CPE. PPP protocol field is set to IP and the destination IP address is the multicast group address for each unique session ID.
This model is shown in Multicast IPv4 address and unicast MAC address in PPPoE subscriber multicast.
Note that in a SRRP setup, the multi-chassis active BNG (SRRP master state) uses the SRRP MAC address as the source MAC for both multicast and multicast control protocol packets. This is only applicable to PPPoE hosts.
IGMP flooding containment
The query function in IGMP can cause some unintended flooding in N:1 IPoE deployment model with AN in the IGMP snooping mode. By maintaining IGMP session states per host, it is assumed that the IGMP interaction between multicast receivers and the BNG are on a one-to-one basis. Upon arrival of an IGMP leave from a host for a specific multicast group, the IGMP querier would normally multicast a group-specific query (fast-leave). In N:1 model with SAP replication mode enabled, the 7750 SR and 7450 ESS sends a group-specific query (fast-leave) only when it receives the IGMP leave message for the last group shared amongst all subscribers on this SAP.
IGMP/MLD timers
IGMP/MLD timers are maintained under the following hierarchy:
IPv4:
config>router>igmp
config>service>vprn>igmp
IPv6:
config>router>mld
config>service>vprn>mld
The IGMP/MLD timers are controlled on a per routing instance (VRF or GRT) level.
The timer values are used to:
determine the interval at which queries are transmitted (query-interval)
determine the amount of time after which a join times out
However, the timers can be different for hosts and redirected interface in case that redirection between VRFs is enabled.
IGMP/MLD query intervals
IGMP/MLD query related intervals (query-interval, query-last-member-interval, query-response-interval, robust-count) are configured on a global router/vprn IGMP/MLD level. They are used to determine the IGMP/MLD timeout states and the rates at which queries are transmitted.
In case of redirection, the subscriber-host IGMP/MLD state determines the IGMP/MLD state on the redirected interface, assuming that IGMP/MLD messages are not directly received on the redirected interface (for example from the AN performing IGMP/MLD forking). For example, if the redirected interface is not receiving IGMP/MLD messages from the downstream node, then the IGMP/MLD state under the redirected interface is removed simultaneously with the removal of the IGMP/MLD state for the subscriber host (because of leave or a timeout).
If the redirected interface is receiving IGMP/MLD message directly from the downstream node, the IGMP/MLD states on that redirected interface are driven by those direct IGMP/MLD messages.
For example, an IGMP/MLD host in VRF1 has an expiry time of 60 seconds and the expiry time defined under the VRF2 where multicast traffic is redirected is set to 90 seconds. The IGMP/MLD state times out for the host in VRF1 after 60s, and if no host has joined the same multicast group in VRF2 (where redirected interface resides), the IGMP state is removed there too.
If a join was received directly on the redirection interface in VRF2, the IGMP/MLD state for that group is maintained for 90s, regardless of the IGMP/MLD state for the same group in VRF1.
HQoS adjustment
HQoS Adjustment is required in the scenarios where subscriber multicast traffic flow is disassociated from subscriber queues. In other words, the unicast traffic for the subscriber is flowing through the subscriber queues while at the same time multicast traffic for the same subscriber is explicitly (through redirection) or implicitly (per-sap replication mode) redirected through a separate non-subscriber queue. In this case HQoS Adjustment can be deployed where preconfigured multicast bandwidth per channel is artificially included in HQoS. For example, bandwidth consumption per multicast group must be known in advance and configured within the 7450 ESS and 7750 SR. By keeping the IGMP state per host, the bandwidth for the multicast group (channel) to which the host is registered is known and is deducted as consumed from the aggregate subscriber bandwidth.
The multicast bandwidth per channel must be known (this is always an approximation) and provisioned in the BNG node in advance.
In PPPoE and in IPoE per-host replication environment, HQoS Adjustment is not needed as multicast traffic is unicast to each subscriber and therefore is flowing through subscriber queues.
For HQoS Adjustment, the channel bandwidth definition and association with an interface is the same as in the MCAC case. This is a departure from the legacy HT channel bandwidth definition which is done by a multicast-info-policy.
Example of HQoS adjustment:
Channel definition:
configure
router
mcac
policy <name>
<channel definition>
Channel bandwidth definition policy can be applied under:
group-interface
configure service vprn <id> igmp group-interface <ip-int-name> mcac policy <mcac-policy-name>
-
plain interface
configure router/service vprn igmp interface <ip-int-name> mcac policy <mcac-policy-name>
-
retailer group-interface
configure service vprn <id> igmp group-interface fwd-service <svc-id> <grp-if-name> mcac policy <mcac-policy-name>
Enabling HQoS adjustment:
configure
subscriber-management
igmp-policy <name>
egress-rate-modify [egress-aggregate-rate-limit | scheduler <name>]
Applying HQoS adjustment to the subscriber:
configure
subscriber-management
sub-profile <subscriber-profile-name>
igmp-policy <name>
To activate HQoS adjustment on the subscriber level, the sub-mcac-policy must be enabled under the subscriber with the following CLI:
configure
subscriber-management
sub-mcac-policy <pol-name>
no shutdown
configure
subscriber-management
sub-profile <subscriber-profile-name>
sub-mcac-policy <pol-name>
The adjusted bandwidth during operation can be verified with the following commands (depending whether agg-rate-limit or scheduler-policy is used):
*B:BNG-1# show service active-subscribers subscriber "sub-1" detail
===============================================================================
Active Subscribers
===============================================================================
-------------------------------------------------------------------------------
Subscriber sub-1
-------------------------------------------------------------------------------
I. Sched. Policy : up-silver
E. Sched. Policy : N/A E. Agg Rate Limit: 4000
I. Policer Ctrl. : N/A
E. Policer Ctrl. : N/A
Q Frame-Based Ac*: Disabled
Acct. Policy : N/A Collect Stats : Enabled
Rad. Acct. Pol. : sub-1-acct
Dupl. Acct. Pol. : N/A
ANCP Pol. : N/A
HostTrk Pol. : N/A
IGMP Policy : sub-1-IGMP-Pol
Sub. MCAC Policy : sub-1-MCAC
NAT Policy : N/A
Def. Encap Offset: none Encap Offset Mode: none
Avg Frame Size : N/A
Preference : 5
Sub. ANCP-String : "sub-1"
Sub. Int Dest Id : ""
Igmp Rate Adj : -2000
RADIUS Rate-Limit: N/A
Oper-Rate-Limit : 2000
...
-------------------------------------------------------------------------------
*B:BNG-1#
Consider a different example with a scheduler instead of agg-rate-limit:
*A:Dut-C>config>subscr-mgmt>sub-prof# info
----------------------------------------------
igmp-policy "pol1"
sub-mcac-policy "smp"
egress
scheduler-policy "h1"
scheduler "t2" rate 30000
exit
exit
----------------------------------------------
*A:Dut-C>config>subscr-mgmt>igmp-policy# info
----------------------------------------------
egress-rate-modify scheduler "t2"
redirection-policy "mc_redir1"
----------------------------------------------
Assume that the subscriber joins a new channel with bandwidth of 1 Mb/s (1000 kb/s).
A:Dut-C>config>subscr-mgmt>sub-prof>egr>sched# show qos scheduler-hierarchy subscrib
er "sub_1" detail
===============================================================================
Scheduler Hierarchy - Subscriber sub_1
===============================================================================
Ingress Scheduler Policy:
Egress Scheduler Policy : h1
-------------------------------------------------------------------------------
Legend :
(*) real-time dynamic value
(w) Wire rates
B Bytes
-------------------------------------------------------------------------------
Root (Ing)
|
No Active Members Found on slot 1
Root (Egr)
| slot(1)
|--(S) : t1
| | AdminPIR:90000 AdminCIR:10000
| |
| |
| | [Within CIR Level 0 Weight 0]
| | Assigned:0 Offered:0
| | Consumed:0
| |
| | [Above CIR Level 0 Weight 0]
| | Assigned:0 Offered:0
| | Consumed:0
| | TotalConsumed:0
| | OperPIR:90000
| |
| | [As Parent]
| | Rate:90000
| | ConsumedByChildren:0
| |
| |
| |--(S) : t2
| | | AdminPIR:29000 AdminCIR:10000(sum) <==== bw 1000 from igmp s
ubstracted
| | |
| | |
| | | [Within CIR Level 0 Weight 1]
| | | Assigned:10000 Offered:0
| | | Consumed:0
| | |
| | | [Above CIR Level 1 Weight 1]
| | | Assigned:29000 Offered:0 <==== bw 1000 from igmp s
ubstracted
| | | Consumed:0
| | |
| | |
| | | TotalConsumed:0
| | | OperPIR:29000 <==== bw 1000 from igmp s
ubstracted
| | |
| | | [As Parent]
| | | Rate:29000 <==== bw 1000 from igmp
substracted
| | | ConsumedByChildren:0
| | |
| | |
| | |--(S) : t3
| | | | AdminPIR:70000 AdminCIR:10000
| | | |
| | | |
| | | | [Within CIR Level 0 Weight 1]
| | | | Assigned:10000 Offered:0
| | | | Consumed:0
| | | |
| | | | [Above CIR Level 1 Weight 1]
| | | | Assigned:29000 Offered:0
| | | | Consumed:0
| | | |
| | | |
| | | | TotalConsumed:0
| | | | OperPIR:29000
| | | |
| | | | [As Parent]
| | | | Rate:29000
| | | | ConsumedByChildren:0
| | | |
*A:Dut-C>config>subscr-mgmt>igmp-policy# show service active-subscribers sub-mcac
===============================================================================
Active Subscribers Sub-MCAC
===============================================================================
Subscriber : sub_1
MCAC-policy : smp (inService)
In use mandatory bandwidth : 1000
In use optional bandwidth : 0
Available mandatory bandwidth : 1147482647
Available optional bandwidth : 1000000000
-------------------------------------------------------------------------------
Subscriber : sub_2
MCAC-policy : smp (inService)
In use mandatory bandwidth : 0
In use optional bandwidth : 0
Available mandatory bandwidth : 1147483647
Available optional bandwidth : 1000000000
-------------------------------------------------------------------------------
-------------------------------------------------------------------------------
Number of Subscribers : 2
===============================================================================
*A:Dut-C#
*A:Dut-C# show service active-subscribers subscriber "sub_1" detail
===============================================================================
Active Subscribers
===============================================================================
-------------------------------------------------------------------------------
Subscriber sub_1 (1)
-------------------------------------------------------------------------------
I. Sched. Policy : N/A
E. Sched. Policy : h1 E. Agg Rate Limit: Max
I. Policer Ctrl. : N/A
E. Policer Ctrl. : N/A
Q Frame-Based Ac*: Disabled
Acct. Policy : N/A Collect Stats : Disabled
Rad. Acct. Pol. : N/A
Dupl. Acct. Pol. : N/A
ANCP Pol. : N/A
HostTrk Pol. : N/A
IGMP Policy : pol1
Sub. MCAC Policy : smp
NAT Policy : N/A
Def. Encap Offset: none Encap Offset Mode: none
Avg Frame Size : N/A
Preference : 5
Sub. ANCP-String : "sub_1"
Sub. Int Dest Id : ""
Igmp Rate Adj : N/A
RADIUS Rate-Limit: N/A
Oper-Rate-Limit : Maximum
...
===============================================================================
*A:Dut-C#
*A:Dut-C# show subscriber-mgmt igmp-policy "pol1"
===============================================================================
IGMP Policy pol1
===============================================================================
Import Policy :
Admin Version : 3
Num Subscribers : 2
Host Max Group : No Limit
Host Max Sources : No Limit
Fast Leave : yes
Redirection Policy : mc_redir1
Per Host Replication : no
Egress Rate Modify : "t2"
Mcast Reporting Destination Name :
Mcast Reporting Admin State : Disabled
===============================================================================
*A:Dut-C#
Host Tracking (HT) considerations
HT is a light version of HQoS Adjustment feature. The use of HQoS Adjustment functionality in place of HT is strongly encouraged.
When HT is enabled, the AN forks off (duplicates) the IGMP messages on the common mcast SAP to the subscriber SAP. IGMP states are not fully maintained per sub-host in the BNG, instead they are only tracked (less overhead) for bandwidth adjustment purposes.
Example of HT
Channel Definition:
configure
mcast-management
multicast-info-policy <policy-name>
<channel to b/w mapping definition>
Applying channel definition policy on a router/VPRN global level:
config>router>multicast-info-policy <policy-name>
config>service>vprn>multicast-info-policy <policy-name>
Defining the rate object on which HT is applied:
configure
subscriber-management
host-tracking-policy <policy-name>
egress-rate-modify [agg-rate-limit | scheduler <sch-name>]
Applying the HT to the subscriber:
configure
subscriber-management
sub-profile <subscriber-profile-name>
host-tracking-policy <policy-name> => mutually exclusive with igmp-
policy
HQoS adjust per Vport
HQoS adjust per Vport can be used in environments where Vport represents a physical medium over which traffic for multiple subscribers is shared. Typical example of this scenario is shown in HQoS adjustment per subscriber and Vport. Multicast traffic within the router is taking a separate path from unicast traffic, only for the two traffic flows to merge later in the PON (represented by a Vport) and ONT (represented by the subscriber in the 7450 ESS and 7750 SR).
A single copy of each channel is replicated on the PON as long as there is at least one subscriber on that PON interested in this channel (has joined the IGMP/MLD group).
The 7450 ESS and 7750 SR monitors IGMP/MLD Joins at the subscriber level and consequently the channel bandwidth is subtracted from the current Vport rate limit only in the case that this is the first channel flowing through the corresponding PON. Otherwise, the Vport bandwidth is not modified. Similarly, when the channel is removed from the last subscriber on the PON, the channel bandwidth is returned to the Vport.
Association between the Vport and the subscriber is performed by an inter-destination-string or svlan during the subscriber setup phase. An inter-destination-string can be obtained either by RADIUS or LUDB. If the association between the Vport and the subscriber is performed based on the svlan (as specified in sub-sla-mgmt under the SAP or MSAP), then the destination string under the Vport must be a number matching the svlan.
The mcac-policy (channel definition bandwidth) can be applied on the group interface under which the subscribers are instantiated or in case of redirection under the redirected-interface.
In a LAG environment, the Vport instance is instantiated per member LAG link on the IOM. For accurate bandwidth control, it is prerequisite for this feature that subscriber traffic hashing is performed per Vport.
The CLI structure is as follows.
configure
port <port-id>
ethernet
access
egress
vport <name>
egress-rate-modify
agg-rate
host-match <destination-string>
port-scheduler-policy <port-scheduler-policy-name>
configure
port <port-id>
sonnet-sdh
path [<sonnet-sdh-index>]
access
egress
vport <name>
egress-rate-modify
agg-rate
host-match <destination-string>
port-scheduler-policy <port-scheduler-policy-name>
The Vport rate that is affected by this functionality depends on the configuration:
When the agg-rate-limit within the Vport is configured, its value is modified based on the IGMP activity associated with the subscriber under this Vport.
When the port-scheduler-policy within the Vport is referenced, the max-rate defined in the corresponding port scheduler policy is modified based on the IGMP activity associated with the subscriber under this Vport.
The Vport rates can be displayed with the following two commands:
- show port 1/1/5 vport name
- qos scheduler-hierarchy port port-id vport vport-name
As an example:
*A:system-1# show port 1/1/7 vport
========================================================================
Port 1/1/7 Access Egress vport
========================================================================
VPort Name : isam1
Description : (Not Specified)
Sched Policy : 1
Rate Limit : Max
Rate Modify : enabled
Modify delta : -14000
In this case, the configured Vport aggregate-rate-limit max value has been reduced by 14 Mb/s.
Similarly, if the Vport had a port-scheduling-policy applied, the max-rate value configured in the port-scheduling-policy would have been modified by the amount shown in the Modify delta output in the above command.
Multi-chassis redundancy
Modified Vport rate synchronization in multi-chassis environment relies on the synchronization of the subscriber IGMP/MLD states between the redundant nodes. Upon the switchover, the Vport rate on the newly active node is adjusted according to the current IGMP/MLD state of the subscribers associated with the Vport.
Scalability considerations
It is assumed that the rate of the IGMP/MLD state change on the Vport level is substantially lower than on the subscriber level.
The reason for this is that the IGMP/MLD Join/Leaves are shared amongst subscribers on the same Vport (PON for example) and therefore, the IGMP/MLD state on the Vport level is changed only for the first IGMP/MLD Join per channel and the last IGMP/MLD leave per channel.
Redirection
Two levels of MCAC can be enabled simultaneously and in such case this is referred as Hierarchical MCAC (H-MCAC). In case that redirection is enabled, H-MCAC per subscriber and the redirected interface is supported. However, mcac per group-interface in this case is not supported. Channel definition policy for the subscriber and the redirected interface is in this case referenced under the igmp>interface (redirected interface) CLI or for IPv6 mld>interface.
If redirection is disabled, H-MCAC for both, the subscriber and the group-interface is supported. The channel definition policy is in this case configured under the config>router>igmp>group interface context or for IPv6 config>router>mld>group interface.
There are two options in multicast redirection. The first option is to redirect all subscriber multicast traffic to a dedicated redirect interface.
Example:
Defining redirection action:
configure
router
policy-options
begin
policy-statement <name>
default-action accept
multicast-redirection [fwd-service <svc id>] <interface name>
exit
exit
exit
exit
The second option is to redirect only specific multicast groups to the redirect interface while the remaining groups remains on the subscriber SAP. This is applicable for both IPv4 and IPv6. For IPv6 host-ip for a policy statement is not supported.
Example:
Defining redirection action:
configure
router
policy-options
begin
prefix-list <name>
prefix <IPv4 multicast groups>
prefix <IPv4 multicast groups
exit
policy-statement <name>
entry 1
from
group-address <prefix-list name>
action accept
multicast-redirection [fwd-service <svc id>] <interface name>
exit
exit
exit
Applying redirection to the subscriber for IGMP and MLD respectively.
configure
subscr-mgmt
igmp-policy <name>
redirection-policy <name>
exit
exit
mld-policy <name>
redirection-policy <name>
Redirection that cross-connects GRT and VPRN is not supported. Redirection can be only performed between interfaces in the GRT, or between the interfaces in any of the VPRN (cross connecting VPRNs is allowed).
Redirection is also supported in a wholesaler/retailer VPRN model where redirected Layer 3 interface resides in the retailer VPRN.
Hierarchical Multicast CAC (H-MCAC)
MCAC is supported on three levels:
per subscriber
per group-interface
per redirected interface
MCAC is supported for the following replication modes:
per-host
per-SAP
per-SPI
redirected interface
Two levels of MCAC can be enabled simultaneously and in such case this is referred as Hierarchical MCAC (H-MCAC). If redirection is enabled, H-MCAC per subscriber and the redirected interface is supported. However, MCAC per group-interface in this case is not supported. Channel definition policy for the subscriber and the redirected interface is in this case referenced under the config>router>igmp>interface (redirected interface) CLI hierarchy or IPv6 config>router>mld>interface.
If redirection is disabled, H-MCAC for both, the subscriber and the group-interface is supported. The channel definition policy is in this case configured under the config>router>igmp>group-interface CLI hierarchy or IPv6 config>router>mld>group-interface.
Examples
Channel definition:
configure
router
mcac
policy <mcac-pol-name>
bundle <bundle-name>
bandwidth <kbps>
channel <start-address> <end-address> bw <bw> [class {high|low}]
[type {mandatory|optional}]
:
:
Channel bandwidth definition policy can be referenced under the:
group-interface (this is used for subscribers when redirection is disabled)
configure service vprn <id> igmp/mld group-interface <ip-int-name> mcac policy <mcac-policy-name> policy <mcac-policy-name>
-
plain interface
configure router igmp/mld interface <ip-int-name> mcac policy <mcac-policy-name> configure service vprn <id> igmp/mld interface <if-name> mcac unconstrained-bw <bandwidth> mandatory-bw <mandatory-bw>
-
retailer’s VPRN reference the group-interface in the wholesaler’s VPRN
configure service vprn <id> igmp/mld group-interface fwd-service <svc-id> <ip-int-name> mcac policy <mcac-policy-name>
Enabling MCAC per subscriber:
configure
subscr-mgmt
sub-mcac-policy <name>
unconstrained-bw <bandwidth> mandatory-bw <mandatory-bw>
configure
subscr-mgmt
sub-profile <subscriber-profile-name>
sub-mcac-policy <name>
Enabling MCAC per-group-interface
configure
service vprn <id>
igmp/mld
group-interface <ip-int-name>
mcac
unconstrained-bw <bandwidth> mandatory-bw <mandatory-bw>
Enabling MCAC per redirected interface
configure
router
igmp/mld
interface <ip-int-name>
mcac
unconstrained-bw <bandwidth> mandatory-bw <mandatory-bw>
configure
service vprn <id>
igmp/mld
interface <ip-int-name>
mcac
unconstrained-bw <bandwidth> mandatory-bw <mandatory-bw>
MCAC bundle bandwidth limit considerations
In addition to multicast bandwidth limit that can be imposed on subscribers, group-interfaces or regular interfaces, there is another multicast bandwidth limit that can be imposed on a group of channels (channel bundle).
The MCAC policy, aside from the channel bandwidth definitions, could optionally contain this bandwidth cap for the group of channels:
config>router>mcac# info
----------------------------------------------
policy "test"
bundle "test" create
bandwidth 100000
channel 239.0.0.10 239.0.0.10 bw 10000 type mandatory
channel 239.0.0.11 239.0.0.15 bw 5000 type mandatory
channel 239.0.0.20 239.0.0.30 bw 5000 type optional
exit
exit
This can be used to prevent a single set of channels from monopolizing MCAC bandwidth allocated to the entire interface. The bandwidth of each individual bundle is capped to some value below the interface MCAC bandwidth limit, allowing each bundle to have its own share of the interface MCAC bandwidth.
In most cases, the bandwidth limit per bundle is not necessary to configure. The aggregate limit per all channels as defined under the subscriber or interface covers majority of scenarios. In case that one wants to explore the bundle bandwidth limits and how they affect MCAC behavior, the following information helps understanding this topic.
To further understand how various MCAC bandwidth limits are applied, one need to understand the concept of the mandatory bandwidth that is pre-allocated in the following way:
Bandwidth of each mandatory channel in a bundle is pre-allocated. The artifacts of this are:
The total mandatory bandwidth in the bundle cannot exceed the bundle cap. For the sake of deterministic behavior, the configured bandwidth of each mandatory channel in the bundle is counted toward the total mandatory bandwidth only once. This means that only one replication of each mandatory channel is assumed. This is normal behavior on a regular interface with a single SAP under it. More than one replication of the same channel per regular interface (or SAP) would lead to packet duplication.
Optional (non-mandatory) channels can use only the difference in bandwidth between the bundle cap and total pre-allocated mandatory bandwidth. They cannot use more bandwidth than that even if the total pre-allocated mandatory bandwidth is not used up (mandatory channels are not being replicated).
Mandatory bandwidth under the interface is pre-allocated and subtracted from the unconstrained bandwidth. In the configuration example below, 2 Mb/s is pre-allocated (guaranteed for mandatory channels) and the remaining 8 Mb/s can be used by the optional channels on a first come first serve basis.
config>router>igmp# info
----------------------------------------------
interface "ge-1/1/1"
mcac
unconstrained-bw 10000 mandatory-bw 2000
exit
exit
The bundle bandwidth limit poses a problem when the MCAC policy is applied under the group interface. The reason is that the group interface represents the aggregation point for the subscribers and their bandwidth. As such it is natural that the any aggregated bandwidth limit under the group interface be larger than the bandwidth limit applied to any individual subscriber under it. Because the MCAC policy, along with the bundle bandwidth limit, is inherited by all subscribers under the group-interface, the exhaustion of the bundle bandwidth limit under the group interface coincides with the exhaustion of the bundle bandwidth limit of any individual subscriber. This results in a single subscriber starving out of multicast bandwidth the remaining subscribers under the same group interface. While it is perfectly acceptable for the subscribers to inherit the multicast channel definition from the group-interface, for the above reasons it is not acceptable that the subscriber inherit the bandwidth cap from the group-interface.
To resolve this situation, the MCAC bandwidth limits are independently configured for the group interface level (aggregated level) and the subscriber level with the unconstrained-bw kbps mandatory-bw kbps command. The unwanted bundle bandwidth cap in the MCAC policy is ignored under the group-interface and under the subscriber. However, the bundle bandwidth cap is applied automatically to each SAP under the group interface. A SAP is a natural place for a bundle bandwidth limit because each channel on a SAP can be replicated only once and therefore the amount of pre-allocated mandatory bandwidth can be pre-calculated. This is not the case for the group interface where single channel can be replicated multiple times (one per each SAP under the grp-if). Similarly, the same channel can be replicated multiple times for the same subscriber in per-host replication mode. Only subscribers in per-sap replication mode warrants a single replication per channel. Therefore, if bundle cap is configured, it is applied to limit the bandwidth of the bundle that is applied to a subscriber in a per-sap replication mode.
MCAC policy inheritance in per-SAP replication mode depicts MCAC related inheritances and MCAC bandwidth allocation model in per-sap replication mode. The MCAC policy is applied to the group interface and inherited by each subscriber as well as each SAP under the same group interface. However, the bundle bandwidth limit in the MCAC policy is ignored on the group-interface and under the subscriber (denoted by the X in the figure). The bundle limit is applied only to each sap under the group-interface.
Overall (non-bundle) MCAC bandwidth limits are independently applied to the group-interface and the subscribers. According to our example, 20 Mb/s of multicast bandwidth in total is allocated per group-interface. 6 Mb/s of the 20 Mb/s is allocated for mandatory channels. This leaves 14 Mb/s of multicast bandwidth for the optional channels combined served on a first come first serve basis. Each physical replication (multiple replications of the same channel can occur, one per each SAP), counts toward the respective group-interface bandwidth limits.
Similar logic applies to the subscriber MCAC bandwidth limits which are applied per sub-profile.
Finally, each SAP can optionally contain the bundle bandwidth limit.
In the example, six subscriber hosts watch the same channel but there are only three active replications (one per SAP). This would yield:
14 Mb/s of available multicast bandwidth under the group-interface. This bandwidth can be used for optional channels on a first come first serve basis. No reserved bandwidth is left.
For Subscriber A, 3 Mb/s is still reserved for mandatory channels and 5 MB/s is available for optional channels (first come first serve). All this assume that the SAP and the grp-if bandwidth checks pass.
For Subscriber B, 2 Mb/s is still reserved for mandatory channels and 6 Mb/s is available for optional channels (first come first serve). All this assume that the SAP and the grp-if bandwidth checks pass.
For Subscriber C, no reserved bandwidth for mandatory channels is left. 3 Mb/s is still left for optional channels. All this assume that the SAP and the grp-if bandwidth checks pass.
For SAPs, considering that 2 Mb/s are currently replicating (ch A, each SAP can still accept 1 Mb/s of the mandatory bandwidth (channel B and 7 Mb/s of the remaining optional channels.
To resolve this, the MCAC bandwidth limits are independently configured for the group interface (aggregated level) and the subscriber using the unconstrained-bw mandatory-bw command. The unwanted bundle bandwidth restriction in the MCAC policy is ignored for the group interface and under the subscriber. However, the bundle bandwidth maximum is applied automatically to each SAP for the group interface.
The bundle bandwidth limit is applied at the SAP level because each channel on the SAP is replicated only once and therefore the amount of pre-allocated mandatory bandwidth is precalculated. This is not the case for the group interface, where a single channel is replicated multiple times (once for each SAP in the group interface). Similarly, the same channel is replicated multiple times for the same subscriber in per-host replication and per-SPI replication modes. Only subscribers in per-SAP replication mode warrant a single replication per channel. Therefore, if the bundle maximum is configured, it limits the bandwidth of the bundle that is applied to a subscriber in per-SAP replication mode.
MCAC policy inheritance in per-HOST replication mode depicts behavior in per-host replication mode. MCAC policy inheritance flow is the same as in the previous example with the difference that the bundle limit has no effect at all. Each host generates its own copy of the same multicast stream that is flowing through subscriber queues and not the SAP queue. Because each of the copies counts toward the subscriber or group-interface bandwidth limits, the multicast bandwidth consumption is higher in this example. This needs to be reflected in the configured multicast bandwidth limits. For example, the group-interface mandatory bandwidth limit is increased to 12 Mb/s.
In our example, 6 subscriber hosts are still watching the same channel but now the number of replications is doubled from previous example. So the final tally for our MCAC bandwidth limit is as follows:
For Subscriber A, 1 Mb/s is still reserved for mandatory channels and 5 Mb/s for optional channels (first come first serve). All this assume that SAP and grp-if bandwidth checks pass.
For Subscriber B, no reserved mandatory bandwidth is left. 6 Mb/s is still left for optional channels (first come first serve). All this assume that SAP and grp-if bandwidth checks pass.
For Subscriber C, no reserved mandatory bandwidth is left. 1 Mb/s is still left for optional channels (first come first serve). All this assume that SAP and grp-if bandwidth checks pass.
No reserved bandwidth is left under the group-interface. 8 Mb/s of available multicast bandwidth under the group-interface is still left for optional channels on a first come first serve basis.
Bundle limit on a SAP is irrelevant in this case.
Determining MCAC policy in effect
Channel bandwidth definition (by a MCAC policy) can be applied under the interface level (group-interface or regular interface):
config>router/service>igmp/mld>interface/grp-if
The following configuration options can lead to the confusion as to which MCAC policy is in effect:
The MCAC policy (channel bandwidth definition) can be applied under two different places (grp-if or regular intf).
The same policy is used for (H)MCAC and HQoS Adjust with redirection enabled or disabled.
The general, the rule is that the MCAC policy under the group-interface is always be in effect in cases where redirection is disabled. This is valid for subscriber or group-interface MCAC, H-MCAC (subscriber and group-interface) and HQoS Adjust in per SAP replication mode.
If redirection is enabled, the MCAC policy under the group-interface is ignored.
If redirection is enabled, but there is no MCAC policy applied under the redirected interface (redirected interface is the interface to which IGMP/MLD Joins are redirected from subscriber hosts), regardless of whether the MCAC policy under the group-interface is applied or not, then:
HQoS adjust has no effect.
MCAC has no effect not only per redirected interface but also per subscriber.
Multicast filtering
Multicast filtering must be done per session (host) for IPoE and PPPoE. There are two types of filters that are supported:
-
IGMP filters on access ingress. Those filters control the flow of IGMP messages between the host and the BNG. They are applied with the import statement in the igmp-policy. The same filters are used for multicast-redirection policy:
For IPv4:
configure subscr-mgmt igmp-policy <name> import <policy-name>
For IPv6:
configure subscr-mgmt mld-policy <name> import <policy-name>
An example of the filter definition is shown below:
configure router policy-options begin prefix-list <pref-name> prefix <pref-definition> policy-statement <name> entry 1 from group-address <pref-name> source-address <ip> protocol igmp exit action accept exit exit default-action reject
-
Regular traffic filters where control multicast traffic flow can be controlled in both directions (ingress/egress). This is supported through ip-filters under the SLA profile.
For IPv6 specifically, up to 15 import policies can be applied to the subscriber IPv6 host. Each of the import policies can be configured as an IPTV package. Depending on the IPTV package the subscriber ordered, a list of import policies is applied at authentication.
At authentication, the first 14 import policies are applied either through LUDB or RADIUS using a list of VSAs. These 14 import policies are not read in any particular order and to make the list deterministic as a black list or a white list, the import policy inside the MLD policy must be configured. The import policy inside the MLD policy is always applied as the last policy which determines whether the list of 14 policies is a white list or a black list. For example, to make the list of 15 import policies a white list, the first 14 policies should have action configured as accept and the last import policy inside the MLD policy should have default action configured as reject. For a black list, the first 14 policies should have action configured as reject and the last import policy inside the MLD policy should have default action configured as accept. It is recommended to configure the MLD import policy with a default action without an import list which is used to determine if the list is a black or white list. If the last import policy is overridden, it changes the MLD policy which in turn overrides the subscriber profile.
It is possible to make a mixed black and white list, where the general rule is that the import policy inside the MLD policy is always the last to be applied to the subscriber.
The list of MLD import policies can be overridden either through RADIUS CoA or through the tools>perform>subscriber-management>coa command. There are two VSAs for overriding the list of import policies. The first VSA, Alc-Mld-Import-Policy, is used to completely override the first 14 MLD policies. A RADIUS CoA can contain up to 14 VSAs to override the existing list. Because of limitations with the length of a CLI command, up to five VSAs can be passed to override the existing list when using the tools command. The second VSA, Alc-Mld-Import-Policy-Modif, allows the addition or subtraction of a single MLD import policy. To add an import policy to the existing list, use a prefix of ‛a:’ and to remove an import policy, use a prefix of ‛s:’, for example, ‛a:sport-pack1-hd’ or ‛s:movie-pack1-hd’. Multiple import policies can be added or removed within a single CoA by using multiple VSAs. Up to five import policies can be added or removed using a single CoA. The last import policy inside the MLD policy can be overridden using the subscriber profile.
If a subscriber is not associated with an MLD policy, the import policies applied either at authentication or overridden in mid-session are always stored while the subscriber is online. The import policy does not take effect as long as the subscriber is not associated with an MLD policy. However, as soon as the subscriber is associated with an MLD policy, the list of import policies stored is instantly applied.
Joining the multicast tree
The delivery of multicast to the subscribers-interface in a VPRN environment depends on the multicast deployment model (PIM, mBGP). In each model, a subscriber-interface is treated as a regular CE-PE interface that has registered v4 multicast listeners.
Wholesale/retail requirements
Multicast support on subscriber interfaces is supported in both wholesale/retail models:
wholesale/retail VPRN (IPoE and PPPoE)
LAC/LNS (PPPoE only)
This model is shown in Wholesale/retail multicast support.
The distinction between these two models is that in the case of LAC/LNS, the replication is performed further up in the network on an LNS node. This means that the traffic between LAC and LNS is multiplied by the amount of replications.
QoS considerations
In per-sap replication mode (which is applicable only to IPoE subscribers), multicast traffic is forwarded through the SAP queue which is outside of the subscriber’s queues and therefore not accounted in subscriber aggregate rate limit. HQoS Adjust is used to remedy this situation.
If the SAP queue is removed from the static SAP in IPoE 1:1 model (with profiled-only-traffic command), multicast traffic flows through internal queues which cannot be tied into a port-scheduler as part of HQoS. Consequently, the port-scheduler max-rate as defined in the port scheduler policy is used only to rate limit unicast traffic. In other words, the max-rate value in the port scheduler policy must be lowered for the amount of anticipated multicast traffic that flows through the port where the port scheduler policy is applied.
A similar logic applies to the per-sap replication mode on dynamic SAPs (MSAPs) even if the SAP queue is not removed. Although the multicast traffic is flowing through the SAP queue in this case, the SAP QoS policy on MSAP cannot be changed from the default one. The default QoS policy on a SAP contains a single queue that is not parented by the port-scheduler.
Those restrictions do not apply to static SAPs where the SAP QoS policy can be customized and its queues consequently tied to the port-scheduler.
Redundancy considerations
The subscriber can receive multicast content through the subscriber SAP, the redirected interface, or a combination of both.
Multicast redundancy is only supported on a MC-LAG topology. Multicast traffic can be delivered over the subscriber SAP, the redirected interface, or a combination of both.
Subscriber IGMP states can be synchronized across multiple routers to ensure minimal interruption of (video delivery) service during network outages. The IGMP/MLD state of a subscriber-host in the system is tied to the state of the underlying MC-LAG protection mechanism. For example, IGMP states is activated only for subscribers that are anchored under the group-interfaces with master SRRP state or under an active MC-LAG port.
For multicast redirection on a MC-LAG topology, it must be ensured that the redirected interface (the interface to which multicast forwarding is redirected) is under the same MC-LAG as the subscriber. Otherwise, IGMP states on the redirected interface is derived independently of the IGMP states for the subscriber from which IGMP/MLD messages are redirected.
The IGMP/MLD synchronization process in conjunction with underlying access protection mechanisms works as follows:
IGMP/MLD states for the subscriber updates only if IGMP/MLD messages (Joins/Leaves/Reports, and so on) are received directly from the downstream node on an active MC-LAG link. This is valid irrespective of the IGMP querier status for the subscriber.
In all other cases, assuming that some protection mechanism in the access is present, the IGMP/MLD messages are discarded and consequently no IGMP/MLD state is updated. Similar logic applies to regular Layer 3 interfaces, where SRRP is replaced with VRRP.
After the subscriber IGMP/MLD state is updated as a result of a directly-received IGMP/MLD message on an active subscriber (SRRP master state or active MC-LAG), the sync IGMP/MLD message is sent to the standby subscriber over the Multi-Chassis Synchronization (MCS) protocol. The synchronized IGMP state is populated in the MCS database in all pairing routers.
If an IGMP/MLD sync (MCS) message is received from the peering node, the IGMP state for the standby subscriber is updated in the MCS database but it is not downloaded into the forwarding plane unless there is a switchover. If the IGMP/MLD sync message is received for the active subscriber, the message is discarded.
IGMP/MLD queries are sent out only by IGMP/MLD querier. In an MC-LAG environment, this is the node with the active MC-LAG link.
Note: MC-LAG is usually configured with SRRP and the SRRP state is derived from the MC-LAG.IGMP/MLD states from the MCS database are:
activated on non-querier subscriber in case that neither SRRP nor MC-LAG is deployed. It is assumed that the querier subscriber has received the original IGMP/MLD message and consequently sent the IGMP/MLD MCS Sync to the non-querier (standby). Non-querier interface accepts the MCS sync message and also it propagate these IGMP/MLD states to PIM.
The querier subscriber does not accept the IGMP/MLD update from the MCS database.
-
aware of the state of MC-LAG. As soon as the standby MC-LAG becomes active, the IGMP/MLD states is activated and is propagated to PIM. Traffic is forwarded as soon as multicast streams are delivered to the node and the IGMP states under the subscriber are activated. On a standby MC-LAG, IGMP states are not propagated from the MCS database to PIM and consequently subscribers.
-
aware of the SRRP state. Because the subscriber with SRRP master state is considered active, the states are propagated to PIM as well. On standby SRRP, IGMP states are be propagated from MCS database to PIM and consequently to subscribers.
For MC-lag setups, after the switchover is triggered by MC-LAG or SRRP, the IGMP/MLD states from MCS database on the newly active MC-LAG node or subscriber under the new SRRP instance in the master state is sent to PIM and consequently to the forwarding plane effectively turning on multicast forwarding.
An active and standby subscriber refers to the state of underlying protection mechanism (active MC-LAG).
The subscribers themselves are always instantiated (or active) on both nodes. However, traffic forwarding over those subscribers is driven by the state of the underlying protection mechanism (MC-LAG). Hence the terms active and standby subscriber.
In subscriber environment, SRRP should be always activated in dual-homing scenario. SRRP in subscriber environment ensures that downstream traffic is forwarded by the same node that is forwarding upstream traffic. In this fashion, accounting and QoS for the subscriber are consolidated within a single node.
To summarize, in multi-chassis environment with subscribers, IGMP synchronization enabled and an access layer protection mechanism in place (MC-LAG), the behavior for is the following:
IGMP/MLD states are synchronized between the chassis.
On a MC-LAG setup, only the SRRP instance in master state or active MC-LAG forwards downstream multicast traffic.
Length of outage during the switchover is determined by the detection and recovery of the underlying protection mechanism (MC-LAG or MCS) in addition to local propagation of IGMP/MLD states from MCS database to PIM and consequently to forwarding plane.
Note: IGMP/MLD states can be statically configured on both redundant nodes to attract multicast traffic from upstream and therefore minimize outage during the switchover.
Redirection considerations
The redirection policy has two options: to redirect only a specific set of multicast groups to the redirect interface or redirect all multicast to the redirect interface. The redirect policy is source agnostic.
On a MC-LAG setup, for redirection and MCS to work simultaneously in predictable manner, the redirected interface and the corresponding subscribers have to be protected by the same MC-LAG. This binds the redirected interfaces and the subscriber-hosts to the same physical ports.
The following describe some guidelines for a MC-LAG setup:
The active subscriber replicates its received IGMP/MLD message to the redirected Layer 3 interface. The Layer 3 redirected interface accepts this message:
independently of the corresponding VRRP state if MC-LAG is not used
only if the Layer 3 interface is IGMP querier
MC-LAG is used and in active state
In all other cases the IGMP message under the Layer 3 redirected interface is be rejected.
Note: Layer 3 redirected interface can also receive IGMP message directly from the downstream node in case that IGMP forking in the access node is activated.The Layer 3 redirected interface does not accept the IGMP state update from the MCS database unless the Layer 3 interface is a non-querier.
In case that the Layer 3 redirected interface is part of MC-LAG, the IGMP state update sent to it by MCS database is accepted only during the transitioning phase from standby to active MC-LAG state.
Briefly, IGMP states on Layer 3 interface are not VRRP aware. However, they are MC-LAG aware.
Query intervals for multicast
The query interval commands—query-interval, query-last-listener-interval, query-last-member-interval, and query-response-interval are configurable in the following command contexts:
router and VPRN service IGMP and MLD command contexts
router and VPRN service IGMP and MLD group interface command contexts
IGMP policy and MLD policy command contexts
The IGMP policy and MLD policy query intervals are only applicable to ESM host-based queries. Group interface query intervals are only applicable when the no sub-hosts-only command is configured. This triggers all group interface SAPs to send multicast queries. IGMP policy, MLD policy, and group interface timers use the router IGMP and MLD timers as a default fallback (see ESM host-based queries and Group interface-based queries).
For IGMP, the query-interval must be configured to be longer than the query-last-member-interval and query-response-interval. For MLD, the query-interval must be configured to be longer than the query-last-listener-interval and query-response-interval.
ESM host-based queries
An ESM host associated with an IGMP policy or an MLD policy is sent host-based queries. A host-based query uses the intervals defined under the IGMP policy or MLD policy first, and if none are configured, it uses the interval values configured under the router IGMP or MLD context as a fallback. The router IGMP and MLD interval values are set to default values if none are configured. It is possible for the IGMP or MLD policy to only configure some of the query intervals while leaving other query intervals to use the values configured under the router IGMP or MLD context. For example, if the IGMP policy only has the query-interval configured to 11, the remaining query-last-member-interval and query-response-interval configurations uses values 1 and 10, respectively, defined under router igmp. It is also possible to configure an invalid combination of query timers when using both the IGMP policy configured intervals and intervals under router igmp. Using the example above, if the IGMP policy only sets the query-interval to 5, while the query-response-interval under router igmp is set to 10, this becomes an invalid combination because the response interval is longer than the query interval. When the system detects an invalid combination, it falls back and use all interval values configured under the router igmp or router mld context and ignores the interval configured under the IGMP or MLD policy. It is highly recommended that all three query intervals be configured together under the IGMP or MLD policy. To see the active values that each host is using for queries, use the show router igmp hosts detail or show router mld host command.
Group interface-based queries
When the group interface is configured with no sub-hosts-only (applicable only to IPoE multicast), each SAP configured under that group interface sends out SAP-based queries. In this case, a single query is sent out of the SAP. The query intervals uses the values configured under the group interface. If there are no values configured under the group interface, the router IGMP or MLD values are used. When using the no sub-hosts-only command, IGMP and MLD policy are not required. To see the active values for the SAP-based queries, use the show router igmp group-interface detail or show router mld group-interface detail command.
ESM multicast replication modes
The following ESM multicast replication modes are supported:
per-host replication
This mode supports multicast packet replication per subscriber host. Each multicast packet destination uses an individual host source MAC address. The multicast traffic is unicast to the host by replacing the destination MAC from a multicast MAC with the host MAC address. A typical use case for per-host replication is for PPoE subscribers.
per-SAP replication
When all subscribers use this replication mode on a single SAP, the system ensures that no more than one multicast packet is sent per (S,G). This mode is used by default when another replication mode is not configured.
Note: In a BNG CUPS deployment, per-SAP replication is not supported and the system automatically defaults to per-host replication.per-SPI replication
In this mode, the multicast packet replication depends on the number of SLA profile instances configured for the subscriber hosts. A single copy of the multicast packet is sent to hosts that use the same SLA profile, and the multicast packet is replicated for each host that uses a different SLA profile. For example, if all subscriber hosts use the same SLA profile, only one multicast packet is sent from a specific (S,G). However, if the subscriber hosts use three different SLA profiles, three multicast packets are sent.
multicast redirection
In this mode, the multicast packets are sent to all the subscribers using a dedicated VLAN.
ESM multicast on BNG CUPS UPF
The User Plane Function (UPF) of BNG Control and User Plane Separation (CUPS) deployments supports ESM multicast as described in ESM multicast on BNG CUPS UPF .
Multicast replication model | Regular ESM model | Acting as UPF for BNG CUPS |
---|---|---|
Per-host |
Supported in:
The following apply:
|
Supported in:
When the wholesale and retail model is used, the following apply:
|
Per-SAP |
Supported in:
When both the wholesale and retail are VPRNs, the multicast source must be injected into the wholesale VPRN. |
Supported in VPRN service When the wholesale and retail model is used and retail is a VPRN, the multicast source must be injected into the retail VPRN. |
Per-SPI |
Supported in:
The following apply:
|
Supported in:
When the wholesale and retail model is used:
|
Mcast redirect |
Supported in:
|
Supported in:
|
ESM multicast support on VSR
ESM multicast redirection is the only supported ESM multicast mode on VSR for both IPv4 and IPv6 subscribers. The subscriber can be in an IES or a VPRN routing instance.
The following features are unsupported with ESM multicast redirection:
MCAC
egress rate adjust
MCS and redundancy with redirect interface
wholesale and retail service including both VPRN and LNS
Configuring Triple Play multicast services with CLI
This section provides information to configure multicast parameters in a Triple Play network using the command line interface.
Configuring IGMP snooping in the BSA
Enabling IGMP snooping in a VPLS service
IGMPv3 multicast routers
When multicast routers use IGMPv3, it is enough to only enable IGMP snooping, without any further modification of parameters.
The following displays an example of an IGMP snooping configuration:
A:ALA-48>config>service>vpls# info
----------------------------------------------
igmp-snooping
no shutdown
exit
no shutdown
----------------------------------------------
A:ALA-48>config>service>vpls#
With IGMPv1/2 multicast routers
When the multicast routers do not support IGMPv3, some timing parameters must be configured locally in the SR-series routers.
The following displays an example of a modified IGMP snooping configuration:
A:ALA-48>config>service>vpls# info
----------------------------------------------
stp
shutdown
exit
igmp-snooping
query-interval 60
robust-count 5
no shutdown
exit
no shutdown
----------------------------------------------
A:ALA-48>config>service>vpls#
Modifying IGMP snooping parameters
For interoperability with some multicast routers, the source IP address of IGMP group reports can be configured. Use the following CLI syntax to customize this IGMP snooping parameter:
The following displays an example of a modified IGMP snooping configuration:
A:ALA-48>config>service>vpls# info
----------------------------------------------
stp
shutdown
exit
igmp-snooping
query-interval 60
robust-count 5
report-src-ip 10.20.20.20
no shutdown
exit
no shutdown
----------------------------------------------
A:ALA-48>config>service>vpls#
Modifying IGMP snooping parameters for a SAP or SDP
Use the following CLI syntax to customize IGMP snooping parameters on an existing SAP. Commands for spoke or mesh SDPs are identical.
CLI syntax:
config>service# vpls service-id
sap sap-id
igmp-snooping
fast-leave
import policy-name
last-member-query-interval interval
max-num-groups max-num-groups
mrouter-port
query-interval interval
query-response-interval interval
robust-count count
send-queries
To enable and customize sending of IGMP queries to the hosts:
Example:
config>service# vpls 1
config>service>vpls# sap 1/1/3:0
config>service>vpls>sap# igmp-snooping
config>service>vpls>sap>snooping# send-queries
config>service>vpls>sap>snooping# query-interval 100
config>service>vpls>sap>snooping# query-response-interval 60
config>service>vpls>sap>snooping# robust-count 5
config>service>vpls>sap>snooping# exit
config>service>vpls>sap# no shutdown
To customize the leave delay:
Example:
config>service# vpls 1
config>service>vpls# sap 1/1/1:1
config>service>vpls>sap# igmp-snooping
config>service>vpls>sap>snooping# last-member-query-interval 10
config>service>vpls>sap>snooping# no fast-leave
config>service>vpls>sap>snooping# exit
config>service>vpls>sap# exit
To enable Fast Leave:
Example:
config>service# vpls 1
config>service>vpls# sap 1/1/1:1
config>service>vpls>sap# igmp-snooping
config>service>vpls>sap>snooping# no last-member-query-interval
config>service>vpls>sap>snooping# fast-leave
config>service>vpls>sap>snooping# exit
config>service>vpls>sap# exit
To limit the number of streams that a host can join:
Example:
config>service# vpls 1
config>service>vpls# sap 1/1/1:1
config>service>vpls>sap# igmp-snooping
config>service>vpls>sap>snooping# max-num-groups 4
config>service>vpls>sap>snooping# exit
config>service>vpls>sap# exit
To enable sending group reports on a SAP to standby multicast routers:
Example:
config>service# vpls 1
config>service>vpls# sap 1/1/1:1
config>service>vpls>sap# igmp-snooping
config>service>vpls>sap>snooping# mrouter-port
config>service>vpls>sap>snooping# exit
config>service>vpls>sap# exit
The following example displays the modified IGMP snooping configuration on a SAP:
A:ALA-48>config>service>vpls>sap>snooping# info detail
----------------------------------------------
no fast-leave
no import
max-num-groups 4
last-member-query-interval 10
no mrouter-port
query-interval 100
query-response-interval 60
robust-count 5
send-queries
----------------------------------------------
A:ALA-48>config>service>vpls>sap>snooping#
Configuring static multicast groups on a SAP or SDP
Use the following CLI syntax to add static group membership entries on an existing SAP (commands for spoke or mesh SDPs are identical):
The following displays an example of a static IGMP snooping configuration on a SAP:
A:ALA-48>config>service>vpls>sap# info
----------------------------------------------
max-nbr-mac-addr 4
igmp-snooping
fast-leave
mrouter-port
static
group 239.0.10.10
source 10.10.10.1
source 10.10.10.2
exit
exit
exit
----------------------------------------------
A:ALA-48>config>service>vpls>sap#
Enabling IGMP group membership report filtering
Routing policies can be defined to limit the multicast channels that can be joined by a host. For example, it is possible to define a policy listing a group of multicast streams (for example, 'basic' containing a basic set of TV channels or 'extended' containing a more extended set of TV channels), and to apply this policy to subscribers of IGMP snooping (SAPs or SDPs).
The following displays an example of a configuration to import a routing policy on a SAP:
A:ALA-48>config>service>vpls# info
----------------------------------------------
stp
shutdown
exit
igmp-snooping
query-interval 60
robust-count 5
report-src-ip 10.20.20.20
no shutdown
exit
sap 1/1/3:0 create
igmp-snooping
query-interval 100
query-response-interval 60
robust-count 5
send-queries
exit
exit
sap 1/1/3:22 create
max-nbr-mac-addr 4
igmp-snooping
fast-leave
import "test_policy"
mrouter-port
static
group 239.0.10.10
source 10.10.10.1
source 10.10.10.2
exit
exit
exit
exit
no shutdown
----------------------------------------------
A:ALA-48>config>service>vpls#
For details configuring a routing policy, see the Configuring Route Policies section in the 7450 ESS, 7750 SR, 7950 XRS, and VSR Router Configuration Guide.
The following shows a sample routing policy configuration accepting IGMP messages for only five multicast channels:
A:ALA-48>config>router>policy-options# info
----------------------------------------------
prefix-list "basic_channels"
prefix 239.10.0.1/32 exact
prefix 239.10.0.2/32 exact
prefix 239.10.0.3/32 exact
prefix 239.10.0.4/32 exact
prefix 239.10.0.5/32 exact
exit
policy-statement "test_policy"
description "basic set of 5 multicast channels"
entry 1
from
group-address "basic_channels"
exit
action accept
exit
exit
default-action reject
exit
----------------------------------------------
A:ALA-48>config>router>policy-options#
Enabling IGMP traffic filtering
For security, it may be advisable to only allow multicast traffic into the SR-series routers from recognized multicast routers and servers. Multicast packets arriving on other interfaces (for example, customer-facing SAPs or spoke SDPs) can be filtered out by defining an appropriate IP filter policy.
For details on how to configure a filter policy, see section Creating an IP Filter Policy in the Router Configuration Guide.
The following example shows a sample IP filter policy configuration dropping all multicast traffic:
A:ALA-48>config>filter>ip-filter# info
----------------------------------------------
ip-filter 1 create
entry 1 create
match
dst-ip 239.0.0.0/24
exit
action accept
exit
entry 2 create
match
dst-ip 239.0.0.0/4
exit
action drop
exit
exit
----------------------------------------------
A:ALA-48>config>filter>ip-filter#
The following example shows how to apply this sample IP filter policy to a SAP:
A:ALA-48>config>service>vpls # info
----------------------------------------------
sap 1/1/1:1
ingress
filter ip 1
exit
exit
exit
----------------------------------------------
A:ALA-48>config>service>vpls>snooping#
Configuring multicast VPLS Registration
Use the following CLI syntax to configure Multicast VPLS Registration (MVR). The first step is to register a VPLS as a multicast VPLS.
CLI syntax:
config>service# vpls service-id
igmp-snooping
mvr
no shutdown
description description
group-policy policy-name
Example:
config>service# vpls 1000
config>service>vpls# igmp-snooping
config>service>vpls>snooping# mvr
config>service>vpls>snooping>mvr# no shutdown
config>service>vpls>snooping>mvr# description "MVR VPLS"
config>service>vpls>snooping>mvr# group-policy ‟basic_channels_policy"
The second step is to configure a SAP to take the multicast channels from the registered multicast VPLS.
CLI syntax:
config>service# vpls service-id
sap sap-id
igmp-snooping
mvr
from-vpls vpls-id
Example:
config>service# vpls 1
config>service>vpls# sap 1/1/1:100
config>service>vpls>sap# igmp-snooping
config>service>vpls>snooping# mvr
config>service>vpls>snooping>mvr# from-vpls 1000
For MVR by proxy, the destination SAP for the multicast channels should a;sp be configured.
CLI syntax:
config>service# vpls service-id
sap sap-id
igmp-snooping
mvr
from-vpls vpls-id
to-sap sap-id
Example:
config>service# vpls 1
config>service>vpls# sap 1/1/1:100
config>service>vpls>sap# igmp-snooping
config>service>vpls>snooping# mvr
config>service>vpls>snooping>mvr# from-vpls 1000
config>service>vpls>snooping>mvr# to-sap 1/1/1:200
Configuring IGMP, MLD, and PIM in the BSR
Configuring IGMP, MLD, and PIM in the BSR is supported by the 7750 SR only.
See the 7450 ESS, 7750 SR, 7950 XRS, and VSR Multicast Routing Protocols Guide for information about multicast and the commands required to configure basic IGMP and PIM parameters.
Enabling IGMP
The following displays an example of enabled IGMP.
A:LAX>>config>router/service>vprn info detail
...
#------------------------------------------
echo "IGMP Configuration"
#------------------------------------------
igmp
query-interval 125
query-last-member-interval 1
query-response-interval 10
robust-count 2
no shutdown
exit
#------------------------------------------
...
A:LAX>>config>system#
Configuring IGMP interface parameters
The following example displays an IGMP configuration:
A:LAX>config>router/service>vprn>igmp# info
----------------------------------------------
interface "lax-sjc"
exit
interface "lax-vls"
exit
group-interface "Mcast-subscribers"
exit
----------------------------------------------
A:LAX>config>router>igmp# exit
Configuring static parameters
The following example displays a configuration to add IGMP a static multicast source:
A:LAX>config>router/service>vprn>igmp# info
----------------------------------------------
interface "lax-sjc"
exit
interface "lax-vls"
static
group 239.255.0.2
source 172.22.184.197
exit
exit
exit
group-interface "mcast-subscribers"
interface "lax-vls"
static
group 239.255.0.2
source 172.22.184.197
exit
exit
exit
exit----------------------------------------------
A:LAX>config>router>igmp#
:The following example displays the configuration to add a IGMP static starg entry:
A:LAX>config>router/service>vprn>igmp# info
----------------------------------------------
interface "lax-sjc"
static
group 239.1.1.1
starg
exit
exit
exit
interface "lax-vls"
static
group 239.255.0.2
source 172.22.184.197
exit
exit
exit
group-interface "mcast-subscribers"
static
group 239.1.1.1
starg
exit
exit
exit
----------------------------------------------
A:LAX>config>router>igmp#
Configuring SSM translation
The following displays an SSM translation configuration:
A:LAX>config>router/service>vprn>igmp# info
----------------------------------------------
ssm-translate
grp-range 239.255.0.1 231.2.2.2
source 10.1.1.1
exit
exit
interface "lax-sjc"
static
group 239.1.1.1
starg
exit
exit
exit
interface "lax-vls"
static
group 239.255.0.2
source 172.22.184.197
exit
exit
exit
group-interface "mcast-subscribers"
static
group 239.1.1.1
starg
exit
exit
exit
----------------------------------------------
A:LAX>config>router/service>vprn>igmp# exit
Enabling MLD
The following displays an example of enabled MLD.
A:LAX>>config>router/service>vprn # info detail
#------------------------------------------
echo "MLD Configuration"
#------------------------------------------
mld
query-interval 125
query-last-member-interval 1
query-response-interval 10
robust-count 2
no shutdown
exit
#------------------------------------------
...
A:LAX>>config>router/service>vprn #
Configuring MLD interface parameters
The following example displays an MLD configuration:
A:LAX>config>router/service>vprn>mld# info
----------------------------------------------
interface "lax-sjc"
exit
interface "lax-vls"
exit
group-interface "mcast-subscribers"
exit
----------------------------------------------
A:LAX>config>router/service>vprn>mld#exit
Configuring static parameters
The following example displays a configuration to add MLD a static multicast source:
A:LAX>config>router/service>vprn>mld# info
----------------------------------------------
interface "lax-sjc"
exit
interface "lax-vls"
static
group ffe8::1
source 2001::1
exit
exit
exit
group-interface "mcast-subscribers"
static
group ffe8::1
source 2001::1
exit
exit
exit
----------------------------------------------
A:LAX>config>router/service>vprn>mld#
The following example displays the configuration to add a MLD static starg entry:
A:LAX>config>router/service>vprn>mld# info
----------------------------------------------
interface "lax-sjc"
static
group ffe8::2
starg
exit
exit
exit
interface "lax-vls"
static
group ffe8::1
source 2001::1
exit
exit
exit
group-interface "mcast-subscribers"
static
group ffe8::2
starg
exit
exit
exit
exit
----------------------------------------------
A:LAX>config>router/service>vprn>mld#
Configuring SSM translation
The following displays an SSM translation configuration:
A:LAX>config>router/service>vprn>mld# info
----------------------------------------------
ssm-translate
grp-range ff31::1 ff32::1
source 2001::2
exit
exit
interface "lax-sjc"
static
group ffe8::2
starg
exit
exit
exit
interface "lax-vls"
static
group ffe8::1
source 2001::1
exit
exit
exit
group-interface "mcast-subscribers"
static
group ff31::20
starg
exit
exit
exit
exit
----------------------------------------------
A:LAX>config>router/service>vprn>mld# exit
Configuring PIM
Enabling PIM
When configuring PIM, make sure to enable PIM on all interfaces for the routing instance, otherwise multicast routing errors can occur.
The following example displays detailed output when PIM is enabled.
A:LAX>>config>router# info detail
...
#------------------------------------------
echo "PIM Configuration"
#------------------------------------------
pim
no import join-policy
no import register-policy
apply-to none
rp
no bootstrap-import
no bootstrap-export
static
exit
bsr-candidate
shutdown
priority 0
hash-mask-len 30
no address
exit
rp-candidate
shutdown
no address
holdtime 150
priority 192
exit
exit
no shutdown
exit
#------------------------------------------
...
A:LAX>>config>system#
Configuring PIM interface parameters
The following displays a PIM interface configuration:
A:LAX>config>router>pim# info
----------------------------------------------
interface "system"
exit
interface "lax-vls"
exit
interface "lax-sjc"
exit
interface "p1-ix"
exit
rp
static
address 10.22.187.237
group-prefix 239.24.24.24/32
exit
address 10.10.10.10
exit
exit
bsr-candidate
shutdown
exit
rp-candidate
shutdown
exit
exit
----------------------------------------------
A:LAX>config>router>pim#
A:SJC>config>router>pim# info
----------------------------------------------
interface "system"
exit
interface "sjc-lax"
exit
interface "sjc-nyc"
exit
interface "sjc-sfo"
exit
rp
static
address 10.22.187.237
group-prefix 239.24.24.24/32
exit
exit
bsr-candidate
shutdown
exit
rp-candidate
shutdown
exit
exit
----------------------------------------------
A:SJC>config>router>pim#
A:MV>config>router>pim# info
----------------------------------------------
interface "system"
exit
interface "mv-sfo"
exit
interface "mv-vlc"
exit
interface "p3-ix"
exit
rp
static
address 210.22.187.237
group-prefix 239.24.24.24/32
exit
exit
bsr-candidate
address 10.22.187.236
no shutdown
exit
rp-candidate
address 10.22.187.236
no shutdown
exit
exit
----------------------------------------------
A:MV>config>router>pim#
A:SFO>config>router>pim# info
----------------------------------------------
interface "system"
exit
interface "sfo-sjc"
exit
interface "sfo-was"
exit
interface "sfo-mv"
exit
rp
static
address 10.22.187.237
group-prefix 239.24.24.24/32
exit
exit
bsr-candidate
address 10.22.187.239
no shutdown
exit
rp-candidate
address 10.22.187.239
no shutdown
exit
exit
----------------------------------------------
A:SFO>config>router>pim#
A:WAS>config>router>pim# info
----------------------------------------------
interface "system"
exit
interface "was-sfo"
exit
interface "was-vlc"
exit
interface "p4-ix"
exit
rp
static
address 10.22.187.237
group-prefix 239.24.24.24/32
exit
exit
bsr-candidate
address 10.22.187.240
no shutdown
exit
rp-candidate
address 10.22.187.240
no shutdown
exit
exit
----------------------------------------------
A:WAS>config>router>pim#
Importing PIM join and register policies
The import command provides a mechanism to control the (*,g) and (sag) state that gets created on a router. Import policies are defined in the config>router>policy-options context. See Configuring PIM join and register policies.
The following example displays the command usage to apply the policy statement does not allow join messages for group 239.50.50.208/32 and source 192.168.0.0/16 but allows join messages for 192.168.0.0/16, 239.50.50.208:
Example:
config>router# pim
config>router>pim# import join-policy "foo"
config>router>pim# no shutdown
The following example displays the PIM configuration:
A:LAX>config>router>pim# info
----------------------------------------------
import join-policy "foo"
interface "system"
exit
interface "lax-vls"
exit
interface "lax-sjc"
exit
interface "p1-ix"
exit
rp
static
address 10.22.187.237
group-prefix 239.24.24.24/3
exit
address 10.10.10.10
exit
exit
bsr-candidate
shutdown
exit
rp-candidate
shutdown
exit
exit
----------------------------------------------
A:LAX>config>router>pim#
Configuring PIM join and register policies
Join policies are used in Protocol Independent Multicast (PIM) configurations to prevent the transportation of multicast traffic across a network and the dropping of packets at a scope at the edge of the network. PIM Join filters reduce the potential for denial of service (DoS) attacks and PIM state explosion—large numbers of Joins forwarded to each router on the RPT, resulting in memory consumption.
*,g or s,g is the information used to forward unicast or multicast packets.
group-address matches the group in join/prune messages
group-address 239.55.150.208/32 exact
source-address matches the source in join/prune messages
source-address 192.168.0.0/16 longer
interface matches any join message received on the specified interface:
interface port 1/1/1
neighbor matches any join message received from the specified neighbor:
neighbor 1.1.1.1
The following configuration example does not allow join messages for group 239.50.50.208/32 and source 192.168.0.0/16 but allows join messages for 192.168.0.0/16, 239.50.50.208.
A:ALA-B>config>router>policy-options# info
----------------------------------------------
...
policy-statement "foo"
entry 10
from
group-address "239.50.50.208/32"
source-address 192.168.0.0
exit
action reject
exit
exit
policy-statement "reg-pol"
entry 10
from
group-address "239.0.0.0/8"
exit
action accept
exit
exit
exit
...
----------------------------------------------
A:ALA-B>config>router>policy-options#
Configuring bootstrap message import and export policies
Bootstrap import and export policies are used to control the flow of bootstrap messages to and from the RP.
The following configuration example specifies that no BSR messages received or sent out of interface port 1/1/1.
:A:ALA-B>config>router>policy-options# policy-statement pim-import
:A:ALA-B>config>router>policy-options>policy-statement$ entry 10
:A:ALA-B>config>router>policy-options>policy-statement>entry$ from
:A:ALA-B>config>router>policy-options>policy-statement>entry>from$ interface port1/
1/1/
:A:ALA-B>config>router>policy-options>policy-statement>entry>from$ exit
:A:ALA-B>config>router>policy-options>policy-statement>entry# action reject
:A:ALA-B>config>router>policy-options>policy-statement>entry# exit
:A:ALA-B>config>router>policy-options>policy-statement# exit
:A:ALA-B>config>router>policy-options# policy-statement pim-export
:A:ALA-B>config>router>policy-options>policy-statement$ entry 10
:A:ALA-B>config>router>policy-options>policy-statement>entry$ to
:A:ALA-B>config>router>policy-options>policy-statement>entry>to$