Help us improve your experience.

Let us know what you think.

Do you have time for a two-minute survey?

 
 

Optimized Intersubnet Multicast in EVPN Networks

SUMMARY Enable intersubnet multicast (OISM) to optimize multicast traffic routing and forwarding in an EVPN edge-routed bridging (ERB) overlay fabric. With OISM, your network can also support multicast traffic flow between devices inside and outside of the EVPN fabric.

Overview of OISM

Optimized intersubnet multicast (OISM) is a multicast traffic optimization feature. This feature operates at L2 and L3 in EVPN-VXLAN edge-routed bridging (ERB) overlay fabrics. The design is based on the IETF draft specification https://datatracker.ietf.org/doc/html/draft-ietf-bess-evpn-irb-mcast. You can apply OISM configuration and operation to multicast traffic but not to broadcast or unknown unicast traffic.

In EVPN ERB overlay fabric designs, the leaf devices in the fabric route traffic between tenant bridge domains (that is, between VLANs). When you enable OISM, the leaf devices route intersubnet multicast traffic locally through IRB interfaces using the control plane multicast states. With local routing between VLANs, the receiver IRB interface doesn't send the routed multicast traffic out into the EVPN core. The local routing model helps minimize the traffic load within the EVPN core. It also avoids traffic hairpinning.

OISM leaf devices also selectively forward traffic into the EVPN core only toward other EVPN devices with interested receivers. Selective forwarding further improves multicast traffic performance in the EVPN fabric.

With OISM enabled, ERB overlay fabrics can efficiently and effectively support multicast traffic flow between devices inside and outside the EVPN fabric. Without OISM, fabric designers must use the centrally routed bridging (CRB) overlay model to support multicast with external sources or receivers. OISM border leaf devices support different methods to route traffic to and from an external PIM domain. These methods use either integrated routing and bridging (IRB) interfaces or Layer 3 (L3) interfaces. The border leaf devices also employ a supplemental bridge domain (SBD) inside the fabric to carry the traffic from external sources toward receivers within the EVPN fabric.

Benefits of OISM

  • Enables EVPN-VXLAN fabrics with the ERB overlay model to support multicast traffic with sources and receivers outside of the fabric.
  • Minimizes multicast control packets and replicated data packets in the EVPN fabric core to optimize fabric multicast performance in scaled designs.

OISM Support in EVPN Instances

We support OISM in EVPN-VXLAN fabrics in the following types of EVPN instances:

  • EVPN in the default switch instance:

    • Starting in Junos OS Release 21.2R1 on QFX5110, QFX5120, and QFX10002 (except QFX10002-60C) switches.

    • Starting in Junos OS Release 22.2R1 on EX4650, QFX10008, and QFX10016 switches.

    • Starting in Junos OS Release 22.3R1 on EX4300-48MP and EX4400 switches.

    • Starting in Junos OS Release 23.4R1 on ACX7024, ACX7100-32C, and ACX7100-48L routers.

  • MAC-VRF EVPN routing instances with vlan-aware and vlan-based service types only (see MAC-VRF Routing Instance Type Overview):

    • Starting in Junos OS Evolved Release 22.1R1 on QFX5130-32CD and QFX5700 switches.

    • Starting in Junos OS Release 22.2R1 on EX4650, QFX5110, QFX5120, QFX10002 (except QFX10002-60C), QFX10008, and QFX10016 switches.

    • Starting in Junos OS Evolved Release 22.3R1 on PTX10001-36MR, PTX10004, PTX10008, and PTX10016 routers.

    • Starting in Junos OS Release 23.4R1 on ACX7024, ACX7100-32C, and ACX7100-48L routers.

On Junos OS Evolved devices, we support EVPN-VXLAN using EVPN configurations with MAC-VRF instances only, and not in the default switch instance. As a result, on these devices we support OISM only in MAC-VRF EVPN instances.

OISM with Multicast Protocols and Other Multicast Optimizations in EVPN Fabrics

OISM works with the following multicast protocols and other EVPN multicast optimization features:

Multicast Protocols Supported with OISM

  • IGMPv2:

    • Starting in Junos OS Release 21.2R1 on QFX5110, QFX5120, and QFX10002 (except QFX10002-60C) switches.

    • Starting in Junos OS Evolved Release 22.1R1 on QFX5130-32CD and QFX5700 switches.

    • Starting in Junos OS Release 22.2R1 on EX4650, QFX10008, and QFX10016 switches.

    • Starting in Junos OS Release 22.3R1 on EX4300-48MP and EX4400 switches.

    • Starting in Junos OS Evolved Release 22.3R1 on PTX10001-36MR, PTX10004, PTX10008, and PTX10016 routers.

    • Starting in Junos OS Release 23.4R1 on ACX7024, ACX7100-32C, and ACX7100-48L routers.
  • IGMPv3:

    • Starting in Junos OS Evolved Release 22.1R1 on QFX5130-32CD and QFX5700 switches.

    • Starting in Junos OS Release 22.2R1 on EX4650, QFX5110, QFX5120, QFX10002 (except QFX10002-60C), QFX10008, and QFX10016 switches.

    • Starting in Junos OS Evolved Release 22.3R1 on PTX10001-36MR, PTX10004, PTX10008, and PTX10016 routers.

    • Starting in Junos OS Release 23.4R1 on ACX7024, ACX7100-32C, and ACX7100-48L routers.

  • MLDv1 and MLDv2:

    • Starting in Junos OS Evolved Release 23.1R1 on QFX5130-32CD and QFX5700 switches.

  • PIM, which facilitates both local routing and external multicast traffic routing.

OISM supports IGMP snooping with both IGMPv2 and IGMPv3 on the same device at the same time only under certain configuration constraints. Similarly, OISM supports MLD snooping with both MLDv1 and MLDv2 at the same time under the same configuration constraints. See IGMPv2 and IGMPv3 (or MLDv1 and MLDv2) in the Same EVPN-VXLAN Fabric for details.

Also see Supported IGMP or MLD Versions and Group Membership Report Modes for information on IGMP or MLD any-source multicast (ASM) and source-specific multicast (SSM) mode support in EVPN-VXLAN fabrics.

Other Multicast Optimization Features That Work with OISM

OISM works with these other multicast optimization features:

  • IGMP snooping or MLD snooping (some platforms) on the access side on the leaf devices.

    With IGMP snooping or MLD snooping enabled, a leaf device that receives multicast traffic forwards it only toward other devices with interested receivers.

  • Multihoming support in an Ethernet segment (ES) using EVPN Type 7 (Join Sync) and Type 8 (Leave Sync) routes.

    EVPN fabric devices advertise these route types to synchronize the multicast state among EVPN devices that are multihoming peers.

    Note: ACX Series OISM leaf devices can be multihoming peer PE devices only with ACX Series devices.
  • Selective multicast Ethernet tag (SMET) forwarding in the EVPN fabric core using EVPN Type 6 routes.

    EVPN devices use Type 6 routes to limit forwarding within the EVPN core only to receivers interested in receiving traffic for a multicast group. You can use OISM to make this optimization work in EVPN ERB overlay fabrics. When you configure IGMP or MLD snooping, the fabric enables SMET forwarding with OISM automatically.

  • Assisted replication (AR).

    You can integrate AR into a fabric running OISM as follows, depending on the platforms that support the different AR and OISM device roles:

    • Starting in Junos OS Release 22.2R1 on EX4650, QFX5110, QFX5120, QFX10002 (except QFX10002-60C), QFX10008, and QFX10016 switches:

      • You can configure the AR leaf role on any of these devices that are also acting as OISM border leaf or server leaf devices.

      • You can configure only QFX10002 (except QFX10002-60C), QFX10008, and QFX10016 switches as AR replicators, in either of these modes:

        Collocated mode: The device acts as both an AR replicator device and an OISM border leaf device.

        Standalone mode: The device is an AR replicator but isn't also an OISM border leaf or server leaf device.

    • Starting in Junos OS Evolved Release 22.2R1 on QFX5130-32CD and QFX5700 switches:

      • You can configure the AR leaf role on OISM border leaf or server leaf devices.

      • You can configure these devices as AR replicators with OISM in standalone mode only. In standalone mode, the AR replicator device doesn't also operate as an OISM border leaf or server leaf.

    Note:

    ACX7024, ACX7100-32C, and ACX7100-48L routers don't support AR with OISM as AR replicator or AR leaf devices.

    See these references for more on using AR and OISM together:

OISM Components

The OISM environment includes:

  • Leaf devices in the EVPN fabric that function in border roles and server access roles.

  • External multicast sources and receivers in an external L3 PIM domain.

  • Bridge domain (VLAN) configurations that enable the fabric to route multicast traffic between internal and external devices.

The EVPN-VXLAN ERB overlay design includes lean spine devices that support L3 transit functions for the leaf devices. The lean spine devices don't usually perform any OISM functions.

The following sections describe these OISM components.

OISM Device Roles

Figure 1 shows a simple EVPN-VXLAN ERB overlay fabric and the OISM device roles in the fabric.

Figure 1: EVPN Fabric with OISMEVPN Fabric with OISM

Table 1 summarizes the device roles.

Table 1: EVPN Fabric OISM Device Roles
Device Role Description

Border leaf (BL)

OISM leaf devices in the EVPN fabric underlay and overlay. Border leaf devices function as gateways interconnecting the EVPN fabric to multicast devices (sources and receivers) outside the fabric in an external PIM domain. These devices serve in the PIM EVPN gateway (PEG) role.

Lean spine (LS)

Spine devices in the underlay of the EVPN fabric. These devices usually operate as lean spines that support the EVPN underlay as IP transit devices. The lean spines might also act as route reflectors in the fabric.

You configure OISM elements on the lean spine devices only in the following use cases:

  • The devices also serve as border devices for external multicast traffic. In this case, you configure the same OISM elements as you configure on border leaf devices.

  • The lean spine device serves as a standalone AR replicator when you integrate AR with OISM in the fabric. In this case, on the AR replicator spine device, you configure the same common OISM elements that you configure on the border leaf and server leaf devices. You don't need to configure any of the PIM or external multicast elements specific to border leaf or server leaf devices. (See AR with Optimized Intersubnet Multicast (OISM).)

Server leaf (Leaf)

OISM leaf devices on the access side in the EVPN fabric underlay and overlay. Server leaf devices are often top-of-rack (ToR) switches. These devices connect the EVPN fabric to multicast sources and multicast receiver hosts on bridge domains or VLANs within the fabric.

See Configuration Elements for OISM Devices (Symmetric Bridge Domains) for details on the configuration elements that are common and those that are different for each device role.

PIM Domain with External Multicast Sources and Receivers

In Figure 1, the OISM border leaf devices connect to multicast sources and receivers outside the EVPN fabric in a representative external PIM domain. The multicast devices in the external PIM domain follow standard PIM protocol procedures; their operation is not specific to OISM. External multicast traffic flows at L3 through the PIM domain.

You can use OISM to route and to forward multicast traffic in an EVPN-VXLAN ERB overlay fabric between devices in the following use cases:

  • Internal multicast sources and receivers

  • Internal multicast sources and external multicast receivers

  • External multicast sources and internal multicast receivers

For simplicity, in this documentation we represent the external PIM domain as:

  • A PIM router (a device such as an MX Series router) that doubles as the PIM rendezvous point (RP).

  • An external source.

  • An external receiver.

Supported Methods for Multicast Data Transfer to or from an External PIM Domain

OISM border leaf devices support one or more methods to route multicast traffic to and from devices outside of the fabric. Supported methods are platform-dependent.

Some platforms don't support the border leaf role. If you don't see a platform listed in Table 2 in the Supported Platforms column for any of the external multicast methods, that means the platform doesn't support the border leaf role.

Table 2: External Multicast Connection Methods
Name Connection Method Supported Platforms

M-VLAN IRB method

IRB interfaces on a multicast VLAN (M-VLAN) that you extend in the EVPN instance. The fabric uses the M-VLAN and corresponding IRB interfaces only for external multicast traffic flow to and from the external PIM domain.

This method supports EVPN Ethernet segment identifier (ESI) multihoming to connect the external PIM router to more than one OISM border leaf device in the fabric.

PTX10001-36MR

PTX10004

PTX10008

PTX10016

QFX10002 (except QFX10002-60C)

QFX10008

QFX10016

Classic L3 interface method

Classic physical L3 interfaces on OISM border leaf devices that connect individually to the external PIM domain on different subnets.

These interfaces aren't associated with a VLAN. You don't configure these interfaces in the EVPN instances. Instead, you assign IP addresses to these interfaces and include them in the tenant L3 virtual routing and forwarding (VRF) instances.

Note:

The L3 interface connection can be an aggregated Ethernet (AE) interface bundle.

ACX7024

ACX7100

EX4650

PTX10001-36MR

PTX10004

PTX10008

PTX10016

QFX5120

QFX5130-32CD

QFX5700

QFX10002 (except QFX10002-60C)

QFX10008

QFX10016

Non-EVPN IRB method

IRB interfaces on an extra VLAN that you don't extend in the EVPN instance. You include these logical interfaces in the tenant L3 VRF instances.

On each border leaf device, you assign a unique extra VLAN ID and subnet for the associated IRB interface.

We call this type of interface a non-EVPN IRB interface for external multicast.

PTX10001-36MR

PTX10004

PTX10008

PTX10016

QFX5130-32CD

QFX5700

See Overview of Multicast Forwarding with IGMP Snooping or MLD Snooping in an EVPN-VXLAN Environment for a general overview of connecting EVPN-VXLAN fabrics to an external PIM domain using a L2 M-VLAN or L3 links.

OISM Bridge Domains (VLANs)

Table 3 summarizes the OISM bridge domains or VLANs and describes how OISM uses them.

Note:

References in this document to all OISM devices correspond to the border leaf and server leaf devices on which you enable OISM.

Table 3: OISM Bridge Domains or VLANs
Bridge Domain/VLAN Description Configure On:

Multicast VLAN (M-VLAN)

(M-VLAN IRB method for external multicast) A VLAN in the EVPN fabric with associated IRB interfaces that connect the fabric to an external multicast router. This VLAN and IRB interface enable traffic flow between devices inside and outside the fabric. To support IGMP snooping with both IGMPv2 and IGMPv3 traffic, you assign separate M-VLANs to carry traffic for each IGMP version.

You extend this VLAN in the EVPN instance. You can also multihome the external multicast router to multiple border leaf device M-VLAN IRB interfaces in the same EVPN ES. The usual EVPN multihoming designated forwarder (DF) rules apply to send only one copy of the traffic in the ES on the M-VLAN.

Configure the M-VLAN as a VLAN that is not the SBD or any of the revenue bridge domains in the EVPN fabric.

Note:

See Supported Methods for Multicast Data Transfer to or from an External PIM Domain for other supported methods to connect to external sources and receivers.

Border leaf devices

Extra non-EVPN VLAN

(Non-EVPN IRB method for external multicast) An extra VLAN that isn't in the EVPN instances in the fabric. You configure associated IRB interfaces in the tenant L3 VRF instances. This VLAN and IRB interface enable multicast traffic flow between devices inside the fabric and devices outside the fabric.

You must assign distinct extra VLANs and corresponding IRB interface subnets on each border leaf device that are unique across the fabric.

Also, to support IGMP snooping with both IGMPv2 and IGMPv3 traffic, use separate distinct extra non-EVPN VLANs to carry traffic for each IGMP version. The same constraints apply if you want to support MLD snooping with both MLDv1 and MLDv2 traffic.

Note:

See Supported Methods for Multicast Data Transfer to or from an External PIM Domain for other supported methods to connect to external sources and receivers.

Border leaf devices

Revenue bridge domains (VLANs)

Bridge domains for subscribers to the services that the fabric provides. You configure the revenue bridge domains as VLANs in the fabric.

The revenue bridge domains correspond to the customer VLANs in the fabric. These VLANs are not specific to OISM, but the multicast sources and receivers in the fabric are in these bridge domains.

For details on how to allocate VLANs for these bridge domains if you want to support IGMP snooping with both IGMPv2 and IGMPv3 receivers in the fabric, see IGMPv2 and IGMPv3 (or MLDv1 and MLDv2) in the Same EVPN-VXLAN Fabric. Note that the same constraints apply if you want to support MLD snooping with both MLDv1 and MLDv2 receivers.

All OISM devices

Supplemental bridge domain (SBD)

Bridge domain that enables support for external multicast traffic and implements SMET optimization in the EVPN core.

You configure the SBD as a regular VLAN that is different from the revenue bridge domains, the M-VLAN or the extra non-EVPN VLANs.

SBD usually serves all OISM leaf devices in the fabric. To support IGMP snooping with both IGMPv2 and IGMPv3 traffic, or to support MLD snooping with both MLDv1 and MLDv2 traffic, you assign separate SBDs to carry traffic for each IGMP or MLD version. The SBD carries:

  • Routed multicast data traffic from external multicast sources to EVPN devices in the fabric with interested receivers.
  • SMET EVPN Type 6 routes from originating leaf devices to other EVPN leaf devices, which enables selective forwarding in the EVPN core.
Note:

The SBD is central to OISM operation for external multicast traffic routing and forwarding. The SBD IRB interface must always be up for OISM to work.

All OISM devices

The OISM implementation uses a symmetric bridge domains model. The following figure shows how you set up the bridge domains and VLANs in the EVPN fabric on the OISM devices with this model.

Figure 2: EVPN Fabric with OISM Symmetric Bridge Domains ModelEVPN Fabric with OISM Symmetric Bridge Domains Model

In the symmetric bridge domains model, you must configure the SBD and all revenue bridge domains on all OISM border leaf and server leaf devices.

You also configure the following VLANs uniformly on the border leaf devices that connect to the external PIM domain:

  • The M-VLAN, if you use the M-VLAN IRB method for external multicast.

  • A unique extra non-EVPN VLAN on each border leaf device, if you use the non-EVPN IRB method for external multicast.

The lean spine devices in the fabric usually serve only as IP transit devices and possibly as route reflectors. As a result, you don't usually need to configure these elements on the lean spine devices. (See the Lean Spine row in Table 1 for some exceptions.)

Configuration Elements for OISM Devices (Symmetric Bridge Domains)

This section summarizes the elements you need to configure on:

  • All OISM devices—Devices in the border leaf role and the server leaf role, and spine devices that also serve as AR replicators in standalone mode when you integrate AR with OISM.

    See Table 4.

  • Server leaf devices only.

    See Table 5.

  • Border leaf devices only, based on the method you use for external multicast:

    • M-VLAN IRB interface

    • Classic L3 interface

    • Non-EVPN IRB interface

    See Table 6.

Some elements are optional, which the description notes.

Note:

EX4650, QFX5110, and QFX5120 switches support enterprise style interface configurations for OISM elements, but not service provider style interface configurations. For more information on these interface configuration styles, see Flexible Ethernet Services Encapsulation and Understanding Flexible Ethernet Services Support With EVPN-VXLAN.

Table 4 lists the elements you configure on all OISM devices.

Table 4: Configuration Elements on all OISM Devices
Configuration Element Description

OISM

Enable OISM globally, and enable OISM routing functions in L3 VRF instances.

A device with OISM enabled advertises EVPN Type 3 inclusive multicast Ethernet tag (IMET) routes as follows:

  • The device advertises an IMET route for each configured revenue bridge domain (VLAN) and the SBD.

  • The IMET routes include the EVPN multicast flags extended community with a flag indicating OISM support.

Revenue bridge domains (customer VLANs) and corresponding IRB interfaces

Configure revenue bridge domains (customer VLANs) according to your data center services requirements. You must configure all revenue bridge domain VLANs and corresponding IRB interfaces on all OISM devices. See OISM Bridge Domains (VLANs) for more information.

See IGMPv2 and IGMPv3 (or MLDv1 and MLDv2) in the Same EVPN-VXLAN Fabric for special considerations to configure the revenue bridge domains if you want to support:

  • IGMP snooping with traffic for both IGMPv2 and IGMPv3 receivers.

  • MLD snooping with traffic for both MLDv1 and MLDv2 receivers,

SBD (VLAN) and corresponding IRB interface

Configure the SBD and its IRB interface on all OISM devices. The SBD can be any VLAN that is distinct from the M-VLAN, any of the non-EVPN VLANs or any revenue bridge domain VLANs in the EVPN fabric. See OISM Bridge Domains (VLANs) for more information.

You identify this VLAN as the SBD in the L3 VRF instance that supports OISM routing. EVPN IMET routes for the SBD IRB interface include the OISM SBD flag in the multicast flags extended community.

L3 multicast protocol—IGMPv2 or IGMPv3

Enable IGMPv2 or IGMPv3 L3 multicast protocols. Receivers send IGMP reports to express interest in receiving traffic for a multicast group.

You can use IGMPv2 or IGMPv3 in any-source multicast (ASM) mode, or IGMPv3 in source-specific multicast (SSM) mode.

Note that you can't enable IGMP snooping for both IGMP versions together for the same VLAN or in the same VRF instance with OISM enabled. However, to support IGMP snooping with IGMPv2 and IGMPv3 receivers in the same fabric, you can enable IGMP snooping with IGMPv2 for specific VLANs in one VRF instance, and enable IGMP snooping with IGMPv3 for other VLANs in another VRF instance. See IGMPv2 and IGMPv3 (or MLDv1 and MLDv2) in the Same EVPN-VXLAN Fabric for details.

IGMPv2 is the default IGMP version. To configure IGMPv3, you must specify the version 3 option.

L3 multicast protocol—MLDv1 or MLDv2

Enable MLDv1 or MLDv2 L3 multicast protocols if you have IPv6 multicast traffic in your fabric. Receivers send MLD reports to express interest in receiving traffic for a multicast group.

You can use MLDv1 or MLDv2 in any-source multicast (ASM) mode, or MLDv2 in source-specific multicast (SSM) mode.

Note that with OISM enabled, you can't enable MLD snooping for both MLD versions together for the same VLAN or in the same VRF instance. However, you can support MLD snooping with MLDv1 and MLDv2 receivers in the same fabric if you:

  • Enable MLD snooping with MLDv1 for specific VLANs in one VRF instance.

  • Enable MLD snooping with MLDv2 for other VLANs in another VRF instance.

See IGMPv2 and IGMPv3 (or MLDv1 and MLDv2) in the Same EVPN-VXLAN Fabric for details.

MLDv1 is the default MLD version. To configure MLDv2, you must specify the version 2 option.

L2 multicast optimizations—IGMP snooping with SMET

Enable IGMP snooping at L2 with IGMPv2 or IGMPv3 protocols as part of the optimizations OISM provides. With IGMP snooping, the device routes or forwards multicast traffic only toward interested access-side receivers. Receivers send IGMP reports to express interest in receiving traffic for a multicast group.

When you enable IGMP snooping, the device also automatically advertises SMET Type 6 routes. With SMET, the device sends copies of the traffic into the EVPN core only toward other devices that have interested receivers.

Configure IGMP snooping as follows:

  • In the EVPN instance(s) for each OISM revenue VLAN and the SBD.

  • On border leaf devices for external multicast only with the M-VLAN IRB method or the non-EVPN IRB method, as follows:

    • M-VLAN IRB method: In the EVPN instance(s) for the M-VLAN IRB multicast router interface. Include the evpn-ssm-reports-only option only with IGMPv3.

    • Non-EVPN IRB method: Globally for the non-EVPN IRB interface configured as a multicast router interface.

      With IGMPv3, you don't need the IGMP snooping evpn-ssm-reports-only option because the external multicast interface isn't extended in the EVPN instance.
    Note:

    With the classic L3 interface method for external multicast, you don't configure IGMP snooping, an L2 optimization, on the pure L3 interfaces.

L2 multicast optimizations—MLD snooping with SMET for IPv6 multicast traffic

Enable MLD snooping at L2 with MLDv1 or MLDv2 protocols if you have IPv6 multicast traffic in the fabric. With MLD snooping, the device routes or forwards multicast traffic only toward interested access-side receivers. Receivers send MLD reports to express interest in receiving traffic for a multicast group.

When you enable MLD snooping, the device also automatically advertises SMET Type 6 routes. With SMET, the device sends copies of the traffic into the EVPN core only toward other devices that have interested receivers.

Configure MLD snooping as follows:

  • In the EVPN instance(s) for each OISM revenue VLAN and the SBD.

  • On border leaf devices for external multicast only with the non-EVPN IRB method with these guidelines:

    • Configure MLD snooping globally with the non-EVPN IRB interface configured as a multicast router interface.

    • With MLDv2 in SSM mode, you don't need the MLD snooping evpn-ssm-reports-only option because the external multicast interface isn't extended in the EVPN instance.

    Note:

    With the classic L3 interface method for external multicast, you don't configure MLD snooping, an L2 optimization, on the pure L3 interfaces.

L3 VRF instance

Configure a routing instance (instance type vrf) that supports the L3 OISM routing functions.

originate-smet-on-revenue-vlan-too

(Optional) Enable the device to originate SMET Type 6 routes for revenue bridge domains (as well as on the SBD) upon receiving local IGMP or MLD reports. By default, OISM devices originate Type 6 routes only on the SBD.

Use this option for compatibility with other vendor devices that don't support OISM. Those devices can't create the right states for the revenue bridge domain VLANs upon receiving Type 6 routes on the SBD.

install-star-g-routes

Enable the Routing Engine (RE) on the device to install (*,G) multicast routes on the Packet Forwarding Engine (PFE) for all of the revenue bridge domain VLANs in the routing instance immediately upon receiving an EVPN Type 6 route. Setting this option helps minimize traffic loss when multicast traffic first arrives.

This option is mutually exclusive with the conserve-mcast-routes-in-pfe option, so you can't set both options together in a routing instance.

We require this option on:

  • The QFX10000 line of switches, QFX5130-32CD switches, and QFX5700 switches when you configure those devices in the AR replicator role.

  • In releases prior to Junos OS and Junos OS Evolved Release 23.4R1, on the QFX10000 line of switches and the PTX10000 line of routers when you configure those devices as OISM server leaf or border leaf devices.

    We no longer require this option in this case starting in Junos OS and Junos OS Evolved Release 23.4R1.

Setting this option is generally not recommended in any use cases other than those listed above where it is required.

See Latency and Scaling Trade-Offs for Installing Multicast Routes with OISM (install-star-g-routes Option) for details on how and when to set this option.

conserve-mcast-routes-in-pfe

(Required on ACX Series routers, QFX5130-32CD switches, and QFX5700 switches when you configure those devices as OISM server leaf or OISM border leaf devices) Configure this option with OISM to conserve PFE table space. The device installs only the L3 multicast routes and avoids installing L2 multicast snooping routes.

Don't set this option on QFX5130-32CD and QFX5700 switches when you configure those devices as standalone AR replicator devices with OISM. This option is mutually exclusive with the install-star-g-routes option, so you can't set both options together in a routing instance.

See ACX Series Routers, QFX5130-32CD Switches, and QFX5700 Switches as Server Leaf and Border Leaf Devices with OISM for details.

Table 5 lists the elements you configure on the server leaf devices.

Table 5: Configuration Elements on OISM Server Leaf Devices
Configuration Element Description

PIM in passive mode on all revenue bridge domains and the SBD

Configure this mode to facilitate local routing without all of the traditional PIM protocol functions. The server leaf device:

  • Doesn't form PIM neighbor relationships with any other devices (avoids sending or receiving PIM hello messages).

  • Acts as a PIM local RP. The device creates the PIM state locally from IGMP or MLD reports. The device also doesn't do source registration. Only the border leaf devices perform source registration toward the external PIM RP.

PIM with the accept-remote-source option on SBD IRB interfaces

This option enables an SBD IRB interface to accept multicast traffic from a source that isn't on the same subnet. The server leaf devices require this setting because:

  • The border leaf devices route the multicast source traffic from the external multicast interfaces to the SBD toward the server leaf devices.

  • The traffic arrives at the server leaf device on the SBD IRB interface, which is not a PIM neighbor. The source is not a local source, so without this setting, the device would otherwise drop the traffic.

OSPF for peer connectivity to support external multicast traffic on SBD

Configure OSPF so the server leaf devices learn routes to external sources when multicast traffic from outside the fabric arrives on the SBD. The device creates the PIM (S,G) entries it needs to forward the traffic from the SBD to the revenue bridge domains.

On the server leaf devices, you configure OSPF on the device loopback interface (lo0) and the SBD IRB interface.

Table 6 lists the elements you configure on the border leaf devices based on the external multicast method you use.

Table 6: Configuration Elements on OISM Border Leaf Devices
Configuration Element External Multicast Method Description

M-VLAN and corresponding IRB interface (in EVPN instance)

M-VLAN IRB method

Configure a VLAN to serve as the M-VLAN, and extend this VLAN in the EVPN instance. This VLAN must be distinct from the SBD or any revenue bridge domain VLANs in the EVPN fabric. Also configure an M-VLAN IRB interface in the EVPN instance. See OISM Bridge Domains (VLANs) for more information about the M-VLAN.

You can link multiple border leaf device M-VLAN IRB interfaces to the external multicast router in the same EVPN ES. The usual EVPN multihoming DF rules apply to prevent sending duplicate traffic on the M-VLAN.

L2 multicast router interface on external multicast ports

M-VLAN IRB method or non-EVPN IRB method

Configure the multicast-router-interface option with IGMP snooping or MLD snooping on the L2 ports that link the border leaf device to the external PIM domain at L2.

With the M-VLAN IRB method, these interfaces support multicast traffic when the external domain router is multihomed to the border leaf devices. As a result, multihomed M-VLAN use cases require this configuration. This setting is also required with the non-EVPN IRB method.

PIM on M-VLAN IRB interface

M-VLAN IRB method

Configure PIM in distributed designated router (DR) mode (distributed-dr) or standard PIM mode on an M-VLAN IRB interface. We recommend using distributed DR mode in most cases, especially on border leaf devices where the external PIM router is multihomed to multiple border leaf devices.

The device uses PIM to:

  • Form PIM neighbor relationships with the other border leaf devices and the external PIM router. The external PIM router might be multihomed in an ES. As a result, EVPN-forwarded traffic might come to a different peer border leaf device.

  • Elect a single last-hop router (LHR) designated router (DR) for the M-VLAN, so only one device does source registration for an internal source toward the PIM RP.

You configure PIM on the M-VLAN IRB interface in the tenant VRF instances. similar to how you configure PIM on the revenue bridge domains.

L3 physical interface with IP address

Classic L3 interface

Configure a physical L3 interface with an IP address for external multicast that connects the border leaf device to the external PIM domain at L3.

Define the external multicast L3 interface in a different subnet on each border leaf device.

Note:

The L3 interface connection can be an AE interface bundle.

PIM on the logical interface for the external multicast physical L3 interface

Classic L3 interface

Configure the logical interface (unit 0) for the external multicast L3 interface in the tenant VRF instances. Configure standard PIM mode on the logical interface.

With this setting, the border leaf device forms a PIM neighbor relationship with the external PIM router to send join messages and transmit or receive external multicast traffic.

Extra VLAN and corresponding IRB interface (not in EVPN instance)

Non-EVPN IRB method

Configure an extra VLAN and IRB interface globally for external multicast without EVPN signaling. This VLAN and the IRB interface subnet must be distinct from the SBD, any revenue bridge domain VLANs, or the extra VLAN on any other border leaf device in the EVPN fabric. See OISM Bridge Domains (VLANs) for more about this extra VLAN and external multicast method.

PIM on non-EVPN IRB interface

Non-EVPN IRB method

Configure PIM on the non-EVPN IRB interface in the tenant VRF instances.

With this setting, the border leaf device forms a PIM neighbor relationship with the external PIM router to send join messages and transmit or receive external multicast traffic.

PIM on SBD IRB interface

All

Configure standard PIM mode on the SBD IRB interface in the tenant VRF instances for SBD routing and forwarding. With this setting, the border leaf device:

  • Routes external multicast source traffic from the external multicast interfaces to the SBD, and forwards copies toward server leaf devices with multicast receivers.

  • Forms PIM neighbor relationships with the other border leaf devices.

  • Elects a single LHR DR on the SBD among the peer border leaf devices. Only this elected PIM DR forwards the external multicast source traffic on the SBD. This election prevents peer border leaf devices from forwarding duplicate traffic into the EVPN core.

PIM EVPN gateway (PEG) role

All

(include the external IRB option for EVPN, external-irb interface-name, only with the M-VLAN IRB method)

Configure the pim-evpn-gateway role on the border leaf device to connect to the external PIM router. In this role, the border leaf device uses traditional PIM routing behavior and does local routing, as follows:

For externally-sourced traffic:

  • Routes external source traffic from the M-VLAN IRB interface, L3 interface, or non-EVPN IRB interface to any local receivers on the revenue bridge domains on the border leaf device.

  • Routes external source traffic from the M-VLAN IRB interface, L3 interface, or non-EVPN IRB interface to the SBD to reach any internal receivers in the fabric. (The device forwards the traffic to the EVPN core only on the SBD to other OISM leaf devices.)

For internally-sourced traffic:

  • Locally routes traffic that arrives on a revenue bridge domain to other revenue bridge domains.

  • Routes the traffic to external multicast receivers through the M-VLAN IRB interface, L3 interface, or non-EVPN IRB interface. The device doesn't forward the traffic back into the EVPN core.

EVPN IMET routes for PEG interfaces include the OISM PEG flag in the multicast flags extended community field of the route.

OSPF for external multicast interface peer connectivity

All

Configure an OSPF area in the tenant L3 VRF instance so the border leaf device learns routes to the multicast sources. The device requires these routes to support forwarding multicast traffic:

  • From sources inside the fabric toward receivers outside the fabric.

  • From sources outside the fabric toward receivers inside the fabric.

The device needs this route information to create the PIM (S,G) entries to forward the traffic on the external multicast interfaces, the SBD, and the revenue bridge domains.

On the border leaf devices, you configure OSPF on the device loopback interface, the SBD IRB interface, and the external multicast interface (M-VLAN IRB interface, L3 interface, or non-EVPN IRB interface).

PIM distributed DR mode on revenue bridge domain IRB interfaces

All

Configure PIM in distributed DR mode (distributed-dr) on the revenue bridge domain IRB interfaces in the tenant VRF instances. In this mode, the border leaf device:

  • Forms PIM neighbor relationships with other PIM devices to support multihomed external PIM router connections.

  • Acts as the LHR DR on its revenue bridge domain IRB interfaces, and creates the PIM state locally from received IGMP or MLD reports. As a result:

    • The device can do local multicast routing between the revenue bridge domains.

    • One peer border leaf device becomes the PIM DR that performs source registration toward the PIM RP. PIM hello messages and the PIM DR election process determine the PIM DR.

PIM accept-join-always-from option and policy on M-VLAN IRB interface M-VLAN IRB method

Set this option on the M-VLAN IRB interface in the tenant VRF instances when the external PIM router is multihomed to more than one EVPN border leaf device. With this option, the device can accept and install the same PIM (S,G) join states on multihoming peer border leaf devices. This option supports sending multicast traffic from sources inside the fabric to receivers in the external PIM domain.

With multihoming on the M-VLAN, the usual EVPN multihoming DF rules apply in an ES to prevent sending duplicate traffic. If peer border leaf devices have the same valid join states in place, any device that is the EVPN DF can forward the multicast traffic.

Configure this statement with policies that specify the interface should always install PIM joins from upstream neighbor addresses that correspond to the external PIM router.

Note:

You don't use this option with the classic L3 interface and non-EVPN IRB methods. Those methods don't extend the external multicast interfaces in the EVPN instance.

See the following sections for more details on configuring OISM devices:

For a full OISM configuration example of a data center fabric use case that includes classic L3 interface connections to the external PIM domain, see Optimized Intersubnet Multicast (OISM) with Assisted Replication (AR) for Edge-Routed Bridging Overlays.

How OISM Works

The following sections describe how OISM works and show how the multicast traffic flows in several common use cases.

Local Routing on OISM Devices

In Figure 3, we illustrate how local routing and forwarding works in general on OISM devices. As the figure shows, OISM local routing forwards the traffic on the source VLAN. Each leaf device routes the traffic locally to its receivers on other VLANS, which avoids hairpinning for intersubnet routing on the same device.

Figure 3: Local Routing with OISMLocal Routing with OISM

In this case, the source traffic comes from Mcast-Src-1 on VLAN-1, the blue VLAN. Server leaf devices use IRB interfaces and PIM in passive mode to route traffic between VLANs. With PIM in passive mode, server leaf devices:

  • Don’t become PIM neighbors with the other leaf devices.

  • Act as a local PIM RP, create local PIM state upon receiving IGMP or MLD reports, and avoid doing source registration.

As a result, the server leaf devices forward and route multicast traffic within the fabric as follows to receivers interested in the multicast group:

  • The ingress leaf device (Leaf-1) forwards the traffic on the source VLAN into the EVPN fabric toward the other leaf devices with interested receivers.

  • All of the server leaf devices don't need to forward the traffic back into the EVPN core to another device that is a designated router. Server leaf devices can locally:

    • Forward the traffic on the source VLAN toward local interested receivers on the source VLAN.

    • Route the traffic from the source VLAN through the IRB interfaces toward local interested receivers in other VLANs.

Multicast Traffic Forwarding and Routing with Source and Receivers Inside the EVPN Data Center

When the multicast source is inside the EVPN fabric, the server leaf devices receive the multicast traffic on the source VLAN. Then they locally route or forward the traffic as described in Local Routing on OISM Devices.

The following figure illustrates OISM local routing and forwarding within an EVPN fabric in detail. The figure also shows how local routing works with EVPN multihoming for a multicast receiver.

Figure 4: OISM with an Internal Multicast Source and Internal Multicast ReceiversOISM with an Internal Multicast Source and Internal Multicast Receivers

In Figure 4, the multicast source, Mcast-Src-1, is single-homed to Leaf-1. The source VLAN is VLAN-1 (the blue VLAN). Multicast control and data traffic flow proceeds as follows:

  1. Receivers on all three server leaf devices sent IGMP or MLD reports (join messages) expressing interest in receiving the traffic for a multicast group.

  2. Leaf-1 forwards the traffic on the source VLAN to both Leaf-2 and Leaf-3 because both leaf devices have interested receivers. In this case, the receivers on Leaf-2 and Leaf-3 use single-homing.

  3. Leaf-2 and Leaf-3 forward or locally route the traffic to their interested receivers (Rcvr-2, Rcvr-3, and Rcvr-4) as described in Local Routing on OISM Devices.

  4. Rcvr-1 on VLAN-2 is multihomed to Leaf-1 and Leaf-2 in an EVPN ES. Rcvr-1 has expressed interest in receiving the multicast traffic, so:

    • Both server leaf devices, Leaf-1 and Leaf-2, receive the IGMP or MLD report.
    • Both Leaf-1 and Leaf-2 locally route the traffic from the source VLAN (VLAN-1) because each device has the PIM passive mode configuration.
    • However, because Leaf-1 is the DF for the EVPN ES, only Leaf-1 forwards the traffic to Rcvr-1.
  5. The border leaf devices receive the multicast traffic through the EVPN fabric on the source VLAN. Note that the border leaf devices could have local receivers, although we don't show that case. With local receivers, the device also locally routes or forwards the traffic to those receivers the same way the server leaf devices do.

Figure 4 also shows that the border leaf devices locally route the traffic from the source VLAN toward any external multicast receivers in the external PIM domain. See Supported Methods for Multicast Data Transfer to or from an External PIM Domain for the available external multicast methods by platform. Later sections describe the multicast control and data traffic flow for external source and external receiver use cases.

Multicast Traffic From an Internal Source to Receivers Outside the EVPN Data Center—M-VLAN IRB Method

In Figure 5, we illustrate the OISM use case where a multicast source inside the EVPN fabric sends multicast traffic to an interested receiver outside the fabric using the M-VLAN IRB method for external multicast. (Supported Methods for Multicast Data Transfer to or from an External PIM Domain lists external multicast method support by platform.)

The border leaf devices you configure in the OISM PEG role receive the multicast traffic on the source VLAN through the EVPN core. Then the border leaf devices replicate the traffic and route it onto the M-VLAN toward the external PIM domain to reach the external receiver.

Note:

PEG border leaf devices only send multicast source traffic received on the revenue bridge domains to the M-VLAN. These devices don't forward the traffic back into the EVPN core toward the other border leaf devices.

This use case also shows an internal multihomed multicast source and local routing to single-homed receivers inside the fabric.
Figure 5: OISM with an Internal Multicast Source and an External Multicast Receiver—M-VLAN IRB MethodOISM with an Internal Multicast Source and an External Multicast Receiver—M-VLAN IRB Method

In Figure 5, the internal source for a multicast group is Mcast-Src-2, the same device as Rcvr-1, which is multihomed to Leaf-1 and Leaf 2. The source sends the multicast traffic on VLAN-2. The external receiver, Ext-Mcast-Rcvr, expresses interest in receiving the multicast traffic for that multicast group (sends a join message). Internal receivers Rcvr-3 (on VLAN-1) and Rcvr-4 (on VLAN-2) also request to join the multicast group and receive the traffic.

Note that the PIM router is multihomed to both BL-1 and BL-2, the PEG devices, in the EVPN fabric. Those connections are in the same ES; the DF election process chooses one of these devices as the DF for the ES. Only the DF will forward traffic (on the M-VLAN) toward external receivers.

The source traffic reaches the interested internal and external receivers as follows:

Traffic Flow from Multihomed Source to Internal Receivers

These steps summarize the multicast control and data traffic flow from the multihomed source to the internal receivers:

  1. Mcast-Src-2 (also labeled Rcvr-1) originates the traffic on VLAN-2, the red VLAN. Because the device is multihomed to Leaf-1 and Leaf-2, the device hashes the traffic on VLAN-2 to one of those server leaf devices. In this case, Leaf-2 receives the traffic.

  2. The red arrows show that Leaf-2 forwards the traffic on the source VLAN, VLAN-2, only to:

    • The other server leaf devices with interested receivers—In this case, only Leaf-3.

    • The border leaf devices, which both act in the OISM PEG role.

    Note that no receivers behind Leaf-1 or Leaf-2 sent an IGMP report to join the multicast group. With IGMP snooping and SMET forwarding enabled, Leaf-2 doesn't forward the traffic to Leaf-1 because Leaf-1 has no interested receivers. Leaf-2 also doesn't locally route the traffic to Rcvr-2 for the same reason.

  3. Leaf-3 receives the source traffic on VLAN-2. Then Leaf-3 routes the traffic locally to VLAN-1 to Rcvr-3. Leaf-3 also forwards the traffic to Rcvr-4 on VLAN-2.

  4. Both border leaf devices BL-1 and BL-2 also receive the source traffic from the EVPN core. We describe the external multicast flow next.

Traffic Flow to External Receiver—M-VLAN IRB Method

These steps summarize the multicast control and data traffic flow in Figure 5 from the border leaf devices toward the external receiver using the M-VLAN IRB method:

  1. In the external PIM domain, the PIM RP enters a PIM (*,G) multicast routing table entry. The entry includes the L3 interface toward Ext-Mcast-Rcvr as the downstream interface.

  2. Both border leaf devices BL-1 and BL-2 receive the source traffic from the EVPN core. The IRB interface on VLAN-2 on one of these border leaf devices is the PIM DR for VLAN-2. In this case, the PIM DR is on BL-1, so BL-1 sends a PIM Register message toward the PIM RP on the M-VLAN IRB interface.

  3. The PIM RP sends a PIM Join message back toward BL-1. BL-1 creates an (S,G) multicast routing table entry as follows:

    • The source address is the IP address of Mcast-Src-2 in VLAN-2.
    • The downstream interface is the M-VLAN IRB interface.
  4. Both BL-1 and BL-2 are PEG devices and configured in PIM distributed DR mode for the revenue bridge domain (VLAN-1 and VLAN-2) IRB interfaces. As a result, both BL-1 and BL-2 receive the PIM Join and create a similar (S,G) state. Both devices route the traffic locally from VLAN-2 to the M-VLAN.

    However, only the DF for the M-VLAN ES actually forwards the data on the M-VLAN to the external PIM domain. In this case, BL-1 is the DF and sends the traffic toward the external receiver. (See the label "M-VLAN ESI DF" and the black arrow between BL-1 and the PIM router in Figure 5.)

  5. The PIM RP receives the traffic from the OISM M-VLAN IRB interface connection. The PIM router sends the traffic to an L3 interface toward the external receiver.

Multicast Traffic from an Internal Source to Receivers Outside the EVPN Data Center—L3 Interface Method or Non-EVPN IRB Method

In Figure 6, we illustrate the OISM use case where a multicast source inside the EVPN fabric sends multicast traffic to a receiver outside the fabric using either of the following methods for external multicast:

  • Classic L3 interface external multicast method:

    On each border leaf device, you configure a classic L3 interface with family inet that connects to the external PIM router. You assign an IP address to that interface on a different subnet than the L3 interface subnets on other border leaf devices in the fabric.

    You enable PIM on the interface and include the interface in the tenant VRF instances that have multicast data receivers. This method differs from the M-VLAN IRB method because you don't extend this interface in the EVPN instance.

    Note:

    The L3 interface connection here can be an individual physical interface or an AE interface bundle that includes multiple physical L3 interfaces.

  • Non-EVPN IRB external multicast method:

    On each border leaf device, you configure a unique extra VLAN that is only for external multicast. You also configure a corresponding L3 IRB interface with an IP address that connects to the external PIM router. The extra VLAN ID can't be the same as the VLAN ID of the revenue bridge domains, SBD, or extra VLAN on any other border leaf device in the fabric. In addition, similar to the L3 interface method, the non-EVPN IRB interfaces on different border leaf devices should connect to the PIM router on different subnets in the fabric.

    You enable PIM on the IRB interface and include the interface in the tenant VRF instances that have multicast data receivers. This method differs from the M-VLAN IRB method because you don't extend this VLAN or IRB interface in the EVPN instance.

Supported Methods for Multicast Data Transfer to or from an External PIM Domain lists the platforms that support these external multicast methods.

Figure 6 includes the same internal multihomed source, internal receivers, and an external receiver as in Multicast Traffic From an Internal Source to Receivers Outside the EVPN Data Center—M-VLAN IRB Method. See the steps in Traffic Flow from Multihomed Source to Internal Receivers for details on the internal multicast traffic flow. In this section we describe only what's different in this case, which is the multicast traffic flow from the border leaf devices to the external receiver.

Figure 6: OISM with an Internal Multicast Source and an External Multicast Receiver—L3 Interface or Non-EVPN IRB MethodOISM with an Internal Multicast Source and an External Multicast Receiver—L3 Interface or Non-EVPN IRB Method

In Figure 6, the internal source for a multicast group is Mcast-Src-2, which is multihomed to Leaf-1 and Leaf-2. The source sends the multicast traffic on VLAN-2, the red VLAN. The external receiver, Ext-Mcast-Rcvr, expresses interest in receiving the multicast traffic for that multicast group (sends a join message).

The external multicast flow in this use case is very similar to the M-VLAN IRB use case. The border leaf devices in the OISM PEG role receive the multicast traffic on the source VLAN through the EVPN core. However, the main difference in this case is that the external multicast interfaces don't use EVPN signaling and don't share an ESI across the border leaf devices. The external multicast interfaces on each border leaf device are distinct, and each have L3 reachability to the external PIM gateway router. The border leaf device that establishes the PIM join state replicates and sends the traffic on the L3 interface or non-EVPN IRB interface to the external PIM domain with the external receiver.

Note:

PEG border leaf devices only send multicast source traffic received on the revenue bridge domains to the external PIM domain. These devices don't forward the traffic back into the EVPN core toward the other border leaf devices.

The following section explains how the source traffic reaches the interested external receiver.

Traffic Flow to External Receiver—L3 Interface or Non-EVPN IRB Method

These steps summarize the multicast control and data traffic flow in Figure 6 from the border leaf devices toward the external receiver using the classic L3 interface method or the non-EVPN IRB method:

  1. In the external PIM domain, the PIM RP enters a PIM (*,G) multicast routing table entry. The entry includes the L3 interface toward Ext-Mcast-Rcvr as the downstream interface.

  2. Both border leaf devices BL-1 and BL-2 receive the source traffic from the EVPN core on VLAN-2. The IRB interface on VLAN-2 on one of these border leaf devices is the PIM DR for VLAN-2. In this case, the PIM DR is on BL-1, so BL-1 sends a PIM Register message toward the PIM RP on its external multicast L3 interface or non-EVPN IRB interface.

  3. The PIM RP sends a PIM Join message back toward BL-1. BL-1 receives the PIM join and creates an (S,G) multicast routing table entry as follows:

    • The source address is the IP address of Mcast-Src-2 in VLAN-2.
    • The downstream interface is the external multicast L3 interface or the non-EVPN IRB interface.
  4. BL-1 routes the traffic from VLAN-2 to its external multicast L3 interface or non-EVPN IRB interface.

  5. The PIM RP receives the traffic from BL-1 on the external multicast interface. The PIM router sends the traffic to an L3 interface toward the external receiver. The external receiver receives the multicast traffic.

Multicast Traffic from an External Source to Receivers Inside the EVPN Data Center—M-VLAN IRB Method

Figure 7 illustrates the OISM use case where a multicast source outside the EVPN fabric sends multicast traffic to receivers inside the fabric. This use case exhibits the two main ways OISM uses the SBD in the EVPN core:

  • To carry external multicast source traffic.

  • To advertise SMET Type 6 routes.

The Type 6 routes ensure that the border leaf devices only forward the traffic toward the EVPN devices with interested receivers.

OISM border leaf devices receive external multicast source traffic through the M-VLAN IRB interfaces. OISM devices use the SBD to forward traffic toward EVPN server leaf devices with interested receivers on the revenue bridge domains. Each leaf device then locally forwards or routes the traffic on the revenue bridge domains to its local receivers.

This use case has an internal receiver that is multihomed to two server leaf devices.

Figure 7: OISM with an External Multicast Source and an Internal Multihomed Multicast Receiver—M-VLAN IRB MethodOISM with an External Multicast Source and an Internal Multihomed Multicast Receiver—M-VLAN IRB Method

In Figure 7, Rcvr-1 within the EVPN fabric is multihomed to server leaf devices Leaf-1 and Leaf-2. Rcvr-1 expresses interest in receiving traffic from a multicast group. The source for the multicast traffic for the group is Ext-Mcast-Src in the external PIM domain.

The external source traffic reaches the interested multihomed receiver, Rcvr-1, as follows:

Multicast Control Flow between the Internal Multihomed Receiver and the External Source—M-VLAN IRB Method

These steps summarize the multicast control flow in this use case:

  1. Rcvr-1 sends an IGMP join message to both multihoming peers Leaf-1 and Leaf-2.

  2. Both Leaf-1 and Leaf-2 generate an EVPN Type 6 route toward the EVPN core on the SBD. The Type 6 (SMET) route advertises that Rcvr-1 is interested in the multicast data.

  3. Both border leaf devices BL-1 and BL-2 receive the Type 6 route on the SBD.

  4. The Type 6 route (on the SBD) signals the border leaf devices to create a PIM join toward the PIM RP (reachable through the M-VLAN). However, to avoid duplicate join messages, only the border leaf device that is the PIM DR for the SBD generates the PIM join message. In this case, the figure shows the PIM DR for the SBD is BL-1. BL-1 sends the PIM join message toward the PIM RP by way of its neighbor, the M-VLAN IRB interface.

  5. The PIM RP receives the join message. Then the PIM RP creates a PIM (*,G) entry in the multicast routing table with the M-VLAN IRB interface as the downstream interface.

  6. The external source Ext-Mcast-Src registers with the PIM RP. The PIM RP has a multicast route for the group with the M-VLAN IRB interface as the downstream interface. As a result, the PIM RP routes the multicast traffic coming in at L3 onto its connection to the M-VLAN IRB toward BL-1 or BL-2. In this case, BL-1 sent the PIM join, so BL-1 receives the traffic on its M-VLAN IRB interface.

Traffic Flow from the Border Leaf Devices to Internal Receivers—M-VLAN IRB Method

In Figure 7, BL-1 is the PIM DR for the SBD and sent the PIM join toward the external PIM domain. BL-1 receives and routes (or forwards) the external source traffic as follows:

  1. BL-1 routes the traffic locally from the M-VLAN to the SBD on its SBD IRB interface because BL-1 is the PIM DR for the SBD. See the small gray arrow from the M-VLAN to the SBD on BL-1.

  2. BL-1 forwards a copy of the traffic on the M-VLAN toward BL-2 because both border leaf devices are in the PEG role. See the black arrow from BL-1 toward BL-2.

    As a PEG device using the M-VLAN IRB method, BL-2 expects to receive external multicast traffic only on the M-VLAN IRB interface. If BL-2 has any local receivers, BL-2 can receive the traffic and route it locally to those receivers.

  3. BL-1 also forwards a copy of the traffic on the SBD into the EVPN core to BL-2. See the green arrow from BL-1 toward BL-2.

    BL-2 drops the traffic because again, as a PEG device using the M-VLAN IRB method, BL-2 expects to receive external source traffic only on the M-VLAN IRB interface. BL-2 doesn't expect external source traffic on the SBD IRB interface from BL-1. In other words, BL-2 sees this case as a source interface mismatch (a reverse path forwarding [RFP] failure).

    Note:

    One reason why the ingress border leaf device also forwards a copy on the SBD to other border leaf devices is to ensure that another border leaf device can receive external source traffic if its M-VLAN interface is down. Then any interested local receivers on the other border leaf device can still get the traffic.

  4. BL-1 selectively forwards copies of the traffic on the SBD to the server leaf devices with interested receivers, based on the advertised Type 6 routes.

    In this case, Leaf-1 and Leaf-2 have a multihomed interested receiver, Rcvr-1, on VLAN-2. As a result, BL-1 sends the traffic toward both leaf devices. See the green arrows from BL-1 toward Leaf-1 and Leaf-2.

    Note:

    In a similar use case with the PIM router multihomed to BL-1 and BL-2, BL-1 might receive the external multicast source traffic, but BL-2 is the PIM DR on the SBD. One reason why BL-1 forwards the incoming external multicast traffic toward BL-2 on the M-VLAN is so that BL-2 can handle this use case. See the black arrow from BL-1 toward BL-2 on the M-VLAN in Figure 7. If BL-2 is the PIM DR on the SBD, upon receiving the traffic on the M-VLAN from BL-1, BL-2 forwards the traffic on the SBD toward Leaf-1 and Leaf-2. In this case, the green arrows in the figure would flow from BL-2 toward the other EVPN devices instead of flowing from BL-1.

  5. Leaf-1 and Leaf-2 locally route the traffic from the SBD IRB interface to the revenue bridge domain IRB interface for VLAN-2 toward the interested (multihomed) receiver. However, with EVPN multihoming, only the EVPN DF in the ES forwards the traffic toward Rcvr-1 so Rcvr-1 doesn't get duplicate traffic.

    In this case, Leaf-1 is the EVPN DF, so only Leaf-1 forwards the traffic to Rcvr-1.

What Happens When a Multihomed External PIM Router Load-Balances the Traffic—M-VLAN IRB Method

In Figure 7, the external PIM gateway router is multihomed to BL-1 and BL-2 on an ES in the EVPN fabric. If the pair of connections on the PIM router side is an AE interface bundle, the PIM router does load balancing among the interfaces in the bundle. In that case, both BL-1 and BL-2 will each receive some of the multicast traffic flow from the external source. However, all receivers should receive all of that traffic. For simplicity, the figure doesn't show traffic arrows for this load balancing, but we'll describe the flow here.

BL-1 and BL-2 each receive part of the external multicast source traffic on their M-VLAN IRB interfaces. However, because BL-1 is the PIM DR on the SBD, only BL-1 will route the traffic onto the SBD into the EVPN fabric, as follows:

  1. BL-1 routes the traffic it receives onto the SBD toward the server leaf devices.

    BL-1 also forwards that traffic on the M-VLAN and routes it on the SBD toward BL-2, in case BL-2 has any local receivers (as described in Traffic Flow from the Border Leaf Devices to Internal Receivers—M-VLAN IRB Method).

  2. BL-2 forwards the traffic it receives from the external PIM domain on the M-VLAN toward BL-1, because BL-1 only expects to receive external source traffic on the M-VLAN.

    Due to DF and split horizon rules, BL-2 won't forward any traffic it receives on the M-VLAN from BL-1 into the EVPN core or back toward the source, BL-1.

  3. BL-1 routes the traffic it receives on the M-VLAN from BL-2 onto the SBD toward the server leaf devices.

What Happens with Local Receivers on Border Leaf Devices—M-VLAN IRB Method

Figure 7 doesn't show local receivers attached to the border leaf devices. However, let's look briefly at PIM join message flow and how external source traffic reaches local receivers on border leaf devices.

Consider that BL-1 or BL-2 has an interested receiver on a revenue bridge domain in the fabric. In that case:

  1. Both devices generate a PIM join on the IRB interfaces for the revenue bridge domain toward the PIM RP.

  2. You configure the border leaf devices with PIM in distributed DR mode on the revenue bridge domain IRB interfaces. That way, neither BL-1 nor BL-2 acts as the PIM DR alone. Both devices locally route external multicast source traffic coming in on the M-VLAN IRB interface to the appropriate revenue bridge domain IRB interface.

Multicast Traffic from an External Source to Receivers Inside the EVPN Data Center—L3 Interface Method or Non-EVPN IRB Method

Figure 8 illustrates the OISM use case where a multicast source outside the EVPN fabric sends multicast traffic to receivers inside the fabric. In this case, the fabric uses the classic L3 interface or non-EVPN IRB external multicast method to connect to the external PIM domain. Also, this case includes the following internal receivers:

  • A receiver that is multihomed to two server leaf devices.

  • A local receiver on one of the border leaf devices.

This use case, like the M-VLAN IRB external source use case in Figure 7, shows the two main ways OISM uses the SBD in the EVPN core—To carry external multicast source traffic and to advertise SMET Type 6 routes. The Type 6 routes ensure that the border leaf devices only forward the traffic toward the EVPN devices with interested receivers.

Figure 8: OISM with an External Multicast Source and an Internal Multihomed Multicast Receiver—L3 Interface or Non-EVPN IRB MethodOISM with an External Multicast Source and an Internal Multihomed Multicast Receiver—L3 Interface or Non-EVPN IRB Method

In Figure 8:

  • Rcvr-1 is multihomed to server leaf devices Leaf-1 and Leaf-2 in the EVPN fabric, and expresses interest in receiving traffic from a multicast group.

  • Rcvr-5 on BL-2 in the EVPN fabric is also interested in receiving the multicast traffic.

  • Ext-Mcast-Src in the external PIM domain is the source for the traffic for the multicast group.

The multicast control flow and data traffic flow for external multicast are similar for the classic L3 interface and the non-EVPN IRB interface methods. As a result, in this section we commonly say external multicast interface when we refer to the border leaf device external connection points.

The external source traffic reaches the interested receivers (Rcvr-1 and Rcvr-5) as follows:

Multicast Control Flow between the Internal Receivers and the External Source—L3 Interface or Non-EVPN IRB Method

These steps summarize the multicast control flow in this use case:

  1. Rcvr-1 sends an IGMP or MLD join message on VLAN-2 to both multihoming peers Leaf-1 and Leaf-2.

  2. Both Leaf-1 and Leaf-2 generate an EVPN Type 6 route toward the EVPN core on the SBD. The Type 6 (SMET) route advertises that Rcvr-1 is interested in the multicast data.

  3. Both border leaf devices BL-1 and BL-2 receive the Type 6 route on the SBD.

  4. The Type 6 route (on the SBD) signals the border leaf devices to create a PIM join toward the PIM RP (reachable through the external multicast interface). However, to avoid duplicate join messages for server leaf devices on the SBD, only the border leaf device that is the PIM DR for the SBD generates the PIM join message. In this case, the figure shows the PIM DR for the SBD is BL-1. BL-1 sends the PIM join message toward the PIM RP by way of its PIM neighbor, the external multicast interface.

  5. The local receiver on BL-2, Rcvr-5, also sends an IGMP or MLD join message on VLAN-2 to BL-2. Note that in this case, BL-1 and BL-2 are not multihoming peers in an EVPN ES. As a result, BL-2 sends a separate PIM join message on its external multicast interface because it has a local interested receiver (Rcvr-5).

  6. The PIM RP receives the join messages. The PIM RP creates PIM (*,G) entries in the multicast routing table with the BL-1 and BL-2 external multicast interfaces as downstream interfaces.

  7. The external source Ext-Mcast-Src registers with the PIM RP. The PIM RP has multicast routes for the group with the BL-1 and BL-2 external multicast interfaces as downstream interfaces. As a result, the PIM RP routes the multicast traffic coming in at L3 toward BL-1 and BL-2.

Both BL-1 and BL-2 receive the multicast traffic. The next section explains how the border leaf devices forward or route the traffic in the EVPN fabric.

Traffic Flow from the Border Leaf Devices to Internal Receivers—L3 Interface or Non-EVPN IRB Method

In Figure 8, BL-1 is the PIM DR for the SBD and sent the PIM join message toward the external PIM domain for server leaf devices on the SBD. BL-2 also sent a PIM join message toward the external PIM domain for its interested local receiver.

BL-1 and BL-2 receive the external source traffic, and route (or forward) it as follows:

  1. BL-1 routes the traffic locally from the external multicast interface to the SBD IRB interface because BL-1 is the PIM DR for the SBD. See the small gray arrow from the external multicast interface to the SBD on BL-1.

  2. BL-1 forwards a copy of the traffic on the SBD into the EVPN core to BL-2. See the green arrow from BL-1 toward BL-2.

    However, BL-2 drops the traffic from the SBD because, as a PEG device using the classic L3 interface or non-EVPN IRB method, BL-2 doesn't expect external source traffic on the SBD IRB interface from BL-1. If BL-2 has interested receivers, it would have sent a PIM join message and should receive the same traffic from its external multicast connection.

    Note:

    One reason why the ingress border leaf device also forwards a copy on the SBD to other border leaf devices is to ensure that another border leaf device can receive external source traffic if its external multicast interface is down. Then any interested local receivers on the other border leaf device can still get the traffic.

  3. BL-2 routes the external multicast traffic to its local receiver, Rcvr-5, on VLAN-2. See the small gray arrow on BL-2 from the external multicast interface to VLAN-2.

    Note:

    The border leaf devices configured in PEG mode that are not the PIM DR on the SBD will still locally route the traffic received from the external multicast interface. These devices don't send traffic from external multicast sources to other PEG border leaf devices on the SBD. These devices also don't forward the traffic on the SBD into the EVPN core.

  4. BL-1 (the PIM DR on the SBD) selectively forwards copies of the traffic on the SBD to the server leaf devices with interested receivers (based on the advertised Type 6 routes). See the green arrows from BL-1 toward Leaf-1 and Leaf-2.

    In this case, Leaf-1 and Leaf-2 have a multihomed interested receiver, Rcvr-1, on VLAN-2. As a result, BL-1 sends the traffic on the SBD toward both leaf devices.

  5. Leaf-1 and Leaf-2 locally route the traffic from the SBD IRB interface to the revenue bridge domain IRB interface for VLAN-2 toward the interested (multihomed) receiver. However, with EVPN multihoming, only the EVPN DF in the ES forwards the traffic toward Rcvr-1 so Rcvr-1 doesn't get duplicate traffic.

    In this case, Leaf-1 is the EVPN DF, so only Leaf-1 forwards the traffic to Rcvr-1. See the red arrow from Leaf-1 toward Rcvr-1.

AR and OISM with an Internal Multicast Source

In Figure 9, we show an OISM use case where you configure the spine devices as standalone AR replicator devices. The OISM server leaf and border leaf devices are AR leaf devices. The AR replicator devices handle replicating the multicast traffic for the OISM server leaf and border leaf devices. This case shows a multicast source and single-homed receivers behind server leaf devices inside the EVPN fabric.

Note:

AR behavior is different when the multicast source is behind a server leaf device that also has a multihomed receiver; see AR and OISM with an Internal Multicast Source and Multihomed Receiver for more on the behavior with that use case.

With OISM traffic from an internal source, the ingress device forwards the traffic in the EVPN fabric on the source VLAN. When you also enable AR, the ingress leaf device forwards one copy of the traffic to an AR replicator. The AR replicator replicates the traffic and sends the copies on the source VLAN to the other leaf devices with interested receivers. Then each leaf device:

  • Locally forwards the traffic toward its receivers on the source VLAN.

  • Locally routes the traffic toward its receivers on the other revenue VLANs.

Figure 9: AR with OISM—Internal Multicast Source and Single-Homed Internal ReceiversAR with OISM—Internal Multicast Source and Single-Homed Internal Receivers

In the use case in Figure 9:

  1. Rcvr-2, Rcvr-3 and Rcvr-4 send IGMP or MLD reports to join the multicast group.

  2. The traffic source for the multicast group, Mcast-Src-1, forwards the traffic on the source VLAN, VLAN-1, to Leaf-1.

  3. Leaf-1 forwards the traffic to one of the available AR replicators on VLAN-1 to replicate the traffic to the other leaf devices with interested receivers. In this case, Leaf-1 forwards the traffic to ARR-1.

    Note:

    See AR Leaf Device Load Balancing with Multiple Replicators for details on how AR leaf devices load-balance among multiple available AR replicators.

  4. ARR-1 replicates the traffic and sends copies on VLAN-1 toward all the leaf devices with interested receivers.

  5. Each server leaf device:

    • Forwards the traffic toward the interested receivers on VLAN-1.

    • Locally routes the traffic to VLAN-2 and forwards it toward the interested receivers on VLAN-2.

    Also note that in Figure 9, Rcvr-1 is multihomed to Leaf-1 and Leaf-2, with Leaf-1 as the ESI DF. As a result, only Leaf-1 forwards the traffic to Rcvr-1 on VLAN-2.

  6. If any external receivers expressed interest in receiving the traffic, the border leaf devices locally route the traffic to the external multicast interface. The external multicast interface sends the traffic toward any interested external receivers based on the external multicast method you configure.

See Configure Assisted Replication for details on how to configure AR.

AR and OISM with an Internal Multicast Source and Multihomed Receiver

In Figure 10, we show an OISM use case similar to the setup in AR and OISM with an Internal Multicast Source. However, in this case, the multicast source is behind a server leaf device that also has a multihomed receiver. In this case, AR operates in extended AR mode by default to efficiently support the multihomed receiver. See Extended AR Mode for Multihomed Ethernet Segments for full details on this mode.

Here is a summary of how multicast traffic ingressing on a server leaf device reaches a multihomed receiver in this case:

  • The ingress server leaf device with an ESI for a multihomed receiver maintains a list of its multihoming peer leaf devices on the ES.

    The AR replicator device also knows which AR leaf devices have multihoming peers.

  • The ingress server leaf device takes care of replicating and forwarding the multicast traffic to any of its multihoming peers that are interested in the traffic.

    The ingress leaf device also sends one copy to an AR replicator device to handle replication and forwarding to any other leaf devices.

Other than this difference in handling replication to the multihoming peers, the traffic flow to an AR replicator and then to the interested receivers is the same as we describe in AR and OISM with an Internal Multicast Source.

Figure 10: AR with OISM—Internal Multicast Source and Multihomed ReceiverAR with OISM—Internal Multicast Source and Multihomed Receiver

In Figure 10:

  1. Rcvr-1, Rcvr-2, Rcvr-3 and Rcvr-4 send IGMP or MLD reports to join the multicast group.

  2. The traffic source for the multicast group, Mcast-Src-1, forwards the traffic on the source VLAN, VLAN-1, to Leaf-1.

  3. Following the extended AR mode for multihoming peers, Leaf-1 forwards the traffic directly to its multihoming peer Leaf-2, which has an interested receiver. Leaf-1 uses the source VLAN, VLAN-1, according to OISM behavior.

  4. Leaf-1 also forwards the traffic to one of the available AR replicators, ARR-1 in this case, on the source VLAN, VLAN-1.

    Note:

    See AR Leaf Device Load Balancing with Multiple Replicators for details on how AR leaf devices load-balance among multiple available AR replicators.

  5. ARR-1 replicates the traffic and sends copies on VLAN-1 only toward the other leaf devices with interested receivers besides Leaf-2. Due to the default extended AR mode behavior (see Step 3 above), ARR-1 skips sending the traffic to Leaf-2, the multihoming peer of the ingress leaf device Leaf-1.

  6. Each server leaf device then forwards or routes the traffic to its interested receivers.

    Note that in Figure 10, Leaf-1 is the ESI DF for multihomed receiver Rcvr-2. As a result, only Leaf-1 forwards the traffic to Rcvr-1 on VLAN-2.

See Configure Assisted Replication for details on how to configure AR.

AR and OISM with an External Multicast Source

In Figure 11, we show an OISM use case where you configure the spine devices as standalone AR replicator devices. The OISM server leaf and border leaf devices are AR leaf devices. The multicast source is outside the EVPN fabric in an external PIM domain. The border leaf devices in this case use the classic L3 interface method to connect to the PIM router and PIM RP.

With OISM traffic from an external source, the ingress border leaf device forwards the traffic in the EVPN fabric on the SBD VLAN. When you also enable AR, the ingress border leaf device forwards one copy of the traffic to an AR replicator. The AR replicator replicates the traffic and sends the copies on the SBD VLAN to the other leaf devices with interested receivers. Then each device locally routes the traffic it receives on the SBD toward its receivers on the revenue bridge domain VLANs.

Figure 11: AR with OISM-External Multicast SourceAR with OISM-External Multicast Source

In the use case in Figure 11:

  1. Rcvr-1 (multihomed to Leaf-1 and Leaf-2) and Rcvr-5 (a local host behind BL-2) send IGMP or MLD reports to join the multicast group.

  2. The external source, Ext-Mcast-Src, sends multicast traffic through the external PIM domain. In this case we use the classic L3 interface external multicast method, and both devices sent a PIM join message, so the PIM router sends the traffic to both BL-1 and BL-2. (See Multicast Traffic from an External Source to Receivers Inside the EVPN Data Center—L3 Interface Method or Non-EVPN IRB Method for a full explanation of this behavior.)

    Note:

    As Figure 11 shows, in this use case, because BL-2 has a local receiver, BL-2 routes the incoming externally sourced traffic directly toward its receiver on VLAN-2. BL-2 doesn't route traffic to the local receiver that it receives on the SBD from ARR-2 because its reverse-path forwarding toward the external source refers to the L3 interface.

    BL-2 also doesn't route the traffic to the SBD because BL-1 is the PIM DR on the SBD (see the next step).

  3. BL-1 is the PIM DR for the SBD, so BL-1 is the border leaf device that routes the externally sourced traffic into the EVPN fabric. With AR enabled, BL-1 forwards the traffic on the SBD to one of the available AR replicators. In this case, BL-1 forwards the traffic to ARR-2.

    Note:

    See AR Leaf Device Load Balancing with Multiple Replicators for details on how AR leaf devices load-balance among multiple AR replicators.

  4. ARR-2 replicates the traffic and sends copies on the SBD toward the leaf devices with interested receivers—in this case, Leaf-1, Leaf-2, and BL-2.

  5. Each leaf device that receives the traffic on the SBD locally routes the traffic toward the interested receivers on the revenue VLANs. In this case:

    • BL-2 routes the traffic toward its receiver on VLAN-2.

    • Both Leaf 1 and Leaf-2 receive the traffic on the SBD. Rcvr-1 is multihomed to Leaf-1 and Leaf-2, and Leaf-1 is the ESI DF. As a result, only Leaf-1 forwards the traffic toward Rcvr-1 on VLAN-2.

See Configure Assisted Replication for details on how to configure AR.

Considerations for OISM Configurations

Before you begin to set up your OISM installation, here are a few considerations in specific use cases.

IGMPv2 and IGMPv3 (or MLDv1 and MLDv2) in the Same EVPN-VXLAN Fabric

You have several options to configure IGMP snooping with IGMPv2, IGMPv3, or both IGMP versions together in an EVPN-VXLAN fabric with OISM. The same is true of MLD snooping with MLDv1, MLDv2, or both MLD versions together. You might also want to mix configuring IGMP and MLD together in the same fabric. This section includes configuration considerations for a few of these options.

If you have traffic for either IGMPv2 or IGMPv3 on a device with OISM enabled, you can enable that IGMP version globally on the device with IGMP snooping. You can alternatively enable that IGMP version only for the interfaces that will handle the multicast traffic. You can enable IGMP snooping with that version of IGMP on all VLANs or on specific VLANs, as needed.

You have the same options for either MDLv1 or MLDv2 with MLD snooping (on the platforms that support MLD with OISM).

You can also enable one version of IGMP with IGMP snooping and one version of MLD with MLD snooping together on the device with OISM.

However, OISM supports IGMP snooping with both IGMPv2 and IGMPv3 traffic together on a device only within the following constraints:

  • You can't enable IGMP snooping with both IGMPv2 and IGMPv3 for interfaces in the same VLAN.

  • You can't enable IGMP snooping with both IGMPv2 and IGMPv3 for VLANs that are part of the same L3 VRF instance with OISM enabled.

The constraints above also apply if you want to enable MLD snooping with both MLDv1 and MLDv2 traffic together on a device.

These constraints don't apply if you use one version of IGMP with one version of MLD together on a device.

To support IGMP snooping with both IGMP versions, or MLD snooping with both MLD versions, you must configure:

  • One tenant VRF instance to support the IGMPv2 or MLDv1 receivers.

  • Another tenant VRF instance to support the IGMPv3 or MLDv2 receivers.

as follows:

  1. In your configuration, define VLANs for the IGMPv2 receivers, and define different VLANs for the IGMPv3 receivers.

    Similarly, for MLD, define VLANs for the MLDv1 receivers, and define different VLANs for the MLDv2 receivers.

  2. Include the IRB interfaces that support IGMPv2 in one VRF instance, and enable IGMPv2 on those IRB interfaces. Enable IGMP snooping on the corresponding VLANs.

    Similarly, for MLD, include the IRB interfaces that support MLDv1 in one VRF instance, and enable MLDv1 on those IRB interfaces. Enable MLD snooping on the corresponding VLANs.

  3. Include the IRB interfaces that support IGMPv3 in another VRF instance, and enable IGMPv3 on those IRB interfaces. Enable IGMP snooping with the evpn-ssm-reports-only option on the corresponding VLANs.

    Similarly, for MLD, include the IRB interfaces that support MLDv2 in another VRF instance, and enable MLDv2 on those IRB interfaces. Enable MLD snooping with the evpn-ssm-reports-only option on the corresponding VLANs.

In this use case, for each IGMP or MLD version, allocate a set of VLANs and IRB interfaces for:

  • The OISM revenue bridge domains.

  • The SBD.

  • Any external multicast VLANs and interfaces (depending on the external multicast method you use).

You also define two L3 VRF instances for each tenant instance you need in your installation, one for each IGMP version or MLD version. If you use MAC-VRF routing instances at L2, you might want to allocate different MAC-VRF EVPN instances for the IGMP snooping or MLD snooping traffic for each IGMP or MLD version.

The next sections show example configurations with both versions of IGMP or both versions of MLD together. You can scale these simple scenarios to support different tenants with different combinations of IGMP or MLD versions.

See Supported IGMP or MLD Versions and Group Membership Report Modes for more information on IGMP any-source multicast (ASM) mode and source-specific multicast (SSM) mode support with IGMPv2, IGMPv3, MLDv1, and MLDv2 in EVPN-VXLAN fabrics.

Example Configuration with IGMPv2 and IGMPv3 Together

Consider a use case with both IGMP versions in a fabric you set up with the M-VLAN IRB method for external multicast. You want to support IGMP snooping with both IGMPv2 and IGMPv3 traffic. In that case, you might configure the following MAC-VRF instances, L3 VRF instances, VLANs, and corresponding IRB interfaces:

  • MAC-VRF2 and L3VRF-A to support IGMPv2 receivers:

    • Revenue bridge domain VLAN-100 with irb.100

    • SBD VLAN-302 with irb.302

    • (Border leaf devices only) M-VLAN VLAN-902 with irb.902

  • MAC-VRF3 and L3VRF-B to support IGMPv3 receivers:

    • Revenue bridge domain VLAN-200 with irb.200

    • SBD VLAN-303 with irb.303

    • (Border leaf devices only) M-VLAN VLAN-903 with irb.903

Then include the IGMPv2 IRB interfaces in L3VRF-A and enable IGMPv2 for those IRB interfaces. Include the IGMPv3 IRB interfaces in L3VRF-B and enable IGMPv3 for those IRB interfaces.

For example:

Finally, enable IGMP snooping at L2 in the EVPN instances as follows:

  • Configure igmp-snooping in MAC-VRF2 for the VLANs corresponding to the IGMPv2 IRB interfaces.

  • Configure igmp-snooping in MAC-VRF3 for the VLANs corresponding to the IGMPv3 IRB interfaces.

    Include the evpn-ssm-reports-only option only when you enable IGMP snooping for IGMPv3 traffic.

For example:

Note:

With the non-EVPN IRB method for external multicast, you don't include the evpn-ssm-reports-only option on the non-EVPN IRB interface. You don't need this option because with the non-EVPN IRB method, you don't extend the external multicast interface in the EVPN instance.

When you use the L3 interface method for external multicast, you don't enable IGMP snooping at all on the L3 interface to the external PIM domain. That interface operates at L3, while IGMP snooping operates at L2.

Example Configuration with MLDv1 and MLDv2 Together

Consider this use case with both MLDv1 and MLDv2 with MLD snooping in a fabric with OISM:

  • MAC-VRF1 and L3VRF-A to support MLDv1 receivers:

    • Revenue bridge domain VLAN-100 with irb.100

    • SBD VLAN-301 with irb.301

  • MAC-VRF2 and L3VRF-B to support MLDv2 receivers:

    • Revenue bridge domain VLAN-200 with irb.200

    • SBD VLAN-302 with irb.302

Note:

In this use case, we don't use the M-VLAN IRB method for external multicast, so we don't configure an M-VLAN IRB interface like we do in the IGMP use case above.

In this case, you configure:

  • MLDv1 on the IRB interfaces for MLDv1 receivers (VLANs 100 and 301).

  • MLDV2 on the IRB interfaces for MLDv2 receivers (VLANs 200 and 302)

  • MLD snooping for MLDv1 VLANs in MAC-VRF1.

  • MLD snooping for MLDv2 VLANs with the evpn-ssm-reports-only option in MAC-VRF2.

For example:

Latency and Scaling Trade-Offs for Installing Multicast Routes with OISM (install-star-g-routes Option)

Devices in an OISM-enabled fabric send EVPN Type 6 routes so other EVPN devices learn about receivers that are interested in the traffic for a multicast group. The receivers are in different OISM revenue bridge domains in an L3 VRF instance. To save bandwidth in the EVPN fabric core, OISM devices send and receive the Type 6 routes only on the OISM SBD in the routing instance.

To help minimizes packet loss at the onset of a multicast flow, we provide the install-star-g-routes option at the [edit <routing-instances name> multicast-snooping-options oism] hierarchy level (see oism (Multicast Snooping Options)). When you configure this option, upon receiving a Type 6 route, the RE on the device immediately installs corresponding (*,G) multicast routes on the PFE for all of the revenue bridge domain VLANs in the routing instance.

With this option, you trade off taking up extra PFE resources to improve network latency. Lower-scale deployments might have fewer multicast flows but have strict network latency requirements. To improve network latency in that case, the device installs the (*,G) routes in the data plane in advance of any incoming multicast traffic.

Configure this option:

  • Globally if you configure EVPN in the default-switch instance, at the [edit multicast-snooping-options oism] hierarchy level.

  • In the MAC-VRF instances if you configure EVPN in instances of type mac-vrf, at the [edit routing-instances instance-name multicast-snooping-options oism] .

We require that you configure install-star-g-routes with OISM on the QFX10000 line of switches, QFX5130-32CD switches, and QFX5700 switches when you configure those devices with AR in the AR replicator role.

In releases prior to Junos OS and Junos OS Evolved Release 23.4R1, you must also configure the install-star-g-routes option on the following devices when you configure them as OISM server leaf or border leaf devices:

  • Switches in the QFX10000 line.

  • PTX10001-36MR, PTX10004, PTX10008, and PTX10016 routers.

Starting in Junos OS and Junos OS Evolved Release 23.4R1, we no longer require you to set this option when you configure those devices as OISM server leaf or border leaf devices.

We don't recommend setting this option other than in the use cases mentioned above.

Consider this option only if you have very stringent latency requirements and can trade off higher scaling to achieve better network latency.

Note:

The functions of the install-star-g-routes option and the conserve-mcast-routes-in-pfe option are mutually exclusive, so you can use only one or the other of these options in a routing instance. See ACX Series Routers, QFX5130-32CD Switches, and QFX5700 Switches as Server Leaf and Border Leaf Devices with OISM for more on when to use the conserve-mcast-routes-in-pfe option.

Default Behavior without the install-star-g-routes Option

By default, without this option, the device prioritizes saving resources on the PFE by not installing multicast routes until the multicast traffic arrives. In this default case:

  1. The PFE receives multicast traffic from source S for multicast group G.

  2. The PFE doesn't have forwarding next-hop information for the traffic, so it signals the RE to get that information.

    Note:

    The PFE drops multicast traffic until it gets the routing information.

  3. The RE learns about the multicast flow for (S,G) from the PFE, and installs that route on the PFE.

  4. The PFE sends the traffic on the next hop in the installed (S,G) route.

Behavior with the install-star-g-routes Option

With the install-star-g-routes option, the device prioritizes having multicast routing information available on the PFE before any traffic arrives. The device consumes extra PFE resources for routes that it isn't using yet (and might never be used). With this option:

  1. The RE receives an EVPN Type 6 route for a receiver subscribing to traffic for a multicast group G on the OISM SBD in a routing instance.

  2. The RE installs corresponding (*,G) routes on the PFE for all of the revenue bridge domains in the L3 VRF instance.

  3. At some later time, the PFE receives multicast traffic from source S for multicast group G.

  4. The PFE has forwarding next hop information for traffic for (*,G). So it forwards the traffic to receivers on any revenue bridge domains using the (*,G) route next hop.

  5. The PFE also signals the RE that it has received multicast traffic from source S for multicast group G.

  6. The RE learns about the multicast flow for (S,G) from the PFE. The RE installs the (S,G) route on the PFE.

  7. The PFE continues sending the traffic, but now uses the (S,G) route and the next hop in that more specific route.

    Note:

    The PFE still retains the (*,G) routes per revenue bridge domain that the RE installed after receiving the Type 6 route.

OISM and AR Scaling with Many VLANs

With OISM and IGMP snooping or MLD snooping enabled in an EVPN-VXLAN fabric, OISM server leaf and border leaf devices send EVPN Type 6 SMET routes into the EVPN core when their receivers join a multicast group.

When an OISM-enabled device receives Type 6 routes on the SBD, the device:

  • Derives multicast states from the Type 6 routes as follows:

    • (*,G) states for IGMPv2 or MLDv1

    • (S,G) states for IGMPv3 or MLDv2

  • Installs the derived states on the OISM SBD and revenue bridge domain VLANs in the MAC-VRF instance for all VLANs that are part of OISM-enabled L3 tenant VRF instances.

  • Uses the derived multicast routes to optimize multicast forwarding by selectively sending the traffic for a group only to other EVPN devices that have receivers subscribed to that group.

On some devices that support OISM, you can also configure the assisted replication (AR) multicast optimization feature with OISM enabled. AR replicator devices use the Type 6 routes the same way OISM devices do.

QFX5130-32CD and QFX5700 switches can serve as OISM server leaf or border leaf devices. They can act as AR replicators only on devices that are not also OISM server leaf or border leaf devices. In that case, the device operates in the standalone AR replicator role.

The next sections describe configuration considerations on these devices when you configure them as OISM server leaf or border leaf devices, or as standalone AR replicators with OISM.

Note:

The use cases and sample configurations in the next sections show IGMP configurations for IPv4 multicast, but also apply in the same ways to MLD configurations for IPv6 multicast.

ACX Series Routers, QFX5130-32CD Switches, and QFX5700 Switches as Server Leaf and Border Leaf Devices with OISM

When you configure ACX Series routers, QFX5130-32CD switches, and QFX5700 switches as server leaf or border leaf devices with OISM, as soon as these devices receive multicast traffic, they use the L3 multicast routes from PIM to forward the traffic. They use the derived multicast snooping states only to learn which receivers are interested in a multicast stream. They don't need to save the multicast snooping derived states in the forwarding plane for forwarding traffic.

Starting in Junos OS Evolved Releases 22.4R2 and 23.1R1, when you configure these devices as OISM server leaf and border leaf devices, we require you to also configure the conserve-mcast-routes-in-pfe option at the [edit routing-instances name multicast-snooping-options oism] hierarchy level. (See oism (Multicast Snooping Options).) With this option, these devices conserve PFE table space by installing only the L3 multicast routes; they avoid installing L2 multicast snooping routes.

Use the following guidelines for setting the conserve-mcast-routes-in-pfe option:

  • You must set this option on ACX Series routers, QFX5130-32CD switches, and QFX5700 switches when you configure them as server leaf or border leaf devices with OISM enabled.

  • Set this option in all OISM-enabled MAC-VRF EVPN routing instances on the device.

  • Don't configure this option if you did not enable OISM on the device.

  • When you disable OISM on a device, you must also delete this setting.

Note:

The functions of the conserve-mcast-routes-in-pfe option and the install-star-g-routes option are mutually exclusive, so you can use only one or the other of these options in a routing instance. See Latency and Scaling Trade-Offs for Installing Multicast Routes with OISM (install-star-g-routes Option) for more on when to use the install-star-g-routes option.

QFX5130-32CD and QFX5700 Switches as Standalone AR Replicators with OISM

QFX5130-32CD and QFX5700 switches can serve as standalone AR replicators in a fabric with OISM. However, in fabrics with many VLANs, QFX5130-32CD and QFX5700 switches might have scaling issues when installing the multicast states on all the OISM VLANs.

As a result, starting in Junos OS Evolved 22.2R1, when you configure these switches as standalone AR replicators with OISM enabled, by default these switches only install multicast states on the SBD VLAN. (This includes multicast (*,G) states for IGMPv2 and multicast (S,G) states for IGMPv3.) These switches don’t install the multicast states on all the revenue bridge domain VLANs.

For example, consider a QFX5130-32CD device where you have a MAC-VRF instance evpn-vxlan-A with 3 VLANs—VLAN_2, VLAN_3, and VLAN_4. The show igmp snooping evpn status detail command shows that you configured VLAN_4 as the SBD (the Supplementary BD output field is Yes), and the other two VLANs are OISM revenue bridge domain VLANs:

The device received Type 6 routes from remote devices for multicast groups 233.252.0.1 and 233.252.0.2:

Due to the scaling behavior difference on QFX5130-32CD or QFX5700 switches, if you run the show multicast snooping route command on these devices, the output shows multicast group entries only on the SBD, and not on any of the revenue bridge domains. For example, with our multicast groups 233.252.0.1 and 233.252.0.2:

QFX5130-32CD or QFX5700 switches that are not running as AR replicators with OISM will install the multicast group entries on all of the revenue bridge domain VLANs and the SBD. When you run the show multicast snooping route command in that case, you see all of the revenue bridge domain VLANs and the SBD . This behavior also applies to all other platforms running OISM whether the device is an AR replicator or not.

Note:

On QFX5130-32CD or QFX5700 switches acting as AR replicators, you should not configure the conserve-mcast-routes-in-pfe option we describe in ACX Series Routers, QFX5130-32CD Switches, and QFX5700 Switches as Server Leaf and Border Leaf Devices with OISM.

PEG DF Election

By default, peer OISM PEG devices use PIM-based DF election—the devices discover their PIM neighbors on the OISM revenue VLANs and the SBD in each L3 VRF, and elect the DF from among those neighbors. Starting in Junos OS Release 23.4R1 and Junos OS Evolved Release 23.4R1, you can instead configure mod-based or preference-based PEG DF election using the peg-df-election statement at the [edit routing-instances name protocols evpn oism pim-evpn-gateway] hierarchy level, as follows:

When you configure PEG DF election, the device maintains an ordinal list of the PEG devices in the fabric that host each revenue VLAN (bridge domain) or the SBD. PEG devices use the EVPN multicast flags extended community in EVPN Type 3 IMET routes to advertise OISM, IGMP snooping, MLD snooping, and PEG device support. In addition, PEG devices include the DF election extended community to communicate the configured DF election method parameters (as defined in the RFC 8584 standard).

The peer PEG DF candidates are the devices that advertise IMET routes for a revenue VLAN or the SBD with the following EVPN multicast flags extended community values:

  • For IPv4 multicast—igmp-snooping-enabled:oism:peg
  • For IPv6 multicast—mld-snooping-enabled:oism:peg
    Note:

    The devices elect a DF for IPv4 multicast traffic and a DF for IPv6 multicast traffic separately based on those advertised multicast flags extended community values.

When the PEG devices use PEG DF election, they don't use the PIM protocol (they don't exchange PIM protocol packets) within the data center. As a result, we recommend you have external L3 redundancy on the PEG devices.

You can configure the PEG devices to use either of the following PEG DF election methods:

  • Mod-based—The default method when you enable PEG DF election at the [edit routing-instances name protocols evpn oism pim-evpn-gateway peg-df-election] hierarchy level without including the mod option or the preference option. You can also explicitly configure the mod option to use this method.

    The algorithm to choose the device in the ordinal list that will be the DF is:

    (mapped VNI for the VLAN) mod (number of entries in the list)

    For example, if you have three peer PEG devices BL1, BL2, and BL3 configured with mod-based PEG DF election for VLAN 1, which is mapped to VNI 100:

    • The three devices each maintain an ordinal list of PEG DF candidates for VLAN 1 and VNI 100, such as:

      Table 7: Sample Mod-based PEG DF Election Candidates List

      Index

      Device

      0

      BL1

      1

      BL2

      2

      BL3

    • In this case, (mapped VNI for the VLAN) mod (number of entries in the list) is (100) mod (3) = 1, so the devices elect the device at index 1 as the PEG DF, which is BL2.

  • Preference-based with a customized preference value—Configure the value preference-value option at the [edit routing-instances name protocols evpn oism pim-evpn-gateway peg-df-election preference] hierachy level. We recommend you set a unique preference value on each peer PEG device. With preference-based PEG DF election:

    • Each device communicates its preference value in the EVPN Type 3 IMET route advertisement.

    • The peer PEG devices elect the device with the highest preference value (by default) as the DF for the VLAN.

    • You can customize the preference-based method to elect the device with the lowest preference value instead of the highest value. To do this, set the use-least option at the [edit routing-instances name protocols evpn oism pim-evpn-gateway peg-df-election preference] hierarchy level.

With either PEG DF election method, you can also:

  • Specify a time period to wait before the devices elect a DF. To do this, set the delay-time num option at the [edit routing-instances name protocols evpn oism pim-evpn-gateway peg-df-election] hierarchy level.

  • Set the maximum number (0-255) of PEG DF election event entries the device maintains in a database for DF election history per VLAN (VNI). To do this, set the peg-df-election-history num option at the [edit <routing-instances name> protocols evpn] hierarchy level.

    The show evpn oism peg-df-status extensive command displays DF election history details.

All peer PEG devices must use the same DF election mechanism. As a result, if you enable PEG DF election, configure the same DF election method symmetrically on all peer PEG devices. If the configured PEG DF election methods don't match, the peer PEG devices all fall back to using the default mod-based PEG DF election method. If you enable PEG DF election on only some of the peer PEG devices, all of the devices fall back to using PIM-based DF election.

How to Check PEG DF Election Status

Use the following commands to see PEG DF election status for a PEG device:

  1. show evpn oism—See if PEG DF election is enabled for each L3 routing instance. Include the extensive option to see the configured PEG DF election method and preference value if you selected the preference option. For example:

    Mod-based PEG DF Election:

    Preference-based PEG DF Election—uses the highest preference value by default, so BL1 would be the elected PEG DF here:

    Preference-based PEG DF Election with the use-least option—uses the lowest preference value, so BL2 would be the elected PEG DF here:

  2. show evpn oism peg-df-status—See PEG DF election status and the DF IP address for all or selected L3 VRF instances per VLAN-mapped VNI. This command displays information only for routing instances and VNIs where you configured PEG DF election. Use the extensive option to see more details such as the DF candidates list and the DF election history information. For example:

    Mod-based PEG DF Election:

    Preference-based PEG DF Election—BL1 has the highest preference value, 200, and is the PEG DF here for VNIs 100 and 200; BL2 with preference value 100 is not elected as the DF (nDF):

    Preference-based PEG DF Election with the use-least option—BL2 has the lowest preference value, 100, so in that case BL2 is the elected PEG DF here for VNIs 100 and 200:

  3. show route table <evpn-instance-name>.evpn.0 match-prefix 3:* extensive—Verify that the PEG devices (the PEG DF election candidates) have the DF election extended community, in addition to the multicast flags extended community, in the EVPN Type 3 IMET routes in the EVPN instance routing table. For example:

    Mod-based PEG DF Election:

    Preference-based PEG DF Election (truncated to show only the extended community values):

  4. show pim interfaces instance vrf-instance-name—Verify the DF election status of the PIM IRB interfaces for each L3 routing instance matches the DF election results from the PEG DF election process. The device relays the PEG DF election status to the PIM protocol processes because OISM relies on PIM to create the multicast routes. For example:

  5. show pim join instance vrf-instance-name extensive—Verify that only the PEG device that is elected as the DF on the SBD sends PIM Join messages to an external PIM router (the PIM RP). The device that sends the PIM Join pulls in externally sourced traffic that is destined for multicast receivers within the EVPN data center. For example:

  6. show pim rps instance vrf-instance-name extensive—Verify that only the PEG device that is elected as the DF for an OISM revenue VLAN or the SBD sends the PIM Register messages to the external PIM RP for multicast sources inside the EVPN data center.

    Look for the elected DF that sends the PIM Register messages for:

    • The source VLAN when you configure the original OISM symmetric bridge domains model. In this case, the PEG devices receive source traffic from inside the data center on the source revenue VLAN.

    • The SBD when you configure enhanced OISM (where you don't need to configure all revenue VLANs on all devices in the fabric). In this case, the PEG devices receive source traffic from inside the data center on the SBD.

Configure Common OISM Elements on Border Leaf Devices and Server Leaf Devices (Symmetric Bridge Domains)

Follow these steps to configure elements common to border leaf and server leaf devices in an EVPN-VXLAN fabric running OISM with the symmetric bridge domains model.

Note:

You also configure these common elements on spine devices that act as standalone AR replicators in a fabric with OISM.

This configuration is based on an EVPN-VXLAN fabric configuration that supports OISM and has:

  • An EBGP underlay with Bidirectional Forwarding Detection (BFD) and Ethernet operations, administration and management (OAM) link detection.

  • An ERB overlay design.

  • Lean spine devices that act only as IP transit nodes in the fabric.

  • Server leaf devices configured as EVPN-VXLAN L2 gateways.

  • Border leaf devices configured as EVPN-VXLAN L2 and L3 gateways.

In the symmetric bridge domains model, you must configure all of the revenue bridge domains and the supplemental bridge domain (SBD) on all OISM devices in the fabric. See Configuration Elements for OISM Devices (Symmetric Bridge Domains) for a summary of what elements you configure on border leaf and server leaf devices, and why you configure those elements.

The sample configuration blocks we provide for these configuration steps use an OISM environment with the following elements:

  • An EVPN instance configured in either the default switch instance (no routing instance specified) OR a MAC-VRF EVPN instance. For example:

    • Default switch EVPN instance:

    • MAC-VRF EVPN instances (for each MAC-VRF instance):

    With a MAC-VRF EVPN instance configuration, you configure some elements in the MAC-VRF instances. In OISM configurations, we support either the vlan-aware or vlan-based MAC-VRF instance service type.

    To illustrate sample configuration steps here, we show a MAC-VRF instance named MAC-VRF1with the vlan-aware service type. The main differences between the two service types are:

    • vlan-aware: You can define more than one VLAN, its corresponding IRB interface, and VXLAN network identifier (VNI) mapping in the instance. As a result, you specify VLAN-related OISM or multicast configuration statements in one vlan-aware MAC-VRF instance for all of the VLANs in that instance.

    • vlan-based: You configure separate MAC-VRF instances in which you define each VLAN and its IRB interface and VNI mapping. As a result, you include similar VLAN-related OISM or multicast configuration statements for each VLAN in the corresponding vlan-based MAC-VRF instance.

  • SBD: VLAN-300

    SBD IRB interface: irb.300

    SBD IRB interface IP address: 10.0.30.1

  • Revenue bridge domains: VLAN-100 and VLAN-200

    Revenue bridge domain IRB interfaces: irb.100 and irb.200

    Revenue bridge domain IRB interface IP addresses: 10.0.10.1 and 10.0.20.1

  • L3 VRF routing instance: L3VRF-1

  • If you use the M-VLAN IRB method for external multicast connectivity:

    M-VLAN: VLAN-900

    M-VLAN IRB interface: irb.900

    M-VLAN IRB interface IP address: 172.16.90.1/24

    Interface name of port connecting to external PIM router: xe-0/0/9

    Note:

    You use the same M-VLAN ID and assign IRB interface IP addresses in the same subnet across any border leaf devices to which the external PIM router is multihomed.

  • If you use the classic L3 interface method for external multicast connectivity:

    L3 interface name: xe-0/0/6

    L3 interface IP address: 172.16.10.1/24

    Note:

    You assign IP addresses in different subnets for the L3 interfaces on each border leaf device connected to the external PIM router.

  • If you use the non-EVPN IRB method for external multicast connectivity:

    Extra VLAN: VLAN-900

    Non-EVPN IRB interface: irb.900

    Non-EVN IRB interface IP address: 172.16.90.1/24

    Interface name of port connecting to external PIM router: xe-0/0/9

    Note:

    With the non-EVPN IRB method, you assign distinct extra VLAN IDs on each border leaf device. You also assign IP addresses in different subnets for the non-EVPN IRB interfaces on each border leaf device connected to the external PIM router.

Configure these OISM statements on both border leaf devices and server leaf devices:

  1. Enable OISM globally on the device.

    Configure the evpn irb oism option at the [edit forwarding-options multicast-replication] hierarchy level. This option enables OISM optimizations and external multicast capabilities in an ERB overlay fabric. This option takes the place of the evpn irb local-only option you would otherwise use in ERB overlay fabrics for multicast without OISM.

    Note:

    Some platforms, such as such as PTX10001-36MR, PTX10004, PTX10008, and PTX10016 routers, only support OISM for multicast traffic. In that case, the device posts a warning in the configuration if you set the evpn irb local-only option or the evpn irb local-remote option, and ignores those configuration items.

  2. Configure all of the revenue bridge domain with an IRB interface and a VNI mapping for each revenue bridge domain. If your configuration uses MAC-VRF EVPN instances, you configure these elements in the MAC-VRF EVPN routing instances.

    For example, if the revenue bridge domains are VLAN-100 and VLAN-200:

    With the default switch instance:

    With a vlan-aware MAC-VRF instance MAC-VRF1:

  3. Configure a VLAN for the SBD with an IRB interface and VXLAN VNI mapping. For example, for SBD VLAN-300:

    For example, if the SBD is VLAN-300:

    With the default switch instance:

    With a vlan-aware MAC-VRF instance MAC-VRF1:

  4. Configure the revenue bridge domain and the SBD IRB interface IP addresses. For example:
  5. Extend the revenue bridge domains and the SBD VNIs in the EVPN-VXLAN overlay. If your configuration uses MAC-VRF EVPN instances, you do this in the MAC-VRF EVPN routing instances.

    For example, if the revenue bridge domains are VLAN-100 and VLAN-200, and the SBD is VLAN-300:

    With the default switch instance:

    With a vlan-aware MAC-VRF instance MAC-VRF1:

  6. Enable IGMPv2 or IGMPv3 on the device for IPv4 multicast traffic, or enable MLDv1 or MLDv2 for IPv6 multicast traffic. Here for simplicity we show how to enable either IGMP version or either MLD version globally on the device.

    You can alternatively enable IGMP or MLD on the specific IRB interfaces included in the tenant L3 VRF instances that handle the multicast traffic.

    In general, on all OISM devices, enable IGMP on the SBD IRB interface and the revenue bridge domain IRB interfaces. On border leaf devices, also enable IGMP on the external multicast interfaces (depending on the external multicast method used).

    See IGMPv2 and IGMPv3 (or MLDv1 and MLDv2) in the Same EVPN-VXLAN Fabric for more on how to configure IGMP snooping with both IGMPv2 and IGMPv2 together (or MLD snooping with MLDv1 and MLDv2 together) on a device.

    Also, to enable IGMP or MLD on individual IRB interfaces that handle multicast traffic, include multiple igmp or mld statements, one for each IRB interface. For example, set protocols igmp interface irb-interface-name<version 3>, or set protocols mld interface irb-interface-name<version 2>.

    1. For IGMPv2, which is the default IGMP version, enable IGMP globally. You can also optionally specify version 2 for clarity if you have both IGMP versions in your configuration. The configuration is the same if you use a default switch EVPN instance or MAC-VRF EVPN instances.
    2. For IGMPv3, enable IGMP globally with the version 3 option. The configuration is the same if you use a default switch EVPN instance or MAC-VRF EVPN instances.
    3. For MLDv1, which is the default MLD version, enable MLD globally. You can also optionally specify version 1 for clarity if you have both MLD versions in your configuration.
    4. For MLDv2, enable MLD globally with the version 2 option.
  7. Enable IGMP snooping for IGMPv2 or IGMPv3, or MLD snooping for MLDv1 or MLDv2, on all the configured OISM VLANs. Here in the common configuration for all OISM leaf devices, we include configuration only for the revenue bridge domains and the SBD. In the configuration steps specific to border leaf devices, we include the IGMP snooping or the MLD snooping configuration specific to the external multicast method you use on those devices.

    When you enable IGMP snooping or MLD snooping, you also automatically enable advertising SMET Type 6 routes in the EVPN core based on received IGMP or MLD reports. If you use MAC-VRF EVPN instances, you enable IGMP snooping or MLD snooping in the MAC-VRF instances.

    1. For IGMPv2, enable igmp-snooping.
      Note:

      You can use individual igmp-snooping commands for each VLAN, or one command with the vlan all option.

      With the default switch instance:

      With a vlan-aware MAC-VRF instance MAC-VRF1:

    2. For IGMPv3, enable igmp-snooping for IGMPv3 source-specific multicast mode (SSM) reports. Include the evpn-ssm-reports-only option for all the configured VLANs.

      With the default switch instance:

      With a vlan-aware MAC-VRF instance MAC-VRF1:

    3. For MLDv1, enable mld-snooping. You can use individual mld-snooping commands for each VLAN, or one command with the vlan all option. We support MLD snooping with OISM only in configurations with MAC-VRF EVPN instances. Here we show configuration with VLAN-aware MAC-VRF instance MAC-VRF1:
    4. For MLDv2, enable mld-snooping for MLDv2 source-specific multicast mode (SSM) reports. Include the evpn-ssm-reports-only option for all the configured VLANs. Like in the previous step for MLDv1, here for MLDv2 we show configuration with VLAN-aware MAC-VRF instance MAC-VRF1:
  8. Configure an L3 VRF instance (instance-typevrf) that you associate with OISM routing functions. Include the revenue bridge domain IRB interfaces and the SBD IRB interface in the routing instance. For example:
  9. In the L3 VRF instance, specify IRB interface that you configured as the OISM SBD. For example:
  10. (Required on the QFX10000 line of switches and PTX10001-36MR, PTX10004, PTX10008, and PTX10016 routers in any OISM role in releases prior to Junos OS and Junos OS Evolved 23.4R1; required on the QFX10000 line of switches, QFX5130-32CD switches, and QFX5700 switches when you configure them in the AR replicator role; optional but not recommended in any use cases other than the required ones) To avoid traffic loss at the onset of multicast flows, enable the install-star-g-routes option at the [edit <routing-instances intance-name> multicast-snooping-options oism] hierarchy level on all OISM devices.

    If you use the default switch EVPN instance, install-star-g-routes is a global option. If you use MAC-VRF EVPN instances, you set this option in each EVPN MAC-VRF routing instance. With this option, upon receiving an EVPN Type 6 route on the SBD, the device immediately installs corresponding (*,G) routes on the PFE for the revenue bridge domain VLANs in the L3 VRF instance.

    For full details on OISM device requirements, recommendations, and behavior with or without this option, see Latency and Scaling Trade-Offs for Installing Multicast Routes with OISM (install-star-g-routes Option).

    With the default switch instance:

    With a vlan-aware or vlan-based MAC-VRF instance called MAC-VRF1:

    You might choose to not enable this option if space on the PFE is at a premium and you don't have stringent latency requirements.

  11. (Required on QFX5130-32CD and QFX5700 switches starting in Junos OS Evolved Releases 22.4R2 and 23.1R1, and ACX Series routers starting in Junos OS Evolved Release 23.4R1, when you configure those switches as OISM server leaf or border leaf devices) Set the conserve-mcast-routes-in-pfe option at the [edit routing-instances name multicast-snooping-options oism] hierarchy level in the MAC-VRF EVPN instances on the device.

    For example, with a MAC-VRF instance called MAC-VRF1:

    Note:

    You don't need to set this option on the device if you configured it as a standalone AR replicator with OISM.

    See oism (Multicast Snooping Options) and ACX Series Routers, QFX5130-32CD Switches, and QFX5700 Switches as Server Leaf and Border Leaf Devices with OISM for details on this option.

Configure Server Leaf Device OISM Elements (Symmetric Bridge Domains)

First configure the OISM elements described in Configure Common OISM Elements on Border Leaf Devices and Server Leaf Devices (Symmetric Bridge Domains) in an EVPN-VXLAN fabric.

Then follow these steps to configure the additional required OISM elements on the server leaf devices. The same EVPN-VXLAN fabric base and sample OISM environment apply to the additional server leaf configuration steps here.

You configure the elements specific to the server leaf functions (like PIM) in the tenant L3 VRF instances.

See Configuration Elements for OISM Devices (Symmetric Bridge Domains) for more information about why OISM server leaf devices require these settings.

  1. Configure PIM passive mode on server leaf devices. For example:
  2. Enable the server leaf device to accept multicast traffic from the SBD IRB interface as the source interface using the accept-remote-source statement at the [edit routing-instances name protocols pim interface irb-interface-name] hierarchy level. For example, for our sample SBD, VLAN-300, the IRB interface is irb.300:
  3. Configure an OSPF area in the L3 VRF instance for peer connectivity to support external multicast traffic on the SBD for server leaf devices.

    The server leaf devices use OSPF to learn routes to external sources when multicast traffic from outside the fabric arrives on the SBD. The device creates the PIM (S,G) entries it requires to forward the traffic from the SBD to the revenue bridge domains.

    On the server leaf devices, configure OSPF on the device loopback interface and the SBD IRB interface. For example:

Configure Border Leaf Device OISM Elements with M-VLAN IRB Method (Symmetric Bridge Domains)

This section covers how to configure border leaf devices that use the OISM M-VLAN IRB method to exchange multicast data with external sources and receivers. See Supported Methods for Multicast Data Transfer to or from an External PIM Domain for more on the available external multicast methods.

First configure the OISM elements described in Configure Common OISM Elements on Border Leaf Devices and Server Leaf Devices (Symmetric Bridge Domains) in an EVPN-VXLAN fabric.

Then follow these steps to configure the additional required OISM elements on the border leaf devices. The same EVPN-VXLAN fabric base and the sample OISM environment in that section apply to the additional border leaf configuration steps here.

You configure different elements that are specific to the border leaf functions either globally, in the EVPN instances, or in the tenant L3 VRF instances.

See Configuration Elements for OISM Devices (Symmetric Bridge Domains) for more information about why OISM border leaf devices require these settings.

  1. Configure the M-VLAN similarly to how you configure the revenue bridge domains and the SBD.
    1. Configure the M-VLAN with an IRB interface and a VXLAN network identifier (VNI) mapping. For example, if the M-VLAN is VLAN-900:

      With the default switch instance:

      With a vlan-aware MAC-VRF instance MAC-VRF1:

    2. Configure the M-VLAN IRB interface IP address. For example:
    3. Extend the M-VLAN in the EVPN-VXLAN overlay. For example:

      With the default switch instance:

      With a vlan-aware MAC-VRF instance MAC-VRF1:

  2. Enable IGMP snooping for the M-VLAN, and configure the multicast-router-interface option on the L2 port that connects to the external multicast PIM router. For example, if xe-0/0/9.0 is the L2 interface that connects the EVPN fabric to the external multicast router on the M-VLAN:
    1. With IGMPv2:

      In the default switch instance:

      In a vlan-aware MAC-VRF instance MAC-VRF1:

    2. With IGMPv3, include the evpn-ssm-reports-only option:

      In the default switch instance:

      In a vlan-aware MAC-VRF instance MAC-VRF1:

  3. In the common leaf device configuration steps, you configure an L3 VRF instance associated with OISM routing functions. On the border leaf devices, also include the M-VLAN IRB interface in that L3 VRF. The following configuration block shows the common L3 VRF instance configuration with the additional M-VLAN IRB interface configuration statement highlighted:
  4. In the L3 VRF instance, set the OISM PEG role on the M-VLAN IRB interface. For example:
  5. Configure PIM in the L3 VRF instance for a border leaf device.
    1. In the L3 VRF instance, configure PIM in distributed DR mode on the revenue bridge domain IRB interfaces using the distributed-dr option at the [edit routing-instances name protocols pim interface irb-interface-name] hierarchy level.

      Configure PIM in standard mode on the SBD IRB interface.

      Configure PIM on the M-VLAN IRB interface in either standard mode or distributed DR mode. A border leaf device works well in standard PIM mode if the external PIM router is single-homed to one border leaf device. However, we strongly recommend to use distributed DR mode in any case, but especially if the external PIM router is multihomed to multiple border leaf devices. Distributed DR mode also helps the device to efficiently do local routing on the M-VLAN to local receivers on border leaf devices. As a result, in the sample configuration here, we show setting PIM with the distributed-dr option on the M-VLAN IRB.

      You also configure a PIM static RP that corresponds to the external PIM RP router. In the sample use cases in this documentation, the external PIM router serves as the PIM RP.

      For example, if the revenue bridge domains are VLAN-100 and VLAN-200, the SBD is VLAN-300, and the M-VLAN is VLAN-900:

    2. In the L3 VRF instance, set the accept-join-always-from option at the [edit routing-instances name protocols pim interface irb-interface-name] hierarchy level on the M-VLAN IRB interface.

      Configure a policy option along with this statement so that the device always installs PIM joins from the external PIM router. See Configuration Elements for OISM Devices (Symmetric Bridge Domains) for more information about why OISM border leaf devices require this configuration.

      This sample configuration block represents the external PIM router as an MX Series router. For the policy prefix list, include the IP address of the M-VLAN interface on the MX Series router that connects to the border leaf device. For example:

  6. Configure an OSPF area in the L3 VRF instance for external multicast peer interface connectivity.

    The border leaf device uses OSPF to learn routes to multicast sources to forward traffic from external sources toward internal receivers, and from internal sources toward external receivers. The device needs these routes to create the PIM (S,G) entries to forward traffic on the revenue bridge domains, the SBD, and the external multicast interfaces.

    On a border leaf device with the M-VLAN IRB method for external multicast, configure the OSPF area to include the device loopback interface, the SBD IRB interface, and the M-VLAN IRB interface. For example, with our sample SBD VLAN-300 and M-VLAN VLAN-900:

Configure Border Leaf Device OISM Elements with Classic L3 Interface Method (Symmetric Bridge Domains)

This section covers how to configure border leaf devices that use the OISM classic L3 interface method to exchange multicast data with external sources and receivers. See Supported Methods for Multicast Data Transfer to or from an External PIM Domain for more on the available external multicast methods.

First configure the OISM elements described in Configure Common OISM Elements on Border Leaf Devices and Server Leaf Devices (Symmetric Bridge Domains) in an EVPN-VXLAN fabric.

Then follow these steps to configure the additional required OISM elements on border leaf devices. The same EVPN-VXLAN fabric base and the sample OISM environment in that section apply to the additional border leaf configuration steps here.

You configure most of the elements that are specific to the border leaf functions at L3 in the tenant L3 VRF instances.

See Configuration Elements for OISM Devices (Symmetric Bridge Domains) for more information about why OISM border leaf devices require these settings.

  1. Configure a physical L3 interface with an IP address for external multicast. For example, for an L3 interface xe-0/0/6 with IP address 172.16.10.1/24:
  2. Include the L3 logical interface in the L3 VRF instance that you configured in the common leaf device configuration steps. For example:
  3. In the L3 VRF instance, set the OISM PEG role on the border leaf device. With the classic L3 interface method, you don't need to specify an external IRB interface for external multicast:
  4. Configure PIM in the L3 VRF instance for a border leaf device.

    For the revenue bridge domain IRB interfaces, configure PIM in distributed DR mode using the distributed-dr option at the [edit routing-instances name protocols pim interface irb-interface-name] hierarchy level.

    Configure PIM in standard mode on the SBD IRB interface and the external multicast L3 logical interface.

    We also configure a PIM static RP that corresponds to the external PIM RP router. In the sample use cases in this documentation, the external PIM router serves as the PIM RP.

    For example, if the revenue bridge domains are VLAN-100 and VLAN-200, the SBD is VLAN-300, and the L3 interface is xe-0/0/6:

  5. Configure an OSPF area in the L3 VRF instance for external multicast peer interface connectivity.

    The border leaf device uses OSPF to learn routes to multicast sources to forward traffic from external sources toward internal receivers, and from internal sources toward external receivers. The device needs these routes to create the PIM (S,G) entries to forward traffic on the revenue bridge domains, the SBD, and the external multicast interfaces.

    On a border leaf device with the classic L3 interface method for external multicast, configure the OSPF area to include the device loopback interface, the SBD IRB interface, and the external multicast logical L3 interface. For example, with our sample SBD IRB interface, irb.300, and L3 interface xe-0/0/6:

Configure Border Leaf Device OISM Elements with Non-EVPN IRB Method (Symmetric Bridge Domains)

This section covers how to configure border leaf devices that use the OISM non-EVPN IRB method to exchange multicast data with external sources and receivers. See Supported Methods for Multicast Data Transfer to or from an External PIM Domain for more on the available external multicast methods.

First configure the OISM elements described in Configure Common OISM Elements on Border Leaf Devices and Server Leaf Devices (Symmetric Bridge Domains) in an EVPN-VXLAN fabric.

Then follow these steps to configure the additional required OISM elements on border leaf devices. The same EVPN-VXLAN fabric base and the sample OISM environment in that section apply to the additional border leaf configuration steps here.

You configure most of the elements that are specific to the border leaf functions (like PIM) in the tenant L3 VRF instances. With this method, you don't extend the extra VLAN in the EVPN instance, so you don't configure related elements in the EVPN instance. The external multicast configuration elements are the same whether you use the default switch instance or MAC-VRF EVPN instances.

See Configuration Elements for OISM Devices (Symmetric Bridge Domains) for more information about why OISM border leaf devices require these settings.

  1. Configure the extra VLAN with an IRB interface for external multicast. For example:
  2. Enable IGMP snooping or MLD snooping for the extra VLAN, and configure the multicast-router-interface option on the port that connects to the external multicast PIM router. For example, if xe-0/0/9.0 is the interface on the border leaf device that connects to the external multicast router on the extra VLAN:
    1. With IGMPv2:
    2. With IGMPv3, you don't need the evpn-ssm-reports only option here because you don't extend the extra VLAN in the EVPN instance:
    3. With MLDv1:
    4. With MLDv2, you don't need the evpn-ssm-reports only option here because you don't extend the extra VLAN in the EVPN instance:
  3. Include the extra VLAN IRB interface in the L3 VRF instance that you configured in the common leaf device configuration steps.

    The following configuration block shows the common L3 VRF configuration and highlights the additional statement:

  4. In the L3 VRF instance, set the OISM PEG role on the border leaf device. For example:
  5. Configure PIM in the L3 VRF instance for a border leaf device.

    For the revenue bridge domain IRB interfaces, configure PIM in distributed DR mode using the distributed-dr option at the [edit routing-instances name protocols pim interface irb-interface-name] hierarchy level.

    Configure PIM in standard mode on the SBD IRB interface and the extra VLAN IRB interface.

    We also configure a PIM static RP that corresponds to the external PIM RP router. In the sample use cases in this documentation, the external PIM router serves as the PIM RP.

    For example, if the revenue bridge domains are VLAN-100 and VLAN-200, the SBD is VLAN-300, and the extra VLAN is VLAN-900:
  6. Configure an OSPF area in the L3 VRF instance for external multicast peer interface connectivity.

    The border leaf device uses OSPF to learn routes to multicast sources to forward traffic from external sources toward internal receivers, and from internal sources toward external receivers. The device needs these routes to create the PIM (S,G) entries to forward traffic on the revenue bridge domains, the SBD, and the external multicast interfaces.

    On a border leaf device with the non-EVPN IRB method for external multicast, configure the OSPF area to include the device loopback interface, the SBD IRB interface, and the non-EVPN IRB interface. For example, with our sample SBD IRB interface (irb.300) and extra non-EVPN IRB interface (irb.900):

CLI Commands to Verify the OISM Configuration

To verify your OISM configuration:
  1. Use the show evpn oism command to view the SBD IRB interface in each L3 VRF instance that you configured on the device for OISM. For example:

    To view information for a specific routing instance or to see more details about the configuration, use the extensive option with this show command. For example:

  2. Use the show route table bgp.evpn.0 ... extensive command to see the OISM capabilities you have enabled on an OISM leaf device. These capabilities are in the EVPN multicast flags extended community in EVPN Type 3 IMET routes. This extended community is displayed in the Communities: ... evpn-mcast-flags: output field as a hexadecimal flags value with the keyword for each enabled function. The OISM–related flags include:
    • igmp-snooping-enabled—You enabled IGMP snooping. The evpn-mcast-flags bit for IGMP snooping without OISM or PEG configuration is 0x01.

    • mld-snooping-enabled—You enabled MLD snooping. The evpn-mcast-flags bit for MLD snooping without OISM or PEG configuration is 0x02.

    • oism—You globally enabled OISM. The evpn-mcast-flags bit for OISM is 0x08. You might see flags values such as the following (without PEG configuration on the device):

      • 0x9—OISM and IGMP snooping

      • 0xa—OISM and MLD snooping

      • 0xb—OISM with both IGMP snooping and MLD snooping

    • peg—You configured PEG mode on the associated interface (for border leaf devices that connect to an external PIM domain). The evpn-mcast-flags bit for PEG mode is 0x10, so with PEG mode enabled, you might see flag values such as the following:

      • 0x19—PEG mode with OISM and IGMP snooping

      • 0x1a—PEG mode with OISM and MLD snooping

      • 0x1b—PEG mode with OISM, IGMP snooping, and MLD snooping

    For example, a route advertised for an M-VLAN IRB interface on a border leaf device has the peg flag set:

    You don't see the peg flag set for a revenue bridge domain interface on a border leaf device or a server leaf device:

    With IGMP snooping and MLD snooping enabled on an OISM server leaf device, you might see:

  3. Enter the show igmp snooping evpn status <vlan name><detail> command to see the EVPN-VXLAN L2 multicast context for the OISM bridge domains (VLANs) on the device. The default output includes the VXLAN VNI mappings of the VLANs. For example:

    If you have MLD snooping enabled, use the show mld snooping evpn status <vlan name><detail> command, which shows the same output as the IGMP snooping version of the command.

    The detail option also displays whether a VLAN is the OISM SBD (Supplementary BD) or the M-VLAN (External VLAN). Both of those output fields display No for revenue bridge domains (VLANs) or if you don't enable OISM.

    Here is an example on a server leaf device in a fabric where the SBD is VLAN-300:

    Here is another example on a border leaf device in a fabric where the M-VLAN is VLAN-900:

  4. Enter the show evpn multicast-snooping status to see if you enabled IGMP snooping or MLD snooping on VLANs you configured for OISM elements.

    For example, in the following sample command, if the revenue bridge domains on a server leaf device are VLAN-100 and VLAN-200, the SBD is VLAN-300, and you enabled IGMP snooping (but not MLD snooping):

    With both IGMP snooping and MLD snooping enabled, you'll also see Multicast Address Family: INET6 with SG Sync: Enabled.

  5. Use EVPN commands, multicast snooping commands, and PIM commands to see details for EVPN multihoming ESIs, learned multicast group routes, multicast traffic flow, and AR replication and forwarding. These commands are not specific to OISM operation, but are helpful to verify OISM operation and troubleshoot OISM issues.

    For example:

    • show igmp snooping membership vlan vlan-name<virtual-switch EVPN-instance-name>

    • show mld snooping membership vlan vlan-name<virtual-switch EVPN-instance-name>

    • show evpn instance <EVPN-instance-name> designated-forwarder <esi esi-number>

    • show pim join summary <instance VRF-instance-name>

    • show pim join detail <instance VRF-instance-name>

    • show evpn multicast-snooping assisted-replication next-hops <instance EVPN-instance-name>

    • show evpn multicast-snooping assisted-replication replicators

    For a full OISM configuration example of a common data center fabric use case, which also shows how to use these commands to verify OISM operation, see Optimized Intersubnet Multicast (OISM) with Assisted Replication (AR) for Edge-Routed Bridging Overlays.

Change History Table

Feature support is determined by the platform and release you are using. Use Feature Explorer to determine if a feature is supported on your platform.

Release
Description
23.4R1
Starting in Junos OS Release 23.4R1, ACX7024, ACX7100-32C, and ACX7100-48L routers support OISM with IGMPv2, IGMPv3, and IGMP snooping. These devices support OISM using MAC-VRF EVPN instances with vlan-aware and vlan-based service types. You can configure these devices in OISM server leaf, border leaf, or lean spine roles. In the border leaf role, these devices support only the classic L3 interface method to connect to an external multicast PIM domain.
23.4R1
Starting in Junos OS Release 23.4R1 and Junos OS Evolved Release 23.4R1, you can customize the DF election method on multihoming peer OISM PEG devices to use mod-based or preference-based PEG DF election. When you configure this feature, the configured DF election method replaces the default PIM-based DF election method.
23.4R1
Starting in Junos OS and Junos OS Evolved Release 23.4R1, we no longer require you to set the install-star-g-routes option on the QFX10000 line of switches or the PTX10000 line of routers when you configure those devices as OISM server leaf or border leaf devices.
23.1R1EVO
Starting in Junos OS Evolved Release 23.1R1, QFX5130-32CD and QFX5700 switches support OISM and AR with MLDv1, MLDv2, and MLD snooping. You can configure MLD and MLD snooping on these devices when they act as OISM server leaf devices, OISM border leaf devices, or standalone AR replicator devices with OISM.
22.3R1-EVO
Starting in Junos OS Evolved Release 22.3R1, PTX10001-36MR, PTX10004, PTX10008, and PTX10016 routers support OISM with IGMPv2 or IGMPv3, IGMP snooping, and SMET route optimization. These devices support OISM using MAC-VRF EVPN instances with vlan-aware and vlan-based service types. You can configure any of these devices in OISM server leaf, border leaf, or lean spine roles. In the border leaf role, these devices support any of the available OISM methods to connect to an external multicast PIM domain: M-VLAN IRB method, classic L3 interface method, or non-EVPN IRB method.
22.3R1
Starting in Junos OS Release 22.3R1, EX4300-48MP and EX4400 switches support forwarding and routing multicast traffic using OISM. These devices support OISM with IGMPv2 in the default switch instance with the VLAN-aware service model. You can configure these devices only in the OISM server leaf role, and not as an OISM border leaf device.
22.2R1-EVO
Starting in Junos OS Evolved Release 22.2R1, you can enable AR with OISM in MAC-VRF EVPN instance configurations on QFX5130-32CD and QFX5700 switches. You can configure the AR leaf role or AR replicator role on these devices. The AR replicator role operates only in standalone mode (the AR replicator role can't be collocated with the OISM border leaf role on the same device). We support AR and OISM with IGMPv2 or IGMPv3, and IGMP snooping.
22.2R1-EVO
Starting in Junos OS Evolved Release 22.2R1, QFX5130-32CD and QFX5700 switches configured as AR replicators with OISM install multicast (*,G) states with IGMPv2 or multicast (S,G) states with IGMPv3 for EVPN Type 6 routes only on the SBD VLAN. On these devices, you only see multicast group routes on the SBD in show multicast snooping route command output.
22.2R1
Starting in Junos OS Release 22.2R1, we support OISM with IGMPv2 or IGMPv3 with the default switch instance or MAC-VRF EVPN instances (vlan-aware and vlan-based service types) on EX4650, QFX5110, QFX5120, QFX10002 (except QFX10002-60C), QFX10008, and QFX10016 switches. You can configure any of these devices in the OISM server leaf role. All of these devices except EX4650 and QFX5110 switches can be OISM border leaf devices. On QFX10000 Series border leaf devices, you can use either the OISM M-VLAN IRB interface method or the classic L3 interface method to connect to an external multicast PIM domain. On EX4650 and QFX5120 border leaf devices, you can use only the classic L3 interface method.
22.2R1
Starting in Junos OS Release 22.2R1, you can enable AR with OISM in the default switch instance or MAC-VRF EVPN instances on EX4650, QFX5110, QFX5120, QFX10002 (except QFX10002-60C), QFX10008, and QFX10016 switches. The AR replicator role can be collocated with the OISM border leaf role on the same device, or you can configure the AR replicator role in standalone mode on a lean spine device in the fabric. (Only switches in the QFX10000 line can be AR replicators.) We support AR and OISM with IGMPv2 or IGMPv3, and IGMP snooping.
22.1R1EVO
Starting in Junos OS Evolved Release 22.1R1, QFX5130-32CD and QFX5700 switches support OISM with IGMPv2 or IGMPv3 in MAC-VRF EVPN instances (vlan-aware and vlan-based service types). These devices can be OISM server leaf or border leaf devices. The border leaf devices support the classic L3 interface model or the non-EVPN IRB model to connect to an external multicast PIM domain.
21.2R1
Starting in Junos OS Release 21.2R1, QFX5110, QFX5120, and QFX10002 (except QFX10002-60C) switches support OISM with IGMPv2 in the default switch instance with the VLAN-aware service model. You can configure any of these devices in the OISM server leaf role, but only QFX10002 switches can be OISM border leaf devices. The border leaf devices support either the OISM M-VLAN IRB model or the classic L3 interface model to connect to an external multicast PIM domain.