Switching - Westcon Security

advertisement
JUNIPER METAFABRIC
Westcon 5 daagse
FEBRUARY, 2014
Washid Lootfun
Sr. Pre-Sales Engineer
wmlootfun@juniper.net
META-FABRIC ARCHITECTURE PILLARS
Simple
Open
Smart
Easy to deploy & use
Maximize flexibility
Save time,improve
performance
•
Mix- and match deployment
•
One OS
•
Universal buidling block for
any network architecture
•
Seamless 1GE  10GE 
40GE  100GE upgrades
•
IT Automation via Open
Interfaces; Vmware, Puppet,
Checf, Python
•
JUNOS Scripting & SDK
•
Standard Optics
•
•
Open Standards-based
interfaces L2,L3 MPLS
Open SDN protocol support,
VxLAN, OVSDB, OpenFlow
•
Elastic (Scale-out) Fabrics
•
Qfabric
•
Virtual Chassis
•
Virtual Chassis Fabric
METAFABRIC ARCHITECTURE PORTFOLIO
Switching
Routing
Management
SDN
Flexible building blocks; simple switching fabrics
Universal data center gateways
Smart automation and orchestration tools
Simple and flexible SDN capabilities
Data Center Security
Adaptive security to counter data center threats
Solutions & Services
Reference architectures and professional services
EX SWITCHES
EX SERIES PRODUCT FAMILY
AGGREGATION/ CORE
Network
Director
One JUNOS
Core/
Aggregation
Switch
Dense Access/
Aggregation
Switch
ACCESS
EX8208
EX8216
EX6210
Entry Level
Access
Switches
EX2200
EX2200-C
EX3300
MODULAR
Versatile
Access
Switch
Proven
Access
Switch
EX4200
FIXED
Powerful
Aggregation
Switch
EX4300
EX9204
EX9208
EX9214
Programmable
Core/Distribution
Switch
EX4300 SERIES SWITCHES
Product Description
•
•
•
•
•
24/48x 10/100/1000 TX access ports
4x 1/10G (SFP/SFP+) uplink ports
4x 40G (QSFP+) VC / uplink ports
PoE / PoE+ options
Redundant / Field Replaceable components (power supplies,
fans, uplinks)
DC power options
•
AFI
AFO

Notable Features
•
L2 and basic L3 (static, RIP) included
•
OSPF, PIM available with enhanced license
•
BGP, ISIS available with advanced license
Virtual Chassis
•
10 members
•
160-320 Gbps VC backplane
12 hardware queues per port
Front to Back & Back to front airflow options
SKU
# Ports
PoE/PoE+
Ports
PoE power budget
EX4300-24P
24
24
550 W
EX4300-24T
24
-
-
EX4300-48P
48
48
900 W
EX4300-48T
48
-
-
Target Applications
EX4300-48T-AFI
48
-
-
•
EX4300-48T-DC
48
-
-
EX4300-48T-DC-AFI
48
-
-
•
•
•
•
•
Campus data closets
Top of Rack data center / High Performance 1G
server attach applications
Small Network Cores
INTRODUCING THE EX9200 ETHERNET SWITCH
AVAILABLE MARCH 2013
EX9214
• Native programmability (Junos image)
• Automation toolkit
• Programmable Control/Management planes and
SDK (SDN, OpenFlow, etc.)
EX9208
• 1M MAC addresses
• 256K IPv4 and 256K IPv6 routes
EX9204
• 32K VLANs (bridge domains)
•
•
•
•
L2, L3 switching
MPLS & VPLS /EVPN*
ISSU
Junos Node Unifier
• 4, 8 & 14 slots; 240G/slot
• 40x1GbE, 32x10GbE, 4x40GbE & 2x100GbE
• Powered by Juniper One Custom Silicon
Juniper One Custom Silicon
 Roadmap
EX9200 LINE CARDS
EX9200-40F/40T
EX9200-32XS
EX9200-4QS
EX9200-2C-8XS
1GbE
Line Cards
 40
x 10/100/1000BASE-T
 40
x 100FX/1000BASE-X SFP
10GbE
Line Card
 32
x 10GbE SFP+
 Up
to 240G throughput
40GbE
Line Card
4
100GbE
Line Card
2
x 40GE QSFP+
 Up
to 120G throughput
x 100G CFP + 8 x 10GbE SFP+
 Up
to 240G throughput
EX9200 FLEXIBILITY VIRTUAL CHASSIS
13.2R2
High Availability
 Redundant RE, switch fabric
 Redundant power /cooling
Management
Require Dual RE’s Per Chassis
Performance and Scale
 Modular configuration
 High-capacity backplane
Easy to Manage
 Single image, single config
 One management IP address
Single Control Plane
 Single protocol peering
 Single RT/FT
Access
Switch
Access
Switch
Virtual Chassis–A Notch Up
 Scale ports/services beyond one chassis
 Physical placement flexibility
 Redundancy beyond one chassis
 One management and control plane
ON ENTERPRISE SWITCHING ARCHITECTURES
Network
Director
Multi-Tier
Collapsed Distribution & Core
Distributed Access
CORE
DISTRIBUTION
ACCESS

Problem: Existing architectures lack scale, flexibility and are operationally complex
Solution: Virtual chassis at both
Access and Distribution layers
Solution: Collapse Core and
Distribution, Virtual chassis at
Access layer
Solution: Virtual chassis at
Access layer across
wiring closets
Benefit: Management
Simplification, Reduced Opex
Benefit: Simplification through
Consolidation, Scale,
Aggregation, Performance
Benefit: Flexibility to expand and
grow, Scale, Simplification
VIRTUAL CHASSIS DEPLOYMENT ON ENTERPRISE
Span Horizontal or Vertical
CONNECT WIRING CLOSETS
COLLAPSE A VERTICAL BUILDING
BUILDING A
CLOSET 1
EXSeries
Virtual Chassis
BUILDING B
EX4300VC-3a
EX6200-1b
10GbE/40GbE
uplinks
WLC
Cluster
WLA
WLA
WLA
WLA
EX4300VC-2a
LAG
Centralized
DHCP and
other services
WLA
EX3300VC-1a
10/40GbE
App Servers
40G VCP
WLA
CLOSET 2
WLA
WLA
WLA
LAG
EX4300
Access
Aggregation/
Core
LAG
SRX Series
Cluster
Internet
LAG
EX4550VC-1a
EX9200VC-1b
DEPLOYING MPLS AND VPN ON ENTERPRISE—
METRO/DISTRIBUTED CAMPUS
Stretch the Connectivity for a Seamless Network
Core
Switch (PE)
MPLS
Core
Switch (PE)
VLAN
Access
Switche (CE)
Core
Switch (PE)
Private MPLS Campus
Core with VPLS
or L3VPN
Core
Switch (PE)
MPLS
VLAN
Access
Switche (CE)
Access
Switche (CE)
Core
Switch (PE)
MPLS
Access
Switche (CE)
Core
Switch (PE)
VLAN
Wireless
Access
Point
Wireless Access
Point
Access
Switches (CE)
Access
Switches (CE)
SITE 1
Wireless
Access Point
Wireless
Access Point
SITE 3
Wireless
Access Point
Wireless
Access Point
SITE 2
VLAN1
Finance/ Business Ops VPN
VLAN2
VLAN3
R&D VPN
Marketing/ Sales VPN
JUNIPER ETHERNET SWITCHING
Simple
Reliable
Secure
 #3 market share in 2 years
 20,000+ switching customers
 Enterprise & Service Providers
 23+ Million ports deployed
QFX5100 PLATFORM
Copyright © 2013 Juniper Networks, Inc.
QFX5100 SERIES
• Next Generation Top of rack switches
– Multiple 10GbE/40GbE port
count options
– Supports multiple data center
switching architectures
• New Innovations:
 Rich L2/L3 features including MPLS
 Low Latency
 SDN ready
– Topology-Independent In-Service Software
Upgrades
– Analytics
– MPLS
– GRE tunneling
QFX5100 NEXT GENERATION TOR
QFX5100-48S
QFX5100-96S
QFX5100-24Q
 48 x 1/10GbE SFP+
 96 x 1/10GbE SFP+
 24 x 40GbE QSFP
 6 x 40GbE QSFP uplinks
 8 x 40GbE QSFP uplinks
 1.44 Tbps throughput
 2.56 Tbps throughput
 8 x 40GbE expansion
slots
 1U fixed form factor
 2U fixed form factor
 2.56 Tbps throughput
 1U fixed form factor
Low latency │ Rich L2/L3 feature set │ Optimized FCoE
Q4CY2013
QFX5100-48S
Front side (port side) view
48 x 1/10GbE SFP+ interfaces
Console
6 x 40GbE QSFP interfaces
USB
Mgmt1 Mgmt0
(SFP) (RJ45)
4+1 redundancy fan tray, color coded (orange:
AFO, blue: AFI), Hot-swappable
1+1 redundancy 650W PS
color coded, hot-swappable
 Each 40GbE QSFP interface can be converted to 4 x 10GbE interfaces without reboot
 Maximum 72 x 10GbE interfaces, 720Gbps
 CLI to change port speed:
set chassis fpc <fpc-slot> pic <pic-slot> port <port-number> channel-speed 10G
set chassis fpc <fpc-slot> pic <pic-slot> port-range <low> <high> channel-speed 10G
QFX5100-96S
Q1CY2014
Front side (port side) view
96 x 1/10GbE SFP+ interfaces
 Supports two port configuration modes:
 96 x 10GbE SFP plus 8 x 40GbE interfaces
 104 x 10GbE interfaces
 1.28Tbps (2.56Tbps full duplex) switching performance
 New 850W 1+1 redundant color-coded hot-swappable power supplies
 2+1 redundant color-coded hot-swappable fan tray
8 x 40GbE QSFP interfaces
QFX5100-24Q
Front side (port side) view
Q1CY2014
(Same FRU side configuration as QFX5100-24S
24 x 40GbE QSFP interfaces
Two hot-swappable 4x40GbE QSFP modules
Port configuration has 4 modes, mode change requires reboot
1.
2.
3.
4.
Default (Fully Subscribed mode):
1.
Doesn’t support QIC
2.
Maximum 24x40GbE interfaces or 96x10GbE interfaces; line rate performance for all packet sizes
104-port mode
1.
Only first 4x40GbE QIC are supported with last 2 40GbE interfaces disabled; first 2 QSFPs work as 8x10GbE
2.
2nd QIC slot cannot be used; no native 40GbE support.
3.
All base ports can be changed to 4x10GbE ports (24x4=96), so total is 104x10GbE interfaces
4x40GbE PIC mode
1.
All base ports can be channelized
2.
Only 4x40GbE QIC is supported; works in both QIC slots but can’t be channelized.
3.
32X40GbE or 96X10GbE + 8X40GbE
Flexi PIC mode
1.
Support all QICs but QIC can’t be channelized
2.
Only base port 4-24 can be channelized. Also supports 32x40GbE configuration
ADVANCED JUNOS SOFTWARE ARCHITECTURE
Provides the foundation for advanced functions
•
•
•
•
ISSU (In-Service Software Upgrade). ENABLE HITLESS UPGRADE
Other Juniper applications for additional service in a single switch
Third-party application
Can bring up the system much faster
JunOS
VM
(Active)
JunOS
VM
(Standby)
Host NW Bridge
3rd Party Application
KVM
Juniper Apps
Linux Kernel (Centos)
QFX5100 HITLESS OPERATIONS
DRAMATICALLY REDUCES MAINTENANCE WINDOWS
Simple
Network Performance
QFX5100 TopologyIndependent ISSU
High-Level QFX5100 Architecture
Junos
Junos VM (Master)
(Master)
Junos
Junos VM
VM (Backup)
(Master)
PFE
PFE
Kernal Based Virtual Machines
Competitive
ISSU Approaches
Network Resiliency
Data Center Efficiency During
Switch Software Upgrade
Linux Kernel
x86 Hardware
Broadcom
BroadcomTrident
Trident IIII
Benefits:
•
Seamless Upgrade
•
•
•
•
No Traffic Loss
No Performance impact
No resilient risk
No port flap
INTRODUCING VCF ARCHITECTURE
Leafs - Integrated L2/L3 gateways
Spines – Integrated L2/L3 switches
 Connects to Virtual and bare metal servers
 Connects leafs , Core, WAN and services
 Local switching
 Any to Any connections
 Single Switch to Manage
Services GW
Spine
Switches
Any to Any connections
Leaf switches
1 RU, 48 SFP+ & 1 QIC
VM
VM
VM
O
VM
VM
VM
O
vSwitch
vSwitch
Virtual Server
Virtual Server
Bare Metal
PLUG-N-PLAY FABRIC
Services GW
WAN/Core

New leafs are auto-provisioned

Auto configuration and image Sync

Any non-factory default node is treated as network device
1 RU, 48 SFP+ & 1 QIC
VM
VM
vSwitch
VM
O
Virtual Server
VM
VM
vSwitch
VM
O
Virtual Server
Bare Metal
VIRTUAL CHASSIS FABRIC DEPLOYMENT OPTION
EX9200
QFX5100-24Q
Virtual Chassis Fabric (VCF) – 10G/40G
QFX5100-48S
QFX3500
EX4300
1 RU, 48 SFP+ & 1 QIC
10G access
Existing 10G access
Existing 1G
access
QFX5100 – SOFTWARE FEATURES
Q1 2014
Q4 2013
• Planned FRS Features*
•
•
•
•
•
•
•
•
•
•
•
•
L2: xSTP, VLAN, LAG, LLDP/MED
L3: Static routing, RIP, OSPF, IS-IS, BGP, vrf-lite, GRE
Multipath: MC-LAG, L3 ECMP
IPv6: Neighbor Discovery, Router advertisement, static
routing, OSPFv3, BGPv6, IS-ISv6, VRRPv3, ACLs
MPLS, L3VPN, 6PE
Multicast: IGMPv2/v3, IGMP snooping/querier, PIMBidir, ASM, SSM, Anycast, MSDP
QoS: Classification, Cos/DSCP rewrite, WRED,
SP/WRR, ingress/egress policing, dynamic buffer
allocation, FCoE/Lossless flow, DCBx, ETS. PFC, ECN
Security: DAI, PACL, VACL, RACL, storm control,
Control Plane Protection
10G/40G FCoE, FIP snooping
Micro-burst Monitoring, analytic
Sflow, SNMP
Python
*Please refer to release notes and manual for latest information
• Planned Post-FRS Features
• Virtual Chassis – Mixed mode
•
•
•
•
•
•
•
• 10 Member Virtual Chassis: Mix of QFX5100,
QFX3500/QFX3600, EX4300
Virtual Chassis Fabric: 20 nodes at FRS with mix of
QFX5100, QFX3500/QFX3600, and EX4300
Virtual Chassis features:
• Parity with standalone
• HA: NSR, NSB, GR for routing protocols,
GRES
ISSU on standalone QFX5100 and all QFX5100
Virtual Chassis, Virtual Chassis Fabric
NSSU in mixed mode of Virtual Chassis or Virtual
Chassis Fabric
64-way ECMP
VXLAN gateway*
OpenStack, Cloudstack integration*
* After Q1 time frame
QFX5100
Virtual Chassis Fabric
Up to 20 members
QFabric
Virtual Chassis
Up to 10 members
…
Up to 128 members
Managed as a Single Switch
Spine-Leaf
Layer 3 Fabric
QFX5100
L3 Fabric
…
VCF OVERVIEW
Flexible
Simple









Single device to manage
Predictable performance
Integrated RE
Integrated control plane
Up to 768 ports
1,10,40G
2-4 spines
10 and 40G spine
L2 , L3 and MPLS
….





Available
4 x Integrated RE
GRES/NSR/NSB
ISSU/NSSU
Any-to-Any connectivity
4 way multi-path



Automated
Plug-n-Play
Analytics for traffic monitoring
Network Director
Hardware
CDBU SWITCHING ROADMAP SUMMARY
2T2013
3T2013
1T2014
2T2014
Future
EX4300
EX9200 2x100G LC
QFX5100 (24QSFP+)
QFX5100 10GBASE-T
Opus PTP
EX4550 10GBASE-T
EX4550 40GbE
Module
QFX5100 (48SFP+)
QFX5100 (24SFP+)
EX9200 6x40GbE LC
EX9200 MACsec
EX9200 400GbE per
slot
QFX5100 (96SFP+)
EX4300 Fiber
Solutions
Software
AnalyticsD
ND 1.5
QFX3000-M/G
L3 Multicast
40GbE
Virtual Chassis w/
QFX Series
QFX3000-M/G
QinQ, MVRP
DC 1.0
Virtualized IT DC
V20
VXLAN Gateway Opus
ISSU on Opus
VXLAN Routing EX9200
ND 2.0
QFX3000-M/G
QFX5100 (48 SFP+)
Node
Campus 1 .0
QFX3000-M/G
10GBASE-T Node
OpenFlow 1.3
DC 1.1
ITaaS & VDI
DC 2.0
IaaS /w Overlay
MX SERIES
Copyright © 2013 Juniper Networks, Inc.
SDN AND THE MX SERIES
Delivering innovation inside and outside of the data center
Flexible SDN enabled silicon to
provide seamless workload
mobility and connections between
private and public cloud infrastructures
USG
EVPN
VMTO
ORE
(Universal
SDN Gateway)
(Ethernet
VPN)
(VM Mobility Traffic
Optimizer)
(Overlay Replication
Engine)
The most advanced and
flexible SDN bridging and
routing gateway
Next-generation technology
for connecting multiple data
centers and providing
seamless workload mobility
Creating the most efficient
network paths for mobile
workloads
A hardware-based, highperformance services
engine for broadcast and
multicast replication within
SDN overlays
VXLAN PART OF UNIVERSAL GATEWAY FUNCTION ON MX
VPLS,
EVPN
L3VPN
IRB.N
IRB.1
Tenant #N,
virtual DC #N
Tenant #0:
virtual DC #0
• - High scale multi-tenancy
– VTEP tunnels per tenant
– P2P, P2MP tunnels
• - Tie to full L2, L3 functions
on MX
IRB.0
Tenant #1,
virtual DC #1
1H 2014
Bridge-Domain.N
VLAN-ID: N
Bridge-Domain.1
VTEP #N
VNID N
LAN interface VLAN-ID:
LAN
interface
1002
#N
#K
Bridge-Domain.0
LAN interfaceVLAN-ID:
LAN
1001interface
#3
#4
LAN interface
#1
LAN interface
#2
VTEP #1
VNID 1
VTEP #0
VNID 0
DC GW
– Unicast, multicast
forwarding
– IPv4, IPv6
– L2: Bridge-domain, virtualswitch
• - Gateway between LAN,
WAN and Overlay
– Ties all media together
– Giving migration options to
the DC operator
NETWORK DEVICES IN THE DATA CENTER
USG
(Universal SDN Gateway)
L4 – 7 Appliances
Bare Metal Servers
Virtualized Servers
SDN Servers
• Databases
• ESX
• NSX ESXi
• Firewalls
• HPC
• ESXi
• NSX KVM
• Load Balancers
• Legacy Apps
• HyperV
• SC HyperV
• NAT
• Non x86
• KVM
• Contrail KVM
• Intrusion Detection
• IP Storage
• ZEN
• Contrail ZEN
• VPN Concentrator
USG (UNIVERSAL SDN GATEWAY)
USG
(Universal SDN Gateway)
Introducing four new options for SDN enablement
Provide SDN-to-non-SDN translation, same IP subnet
SDN to IP (Layer 2)
Layer2 USG
Provide SDN-to-non-SDN translation, different IP subnet
SDN to IP (Layer 3)
Layer3 USG
Provide SDN-to-SDN translation, same or different IP subnet, same or different overlay
SDN to SDN
SDN USG
Provide SDN-to-WAN translation, same or different IP subnet, same or different encapsulation
SDN to WAN
WAN USG
Remote
Data
Center
Branch
Offices
Internet
USGs INSIDE THE DATA CENTER
USG
(Universal SDN Gateway)
DATA CENTER 1
Layer2 USG
VxLAN VxLAN VxLAN VxLAN
VxLAN
VxLAN VxLAN VxLAN VxLAN
Using Layer 2 USGs to bridge between
devices that reside within the same IP
subnet:
SDN USG
1. Bare metal servers like high-performance databases,
non-x86 compute, IP storage, non-SDN VMs
2. Layer 4–7 services such as load balancers, firewalls,
Application Device Controllers, and Intrusion
Detection/Prevention gateways.
WAN USG
Native IP L2 Native IP L2 Native IP L2 Native IP L2 Native IP L2 Native IP L2
Native IP L2 Native IP L2 Native IP L2 Native IP L2
Native IP L2 Native IP L2 Native IP
Layer3 USG
SDN
Pod 1
VxLAN
Legacy Pods
L2 Native IP L2 Native
L4 – 7
Services
USGs INSIDE THE DATA CENTER
USG
(Universal SDN Gateway)
DATA CENTER 1
Layer2 USG
VxLAN VxLAN VxLAN VxLAN
VxLAN
VxLAN VxLAN VxLAN VxLAN
Using Layer 3 USGs to route between
devices that reside within different IP
subnets:
SDN USG
1. Bare metal servers like high-performance databases,
non-x86 compute, IP storage, non-SDN VMs
2. Layer 4–7 services such as load balancers, firewalls,
Application Device Controllers, and Intrusion
Detection/Prevention gateways.
WAN USG
Native IP L3 Native IP L3 Native IP L3 Native IP L3 Native IP L3 Native IP L3
Native IP L3 Native IP L3 Native IP L3 Native IP L3
Native IP L3 Native IP L3 Native IP
Layer3 USG
SDN
Pod 1
VxLAN
Legacy Pods
L3 Native IP L3 Native
L4 – 7
Services
USGs INSIDE THE DATA CENTER
USG
(Universal SDN Gateway)
DATA CENTER 1
Layer2 USG
VxLAN
VxLAN VxLAN VxLAN VxLAN
Using SDN USGs to communicate
between islands of SDN:
1. NSX to NSX – Risk, scale, change control,
administration
2. NSX to Contrail – Multi-vendor, migrations
WAN USG
VxLAN VxLAN VxLAN VxLAN
GRE MPLSoverGRE MPLSoverGRE MPLSoverGRE MP
NSX
SDN Pod 2
LSoverGRE MPLSoverGRE MPLS
SDN USG
VxLAN VxLAN VxLAN VxLAN
VxLAN VxLAN VxLAN VxLAN VxLAN VxLAN
Layer3 USG
SDN
Pod 1
VxLAN
Contrail
SDN Pod 1
USGs FOR REMOTE CONNECTIVITY
USG
(Universal SDN Gateway)
DATA CENTER 1
Internet
Layer2 USG
Layer3 USG
VxLAN
VxLAN VxLAN VxLAN VxLAN
SDN
Pod 1
SDN USG
1. Data Center Interconnect – SDN to [VPLS, EVPN,
L3VPN]
2. Branch Offices – SDN to [GRE, IPSec]
GRE GRE GRE GRE GRE GRE GRE GRE GRE GRE GRE GRE GRE
BRANCH OFFICES
EVPN EVPN EVPN EV
PN EVPN EVPN EVPN EVPN
Using SDN USGs to communicate to
resources outside the local data center:
Native IP L3 Native IP L3 Native IP L3 Native IP L3 Native IP L3 Native IP L3
NSX SDN Pod 2
EVPN EVPN
VxLAN
VxLAN VxLAN VxLAN VxLAN
3. Internet – SDN to IP (Layer 3)
DATA CENTER 2
WAN USG
UNIVERSAL GATEWAY SOLUTIONS
USG
(Universal SDN Gateway)
DATA CENTER 1
Native IP L2 Native IP L2 Native IP L2 Native IP L2 Native IP L2 Native IP L2
Layer2 USG
VxLAN
VxLAN VxLAN VxLAN VxLAN
Native IP L3 Native IP L3 Native IP L3 Native IP L3 Native IP L3 Native IP L3
VxLAN VxLAN VxLAN VxLAN VxLAN
VxLAN VxLAN VxLAN VxLAN
MPLSoverGRE MPLSoverGRE MPLSoverGRE
VxLAN VxLAN VxLAN VxLAN
Native IP L3 Native IP L3 Native IP L3 Native IP L3
Native IP L3 Native IP
VxLAN
GRE GRE GRE GRE GRE GRE GRE
EVPN
GRE GRE GRE
SDN Pod 2
SDN USG
Native IP L2 Native IP L2 Native IP L2 Native IP L2
Internet
LSoverGRE MPLSoverGRE MPLS
Native
IP L2 IPNative
IP L2 IPNative
IP IP
Native
L3 Native
L3 Native
Layer3 USG
VxLAN
SDN
Pod 1
Legacy Pods
L2 Native IP L2 Native
VxLAN VxLAN VxLAN
L3 Native IP L3 Native
BRANCH OFFICES
WAN USG
DATA CENTER 2
NSX
SDN Pod 2
Contrail
SDN Pod 1
L4–7
Services
USG COMPARISONS
USG
(Universal SDN Gateway)
Layer 2
Layer 3
SDN
WAN
USG
USG
USG
USG
Description
Provide SDN-to-non-SDN
translation, same IP subnet
Provide SDN-to-non-SDN
translation, different IP
subnet
Provide SDN-to-SDN translation,
same or different IP subnet,
same or different Overlay
Provide SDN-to-WAN translation,
same or different IP subnet
QFX5100
✔
MX Series/EX9200
✔
✔
✔
✔
X86 Appliance
✔
✔
Competing ToRs
✔
Competing Chassis
✔
Use Cases
NSX or Contrail talk Layer
2 to non-SDN VMs, bare
metal and L4-7 services
NSX or Contrail talk to
other PODs of NSX or
Contrail
NSX or Contrail talk to
other remote locations –
branch, DCI
Description
NSX or Contrail talk Layer
3 to non-SDN VMs, bare
metal and L4-7 services
and Internet
EVPN (Ethernet VPN)
Next-generation technology for connecting multiple data
centers and providing seamless workload mobility
PRE-EVPN: LAYER 2 STRETCH BETWEEN
DATA CENTERS
EVPN
(Ethernet VPN)
Without EVPN
Data
Plane
Control
Plane
•
•
Only one path can be active at a given time
Remaining links are put into standby mode
•
Layer 2 MAC tables are populated via the data plane
(similar to a traditional L2 switch)
Results in flooding of packets across WAN due to
out of sync MAC tables
•
DATA CENTER 1
MAC
VLAN
Interfaces
MAC
VLAN
Interfaces
AA
10
xe-1/0/0.10
BB
10
xe-1/0/0.10
Router 1’s MAC Table
Server 1
MAC: AA
xe-1/0/0.10
DATA CENTER 2
Router 2’s MAC Table
ge-1/0/0.10
ge-1/0/0.10
xe-1/0/0.10
Server 2
PRIVATE MPLS WAN without EVPN
xe-1/0/0.10
VLAN 10
xe-1/0/0.10
ge-1/0/0.10
✕
ge-1/0/0.10
VLAN 10
MAC: BB
POST-EVPN: LAYER 2 STRETCH BETWEEN
DATA CENTERS
EVPN
(Ethernet VPN)
With EVPN
Data
Plane
Control
Plane
•
•
All paths are active
Inter-data center traffic is load-balanced across all
WAN links
•
Layer 2 MAC tables are populated via the control
plane (similar to QFabric)
Eliminates flooding by maintaining MAC table
synchronization between all EVPN nodes
•
DATA CENTER 1
MAC
VLAN
Interfaces
MAC
VLAN
Interfaces
AA
10
xe-1/0/0.10
BB
10
xe-1/0/0.10
BB
10
ge-1/0/0.10
AA
10
ge-1/0/0.10
Router 1’s MAC Table
Server 1
MAC: AA
xe-1/0/0.10
DATA CENTER 2
Router 2’s MAC Table
ge-1/0/0.10
ge-1/0/0.10
xe-1/0/0.10
Server 2
PRIVATE MPLS WAN without EVPN
xe-1/0/0.10
VLAN 10
xe-1/0/0.10
ge-1/0/0.10
ge-1/0/0.10
VLAN 10
MAC: BB
VMTO
(VM Mobility Traffic Optimizer)
Creating the most efficient network paths for mobile
workloads
THE NEED FOR L2 LOCATION AWARENESS
Scenario without VMTO
DC1
(VM Mobility
Traffic Optimizer)
Scenario with VMTO enabled
PRIVATE MPLS WAN
VLAN 10
VMTO
PRIVATE MPLS WAN
VLAN 10
VLAN 10
DC2
DC1
VLAN 10
DC2
VMTO
WITHOUT VMTO: EGRESS TROMBONE EFFECT
(VM Mobility
Traffic Optimizer)
20.20.20.100/24
Server 1
VLAN 20
DC 1
PRIVATE MPLS WAN
Standby VRRP
DG: 10.10.10.1
DC 2
Active VRRP
DG: 10.10.10.1
VLAN 10
Server 2
10.10.10.100/24
Task:
Server 3 in Data Center 3 needs to send packets to Server 1 in Data
Center 1.
Problem:
Server 3’s active Default Gateway for VLAN 10 is in Data Center 2.
Effect:
1. Traffic must travel via Layer 2 from Data Center 3 to Data Center
2 to reach VLAN 10’s active Default Gateway.
2. The packet must reach the Default Gateway in order to be routed
towards Data Center 1. This results in duplicate traffic on WAN
links and suboptimal routing – hence the “Egress Trombone
Effect.”
Standby VRRP
DG: 10.10.10.1
DC 3
Standby VRRP
DG: 10.10.10.1
VLAN 10
Server 3
10.10.10.200/24
VMTO
WITH VMTO: NO EGRESS TROMBONE EFFECT
(VM Mobility
Traffic Optimizer)
20.20.20.100/24
Server 1
VLAN 20
DC 1
PRIVATE MPLS WAN
Active IRB
DG: 10.10.10.1
DC 2
Active IRB
DG: 10.10.10.1
Task:
Server 3 in Datacenter 3 needs to send packets to Server 1 in
Datacenter 1.
Solution:
Virtualize and distribute the Default Gateway so it is active on every
router that participates in the VLAN.
VLAN 10
Server 2
10.10.10.100/24
Effect:
1. Egress packets can be sent to any router on
VLAN 10 allowing the routing to be done in the
local datacenter. This eliminates the “Egress
Trombone Effect” and creates the most optimal
forwarding path for the Inter-DC traffic.
Active IRB
DG: 10.10.10.1
DC 3
Active IRB
DG: 10.10.10.1
VLAN 10
Server 3
10.10.10.200/24
WITHOUT VMTO: INGRESS TROMBONE EFFECT
VMTO
(VM Mobility
Traffic Optimizer)
20.20.20.100/24
Server 1
VLAN 20
DC 1
Route
Mask
Cost
Next Hop
10.10.10.0
24
5
Datacenter 2
10.10.10.0
24
10
Datacenter 3
DC 1’s Edge Router Table Without VMTO
PRIVATE MPLS WAN
10.10.10.0/24 Cost 10
10.10.10.0/24 Cost 5
DC 2
VLAN 10
Server 2
10.10.10.100/24
Task:
Server 1 in Datacenter 1 needs to send packets to
Server 3 in Datacenter 3.
Problem:
Datacenter 1’s edge router prefers the path to
Datacenter 2 for the 10.10.10.0/24 subnet. It has
no knowledge of individual host IPs.
Effect:
1. Traffic from Server 1 is first routed across the
WAN to Datacenter 2 due to a lower cost route
for the 10.10.10.0/24 subnet.
2. Then the edge router in Datacenter 2 will send
the packet via Layer 2 to Datacenter 3.
DC 3
VLAN 10
Server 3
10.10.10.200/24
VMTO
WITH VMTO: NO INGRESS TROMBONE EFFECT
(VM Mobility
Traffic Optimizer)
20.20.20.100/24
Server 1
VLAN 20
DC 1
Route
Mask
Cost
Next Hop
10.10.10.0
24
5
Datacenter 2
10.10.10.0
24
10
Datacenter 3
10.10.10.100
32
5
Datacenter 2
DC 1’s Edge 32
Router Table
VMTO
10.10.10.200
5 WITH Datacenter
3
10.10.10.200/32 Cost 5
10.10.10.100/32 Cost 5
PRIVATE MPLS WAN
10.10.10.0/24 Cost 10
10.10.10.0/24 Cost 5
Task:
Server 1 in Datacenter 1 needs to send packets to Server 3 in
Datacenter 3.
DC 2
Solution:
In addition to sending a summary route of 10.10.10.0/24 the
datacenter edge routers also send host routes which represent the
location of local servers.
VLAN 10
Server 2
10.10.10.100/24
Effect:
1. Ingress traffic destined for Server 3 is sent directly
across the WAN from Datacenter 1 to Datacenter
3. This eliminates the “Ingress Trombone Effect”
and creates the most optimal forwarding path for
the Inter-DC traffic.
DC 3
VLAN 10
Server 3
10.10.10.200/24
NETWORK DIRECTOR
SMART NETWORK MANAGEMENT FROM A SINGLE PANE OF GLASS
Visualize
Physical and virtual visualization
API
Network
Director
Analyze
Smart and proactive networks
Physical
Networks
Virtual
Networks
Control
Lifecycle and workflow automation
CONTRAIL
SDN CONTROLLER
OVERLAY ARCHITECTURE
Orchestrator
SDN
CONTROLLER
REST
Horizontally
scalable
Highly available
Federated
BGP
Clustering
SDN Controller
BGP
Federation
Control
Configuration
Analytics
Control
XMPP
BGP + Netconf
Virtualized Server
VM
VM
VM
JunosV Contrail Controller
XMPP
Virtualized Server
IP fabric
(underlay network)
VM
VM
VM
Tenant VMs
KVM Hypervisor +
JunosV Contrail vRouter/Agent
(L2 & L3)
Juniper Qfabric/QFX/EX
or 3rd party underlay switches
Juniper MX
or 3rd party gateway routers
METAFABRIC ARCHITECTURE:
WHAT WILL IT ENABLE?
SIMPLE
OPEN
VM
VM
VM
VM
VM
VM
SMART
VM
VM
VM
VM
VM
VM
VM
VM
VM
Accelerated time to value and increased value over time
www.juniper.net/metafabric
THANK YOU
Download