Uploaded by MIKIELA Karter

216891445-Evo-Controller-8200-IP-Configuration-KO

advertisement
Evo Controller 8200/RNC
IP configuration
Ivan Guo
Agenda
•
•
•
•
1. Overview
2. IP transmission boards
3. IP Configuration
4. Learn by instance
• References.
Overview
HW architecture
TN Infra Structure
IP/Ethernet
ATM
IP/Ethernet
Traffic
(e.g. Mub/Mut)
OSS-RC
Ext. Mgmt Sys
Main Subrack
1..1
CMXB
SCXB
EPB
ET
Synch. Ref.
Mur
-48V/DC
Mur
Thin Client/
Console
Extension Subrack
Iu
CN
CMXB
SCXB
EPB
ET
0..2
Iur
RNC
-48V/DC
Iub
RBS
RNC
HW architecture
TN Infra Structure
IP/Ethernet
IP/Ethernet
Traffic
(e.g. Mub/Mut)
Main Subrack
CMXB
OSS-RC
Ext. Mgmt Sys
1..1
Synch. Ref.
SCXB
Mur
-48V/DC
EPB
Mur
Thin Client/
Console
Extension Subrack
Iu
CMXB
CN
Iur
EPB
RNC
Iub
RBS
RNC
0..2
SCXB
-48V/DC
Evo Processor Board
• Evo Processor Board (EPB1)
– General processing
– 2 X 10 Gbps Ethernet ports in the back plane
– 2 X RS232 ports in one connector on the front panel
(RS232)
– 2 X 1 Gbps Ethernet ports in one connector on the front
panel (DBG)
25 (GE)
22 (RS232)
EPB1
Evo Processor Board
•
•
•
The EPB implements IP and Ethernet layer functions, such as: 10
Gigabit Ethernet termination, VLAN termination and IP layer
termination, Performance Management (PM), and Fault Management
(FM).
The two 10 Gigabit Ethernet links (interfaces) at the back are accessed
from the Ethernet switch board through the backplane
Local Access to the node is only supported through the Gigabit
Ethernet interface on the front.
Evo Processor Board
•
•
Ethernet interface redundancy (1+1) is handled by the embedded
Ethernet forwarder. The embedded Ethernet forwarder does not
require any configuration.
Each GigabitEthernet MO is configured to use the 10 Gigabit Ethernet
Interface 1 or the 10 Gigabit Ethernet Interface 2 by using the
primaryLink attribute.
IP termination on EPB
3810/3820
ET-IPG
SPB
Iu-ps
(IP)
AAL5
AAL5
Ext. host
ET-IPG
PDR
AAL2
DC
Gateway
Ext. host
Gateway
Iub
(IP)
EVO
UDP
Iu-ps
(IP)
Ext. host
UDP
PDR
EPB
Int. host
UDP
UDP
DC
Ext. host
Iub
(IP)
IP termination on EPB
• Advantage & disadvantage
– Increase call setup capacity and throughput
– Large increase in external IP addresses
• 3 types of Software Allocations (SWAs) for EPB:
– EPB_C1 RNC Central Processor 1 placed only in Main Subrack
in slots 3 and 25.
– EPB_C2 RNC Central Processor 2 placed only in Main Subrack
in slots 4 and 26.
– EPB_BLADE_A Module Controller (MC) functionality resides on
the primary processor and Device functionality on the primary
and secondary processor.
IP termination on EPB
RNC Software deployment
C1 (1+1)
C2 (1+1)
- 2 x SCTP Front End
- O&M
- RFN server (moved from TUB)
- 2 x SCTP Front End
- Central device handling
- UE register
“Blade” (2 – 68)
Primary processor
Secondary processor
CPP
PDR device + CPP
IP (bare metal)
IP (bare metal)
CC device
DC device
RNC Module + RNSAP
DC device
RNC Module + RANAP
DC device
RNC Module + PCAP
DC device
DC device
DC device
DC device
DC device
IP transmission boards
Cabinet
EvoC 8200/RNC
EPB
ES-2
ES2
Power & Fan Module
EPB
ES1
EPB
ES-1
MS
Power & Fan Module
APP2
APP1
MS
Power & Fan Module
R
R
Iu/Iur
APP
APP
Cable shelf
R
R
Iub
System Control Switch Board
• System Control Switch Board (SCXB3)
46 (SYNC2)
42 (E-DBG)
40 (RS232)
37 (GPS)
35 (SYNC1)
32 (SC)
27 (RP-B)
25 (GE 3)
22 (GE 2)
20 (GE 1)
17 (E 4)
15 (E 3)
12 (E 2)
10 (E 1)
SCXB3
– Ethernet switch
– Timing unit
– 28 X 1 Gbps Ethernet port in the back plane (One
is looped back to the same board.)
– 3 X Clock synchronization reference input
connector (SYNC1, SYNC2, GPS)
– 10/100/1000BASE-T debug connector (E-DBG)
– RS232 debug connector (RS232)
– Serial Control connector (SC)
– Connector to regional processor in AXE, not used
in RNC (RP-B)
– 3 X 1 Gbps Ethernet ports on the front panel (GE1,
GE2, GE3)
– 4 X 10 Gbps Ethernet ports on the front panel (E1,
E2, E3, E4)
Common Main Switch Board
• Common Main Switch Board (CMXB3)
– Ethernet switch
– 24 X 10 Gbps Ethernet ports in the back plane
– 10/100/1000BASE-T debug connector on the front
panel (DBG)
– 4 X 1 Gbps Ethernet ports on the front panel (GE1
- GE4)
– 4 X 10 Gbps Ethernet ports on the front panel (E5
– E8)
– 4 X 40 Gbps or 10 Gbps Ethernet ports on the front
panel (E1 – E4)
42 (DBG)
40 (GE 4)
37 (GE 3)
35 (GE 2)
32 (GE 1)
27 (E 8)
25 (E 7)
22 (E 6)
20 (E 5)
17 (E 4)
15 (E 3)
12 (E 2)
10 (E 1)
CMXB3
evel 2
Evo Controller Provides 10G and 1G
Ethernet to all Slot Positions
White fronts and new LEDs
SCXB CMXB3
SCXB CMXB3
Dbg
APP
ExtA
Sync
RPBS
1GE
1GE
1GE
1GE
10GE
1GE
1GE
Dbg
APP
ExtA
Sync
RPBS
Dbg
10GE
10GE
SCXB provides 28x1G backplane ports and
4x10G + 3x1G front ports
Dbg
1GE
1GE
1GE
1GE
10GE
1GE
1GE
10GE
10GE
1GE
40GE
1GE
40GE
10GE
40GE
10GE
40GE
10GE
40GE
10GE
40GE
10GE
40GE
10GE
40GE
10GE
40GE
10GE
40GE
CMXB3 allow 24 * 10GE backplane ports
3 * 10GE + 5 * 40GE + 4 * 1GE front ports
Active Patch Panel
• Active Patch Panel (APP).
– Each APP contains a row of connectors and slots for optoelectrical
converters (SFPs). Two APPs are placed below the lowest subrack in Evo
RNC. Each row contains
– 4 SFP slots (pos. 18, 28, 32 and 42) for maximum 10 Gbps Ethernet
4 DensiShield connectors (pos. 21, 24, 35 and 38) that are connected to
the SFP slots
– 4 RJ-45 connectors (pos. 47, 51, 61 and 65) for Ethernet
2 DensiShield connectors (pos. 55 and 68) that are connected to the RJ-45
connectors (two RJ-45 connected to each DensiShield connector)
2 DensiShield connectors for Supervision and control (pos. 74 and 77)
– 1 connector for power supply (pos. 82)
•
Note that the lower has position designators starting with
“02-“ while the higher APP has position designators starting with “03-“,
Connections APP to CMXB3
Cables between APP and Main subrack.
APP Pos
Type
Connect
or
Pos
CMXB3
E5
5-20
2-21
CMXB3
E6
5-22
2-24
CMXB3
E7
5-25
2-35
CMXB3
E8
5-27
2-38
CMXB3
E5
83-20
3-21
CMXB3
E6
83-22
3-24
CMXB3
E7
83-25
3-35
CMXB3
E8
83-27
3-38
CMXB3
CMXB3
Board connector
Non-CMXB connections between MS and APPs
Cables for management and control.
Connector
Pos
Unit
Pos
SCXB3
SC
02-32
APP
02-74
SCXB3
SC
80-32
APP
03-74
EPB1
GE
08-25
APP
02-68
EPB1
GE
74-25
APP
03-68
SCXB3
Unit
EPB1
One end
EPB1
SCXB3
Other end
ESL & CISL
SCXB3
SCXB3
SCXB3
CMXB3
SCXB3
CMXB3
CMXB3
SCXB3
CMXB3
CMXB3
SCXB3
CMXB3
IP Configuration
Ethernet Infrastructure
40G
SCXB
CMXB
EPB
CMXB
SCXB
40G
40G
40G
SCXB
CMXB
EPB
CMXB
SCXB
40G
40G
40G
4x10G
SCXB
APP
CMXB
EPB
CMXB
SCXB
4x10G
4x10G
APP
4x10G
IP Transport in Evo
•
•
•
•
IP Host User Plane Termination
The EPB1 board is equipped with two processors with 8 cores each. Each processor
will that uses external or internal communication needs an IP host in order to
communicate. Each processor supports up to 16 external and 1 internal IP host
terminations.
RNC user plane software (DC, CC and PDR) will be located on both processors
which results in that IP interfaces will be needed on both processors.
The external IP hosts will be configured in a similar way as the IpAccessHostEt using
the MOM. However the internal IP host will not be shown in the MOM and will not be
possible to configure.
IP Transport in Evo
•
•
•
IP Host Control Plane Termination
The control plane (SCTP) can be terminated in two different ways; via ipAccessSctp
or directly to the ipAccessHostEt. The main benefit of using the IpAccessSctp MO is
that the CP and UP can share the same ipAccessHostEt (IP address). The SCTP
functionality supports multi-homing; see MO structure for Iub CP below.
The proposed default configuration is to configure the Iub CP without any multihoming and share IP address with the Iub UP.
IP Addressing
• External and internal IP hosts
• The IP termination on the EPB consists of a number of external and
internal IP hosts.
– Up to 16 external hosts and up to 64 internal hosts are supported
for each processor.
– Each external host is connected to a separate IpInterface MO
and all the internal hosts are connected to the same IpInterface.
– The internal IP hosts require no configuration and are not visible
in the MO structure.
– It is important to make sure that the VLAN ID is not colliding with
any planned external VLANs.
IP Addressing
• The PIU processor contains MCs, where the IUB CP is terminated. It
also contains a CC device (terminating IUB common channels) and
DC device (terminating the IUB and IUR UP for dedicated channels).
• The Piudevice processor contains PDR device (terminating the IUPS UP) and DC device (terminating the IUB and IUR UP for
dedicated channels).
• A typical configuration:
different VLANs/subnets used for all different interfaces.
– 3 IP hosts per blade on PIU, for IUB UP/CP, IU-CS UP, IUR-UP
– 4 IP hosts per blade on Piudevice, for IUB UP, IU-CS UP, IUR-UP, plus
IU-PS UP.
– For the central boards(C1, C2), one IP host is needed per board for
termination of SCTP for IU-CS, IUPS and IUR CP and for IU-PC.
– In addition, IP hosts are needed on C1 for Mur and IU-BC.
IP Addressing
Iub:
2*68 (EPB blade)+2 (C1 EPB) =138 IP hosts
One /24 network with 256 addresses
IuCS/Iur UP:
2*68 (EPB balde) =136 IP hosts
One /24 network with 256 addresses
E
E E E E
E
Up to 24 EPB blades
E
E E E E
E
Up to 24 EPB blades
IuPS UP:
68 (EPB blade) IP hosts
One /25 network with 128 addresses
C1 C2 E
E E
E C1 C2
2 C1, 2 C2 and 2 – 20 EPB blades
Iu/Iur CP:
2*4 IP hosts
Two /29 network with 8 addresses each
H
Iub UP/CP - two IpAccessHostEt per EPB blade + 2 for NTP Server on C1 device processor
H
IuCS/Iur UP - two IpAccessHostEt per EPB blade
H
IuPS UP - one IP host per EPB blade, only on secondary (PIU Device)
processor
Iu/Iur CP – one (multi homed) IP host on each c1 and c2 EPB (two VLANs, CP-A and CP-B)
H
Vlan Configuration
•
The configuration of node Ethernet infrastructure (VLAN, CoS, RSTP etc)
remains unchanged compared to baseline (3820).
•
•
A number of subnetworks and VLANs should be planned and Vlan MOs
configured before starting the configuration of the external interfaces.
•
For singlehoming, five subnetworks are created, as described below:
–
–
–
–
–
•
One for Iu and Iur control plane
One for Iub user plane and control plane
One for Iu-CS user plane
One for Iur user plane
One for Iu-PS user plane
For multihoming, six subnetworks are created:
–
–
–
–
–
Two for Iu and Iur control plane
One for Iub user plane and control plane
One for Iu-CS user plane
One for Iur user plane
One for Iu-PS user plane
CMXB –VLAN configuration
CP-B on right hand
side CMXB
E
to ES1
to ES2
E1
E2
CP-A
IuPS
Up to 24 EPB blades
E3
IuCS/Iur
external (Iub)
E E E E
E
Up to 24 EPB blades
B
C1 C2 E
E5
E
crosslink MS
E
Iub
E E E E
E E
E C1 C2
2 C1, 2 C2 and 2 – 20 EPB blades
E7
external (Iu/Iur)
H
Iub UP/CP - two IpAccessHostEt per EPB blade + 2 for NTP Server on C1 device processor
H
IuCS/Iur UP - two IpAccessHostEt per EPB blade
H
IuPS UP - one IP host per EPB blade, only on secondary (PIU Device) processor
H
Iu/Iur CP – one (multi homed) IP host on each c1 and c2 EPB (two VLANs, CP-A and CP-B)
Node Internal VLAN
•
A node internal VLAN needs to be defined on all backplane and ISL ports (not
external ports). This is performed by creating a internal VLAN and setting the
vlanType = INTERNAL_IP_TRANSPORT in the VLAN MO.
IP Configuration on EPB Blade
•
•
•
•
The IP Configuration and number of IP Hosts needed on EPB blades can
vary depending on traffic separation. But the IP configuration on all EPB
blades in a node should be identical.
A typical configuration (used in the following examples) would be:
– IuPS User Plane in one VLAN
– IuCS User Plane and Iur User Plane in one VLAN
– Iub Control Plane and Iub User Plane in same VLAN
There are two processors on each EPB and IP termination is required on
each processor for the functionality allocated to that processor.
This gives the following need:
– Primary (PIU) Processor: Iub CP, Iub UP, IuCS UP and Iur UP
– Secondary (PIU Device) Processor: Iub UP, IuCS UP, Iur UP and IuPS
UP
IP Configuration on EPB blade: Iub CP/UP, IuPS UP and IuCS/Iur UP
RncFunction
IubLink
Sctp
rpuId
ipAccessSctpRef
controlPlaneTranportOption
remoteCpIpAddress1
sctpRef
ReliableProgramUniter
admActiveSlot
admPassiveSlot
IpAccessSctp
ipAccessHostEtRef1
ipAccessHostEtRef2
userPlaneTransportOption
userPlaneIpResourceRef
IpAccessHostPool
ipAccessHostRef
IpAccessHostEt
ipAddress
ipInterfaceMoRef
IpAccessHostEt
ipAddress
ipInterfaceMoRef
IurLink
userPlaneTransportOption
userPlaneIpResourceRef
Slot
PlugInUnit
ExchangeTerminalIp
GigaBitEthernet
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef
PiuDevice
IpAccessHostEt
ipAddress
ipInterfaceMoRef
CnOperator
IuLink
userPlaneTransportOption
userPlaneIpResourceRef
IpAccessHostPool
ipAccessHostRef
Ranap
cnDomainInd=CS
IpAccessHostEt
ipAddress
ipInterfaceMoRef
GigaBitEthernet
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef
IuLink
userPlaneTransportOption
userPlaneIpResourceRef
Ranap
cnDomainInd=PS
ExhangeTerminalIp
IpAccessHostPool
ipAccessHostRef
IpAccessHostEt
ipAddress
ipInterfaceMoRef
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef
IP Configuration on EPB Blade
IuPS (1..)
IuLink
IuLink
IpAccessHostPool (IuPS UP)
IuLink
IuLink
Rpu
Slot
Slot
IpAccessHostEt
IpInterface
PiuDevice
IpAccessHostEt
IpInterface
Slot
IpAccessHostEt
IpInterface
IpAccessHostEt
IpAccessHostEt
PiuDevice
IpInterface
IpAccessHostEt
IpAccessHostEt
EPB 2
PlugInUnit
IpInterface
IpInterface
IpInterface
IpInterface
IpInterface
Rpu
PlugInUnit
IpInterface
EPB 1
Sctp
IpAccessSctp
Rpu
PlugInUnit
IpAccessHostEt
IpAccessHostPool (Iub UP)
IpAccessSctp
IpAccessHostEt
IpAccessHostEt
IubLink
IubLink
IubLink
Sctp
IpAccessSctp
IpAccessHostEt
Iub (1..n)
IurLink
IurLink
IpAccessHostPool (IuCS/Iur UP)
Sctp
IpAccessHostEt
Iur (1..n)
IuCS (1..)
IpAccessHostEt
PiuDevice
IpInterface
IpAccessHostEt
IpInterface
IpAccessHostEt
IpInterface
EPB n
IpInterface
IpInterface
IP Configuration on C1 and C2 EPB
•
The IP Configuration and number of IP Hosts needed on C1 and C2 EPBs
depends on Sctp and NTP Server configuration
•
A typical configuration (used in the following examples) would be:
– Four SCTP Front Ends (FE)
• One on each C1 or C2 EPB (non robust)
– Use multihoming with two IP addresses for each Sctp FE
• Does not provide board redundancy; both IP terminations are on
same processor
• Only used for remote-end redundancy and network path diversity
• Two VLANs for the Iu/Iur signalling traffic
• NOTE: Also possible to use other variants
•
IP termination for NTP Server on C1 EPBs
– NTP Server is to be located on the secondary (PIU Device) Processor of
each C1 EPB (two NTP Server instances provide redundancy)
– These two NTP Servers can be used by all RBS nodes connected to the
EvoC node.
Iu/Iur CP – one-legged RPUs, multihoming on same EPB
Sctp
rpuId
ipAccessSctpRef
Sctp
rpuId
ipAccessSctpRef
ReliableProgramUniter
admActiveSlot
admPassiveSlot
IpAccessSctp
ipAccessHostEtRef1
ipAccessHostEtRef2
Sctp
rpuId
ipAccessSctpRef
ReliableProgramUniter
admActiveSlot
admPassiveSlot
IpAccessSctp
ipAccessHostEtRef1
ipAccessHostEtRef2
IpAccessHostEt
ipAddress
ipInterfaceMoRef
ReliableProgramUniter
admActiveSlot
admPassiveSlot
IpAccessSctp
ipAccessHostEtRef1
ipAccessHostEtRef2
IpAccessHostEt
ipAddress
ipInterfaceMoRef
IpAccessHostEt
ipAddress
ipInterfaceMoRef
ReliableProgramUniter
admActiveSlot
admPassiveSlot
IpAccessSctp
ipAccessHostEtRef1
ipAccessHostEtRef2
IpAccessHostEt
ipAddress
ipInterfaceMoRef
IpAccessHostEt
ipAddress
ipInterfaceMoRef
IpAccessHostEt
ipAddress
ipInterfaceMoRef
Slot
Sctp
rpuId
ipAccessSctpRef
Slot
IpAccessHostEt
ipAddress
ipInterfaceMoRef
IpAccessHostEt
ipAddress
ipInterfaceMoRef
Slot
Slot
PlugInUnit
PlugInUnit
PlugInUnit
PlugInUnit
ExchangeTerminalIp
ExchangeTerminalIp
ExchangeTerminalIp
ExchangeTerminalIp
GigaBitEthernet
GigaBitEthernet
GigaBitEthernet
GigaBitEthernet
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef=A
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef=A
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef=A
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef=A
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef=B
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef=B
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef=B
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef=B
PiuDevice
PiuDevice
C1 (Slot 3)
C2 (Slot 4)
PiuDevice
C1 (Slot 25)
PiuDevice
C2 (Slot 26)
Iu/Iur CP – one-legged RPUs, no multihoming – Possible configuration
Sctp
rpuId
ipAccessSctpRef
Sctp
rpuId
ipAccessSctpRef
ReliableProgramUniter
admActiveSlot
admPassiveSlot
IpAccessSctp
ipAccessHostEtRef1
ipAccessHostEtRef2
Sctp
rpuId
ipAccessSctpRef
ReliableProgramUniter
admActiveSlot
admPassiveSlot
IpAccessSctp
ipAccessHostEtRef1
ipAccessHostEtRef2
IpAccessHostEt
ipAddress
ipInterfaceMoRef
ReliableProgramUniter
admActiveSlot
admPassiveSlot
IpAccessSctp
ipAccessHostEtRef1
ipAccessHostEtRef2
IpAccessHostEt
ipAddress
ipInterfaceMoRef
Slot
Sctp
rpuId
ipAccessSctpRef
IpAccessSctp
ipAccessHostEtRef1
ipAccessHostEtRef2
IpAccessHostEt
ipAddress
ipInterfaceMoRef
Slot
ReliableProgramUniter
admActiveSlot
admPassiveSlot
IpAccessHostEt
ipAddress
ipInterfaceMoRef
Slot
Slot
PlugInUnit
PlugInUnit
PlugInUnit
PlugInUnit
ExchangeTerminalIp
ExchangeTerminalIp
ExchangeTerminalIp
ExchangeTerminalIp
GigaBitEthernet
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef
PiuDevice
C1 (Slot 3)
GigaBitEthernet
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef
PiuDevice
C2 (Slot 4)
GigaBitEthernet
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef
PiuDevice
GigaBitEthernet
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef
PiuDevice
C1 (Slot 25)
C2 (Slot 26)
SS7/SCTP Configuration – Four single
SCTP Front Ends
Four FE defined
• One on each EPB with C1 and C2 role (Slots 3,4,25 and 26 in MS)
Redundancy and load sharing must be handled above SCTP level
• Iu links must use at least two SCTP Front Ends (FE) each
– At least two M3uAssocation per CN node
• For board redundancy handled on SCCP/M3UA level
– Up to four M3uAssocation per CN node
• For load sharing on all SCTP FE
• Load distributed evenly on all M3UA/SCTP in each Signalling route set
(Srs)
• Alternative is to use only some SCTP FE for each CN node but spread
traffic for different CN nods on all SCTP FE evenly
• Iur links must use two SCTP hosts each
– At least two M3uAssocation per connected RNC node
• Spread traffic for different RNCs evenly on all SCTP FE
• Possible, but not needed to use more than two SCTP FE for Iur
Iu CP – Ranap/Sccp/Mtp3b
RncFunction
CnOperator
IuLink
Ranap
localSccpApRef
remoteSccpApRef
SccpSp
mtp3bSpId
SccpScrc
SccpApLocal
ssN
SccpApRemote
ssN
mtp3bApId
Mtp3bSpItu
Mtp3bAp
serviceIndicator
routeSetId
Mtp3bSrs
destPointCode
Mtp3bSr
linkSetM3UId
priority
M3uAssociation
localIpMask
localPortNumber
mtp3bSrsId
remoteIpAddress1
remoteIpAddress2
scptId
Sctp
ipAccessSctpRef
rpuId
Mtp3bSr
linkSetM3UId
priority
M3uAssociation
localIpMask
localPortNumber
mtp3bSrsId
remoteIpAddress1
remoteIpAddress2
scptId
Sctp
ipAccessSctpRef
rpuId
Mtp3bSr
linkSetM3UId
priority
Mtp3bSr
linkSetM3UId
priority
M3uAssociation
localIpMask
localPortNumber
mtp3bSrsId
remoteIpAddress1
remoteIpAddress2
scptId
M3uAssociation
localIpMask
localPortNumber
mtp3bSrsId
remoteIpAddress1
remoteIpAddress2
scptId
Sctp
ipAccessSctpRef
rpuId
Sctp
ipAccessSctpRef
rpuId
Iu CP – Ranap/Sccp/Mtp3b – Full loadsharing
RncFunction
CnOperator
IuLink
Ranap
localSccpApRef
remoteSccpApRef
IuLink
Ranap
localSccpApRef
remoteSccpApRef
SccpSp
mtp3bSpId
SccpScrc
SccpApLocal
ssN
SccpApRemote
ssN
mtp3bApId
Mtp3bSpItu
Mtp3bAp
serviceIndicator
routeSetId
Mtp3bAp
serviceIndicator
routeSetId
Mtp3bSrs
destPointCode
Mtp3bSrs
destPointCode
SccpApRemote
ssN
mtp3bApId
Mtp3bSr Mtp3bSr Mtp3bSr Mtp3bSr
M3uA
M3uA
Sctp
ipAccessSctpRef
rpuId
M3uA
M3uA
Sctp
ipAccessSctpRef
rpuId
Mtp3bSr Mtp3bSr Mtp3bSr Mtp3bSr
M3uA
M3uA
Sctp
ipAccessSctpRef
rpuId
M3uA
M3uA
Sctp
ipAccessSctpRef
rpuId
Iur CP – Ranap/Sccp/Mtp3b – Partial load sharing
RncFunction
IurLink
Rnsap
localSccpApRef
remoteSccpApRef
IurLink
Rnsap
localSccpApRef
remoteSccpApRef
SccpSp
mtp3bSpId
SccpScrc
SccpApLocal
ssN
SccpApRemote
ssN
mtp3bApId
Mtp3bSpItu
Mtp3bAp
serviceIndicator
routeSetId
Mtp3bAp
serviceIndicator
routeSetId
Mtp3bSrs
destPointCode
Mtp3bSrs
destPointCode
SccpApRemote
ssN
mtp3bApId
Mtp3bSr
linkSetM3UId
priority
M3uAssociation
localIpMask
localPortNumber
mtp3bSrsId
remoteIpAddress1
remoteIpAddress2
scptId
Sctp
ipAccessSctpRef
rpuId
Mtp3bSr
linkSetM3UId
priority
M3uAssociation
localIpMask
localPortNumber
mtp3bSrsId
remoteIpAddress1
remoteIpAddress2
scptId
Sctp
ipAccessSctpRef
rpuId
Mtp3bSr
linkSetM3UId
priority
M3uAssociation
localIpMask
localPortNumber
mtp3bSrsId
remoteIpAddress1
remoteIpAddress2
scptId
Sctp
ipAccessSctpRef
rpuId
Mtp3bSr
linkSetM3UId
priority
M3uAssociation
localIpMask
localPortNumber
mtp3bSrsId
remoteIpAddress1
remoteIpAddress2
scptId
Sctp
ipAccessSctpRef
rpuId
EvoC 8200/RNC – NTP Server for Network Synchronization (Iub)
Slot=MS3
Slot=MS25
PlugInUnit
PlugInUnit
ExchangeTerminalIp
ExchangeTerminalIp
GigaBitEthernet
GigaBitEthernet
PiuDevice
IpAccessHostEt
ipAddress
ipInterfaceMoRef
ntpServerMode=ENABLED
ExhangeTerminalIp
GigaBitEthernet
PiuDevice
IpAccessHostEt
ipAddress
ipInterfaceMoRef
ntpServerMode=ENABLED
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef
ExhangeTerminalIp
GigaBitEthernet
IpInterface
defaultRouter0
networkPrefixLength
mtu, vlanRef
C1 (MS, Slot 3)
C1 (MS, Slot 25)
IP hosts configured with NTP servers
enabled on both C1 boards
Learn by instance
IP Configuration Instance
• Contents
– Iub User and Control Plane
– IU/IUR Control Plane
– IUPS User Plane
– IUCS/IUR User Plane
Iub User and Control Plane
Iub User and Control Plane
• In this configuration procedure Iub control plane is configured by defining
IpAccessSctp and IpAccessHostEt MOs for SCTP so that the IP hosts can be
used for Iub user plane configuration.
•
•
•
•
•
•
•
•
•
•
For configuration of Iub control plane and user plane the following data must
be known:
• ipAddress: The IP address for the host of the Iub control plane and user
plane. Two for each EPB are needed.
• routerIpAddress: The IP address to the default router. One mandatory
address can be configured. If no router is going to be used in the network a
fictitious address must be used, because this attribute is used to determine
the subnet mask.
• networkPrefixLength: Used to specify which part of the routerIpAddress is
the subnet address.
IUB User Plane
• When the Iub control plane is configured using
IpAccessHostEt on EPB, the additional
IpInterface and IpAccessHostEt are required for
the Iub user plane.
• The typical situation is to have one
IpAccessHostPool for Iub user plane. One
IpAccessHostEt is defined on primary processor
for Iub Control Plane and Iub User Plane, and
one IpAccessHostEt on secondary processor
used only for Iub User Plane.
Iub User and Control Plane
•
•
•
•
•
•
1. Create one IpInterface MO for primary processor (PIU) and one for
secondary processor (PiuDevice).
2. Create two IpAccessHostEt for each EPB
SET “ipInterfaceMoRef” as the interface on the EPB created.
3. Create one IpAccessHostPool for Iub, including all IpAccessHostEt
defined on EPB_BLADE_A in both processors created.
4. Create IpAccessSctp MO for each EPB
Set “ipAccessHostEtRef1” to “IpAccessHostEt” defined on Primary processor.
5. Create one Sctp MO for each EPB
Set “ipAccessSctpRef” to “IpAccessSctp” defined above.
6. Add In the IubLink MO created above add a reference to the
IpAccessHostPool MO, by setting the userPlaneIpResourceRef attribute.
Set the IP addresses for the Iub control plane and the attributes:
userPlaneTransportOption and controlPlaneTransportOption to IPv4.
Iu/Iur Control Plane
•
•
•
•
Horizontal Distribution (HD) stack is divided into two parts, Front End (FE) and Back
End (BE).
The FE handles the communication with the CN and containsthe SCTP-stack.
The BE contains the SCC server. One SCTP (FE) executes on each C1 and C2 EPB
(four EPBs in total).
On each EPB used as Front End (FE) the following MOs are needed.
For single-homed configuration:
• one IpInterface
• one IpAccessHostEt
• one IpAccessSctp
• one Sctp
For multi-homed configuration:
• two IpInterface
• two IpAccessHostEt
• one IpAccessSctp
• one Sctp
The Iu and Iur control plane is not affected by the introduction of the Evo.
There is however a conceptual difference since the multi-homing, both ipAccessHostEt’s
will be on the same processor.
This will lead to the difference; there is no ET board redundancy but path redundancy is
still supported.
RPU for the Iu/Iur control plane is however still supported for board redundancy as in
baseline.
Iu/Iur Control Plane
•
•
•
•
•
•
1. Create one IpInterface MO for each EPB with FE for the singlehomming.
For multihoming, create two IpInterface on the same processor.
2. Create one IpAccessHostEt MO for each EPBs with C1, C2 role for the
singlehoming and two IpAccessHostEt on the same processor, for
multihoming.
3. Create one IpAccessSctp MO for each EPB. For single-homing set the
attribute ipAccessHostEtRef1 as a reference to IpAccessHostEt MO.
In case of multi-homing, set the attributes ipAccessHostEtRef1 and
ipAccessHostEtRef2.
4. Find the RPU to be used when creating the Sctp MO.
The RPU ID depends on the subrack and slot number. The principle rule for
naming the RPUs for SCTP FE on C1 and C2 EPBs is to use the name
reliableProgramLabel_<subrackName>-<slotId>-0 for each four EPBs.
5. Create one Sctp MO for each EPB with C1 and C2 role.
IUPS User Plane
IUPS User Plane
• 1. Configure one IpInterface MO on PlugInUnit (primary processor)
or PiuDevice (secondary processor) for each EPB blade.
• 2. Configure one IpAccessHostEt MO for each EPB blade.
• 3. Create one IpAccessHostPool for Iu-PS user plane with
references to all IpAccessHostEt created in Step 2.
• 4. Configure one IpEthPacketDataRouter instance for each Pdr
Device on the EPB.
• Name “IpEthPacketDataRouterId” as “Pdr_<subrack>-<slot>-<1>”
• 5. For each CN (Core Network) node that is connected to the EvoC
8200/RNCwith IP PS connection create the IuLink MO.
Set “userPlaneIpResourceRef” to the “IpAccessHostPool” created in Step 3.
• 6. Verify new Iu-PS connections by checking whether the
pmNoRoutedIpPacketsUl and pmNoRoutedIpPacketsDl counters
are stepped in the new IpEthPacketDataRouter MOs.
Iu-CS and Iur User Plane
Iu-CS and Iur User Plane
• 1. Create one IpInterface on primary processor (PlugInUnit) and one
on secondary processor (PiuDevice).
Set VlanRef to Vlan MO created for Iu-CS and Iur user plane
• 2. Create two IpAccessHostEt for each EPB board to use.
• 3. Create one IpAccessHostPool MO for Iu-CS and Iur.
Name IpAccessHostPoolId as “Iu_Iur_xx”
There is one common IpAccessHost used for both Iu and Iur.
• 4. Create the IuLink MO for IU CS
Set userPlaneIpResourceRef tp IpAccessHostPool created in Step 3.
• 5. Create the IurLinkMO
Set userPlaneIpResourceRef tp IpAccessHostPool created in Step 3.
External Interface Improvement
• 1 Reconfiguring the RNC Ethernet External Transmission Speed
• The Evo Controller 8200/RNC can be configured to support 10G
Ethernet transmission speed. For instruction on how to reconfigure
transmission speed refer to Reconfiguring External Interfaces.
• 2 Configuring Additional External Interface
• In the Evo Controller 8200/RNC it is possible to split the Iu/Iur and
Iub interfaces into separate physical interfaces. For instruction on
how to activate and configure additional external Ethernet interface
refer to Reconfiguring External Interfaces
•
References
•
•
•
•
1. Initial IP Configuration, 57/1543-AXD 105 03/1 Uen A
2. Hardware Configurations for Evo Controller 8200 for RNC, 2/155 18-COA
109 400 Uen
3. NDS IP Termination on EPB1 in Evo, 50/102 68-FCP 101 8201 Uen
4. IP EPB SW, IP Termination, 20/159 41-FCP 105 1096 Uen
Download