SRX Services Gateway
SRX Services Gateway

SRX5600 dual manment modules

3 weeks ago

i have 2 srx5600 with HA each node has dual managment module "RE" . that mean i have 4 modules in my setup.

 

one of this module has HW.failure. question is why HA has been broken between them and is it normal ? 

6 REPLIES 6
SRX Services Gateway

Re: SRX5600 dual manment modules

3 weeks ago

Hi,

 

No SRX has support for dual RE. I say this in the way that if the primary RE on a node fails the other RE (on the same node) wont take over. Installing a second RE in a node is only used for activating the posibility of having a redundant control-link:

 

"For the SRX5600 and SRX5800 Services Gateways, dual control link functionality requires a second Routing Engine, as well as a second Switch Control Board (SCB) to house the Routing Engine, to be installed on each device in the cluster. The purpose of the second Routing Engine is only to initialize the switch on the SCB. The second Routing Engine, to be installed on SRX5000 line devices only, does not provide backup functionality."

 

Ref: https://www.juniper.net/documentation/en_US/junos/topics/topic-map/security-chassis-cluster-dual-con...

 

The RE that is working on each node is the one installed on slot 0. If the RE that failed in your scenario was the one in slot 0 then it is expected that your cluster was broken. If the backup RE (the one on slot 1) has the same Junos version you could insert it on slot 0 and recover the cluster.

 

 

Please mark my answer as the Solution if it applies.
SRX Services Gateway

Re: SRX5600 dual manment modules

3 weeks ago

Hi Amr,

 

As Rojas correctly stated, the second Routing Engine is only for Dual Control link functionality.

 

I believe in your case the master RE was faulty and that is the reason HA has been broken. In this case, even though the Master RE has failed the Second RE won't take over the mastership. This is the expected behavior.

 

The second Routing Engine can only become the master when the master Routing Engine is not present.



Thanks,
π00bm@$t€®.
Please, Mark My Solution Accepted if it Helped, Kudos are Appreciated too!!!
SRX Services Gateway

Re: SRX5600 dual manment modules

a week ago

The RE that is working on each node is the one installed on slot 0. If the RE that failed in your scenario was the one in slot 0 then it is expected that your cluster was broken. If the backup RE (the one on slot 1) has the same Junos version you could insert it on slot 0 and recover the cluster.....

 

 

did you mean taht even if i have 2 RE in each cluster and one is fail, HA will broken and to reset the HA i have to insert RE in the defective ine slot??



SRX Services Gateway

Re: SRX5600 dual manment modules

a week ago

be infomred that i have 2 RE each node. and 1 control link not 2 control link .

 

Hardware inventory:
Item Version Part number Serial number Description
Chassis JN124E9FDAGB SRX5600
Midplane REV 01 710-024804 ACRD7486 SRX5600 Midplane
FPM Board REV 01 760-058098 CADR6906 Front Panel Display
PEM 0 Rev 03 740-034701 QCS150809011 PS 1.4-2.6kW; 90-264V AC in
PEM 1 Rev 03 740-034701 QCS15080907B PS 1.4-2.6kW; 90-264V AC in
Routing Engine 0 REV 02 740-056658 9009220586 SRX5k RE-1800X4
ad0 3998 MB Virtium - TuffDrive VCF P1T0200343171019 57 Compact Flash
ad1 114304 MB VSFA18PI128G-KC 34505-175 Disk 1
usb0 (addr 1) EHCI root hub 0 Intel uhub0
usb0 (addr 2) product 0x0020 32 vendor 0x8087 uhub1
DIMM 0 VL31B5463F-F8SD-KC DIE REV-0 PCB REV-0 MFR ID-ce80
DIMM 1 VL31B5463F-F8SD-KC DIE REV-0 PCB REV-0 MFR ID-ce80
DIMM 2 VL31B5463F-F8SD-KC DIE REV-0 PCB REV-0 MFR ID-ce80
DIMM 3 VL31B5463F-F8SD-KC DIE REV-0 PCB REV-0 MFR ID-ce80
CB 0 REV 01 750-056587 CAEA5698 SRX5k SCB II
FPC 0 REV 14 750-043157 CADW2770 SRX5k IOC II
CPU REV 05 711-043360 CADX1948 SRX5k MPC PMB
MIC 0 REV 07 750-049488 CADV7874 10x 10GE SFP+
PIC 0 BUILTIN BUILTIN 10x 10GE SFP+
Xcvr 0 REV 01 740-031980 CF08KN0N2 SFP+-10G-SR
Xcvr 1 REV 01 740-031980 CF08KN0MZ SFP+-10G-SR
Xcvr 2 REV 01 740-031980 CF08KN0NQ SFP+-10G-SR
Xcvr 3 REV 01 740-031980 CF08KN0NS SFP+-10G-SR
Xcvr 4 REV 01 740-031980 CF08KN0M6 SFP+-10G-SR
Xcvr 5 REV 01 740-031980 CF08KN0ME SFP+-10G-SR
MIC 1 REV 07 750-049488 CAEE5386 10x 10GE SFP+
PIC 2 BUILTIN BUILTIN 10x 10GE SFP+
FPC 4 REV 12 750-056758 CADW7187 SRX5k SPC II
CPU BUILTIN BUILTIN SRX5k DPC PPC
PIC 0 BUILTIN BUILTIN SPU Cp
PIC 1 BUILTIN BUILTIN SPU Flow
PIC 2 BUILTIN BUILTIN SPU Flow
PIC 3 BUILTIN BUILTIN SPU Flow
FPC 5 REV 12 750-056758 CADW7383 SRX5k SPC II
CPU BUILTIN BUILTIN SRX5k DPC PPC
PIC 0 BUILTIN BUILTIN SPU Flow
PIC 1 BUILTIN BUILTIN SPU Flow
PIC 2 BUILTIN BUILTIN SPU Flow
PIC 3 BUILTIN BUILTIN SPU Flow
Fan Tray Enhanced Fan Tray

 

this is hardware you can confirem that  1 controle link only and 2 RE

SRX Services Gateway

Re: SRX5600 dual manment modules

a week ago

Hello Amr,

 

The RE sits as a daughter card on the SCB (Switch control board). RE0 in CB0 and RE1 in CB1.

 

In the output you have shared I see neither CB1 nor RE1. Am I missing something here?

 

This could be useful: https://www.juniper.net/documentation/en_US/junos/topics/topic-map/security-chassis-cluster-dual-con...

 

Regards,

 

Vikas

SRX Services Gateway

Re: SRX5600 dual manment modules

a week ago

amr,

 

Please keep in mind that even if you have dual RE in one node, only one RE (slot 0)  will be working; the second RE is only for activating the second control-link. If currently you dont have a second control-link then that second RE is doing nothing (on each node).

 

As mrojas stated "No SRX has support for dual RE. I say this in the way that if the primary RE on a node fails the other RE (on the same node) wont take over. Installing a second RE in a node is only used for activating the posibility of having a redundant control-link"

 

Asnwering your question, if the primary RE (slot 0) fails in any of the nodes, the HA cluster will trigger a failover because the only RE that was operational on one of the nodes failed.