Support Support Downloads Knowledge Base Service Request Manager My Juniper Community

Knowledge Base

Search our Knowledge Base sites to find answers to your questions.

Ask All Knowledge Base Sites All Knowledge Base Sites JunosE Defect (KA)Knowledge BaseSecurity AdvisoriesTechnical BulletinsTechnotes Sign in to display secure content and recently viewed articles

Autorecovery of fabric link is supported in SRX

0

0

Article ID: KB30928 KB Last Updated: 04 Mar 2017Version: 2.0
Summary:

This article describes the autorecovery function of fabric link, which is supported from Junos 12.1X46-D20. There are also differences of Chassis Cluster RG(s) action in Junos 12.1X44-D45 and 12.X46-D20 after the fabric link status changes.

This feature is supported in the SRX-Branch and SRX-HE devices.

Symptoms:

  • Before Junos 12.1X46-D20, If the fabric link goes down, all the RGs status will change to 'disable' on the secondary node. The secondary node will not restore to normal status until after a reboot.
  • Begining with Junos 12.1X46-D20, if the fabric link goes down, RG1+ becomes ineligible on either the secondary node or the node with failures, by default.  The node remains in this state until the fabric link comes up or the other node goes away. And once the link is up, it can return to normal without a reboot. However, all the FPC cards will auto soft-restart.
  • If the fabric link goes down followed by the control link, then after approximately 66 seconds, the secondary node (or the node with failures) assumes that the remote node is dead and takes over as the primary node.

Solution:

Junos 12.1X44

{primary:node0}[edit]
root# run show chassis cluster status
Cluster ID: 1
Node Priority Status Preempt Manual failover

Redundancy group: 0 , Failover count: 1
node0 200 primary no no
node1 100 secondary no no

Redundancy group: 1 , Failover count: 1
node0 200 primary no no
node1 100 secondary no no


After disconnect the fabric link:
{primary:node0}[edit]
root# run show chassis cluster status
Cluster ID: 1
Node Priority Status Preempt Manual failover

Redundancy group: 0 , Failover count: 1
node0 200 primary no no
node1 100 disabled no no <---

Redundancy group: 1 , Failover count: 1
node0 200 primary no no
node1 100 disabled no no
<---

After re-connect the fabric link, and do one reboot on the secondary node:
{primary:node0}[edit]
root# run show chassis cluster status
Cluster ID: 1
Node Priority Status Preempt Manual failover

Redundancy group: 0 , Failover count: 1
node0 200 primary no no
node1 100 secondary no no

Redundancy group: 1 , Failover count: 1
node0 200 primary no no
node1 100 secondary no no

Junos 12.1X46-D20

root# run show chassis cluster status
Monitor Failure codes:
CS Cold Sync monitoring FL Fabric Connection monitoring
GR GRES monitoring HW Hardware monitoring
IF Interface monitoring IP IP monitoring
LB Loopback monitoring MB Mbuf monitoring
NH Nexthop monitoring NP NPC monitoring
SP SPU monitoring SM Schedule monitoring

Cluster ID: 1
Node Priority Status Preempt Manual Monitor-failures

Redundancy group: 0 , Failover count: 1
node0 200 primary no no None
node1 100 secondary no no None

Redundancy group: 1 , Failover count: 1
node0 200 primary no no None
node1 100 secondary no no None


After disconnect the fabric link:
root# run show chassis cluster status
Monitor Failure codes:
CS Cold Sync monitoring FL Fabric Connection monitoring
GR GRES monitoring HW Hardware monitoring
IF Interface monitoring IP IP monitoring
LB Loopback monitoring MB Mbuf monitoring
NH Nexthop monitoring NP NPC monitoring
SP SPU monitoring SM Schedule monitoring

Cluster ID: 1
Node Priority Status Preempt Manual Monitor-failures

Redundancy group: 0 , Failover count: 1
node0 200 primary no no None
node1 0 secondary no no FL <---

Redundancy group: 1 , Failover count: 1
node0 200 primary no no None
node1 0 ineligible no no FL
<---

After re-connect the fabric link:
{primary:node0}[edit]
root# run show chassis fpc detail
node0:
---------------------------------------
Slot 0 information:
State Online
Total CPU DRAM ---- CPU less FPC ----
Start time 2016-05-16 05:58:27 UTC
Uptime 11 minutes, 23 seconds

node1:
---------------------------------------
Slot 0 information:
State Present <---
Total CPU DRAM ---- CPU less FPC ----

root# run show chassis fpc detail
node0:
---------------------------------------
Slot 0 information:
State Online
Total CPU DRAM ---- CPU less FPC ----
Start time 2016-05-16 05:58:27 UTC
Uptime 11 minutes, 38 seconds

node1:
---------------------------------------
Slot 0 information:
State Online <---
Total CPU DRAM ---- CPU less FPC ----
Start time 2016-05-16 06:09:37 UTC
Uptime 1 second

root# run show chassis cluster status
Monitor Failure codes:
CS Cold Sync monitoring FL Fabric Connection monitoring
GR GRES monitoring HW Hardware monitoring
IF Interface monitoring IP IP monitoring
LB Loopback monitoring MB Mbuf monitoring
NH Nexthop monitoring NP NPC monitoring
SP SPU monitoring SM Schedule monitoring

Cluster ID: 1
Node Priority Status Preempt Manual Monitor-failures

Redundancy group: 0 , Failover count: 1
node0 200 primary no no None
node1 100 secondary no no None

Redundancy group: 1 , Failover count: 1
node0 200 primary no no None
node1 100 secondary no no None

Related Links

Comment on this article > Affected Products Browse the Knowledge Base for more articles related to these product categories. Select a category to begin.

Security Alerts and Vulnerabilities

Security Alerts and Vulnerabilities Product Alerts and Software Release Notices Problem Report (PR) Search Tool EOL Notices and Bulletins JTAC User Guide Customer Care User Guide Pathfinder SRX High Availability Configurator SRX VPN Configurator Training Courses and Videos End User Licence Agreement Global Search