During MCC-IP initial deployment, node is unable to boot up to Cluster Setup
Applies to
- 4-node MCC-IP deployment
- ONTAP 9
- Non-configured system
- AFF A30
Issue
- During initial deployment of 4-node MCC-IP, one of the nodes is not able to boot up correctly after running options 9a and 9b at the Boot Menu
- Options 9a and 9b complete successfully on both affected node and HA partner
- When the node is finishing boot up, the following messages are observed:
wrote key file "/tmp/rndc.key"Unable to determine local cluster UUID ('bootarg.mgwd.cluster_uuid')Unable to determine local cluster UUID ('bootarg.vm.mgwd.cluster_uuid')Unable to generate /var/etc/iscsi.conf, status: 1Unable to determine local cluster UUID ('bootarg.mgwd.cluster_uuid')Unable to determine local cluster UUID ('bootarg.vm.mgwd.cluster_uuid')Unable to generate /var/etc/ctl.conf, status: 1Apr 07 09:18:57 [localhost:spm.vldb.process.exit:EMERGENCY]: Volume Location Database(VLDB) subsystem with ID 8821 exited as a result of signal signal 6. The subsystem will attempt to restart.Apr 07 09:18:57 [localhost:spm.vifmgr.process.exit:EMERGENCY]: Logical Interface Manager(VifMgr) with ID 8713 aborted as a result of signal signal 6. The subsystem will attempt to restart.Apr 07 09:18:59 [localhost:spm.mgwd.process.exit:EMERGENCY]: Management Gateway (mgwd) subsystem with ID 7084 exited as a result of signal normal exit (0). The subsystem will attempt to restart.Apr 07 09:23:55 [localhost:spm.mgwd.process.exit:EMERGENCY]: Management Gateway (mgwd) subsystem with ID 31724 exited as a result of signal signal 6. The subsystem will attempt to restart.Unable to connect to the management gateway server... The management gateway server restarted. Waiting to see if the connection can be reestablished... The management gateway server restarted. Waiting to see if the connection can be reestablished...The connection with the management gateway server has been reestablished. If the root cause of the interruption was a process core, you can see the core file details by issuing the following command:system node coredump show -node local -type application -corename mgwd.* -instance *********************** ** SYSTEM MESSAGES ** *********************** Automatic replicated database recovery is in progress. This node is not fully operational. Use the command "event log show -event rdb.recovery.* -severity *" to monitor RDB recovery. On successful RDB auto recovery completion, the event 'rdb.recovery.succeeded' will be generated. Should RDB auto recovery fail, the event 'rdb.recovery.failed' will be generated. Failure will also generate a 'callhome.root.vol.recovery.reqd' event. The management gateway server restarted. Waiting to see if the connection can be reestablished...
The connection with the management gateway server could not be reestablished.The most recent coredump file is:
/mroot/etc/crash/mgwd.33238.539041256.2025-04-07.09_25_17.ucore.bz2
For further assistance, contact technical support.Do you want to login to the systemshell as the 'diag' user? {yes, no}:- It is identified that the printenv output for the affected node shows bootarg.rdb_corrupt argument with a defined value:
LOADER-A> printenvVariable Name Value
-------------------- --------------------------------------------------
bootarg.rdb_corrupt RRR00RR000- Node personality is correctly set as an AFF A30
- HA Configuration is set as MCC-IP
- Partner node and DR Cluster are able to successfully complete the boot process and reach Cluster Setup
