To replace a unit in a SPS cluster with a new appliance, complete the following steps.
Verify the HA status on the working node. Select Basic Settings > High Availability. If one of the nodes has broken down or is missing, the Status field displays DEGRADED.
Perform a full system backup. Before replacing the node, create a complete system backup of the working node. For details, see Data and configuration backups.
Check which firmware version is running on the working node. Select Basic Settings > System > Version details and write down the exact version numbers.
Log in to your support portal and download the CD ISO for the same SPS version that is running on your working node.
Without connecting the replacement unit to the network, install the replacement unit from the ISO file. Use the IPMI interface if needed.
When the installation is finished, connect the two SPS units with an Ethernet cable via the Ethernet connectors labeled as 4 or HA.
Reboot the replacement unit and wait until it finishes booting.
Login to the working node and verify the HA state. Select Basic Settings > High Availability. The Status field should display HALF.
Click Other node > Join HA.
Click Other node > Reboot.
The replacement unit will reboot and start synchronizing data from the working node. The Basic Settings > High Availability > Status field will display DEGRADED SYNC until the synchronization finishes. Depending on the size of the hard disks and the amount of data stored, this can take several hours.
After the synchronization is finished, connect the other Ethernet cables to their respective interfaces (external to 1 or EXT,
A node of the SPS cluster is replaced with a new appliance.
The IP addresses of the HA interfaces connecting the two nodes are detected automatically, during boot. When a node comes online, it attempts to connect to the IP address 18.104.22.168. If no other node responds until timeout, then it sets the IP address of its HA interface to 22.214.171.124, otherwise (if there is a responding node on 126.96.36.199) it sets its own HA interface to 188.8.131.52.
Replaced nodes do not yet know the HA configuration (or any other HA settings), and will attempt to negotiate it automatically in the same way. If the network is, for any reason, too slow to connect the nodes on time, the replacement node boots with the IP address of 184.108.40.206, which can cause an IP conflict if the other node has also set its IP to that same address previously. In this case, the replacement node cannot join the HA cluster.
To manually assign the correct IP address to the HA interface of a node, perform the following steps:
Log in to the node using the IPMI interface or the physical console.
Configuration changes have not been synced to the new (replacement) node, as it could not join the HA cluster. Use the default password of the root user of SPS, see "Installing the SPS hardware" in the Installation Guide.
From the console menu, choose 10 HA address.
Figure 290: The console menu
Choose the IP address of the node.
Figure 291: The console menu
Reboot the node.
This section explains the possible statuses of the SPS RAID device and the underlying hard disks. SPS displays this information on the Basic Settings > High Availability page. The following statuses can occur:
Optimal: The hard disks are working as expected.
Degraded: One or more hard disk has reported an error, and might have to be replaced. Contact the One Identity Support Team for help. For contact details, see About us.
Failed stripes: One or more stripes of data failed on the RAID device. It is possible that data loss occurred, but unfortunately there is no way to find out the extent of the data loss (if any).
If you have a single SPS node: You must reinstall SPS and restore the data from the latest backup. For details, see "One Identity Safeguard for Privileged Sessions Software Installation Guide" in the Installation Guide and Restoring SPS configuration and data. If you do not have backup, contact the One Identity Support Team for help. For contact details, see About us.
If you have a high-availability SPS cluster: Shut the node down. Do NOT disconnect its HA interface. Reinstall the node (for details, see "One Identity Safeguard for Privileged Sessions Software Installation Guide" in the Installation Guide), power it on, then navigate to Basic Settings > High Availability, and click Join HA. In case you need assistance, contact the One Identity Support Team for help. For contact details, see About us.
Offline: The RAID device is not functioning, probably because several disks have broken down. SPS cannot operate properly in this case. Contact the One Identity Support Team for help. For contact details, see About us.
The following procedure describes how to restore the configuration and data of SPS from a complete backup, for example, after a hardware replacement.
Do not enable audited traffic to SPS restoring the system backup is complete.
During the restore process, the REST-based search might not function properly, since the data to search in might still be incomplete.
To make sure that the import process has finished, check the logs.
Navigate to Basic Settings > Troubleshooting > View log files. Select syslog as Logtype, the day of the upgrade process as Day and enter Run metadb_importer.py in the Show only messages containing field. Click View.
If the import process has finished, the following line is displayed:
systemd: Started Run metadb_importer.py to import data from metadb to elasticsearch if necessary...
Connect to your backup server and locate the directory where SPS saves the backups. The configuration backups are stored in the config subdirectory in timestamped files. Find the latest configuration file (the configuration files are called PSM-timestamp.config).
Connect to SPS.
If you have not yet completed the Welcome Wizard, click Browse, select the configuration file, and click Import.
If you have already completed the Welcome Wizard, navigate to Basic Settings > System > Import configuration > Browse, select the configuration file, and click Import.
Navigate to Policies > Backup & Archive/Cleanup. Verify that the settings of the target servers and the backup protocols are correct.
Navigate to Basic Settings > Management > System backup, click Restore now and wait for the process to finish. Depending on the amount of data stored in the backup, and the speed of the connection to the backup server, this may take a long time.
Navigate to SSH Control > Connections, and click Restore ALL. Repeat this step for other traffic types. Depending on the amount of data stored in the backup, and the speed of the connection to the backup server, this may take a long time.