This document guides an administrator through the patch update to Cray Systems Management v1.4.3
from an earlier patch version of CSM v1.4
. If upgrading from CSM v1.3.x
, then follow the procedures
described in CSM major/minor version upgrade instead.
In the unusual situation of upgrading from a pre-release version of CSM v1.4.0
, then follow the procedures
described in CSM major/minor version upgrade instead.
Also note that there is no need to perform intermediate CSM v1.4
patch upgrades. Instead,
consider upgrading to the latest CSM v1.4
patch release. See
CSM patch version upgrade for the full list of patch versions.
CVE-2023-20593 Zenbleed
on NCNs
cray-sat
containerSAT
productCSM
patch install procedure to update CFS configurationsgoss
package versiongoss-servers
after updategoss
tests to use goss
hostname variable rather than HOSTNAME
environment variableIMS
importer to restore missing S3
artifacts for existing images or recipesRPMs
are installed on the correct nodesBOS V1
references to BOS V2
in documentationHMS
query handling to prevent errors when queried with an empty node listwait-for-unbound.sh
postgresql
restore step from spire
recovery procedureNCN
reboot procedure to recommend it be used to pick up CVE
patched container imagesNCN
reboot procedure in patch upgrade instructionspostgres pooler
CANU
returning 0
exit code despite errorsSLS
document with procedure to retrieve an authentication tokenHPE
nodes with four HSN NICs
FRU
tracking not creating a detected event after a removed eventcheck_bgp_neighbors_established.sh
scriptbootprep
directories provided to IUF
are the samegoss
test in case where admin keyring
is not on ncn-s004
and aboveceph
for enabling smartmon
on storage nodesSMART
data collection on UANs
that support ithmcollector
kafka
topics to handle new telemetry datahelm
example in documentation for reset of console servicesBOS
to prevent live lock scenariografana
dashboardCVE-2023-38545
for curl
and libcurl
on NCNs
OPA
to cache spire-jwks
requests going through Kuberenetes
network to avoid port exhaustionIMS
backup and restore to preserve IMS image IDs
smartmon
metrics on storage NCNsValidate CSM health.
See Validate CSM Health.
Run the CSM health checks to ensure that everything is working properly before the upgrade starts. After the upgrade is completed, another health check is performed. It is important to know if any problems observed at that time existed prior to the upgrade.
IMPORTANT: See the CSM Install Validation and Health Checks procedures in the documentation for the CURRENT CSM version on the system. The validation procedures in the CSM documentation are only intended to work with that specific version of CSM.
(ncn-m001#
) Start a typescript on ncn-m001
to capture the commands and output from this procedure.
script -af csm-update.$(date +%Y-%m-%d).txt
export PS1='\u@\H \D{%Y-%m-%d} \t \w # '
Download and extract the CSM v1.4.3
release to ncn-m001
.
(ncn-m001#
) Set CSM_DISTDIR
to the directory of the extracted files.
IMPORTANT: If necessary, change this command to match the actual location of the extracted files.
export CSM_DISTDIR="$(pwd)/csm-1.4.3"
echo "${CSM_DISTDIR}"
(ncn-m001#
) Set CSM_RELEASE_VERSION
to the CSM release version.
export CSM_RELEASE_VERSION="$(${CSM_DISTDIR}/lib/version.sh --version)"
echo "${CSM_RELEASE_VERSION}"
Download and install/upgrade the latest documentation on ncn-m001
.
(ncn-m001#
) Run lib/setup-nexus.sh
to configure Nexus and upload new CSM RPM repositories, container images, and
Helm charts:
cd "$CSM_DISTDIR"
./lib/setup-nexus.sh ; echo "RC=$?"
On success, the output should end with the following:
+ Nexus setup complete
setup-nexus.sh: OK
RC=0
In the event of an error,
consult Troubleshoot Nexus
to resolve potential problems and then try running setup-nexus.sh
again. Note that subsequent runs of setup-nexus.sh
may report FAIL
when uploading duplicate assets. This is okay as long as setup-nexus.sh
outputs setup-nexus.sh: OK
and exits with status code 0
.
(ncn-m001#
) Run the following script in preparation for 1.4.3 patch upgrade:
for c in $(kubectl get crd |grep argo | cut -d' ' -f1)
do
kubectl label --overwrite crd $c app.kubernetes.io/managed-by="Helm"
kubectl annotate --overwrite crd $c meta.helm.sh/release-name="cray-nls"
kubectl annotate --overwrite crd $c meta.helm.sh/release-namespace="argo"
done
(ncn-m001#
) Run upgrade.sh
to deploy upgraded CSM applications and services:
cd "$CSM_DISTDIR"
./upgrade.sh
It is important to upload NCN images to IMS and to edit the cray-product-catalog
. This is necessary when updating
products with IUF. If this step is skipped, IUF will fail when updating or upgrading products in the future.
(ncn-m001#
) Execute script to upload CSM NCN images and update the cray-product-catalog
.
/usr/share/doc/csm/upgrade/scripts/upgrade/upload-ncn-images.sh
Note: This step may not be necessary if it was already completed by the CSM v1.3.5
patch or v1.4.2
patch. If it was already run,
the following steps can be re-executed to verify that Ceph daemons are using images in Nexus and the local Docker
registries have been stopped.
These steps will upgrade Ceph to v16.2.13
. Then the Ceph monitoring daemons’ images will be pushed to Nexus and the
monitoring daemons will be redeployed so that they use these images in Nexus. Once this is complete, all Ceph daemons
should be using images in Nexus and not images hosted in the local Docker registry on storage nodes. The third step
stops the local Docker registry on all storage nodes.
(ncn-m001#
) Run Ceph upgrade to v16.2.13
.
/usr/share/doc/csm/upgrade/scripts/ceph/ceph-upgrade-tool.py --version "v16.2.13"
(ncn-m001#
) Redeploy Ceph monitoring daemons so they are using images in Nexus.
scp /usr/share/doc/csm/scripts/operations/ceph/redeploy_monitoring_stack_to_nexus.sh ncn-s001:/srv/cray/scripts/common/redeploy_monitoring_stack_to_nexus.sh
ssh ncn-s001 "/srv/cray/scripts/common/redeploy_monitoring_stack_to_nexus.sh"
(ncn-m001#
) Stop the local Docker registries on all storage nodes.
scp /usr/share/doc/csm/scripts/operations/ceph/disable_local_registry.sh ncn-s001:/srv/cray/scripts/common/disable_local_registry.sh
ssh ncn-s001 "/srv/cray/scripts/common/disable_local_registry.sh"
smartmon
metrics on storage NCNsThis step will install the smart-mon
rpm on storage nodes, and reconfigure the node-exporter
to provide smartmon
metrics.
(ncn-m001#
) Execute the following script.
/usr/share/doc/csm/scripts/operations/ceph/enable-smart-mon-storage-nodes.sh
This step updates the CFS configuration which is set as the desired configuration for the management nodes (NCNs). It ensures that the CFS configuration layers reference the correct commit hash for the version of CSM being installed. It then waits for the components to reach a configured state in CFS.
(ncn-m001#
)
cd "$CSM_DISTDIR"
./update-mgmt-ncn-cfs-config.sh --base-query role=management \
--save --create-backups --clear-error
The output will look similar to the truncated output shown below.
INFO: Querying CFS configurations for the following NCNs: x3000c0s5b0n0, ...
INFO: Found configuration "management-csm-1.4.0" for component x3000c0s5b0n0
...
INFO: Updating existing layer with repo path /vcs/cray/csm-config-management.git and playbook site.yml
INFO: Property "commit" of layer with repo path /vcs/cray/csm-config-management.git and playbook site.yml updated ...
INFO: Property "name" of layer with repo path /vcs/cray/csm-config-management.git and playbook site.yml updated ...
INFO: No layer with repo path /vcs/cray/csm-config-management.git and playbook ncn-initrd.yml found.
INFO: Adding a layer with repo path /vcs/cray/csm-config-management.git and playbook ncn-initrd.yml to the end.
INFO: Successfully saved CFS configuration "management-csm-1.4.0-backup-20230918T205149"
INFO: Successfully saved CFS configuration "management-csm-1.4.0"
INFO: Successfully saved 1 changed CFS configuration(s) to CFS.
INFO: Updated 9 CFS components.
INFO: Waiting for 9 component(s) to finish configuration
INFO: Summary of number of components in each status: pending: 9
INFO: Waiting for 9 pending component(s)
INFO: Sleeping for 30 seconds before checking status of 9 pending component(s).
...
INFO: Sleeping for 30 seconds before checking status of 9 pending component(s).
INFO: 9 pending components transitioned to status configured: x3000c0s5b0n0, ...
INFO: Finished waiting for 9 component(s) to finish configuration.
INFO: Summary of number of components in each status: configured: 9
====> Completed update of CFS configuration(s)
====> Cleaning up install dependencies
When configuration of all components is successful, the summary line will show all components with status “configured”.
(ncn-m001#
) Update select RPMs on the NCNs.
/usr/share/doc/csm/upgrade/scripts/upgrade/util/upgrade-test-rpms.sh
Verify that the new CSM version is in the product catalog.
(ncn-m001#
) Verify that the new CSM version is listed in the output of the following command:
kubectl get cm cray-product-catalog -n services -o jsonpath='{.data.csm}' | yq r -j - | jq -r 'to_entries[] | .key' | sort -V
Example output that includes the new CSM version (1.4.3
):
0.9.2
0.9.3
0.9.4
0.9.5
0.9.6
1.0.1
1.0.10
1.2.0
1.2.1
1.2.2
1.3.0
1.3.1
1.4.0
1.4.1
1.4.2
1.4.3
Confirm that the product catalog has an accurate timestamp for the CSM upgrade.
(ncn-m001#
) Confirm that the import_date
reflects the timestamp of the upgrade.
kubectl get cm cray-product-catalog -n services -o jsonpath='{.data.csm}' | yq r - '"1.4.3".configuration.import_date'
(ncn-m001#
) Execute the following script to take a manual backup of the Etcd clusters.
/usr/share/doc/csm/scripts/operations/etcd/take-etcd-manual-backups.sh post_patch
These clusters are automatically backed up every 24 hours, but taking a manual backup at this stage in the upgrade enables restoring from backup later in this process if needed.
This is an optional step but is strongly recommended. As each patch release includes updated container images that may contain CVE fixes, it is recommended to reboot each NCN to refresh cached container images. For detailed instructions on how to gracefully reboot each NCN, refer to Reboot NCNs.
(ncn-m001#
) Remember to exit the typescript that was started at the beginning of the upgrade.
exit
It is recommended to save the typescript file for later reference.