Pre-Installation

The page walks a user through setting up the Cray LiveCD with the intention of installing Cray System Management (CSM).

  1. Boot installation environment
    1. Prepare installation environment server
    2. Boot the LiveCD
    3. First log in
    4. Prepare the data partition
    5. Set reusable environment variables
    6. Exit the console and log in with SSH
  2. Download and extract the CSM tarball
  3. Create system configuration
  4. Import the CSM Tarball
  5. Validate the LiveCD
  6. Next topic

1. Boot installation environment

This section walks the user through booting and connecting to the LiveCD.

Before proceeding, the user must obtain the CSM tarball containing the LiveCD.

NOTE: Each step denotes where its commands must run; external# refers to a server that is not the Cray, whereas pit# refers to the LiveCD itself.

Any steps run on an external server require that server to have the following tools:

  • ipmitool
  • ssh
  • tar

NOTE: The CSM tarball will be fetched from the external server in the download and extract the CSM tarball step using curl or scp. If a web server is not installed, then scp is the backup option.

1.1 Prepare installation environment server

  1. (external#) Download the CSM software release from the public Artifactory instance.

    NOTES:

    • -C - is used to allow partial downloads. These tarballs are large; in the event of a connection disruption, the same curl command can be used to continue the disrupted download.
    • If air-gapped or behind a strict firewall, then the tarball must be obtained from the medium delivered by Cray-HPE. For these cases, copy or download the tarball to the working directory and then proceed to the next step. The tarball will need to be fetched with scp during the download and extract the CSM tarball step.
    1. (external#) Set the CSM RELEASE version

      Example release versions:

      • An alpha build: CSM_RELEASE=1.5.0-alpha.99
      • A release candidate: CSM_RELEASE=1.5.0-rc.1
      • A stable release: CSM_RELEASE=1.5.0
      CSM_RELEASE=<value>
      
    2. (external#) Download the CSM tarball

      NOTE: CSM does NOT support the use of proxy servers for anything other than downloading artifacts from external endpoints. Using http_proxy or https_proxy in any way other than the following examples will cause many failures in subsequent steps.

      • Without proxy:

        curl -C - -f -O "https://release.algol60.net/$(awk -F. '{print "csm-"$1"."$2}' <<< ${CSM_RELEASE})/csm/csm-${CSM_RELEASE}.tar.gz"
        
      • With https proxy:

        https_proxy=https://example.proxy.net:443 curl -C - -f -O "https://release.algol60.net/$(awk -F. '{print "csm-"$1"."$2}' <<< ${CSM_RELEASE})/csm/csm-${CSM_RELEASE}.tar.gz"
        
  2. (external#) Extract the LiveCD from the tarball.

    OUT_DIR="$(pwd)/csm-temp"
    mkdir -pv "${OUT_DIR}"
    tar -C "${OUT_DIR}" --wildcards --no-anchored --transform='s/.*\///' -xzvf "csm-${CSM_RELEASE}.tar.gz" 'pre-install-toolkit-*.iso'
    

1.2 Boot the LiveCD

  1. (external#) Start a typescript and set the PS1 variable to record timestamps.

    NOTE: Typescripts help triage if problems are encountered.

    script -a "boot.livecd.$(date +%Y-%m-%d).txt"
    export PS1='\u@\H \D{%Y-%m-%d} \t \w # '
    
  2. (external#) Follow one of the procedures below based on the vendor for the ncn-m001 node:

    • HPE iLO BMCs

      Prepare a server on the network to host the pre-install-toolkit ISO file, if the current server is insufficient. Then follow the HPE iLO BMCs to boot the RemoteISO before returning here.

    • Gigabyte BMCs and Intel BMCs

      Create a USB stick using the following procedure.

      1. (external#) Get cray-site-init from the tarball.

        OUT_DIR="$(pwd)/csm-temp"
        mkdir -pv "${OUT_DIR}"
        tar -C "${OUT_DIR}" --wildcards --no-anchored --transform='s/.*\///' -xzvf "csm-${CSM_RELEASE}.tar.gz" 'cray-site-init-*.rpm'
        
      2. (external#) Install the write-livecd.sh script:

        • RPM-based systems:

          rpm -Uvh --force ${OUT_DIR}/cray-site-init*.rpm
          
        • Non-RPM-based systems (requires bsdtar):

          bsdtar xvf "${OUT_DIR}"/cray-site-init-*.rpm --include *write-livecd.sh -C "${OUT_DIR}"
          mv -v "${OUT_DIR}"/usr/local/bin/write-livecd.sh "./${OUT_DIR}"
          rmdir -pv "${OUT_DIR}/usr/local/bin/"
          
        • Non-RPM-based distros (requires rpm2cpio):

          rpm2cpio cray-site-init-*.rpm | cpio -idmv
          mv -v ./usr/local/bin/write-livecd.sh "./${OUT_DIR}"
          rm -vrf ./usr
          
      3. Follow Bootstrap a LiveCD USB and then return here.

1.3 First log in

On first login, the LiveCD will prompt the administrator to change the password.

  1. (pit#) Log in.

    NOTE: The initial password is empty.

    At the login prompt, enter root as the username. Because the initial password is blank, press return twice at the first two password prompts. The LiveCD will force a new password to be set.

    Password:           <-------just press Enter here for a blank password
    You are required to change your password immediately (administrator enforced)
    Changing password for root.
    Current password:   <------- press Enter here, again, for a blank password
    New password:       <------- type new password
    Retype new password:<------- retype new password
    Welcome to the CRAY Pre-Install Toolkit (LiveOS)
    
  2. (pit#) Configure the site-link (lan0), DNS, and gateway IP addresses.

    NOTE: The site_ip, site_gw, and site_dns values must come from the local network administration or authority.

    1. Set site_ip variable.

      Set the site_ip value in CIDR format (A.B.C.D/N):

      site_ip=<IP CIDR>
      
    2. Set the site_gw and site_dns variables.

      Set the site_gw and site_dns values in IPv4 dotted decimal format (A.B.C.D):

      site_gw=<Gateway IP address>
      site_dns=<DNS IP address>
      
    3. Set the site_nics variable.

      The site_nics value or values are found while the user is in the LiveCD (for example, site_nics='p2p1 p2p2 p2p3' or site_nics=em1).

      site_nics='<site NIC or NICs>'
      
    4. Set the SYSTEM_NAME variable.

      SYSTEM_NAME is the name of the system. This will only be used for the PIT hostname. This variable is capitalized because it will be used in a subsequent section.

      SYSTEM_NAME=<system name>
      
    5. Run the csi-setup-lan0.sh script to set up the site link and set the hostname.

      NOTES:

      • All of the /root/bin/csi-* scripts can be run without parameters to display usage statements.
      • The hostname is auto-resolved based on reverse DNS.
      /root/bin/csi-setup-lan0.sh "${SYSTEM_NAME}" "${site_ip}" "${site_gw}" "${site_dns}" "${site_nics}"
      
  3. (pit#) Verify that the assigned IP address was successfully applied to lan0 .

    wicked ifstatus --verbose lan0
    

    NOTE: The output from the above command must say leases: ipv4 static granted. If the IPv4 address was not granted, then go back and recheck the variable values. The output will indicate the IP address failed to assign, which can happen if the given IP address is already taken on the connected network.

1.4 Prepare the data partition

  1. (pit#) Mount the PITDATA partition.

    Use either the RemoteISO or the USB option below, depending how the LiveCD was connected in the Boot the LiveCD step.

    • RemoteISO

      Use a local disk for PITDATA:

      disk="$(lsblk -l -o SIZE,NAME,TYPE,TRAN -e7 -e11 -d -n | grep -v usb | sort -h | awk '{print $2}' | xargs -I {} bash -c "if ! grep -Fq {} /proc/mdstat; then echo {}; fi" | head -n 1)"
      echo "Using ${disk}"
      parted --wipesignatures -m --align=opt --ignore-busy -s "/dev/${disk}" -- mklabel gpt mkpart primary ext4 2048s 100%
      partprobe "/dev/${disk}"
      mkfs.ext4 -L PITDATA "/dev/${disk}1"
      mount -vL PITDATA
      
    • USB

      Mount the USB data partition:

      mount -vL PITDATA
      

1.5 Set reusable environment variables

These variables will need to be set for many procedures within the CSM installation process.

NOTE: This sets some variables that were already set. These should be set again anyway.

  1. (pit#) Set the variables.

    1. Set the PITDATA variable.

      export PITDATA="$(lsblk -o MOUNTPOINT -nr /dev/disk/by-label/PITDATA)"
      
    2. Set the CSM_RELEASE variable.

      The value is based on the version of the CSM release being installed.

      Example release versions:

      • An alpha build: CSM_RELEASE=1.5.0-alpha.99
      • A release candidate: CSM_RELEASE=1.5.0-rc.1
      • A stable release: CSM_RELEASE=1.5.0
      export CSM_RELEASE=<value>
      
    3. Set the CSM_PATH variable.

      After the CSM release tarball has been expanded, this will be the path to its base directory.

      export CSM_PATH="${PITDATA}/csm-${CSM_RELEASE}"
      
    4. Set the SYSTEM_NAME variable.

      This is the user friendly name for the system. For example, for eniac-ncn-m001, SYSTEM_NAME should be set to eniac.

      export SYSTEM_NAME=<value>
      
  2. (pit#) Update /etc/environment.

    This ensures that these variables will be set in all future shells on the PIT node.

    cat << EOF >/etc/environment
    CSM_RELEASE=${CSM_RELEASE}
    CSM_PATH=${PITDATA}/csm-${CSM_RELEASE}
    GOSS_BASE=${GOSS_BASE}
    PITDATA=${PITDATA}
    SYSTEM_NAME=${SYSTEM_NAME}
    EOF
    

1.6 Exit the console and log in with SSH

  1. (pit#) Create the admin directory for the typescripts and administrative scratch work.

    mkdir -pv "$(lsblk -o MOUNTPOINT -nr /dev/disk/by-label/PITDATA)/prep/admin"
    ls -l "$(lsblk -o MOUNTPOINT -nr /dev/disk/by-label/PITDATA)/prep/admin"
    
  2. (pit#) Exit the typescript and log out.

    exit
    exit
    
  3. (pit#) Exit the console.

    This is done by typing the key-sequence: tilde, period. That is, ~.

    If the console was accessed over an SSH session (that is, the user used SSH to log into another server, and from there used ipmitool to access the console), then press tilde twice followed by a period, in order to prevent exiting the parent SSH session. That is, ~~.

  4. (external#) Copy the typescript to the running LiveCD.

    scp boot.livecd.*.txt root@eniac-ncn-m001:/tmp/
    
  5. (pit#) SSH into the LiveCD.

    livecd=eniac-ncn-m001.example.company.com
    ssh root@"${livecd}"
    
  6. (pit#) Copy the previous typescript and start a new one.

    cp -pv /tmp/boot.livecd.*.txt "${PITDATA}/prep/admin"
    script -af "${PITDATA}/prep/admin/csm-install.$(date +%Y-%m-%d).txt"
    export PS1='\u@\H \D{%Y-%m-%d} \t \w # '
    
  7. (pit#) Print information about the booted PIT image for logging purposes.

    Having this information in the typescript can be helpful if problems are encountered during the install.

    /root/bin/metalid.sh
    

    Expected output looks similar to the following (the versions in the example below may differ). There should be no errors.

     = PIT Identification = COPY/CUT START =======================================
     VERSION=ed97205-1706718622724
     TIMESTAMP=2024-01-31_16:30:22
     CRAY-Site-Init build signature...
     Build Commit   : c9a07e366151a72be71d168061d08bd97da5344c-heads-v1.32.4
     Build Time     : 2023-10-20T14:53:30Z
     Go Version     : go1.19
     Version        : v1.32.4
     Platform       : linux/amd64
     canu-1.8.0-1.x86_64
     ilorest-4.2.0.0-20.x86_64
     metal-basecamp-1.2.6-1.x86_64
     metal-ipxe-2.4.7-1.noarch
     metal-init-1.4.6-1.noarch
     metal-nexus-1.3.1-3.38.0_1.x86_64
     metal-observability-1.0.9-1.x86_64
     = PIT Identification = COPY/CUT END =========================================
    

2. Download and extract the CSM Tarball

  1. Download and install the latest documentation and scripts RPMs, see Check for latest documentation.

  2. (pit#) Download the CSM tarball.

    • From Cray using curl:

      • -C - is used to allow partial downloads. These tarballs are large; in the event of a connection disruption, the same curl command can be used to continue the disrupted download.
      • CSM does NOT support the use of proxy servers for anything other than downloading artifacts from external endpoints. Using http_proxy or https_proxy in any way other than the following examples will cause many failures in subsequent steps.

      Without proxy:

      curl -C - -f -o "/var/www/ephemeral/csm-${CSM_RELEASE}.tar.gz" \
        "https://release.algol60.net/$(awk -F. '{print "csm-"$1"."$2}' <<< ${CSM_RELEASE})/csm/csm-${CSM_RELEASE}.tar.gz"
      

      With HTTPS proxy:

      https_proxy=https://example.proxy.net:443 curl -C - -f -o "/var/www/ephemeral/csm-${CSM_RELEASE}.tar.gz" \
        "https://release.algol60.net/$(awk -F. '{print "csm-"$1"."$2}' <<< ${CSM_RELEASE})/csm/csm-${CSM_RELEASE}.tar.gz"
      
    • scp from the external server used in Prepare installation environment server:

      scp "<external-server>:/<path>/csm-${CSM_RELEASE}.tar.gz" /var/www/ephemeral/
      
  3. (pit#) Extract the tarball.

    tar -zxvf  "${PITDATA}/csm-${CSM_RELEASE}.tar.gz" -C ${PITDATA}
    
  4. (pit#) HOTFIX: 1.5.0 through 1.5.2 At this time, a newer metal-ipxe RPM is required. Please review the current field notices for the hotfix(es). The hotfix will insert a new metal-ipxe RPM into the extracted tarball.

    NOTE Anything newer than 1.5.2 may continue without needing any additional step.

  5. (pit#) Install/update the RPMs necessary for the CSM installation.

    NOTE --no-gpg-checks is used because the repository contained within the tarball does not provide a GPG key.

    1. Update cray-site-init and metal-init.

      NOTES

      • cray-site-init provides csi, a tool for creating and managing configurations, as well as orchestrating the handoff and deploy of the final non-compute node.
      • metal-init provides several scripts in /root/bin used for fresh installations.
      • metal-ipxe provides boot parameters for the NCNs, as well as EFI binaries for PXE/iPXE/HTTP booting.
      zypper --plus-repo "${CSM_PATH}/rpm/cray/csm/noos" --no-gpg-checks update -y cray-site-init metal-init metal-ipxe
      
  6. (pit#) Get the artifact versions.

    KUBERNETES_VERSION="$(find ${CSM_PATH}/images/kubernetes -name '*.squashfs' -exec basename {} .squashfs \; | awk -F '-' '{print $(NF-1)}')"
    echo "${KUBERNETES_VERSION}"
    CEPH_VERSION="$(find ${CSM_PATH}/images/storage-ceph -name '*.squashfs' -exec basename {} .squashfs \; | awk -F '-' '{print $(NF-1)}')"
    echo "${CEPH_VERSION}"
    
  7. (pit#) Copy the NCN images from the expanded tarball.

    NOTE This hard-links the files to do this copy as fast as possible, as well as to mitigate space waste on the USB stick.

    mkdir -pv "${PITDATA}/data/k8s/" "${PITDATA}/data/ceph/"
    rsync -rltDP --delete "${CSM_PATH}/images/kubernetes/" --link-dest="${CSM_PATH}/images/kubernetes/" "${PITDATA}/data/k8s/${KUBERNETES_VERSION}"
    rsync -rltDP --delete "${CSM_PATH}/images/storage-ceph/" --link-dest="${CSM_PATH}/images/storage-ceph/" "${PITDATA}/data/ceph/${CEPH_VERSION}"
    
  8. (pit#) Modify the NCN images with SSH keys and root passwords.

    The following substeps provide the most commonly used defaults for this process. For more advanced options, see Set NCN Image Root Password, SSH Keys, and Timezone on PIT Node.

    1. Generate SSH keys.

      NOTE The code block below assumes there is an RSA key without a passphrase. This step can be customized to use a passphrase if desired.

      ssh-keygen -N "" -t rsa
      
    2. Export the password hash for root that is needed for the ncn-image-modification.sh script.

      This will set the NCN root user password to be the same as the root user password on the PIT.

      export SQUASHFS_ROOT_PW_HASH="$(awk -F':' /^root:/'{print $2}' < /etc/shadow)"
      
    3. Inject these into the NCN images by running ncn-image-modification.sh from the CSM documentation RPM.

      NCN_MOD_SCRIPT=$(rpm -ql docs-csm | grep ncn-image-modification.sh)
      echo "${NCN_MOD_SCRIPT}"
      "${NCN_MOD_SCRIPT}" -p \
         -d /root/.ssh \
         -k "/var/www/ephemeral/data/k8s/${KUBERNETES_VERSION}/kubernetes-${KUBERNETES_VERSION}-$(uname -i).squashfs" \
         -s "/var/www/ephemeral/data/ceph/${CEPH_VERSION}/storage-ceph-${CEPH_VERSION}-$(uname -i).squashfs"
      
  9. (pit#) Log the currently installed PIT packages.

    Having this information in the typescript can be helpful if problems are encountered during the install. This command was run once in a previous step – running it again now is intentional.

    /root/bin/metalid.sh
    

    Expected output looks similar to the following (the versions in the example below may differ). There should be no errors.

     = PIT Identification = COPY/CUT START =======================================
     VERSION=ed97205-1706718622724
     TIMESTAMP=2024-01-31_16:30:22
     CRAY-Site-Init build signature...
     Build Commit   : c9a07e366151a72be71d168061d08bd97da5344c-heads-v1.32.4
     Build Time     : 2023-10-20T14:53:30Z
     Go Version     : go1.19
     Version        : v1.32.4
     Platform       : linux/amd64
     canu-1.8.0-1.x86_64
     ilorest-4.2.0.0-20.x86_64
     metal-basecamp-1.2.6-1.x86_64
     metal-ipxe-2.4.8-1.noarch
     metal-init-1.4.6-1.noarch
     metal-nexus-1.3.1-3.38.0_1.x86_64
     metal-observability-1.0.9-1.x86_64
     = PIT Identification = COPY/CUT END =========================================
    

3. Create system configuration

Create the system configuration using one of the following options:

4 Import the CSM Tarball

The following steps require create system configuration to have completed successfully.

  1. (pit#) Upload the CSM tarball’s RPMs and container images to the local Nexus instance.

    /srv/cray/metal-provision/scripts/nexus/setup-nexus.sh -s
    
  2. Add the local Zypper repositories for noos and the current SLES distribution.

    NOTE The ${releasever_major} and ${releasever_minor} variables are interpolated by Zypper, the URI is intentionally wrapped with single-quotes to prevent the shell from interpolating them. Zypper will replace these variables with the currently running distributions major and minor version numbers.

    zypper addrepo --no-gpgcheck --refresh http://packages/repository/csm-noos csm-noos
    zypper addrepo --no-gpgcheck --refresh 'http://packages/repository/csm-sle-${releasever_major}sp${releasever_minor}' 'csm-sle'
    
  3. (pit#) Ensure any new, updated packages pertinent to the CSM install are installed.

    NOTES

    • csm-testing package provides the necessary tests and their dependencies for validating the pre-installation, installation, and more.
    • This provides iuf, a command line interface to the Install and Upgrade Framework.
    zypper --no-gpg-checks install -y canu craycli csm-testing hpe-csm-goss-package iuf-cli platform-utils
    

5 Validate the LiveCD

  1. (pit#) Verify that the LiveCD is ready by running the preflight tests.

    csi pit validate --livecd-preflight
    

    If any tests fail, they need to be investigated. After actions have been taken to rectify the tests (for example, editing configuration or CSI inputs), then restart from the beginning of the Initialize the LiveCD procedure.

    The following test failure may be ignored if the management network switches have not been configured. This is often the case when the system is being installed with CSM for the first time. Configuring switches is covered in the next topic.

    Result: FAIL
    Source: /opt/cray/tests/install/livecd/suites/livecd-preflight-tests.yaml
    Test Name: sls_input.json IPs Correct
    Description: Extracts the switch IP addresses from sls_input.json and pings them to ensure they are accurate.
    Test Summary: check_sls_file_ips: exit-status: Error: Command execution timed out (20s)
    Execution Time: 0.000002214 seconds
    Node: eniac-pit
    
  2. Save the prep directory for re-use.

    This needs to be copied off the system and either stored in a secure location or in a secured Git repository. There are secrets in this directory that should not be accidentally exposed.

Next topic

After completing this procedure, proceed to configure the management network switches.

See Configure management network switches.