Installing Lamini Platform on Kubernetes
Note
The Lamini installer is only available when self-managing Lamini Platform. Contact us to learn more.
Lamini Platform on Kubernetes enables multi-node, multi-GPU inference and training running on your own GPUs, in the environment of your choice.
Prerequisites
Tools
You need to have a working Kubernetes cluster, and python, helm, kubectl installed.
Lamini Self-Managed license
Contact us for access to the Kubernetes installer to host Lamini Platform on your own GPUs or in your cloud VPC.
Hardware system requirements
- Storage: 1 TB or more of high-speed network storage with low latency (around 1 ms), such as a Network File System (NFS).
- GPU: 48 GB or more of memory per GPU.
PVC
Lamini requires a RWX PVC for storing all runtime data.
You can use NFS and other storage solutions.
For example, you can set up a simple provisioner using nfs-subdir-external-provisioner:
helm repo add nfs-subdir-external-provisioner https://kubernetes-sigs.github.io/nfs-subdir-external-provisioner/
helm install nfs-subdir-external-provisioner nfs-subdir-external-provisioner \
--set nfs.server=<NFS_IP> \
--set nfs.path=<NFS_SUBFOLDER_PATH>
Then proceed to create a PVC lamini-volume with ReadWriteMany access for installing Lamini Platform.
GPU Operator
-
For AMD:
-
For NVIDIA:
Installation Steps
Obtain the installer
Ask your Sales Engineer for the installer.
The installer is a .tar.gz compressed file with all helm charts and scripts for installing Lamini Platform.
You should work with your Sales Engineer for each installation or upgrade of Lamini Platform.
You should keep any changes to the installer in a private repository for tracking purposes. Also ask your Sales Engineer to keep track of such changes.
After obtaining the installer, extract it to a directory of your choice:
# Make sure the installer name is used for directory name
# so that you can track the version of the installer.
INSTALLER_NAME="<name>"
mkdir -p ${INSTALLER_NAME}
tar -xzf ${INSTALLER_NAME}.tar.gz -C ${INSTALLER_NAME}
The rest of the instructions are in the INSTALL.md file in the installer. You should operate under the directory of the installer.
# Change to the installer directory
cd ${INSTALLER_NAME}/lamini-kube-installer
# Read the INSTALL.md file, open with your favorite editor
vi INSTALL.md
Update helm_config.yaml
- Optional: If you already have
nfssubdir-external-provisionerinstalled, set thepvc_provisionerto thestorageclassname of defined by your installednfs-subdir-external-provisioner.
- If you opt to use Lamini Platform provided NFS pvc provisioner, set the
pvcLamini.nameto the name of the PVC you want to use, and setcreatetoTrue, and setsizeto the recommended200Gi, or work with your Sales Engineer to determine the size:
if you have already created a PVC, set name to the name of the PVC, set create to False, you can
omit size:
- Confirm the top-level platform
type(one of:amd,nvidia, orcpu) matches your hardware.
- Update the distribution of inference pods.
The example above would create 4 pods using 4 GPUs in total. Each pod has 1 GPU. The example shows 1 inference pod allocated to batch inference, 1 pod dedicated only to streaming inference, 1 dedicated only to embedding inference (also used in classification), and 1 for the catchall pod, which is intended to handle requests for models that have not been preloaded on the batch pod. See Model Management for more details.
- Update the number of training pods and number of GPUs per pod:
We recommend minimizing the number of pods per node. For example, instead of 2 pods with 4 GPUs, it's better to create 1 pod with all 8 GPUs.
- Update the node affinity for the Lamini deployment. These are the nodes where Lamini pods will be deployed:
- (Optional) If you want to use a custom ingress pathway, update the
ingressfield:
Generate Helm Charts and install Lamini Platform
Follow the INSTALL.md included in the installer for the detailed steps. The general steps are:
- Generate Helm charts with the provided shell script
- Install Lamini Platform with
helm installor upgrade withhelm upgrade