Installation Guide
This guide provides quickstart instructions for deploying the MaaS Platform infrastructure.
Note
For more detailed instructions, please refer to Installation under the Administrator Guide.
Prerequisites
- OpenShift cluster (4.19.9+) with kubectl/oc access
- Recommended 16 vCPUs, 32GB RAM, 100GB storage
- ODH/RHOAI requirements:
- RHOAI 3.0 +
- ODH 3.0 +
- RHCL requirements (Note: This can be installed automatically by the script below):
- RHCL 1.2 +
- Cluster admin or equivalent permissions
- Required tools:
oc(OpenShift CLI)kubectljqkustomize(v5.7.0+)
Quick Start
Automated OpenShift Deployment (Recommended)
For OpenShift clusters, use the automated deployment script:
Verify Deployment
The deployment script creates the following core resources:
- Gateway:
maas-default-gatewayinopenshift-ingressnamespace - HTTPRoutes:
maas-api-routein theopenshift-ingressnamespace - Policies:
AuthPolicy,TokenRateLimitPolicy,RateLimitPolicy,TelemetryPolicy - MaaS API: Deployment and service in
maas-apinamespace - Operators: Cert-manager, LWS, Red Hat Connectivity Link and Red Hat OpenShift AI.
Check deployment status:
# Check all namespaces
kubectl get ns | grep -E "maas-api|kuadrant-system|kserve|opendatahub|redhat-ods-applications|llm"
# Check Gateway status
kubectl get gateway -n openshift-ingress maas-default-gateway
# Check policies
kubectl get authpolicy -A
kubectl get tokenratelimitpolicy -A
kubectl get ratelimitpolicy -A
# Check MaaS API
kubectl get pods -n maas-api
kubectl get svc -n maas-api
# Check Kuadrant operators
kubectl get pods -n kuadrant-system
# Check KServe (if deployed)
kubectl get pods -n kserve
kubectl get pods -n opendatahub
kubectl get pods -n redhat-ods-applications
Model Setup (Optional)
Deploy Sample Models (Optional)
Simulator Model (CPU)
PROJECT_DIR=$(git rev-parse --show-toplevel)
kustomize build ${PROJECT_DIR}/docs/samples/models/simulator/ | kubectl apply -f -
Facebook OPT-125M Model (CPU)
PROJECT_DIR=$(git rev-parse --show-toplevel)
kustomize build ${PROJECT_DIR}/docs/samples/models/facebook-opt-125m-cpu/ | kubectl apply -f -
Qwen3 Model (GPU Required)
Warning
This model requires GPU nodes with nvidia.com/gpu resources available in your cluster.
PROJECT_DIR=$(git rev-parse --show-toplevel)
kustomize build ${PROJECT_DIR}/docs/samples/models/qwen3/ | kubectl apply -f -
Verify Model Deployment
# Check LLMInferenceService status
kubectl get llminferenceservices -n llm
# Check pods
kubectl get pods -n llm
Update Existing Models (Optional)
To update an existing model, modify the LLMInferenceService to use the newly created maas-default-gateway gateway.
kubectl patch llminferenceservice my-production-model -n llm --type='json' -p='[
{
"op": "add",
"path": "/spec/gateway/refs/-",
"value": {
"name": "maas-default-gateway",
"namespace": "openshift-ingress"
}
}
]'
apiVersion: serving.kserve.io/v1alpha1
kind: LLMInferenceService
metadata:
name: my-production-model
spec:
gateway:
refs:
- name: maas-default-gateway
namespace: openshift-ingress
Next Steps
After installation, proceed to Validation to test and verify your deployment.