Overview
Orka Burst provides dedicated, on-demand access to elastic cluster capacity when needed. It’s designed to grow your cluster size during heavy workloads, without the commitment of annual contracts. When required, you can enable burst capacity directly from your MacStadium Admin Portal.
Orka Burst requires a one-time setup maintenance window to provision and configure the burst nodes. Once configured (you’ll see which nodes are burst-enabled in your IP Plan), additional burst nodes are available within 15 minutes and billed only for the time used.
Key Concepts
- A burst node is an on-demand, contractually provisioned node you can add to your cluster during high-workload periods.
- Only administrators can start and stop burst nodes via the MacStadium Admin Portal.
- Burst nodes are powered on/off manually by an admin user in the portal.
- You cannot manage burst nodes directly within the Kubernetes cluster; node management commands are restricted for security.
- Orka tags (similar to Kubernetes labels) can be used for monitoring burst nodes via Prometheus (
kube-node-labels metric) or other tools.
First Time Setup
- Contact your MacStadium sales team to enable Orka Burst on your cluster.
- Log in to the MacStadium Admin Portal with an admin-level account.
- Click the “Enable Burst” toggle and confirm to provision your burst nodes.
- Click ‘Confirm’ to accept the use of the contracted service Orka Burst.
- When all scheduled burst workloads have completed and no remaining CI jobs are scheduled on burst runners, manually drain the Orka Burst nodes and move them to a different namespace.
- After the empty Orka Burst nodes have been successfully moved to a different namespace, navigate back to the Macstadium Portal and click ‘Disable Burst’.
Enable/Disable Burst
Enable Orka Burst
After MacStadium support configures your burst nodes, you can turn them on via the portal. Enabling burst powers on the nodes and adds them to your cluster (this typically completes within 15 minutes).
Disable Orka Burst
Orka Burst has no knowledge of currently executing workloads, and as a result, any currently running jobs will fail and need to be rescheduled by the preferred scheduler controller. You may need to clean up failed workloads if the scheduled workloads are executing on disabled burst nodes, as Orka Burst is being disabled.
When Burst is disabled, the nodes are powered down and removed from the cluster.
Plan to disable burst during a period of low cluster utilization, and coordinate with your engineering team to drain or move any active workloads first.
If your burst nodes are in the same namespace as your normal Orka nodes, move them to a separate namespace before disabling Orka Burst. This sandboxes the node and ensures no additional workloads are scheduled on it. You are responsible for monitoring burst node usage and moving nodes when no workloads are running on them. Once your cluster has passed its peak utilization, work with your engineering team to move idle burst machines to a separate namespace.
Removing Failing Workloads
- Identify the failing workload:
kubectl get <RESOURCE> --namespace=<NAMESPACE> and look for pods in CrashLoopBackoff, Failed, or Error status.
- Delete the workload:
kubectl delete pod <pod-name> --namespace=<NAMESPACE>
- Graceful deletion: use a
preStop hook to allow pods to clean up before SIGTERM.
- Forced deletion: add
--force --grace-period=0 to the delete command.
- Confirm cleanup and reschedule your CI jobs once burst nodes are back online.