In this section we will create a new Ocean cluster, associated with your existing EKS cluster.
The Cluster Identifier for your Ocean cluster should be unique within the account, and defaults to the Ocean Cluster Name.
When importing a cluster, Ocean will clone your cluster and node pools configuration. New instances will then be launched and registered directly to your cluster, and will not be visible via your node pools. Your existing instances and applications will remain unchanged.
By default, Ocean will use as wide a selction of instance types as possible, in order to ensure optimal pricing and availabilty for your worker nodes by tapping into many EC2 Spot capacity pools. If you wish, you can exclude cerain types from the pool of instances used by the cluster, by clicking on “Customize” under “Machine Types”.
In order to fully migrate any existing workloads to Ocean, the original EKS Auto Scaling group/s should be gradually drained and scaled down, while replacement nodes should be launched by Ocean. In order to make this process automatic and safe, and gradually migrate all workloads while maintaining high availablity for the application, Ocean has the “Workload Migration” feature. You can read about it here, or watch the video tutorial here.
In the interest of stability, the Workload Migration process is very gradual, and therefore takes a while (up to half an hour), even for small workloads. So for the purposes of this workshop we will assume that our workloads can tolerate a more aggesive rescheduling. Therefore, proceed with the following steps:
If you have several node groups configured, with different sets of labels, taints or launch specifications, before scaling them down make sure to configure matching “Launch Specifications” in Ocean. Have a look at the next page in the workshop to see how.
You’re all set! Ocean will now ensure your EKS cluster worker nodes are running on the most cost-effective, optimally sized instances possible.