1. Packages
  2. AWS Classic
  3. How-to Guides
  4. Amazon EKS Cluster

Try AWS Native preview for resources not in the classic version.

AWS Classic v6.32.0 published on Friday, Apr 19, 2024 by Pulumi

Amazon EKS Cluster

aws logo

Try AWS Native preview for resources not in the classic version.

AWS Classic v6.32.0 published on Friday, Apr 19, 2024 by Pulumi

    View Code Deploy

    This example deploys an EKS Kubernetes cluster with an EBS-backed StorageClass and deploys the Kubernetes Dashboard into the cluster.

    Deploying the App

    To deploy your infrastructure, follow the below steps.

    Prerequisites

    1. Install Pulumi
    2. Install Node.js
    3. Configure AWS Credentials
    4. Install aws-iam-authenticator

    If you’d like to follow the optional instructions in step 7 in order to deploy a Helm chart into your cluster, you’ll also need to set up the Helm client:

    1. Install the Helm client binaries

    2. If you are using Helm v2, initialize the Helm client:

      $ helm init --client-only
      

    Steps

    After cloning this repo, from this working directory, run these commands:

    1. Install the required Node.js packages:

      $ npm install
      
    2. Create a new stack, which is an isolated deployment target for this example:

      $ pulumi stack init
      
    3. Set the required configuration variables for this program:

      $ pulumi config set aws:region us-west-2
      

      We recommend using us-west-2 to host your EKS cluster as other regions (notably us-east-1) may have capacity issues that prevent EKS clusters from creating:

      Diagnostics:
        aws:eks:Cluster: eksCluster
          error: Plan apply failed: creating urn:pulumi:aws-ts-eks-example::aws-ts-eks::EKSCluster$aws:eks/cluster:Cluster::eksCluster: error creating EKS Cluster (eksCluster-233c968): UnsupportedAvailabilityZoneException: Cannot create cluster 'eksCluster-233c968' because us-east-1a, the targeted availability zone, does not currently have sufficient capacity to support the cluster. Retry and choose from these availability zones: us-east-1b, us-east-1c, us-east-1d
              status code: 400, request id: 9f031e89-a0b0-11e8-96f8-534c1d26a353
      

      We are tracking enabling the creation of VPCs limited to specific AZs to unblock this in us-east-1: pulumi/pulumi-awsx#32

    4. Stand up the EKS cluster, which will also deploy the Kubernetes Dashboard:

      $ pulumi up
      
    5. After 10-15 minutes, your cluster will be ready, and the kubeconfig JSON you’ll use to connect to the cluster will be available as an output. You can save this kubeconfig to a file like so:

      $ pulumi stack output kubeconfig --show-secrets >kubeconfig.json
      

      Once you have this file in hand, you can interact with your new cluster as usual via kubectl:

      $ KUBECONFIG=./kubeconfig.json kubectl get nodes
      
    6. You can now connect to the Kubernetes Dashboard by fetching an authentication token and starting the kubectl proxy.

      • Fetch an authentication token:

        $ KUBECONFIG=./kubeconfig.json kubectl -n kube-system get secret | grep eks-admin | awk '{print $1}'
        eks-admin-token-b5zv4
        $ KUBECONFIG=./kubeconfig.json kubectl -n kube-system describe secret eks-admin-token-b5zv4
        Name:         eks-admin-token-b5zv4
        Namespace:    kube-system
        Labels:       <none>
        Annotations:  kubernetes.io/service-account.name=eks-admin
                      kubernetes.io/service-account.uid=bcfe66ac-39be-11e8-97e8-026dce96b6e8
        
        Type:  kubernetes.io/service-account-token
        
        Data
        ====
        token:      <authentication_token>
        ca.crt:     1025 bytes
        namespace:  11 bytes
        
      • Run the kubectl proxy:

        $ KUBECONFIG=./kubeconfig.json kubectl proxy
        
      • Open http://localhost:8001/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy/ in a web browser.

      • Choose Token authentication, paste the token retrieved earlier into the Token field, and sign in.

    7. From there, feel free to experiment. Make edits and run pulumi up to incrementally update your stack. For example, in order to deploy a Helm chart into your cluster, import the @pulumi/kubernetes/helm package, add a Chart resource that targets the EKS cluster to index.ts, and run pulumi up. Note that the Helm client must be set up in order for the chart to deploy. For more details, see the Prerequisites list.

      import * as helm from "@pulumi/kubernetes/helm";
      
      // ... existing code here ...
      
      const myk8s = new k8s.Provider("myk8s", {
          kubeconfig: cluster.kubeconfig.apply(JSON.stringify),
      });
      
      const postgres = new helm.v2.Chart("postgres", {
          // stable/postgresql@0.15.0
          repo: "stable",
          chart: "postgresql",
          version: "0.15.0",
          values: {
              // Use a stable password.
              postgresPassword: "some-password",
              // Expose the postgres server via a load balancer.
              service: {
                  type: "LoadBalancer",
              },
          },
      }, { providers: { kubernetes: myk8s } });
      

      Once the chart has been deployed, you can find its public, load-balanced endpoint via the Kubernetes Dashboard.

    8. Once you’ve finished experimenting, tear down your stack’s resources by destroying and removing it:

      $ pulumi destroy --yes
      $ pulumi stack rm --yes
      
    aws logo

    Try AWS Native preview for resources not in the classic version.

    AWS Classic v6.32.0 published on Friday, Apr 19, 2024 by Pulumi