aws logo
AWS Classic v5.29.0, Feb 1 23

Amazon EKS Cluster | TypeScript

View Code Deploy

This example deploys an EKS Kubernetes cluster with an EBS-backed StorageClass and deploys the Kubernetes Dashboard into the cluster.

Deploying the App

To deploy your infrastructure, follow the below steps.


  1. Install Pulumi
  2. Install Node.js
  3. Configure AWS Credentials
  4. Install aws-iam-authenticator

If you’d like to follow the optional instructions in step 7 in order to deploy a Helm chart into your cluster, you’ll also need to set up the Helm client:

  1. Install the Helm client binaries

  2. If you are using Helm v2, initialize the Helm client:

    $ helm init --client-only


After cloning this repo, from this working directory, run these commands:

  1. Install the required Node.js packages:

    $ npm install
  2. Create a new stack, which is an isolated deployment target for this example:

    $ pulumi stack init
  3. Set the required configuration variables for this program:

    $ pulumi config set aws:region us-west-2

    We recommend using us-west-2 to host your EKS cluster as other regions (notably us-east-1) may have capacity issues that prevent EKS clusters from creating:

      aws:eks:Cluster: eksCluster
        error: Plan apply failed: creating urn:pulumi:aws-ts-eks-example::aws-ts-eks::EKSCluster$aws:eks/cluster:Cluster::eksCluster: error creating EKS Cluster (eksCluster-233c968): UnsupportedAvailabilityZoneException: Cannot create cluster 'eksCluster-233c968' because us-east-1a, the targeted availability zone, does not currently have sufficient capacity to support the cluster. Retry and choose from these availability zones: us-east-1b, us-east-1c, us-east-1d
            status code: 400, request id: 9f031e89-a0b0-11e8-96f8-534c1d26a353

    We are tracking enabling the creation of VPCs limited to specific AZs to unblock this in us-east-1: pulumi/pulumi-awsx#32

  4. Stand up the EKS cluster, which will also deploy the Kubernetes Dashboard:

    $ pulumi up
  5. After 10-15 minutes, your cluster will be ready, and the kubeconfig JSON you’ll use to connect to the cluster will be available as an output. You can save this kubeconfig to a file like so:

    $ pulumi stack output kubeconfig --show-secrets >kubeconfig.json

    Once you have this file in hand, you can interact with your new cluster as usual via kubectl:

    $ KUBECONFIG=./kubeconfig.json kubectl get nodes
  6. You can now connect to the Kubernetes Dashboard by fetching an authentication token and starting the kubectl proxy.

    • Fetch an authentication token:

      $ KUBECONFIG=./kubeconfig.json kubectl -n kube-system get secret | grep eks-admin | awk '{print $1}'
      $ KUBECONFIG=./kubeconfig.json kubectl -n kube-system describe secret eks-admin-token-b5zv4
      Name:         eks-admin-token-b5zv4
      Namespace:    kube-system
      Labels:       <none>
      token:      <authentication_token>
      ca.crt:     1025 bytes
      namespace:  11 bytes
    • Run the kubectl proxy:

      $ KUBECONFIG=./kubeconfig.json kubectl proxy
    • Open http://localhost:8001/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy/ in a web browser.

    • Choose Token authentication, paste the token retrieved earlier into the Token field, and sign in.

  7. From there, feel free to experiment. Make edits and run pulumi up to incrementally update your stack. For example, in order to deploy a Helm chart into your cluster, import the @pulumi/kubernetes/helm package, add a Chart resource that targets the EKS cluster to index.ts, and run pulumi up. Note that the Helm client must be set up in order for the chart to deploy. For more details, see the Prerequisites list.

    import * as helm from "@pulumi/kubernetes/helm";
    // ... existing code here ...
    const myk8s = new k8s.Provider("myk8s", {
        kubeconfig: cluster.kubeconfig.apply(JSON.stringify),
    const postgres = new helm.v2.Chart("postgres", {
        // stable/postgresql@0.15.0
        repo: "stable",
        chart: "postgresql",
        version: "0.15.0",
        values: {
            // Use a stable password.
            postgresPassword: "some-password",
            // Expose the postgres server via a load balancer.
            service: {
                type: "LoadBalancer",
    }, { providers: { kubernetes: myk8s } });

    Once the chart has been deployed, you can find its public, load-balanced endpoint via the Kubernetes Dashboard.

  8. Once you’ve finished experimenting, tear down your stack’s resources by destroying and removing it:

    $ pulumi destroy --yes
    $ pulumi stack rm --yes