Lokomotive AWS quickstart guide

    Introduction

    This quickstart guide walks through the steps needed to create a Lokomotive cluster on AWS.

    Lokomotive runs on top of Flatcar Container Linux . This guide uses the stable channel.

    The guide uses Amazon Route 53 as a DNS provider. For more information on how Lokomotive handles DNS, refer to this document.

    Lokomotive stores Terraform state locally by default, or remotely within an AWS S3 bucket .

    Lokomotive components complement the “stock” Kubernetes functionality by adding features such as load balancing, persistent storage and monitoring to a cluster. To keep this guide short, you will deploy a single component - httpbin - which serves as a demo application to verify the cluster behaves as expected.

    By the end of this guide, you’ll have a production-ready Kubernetes cluster running on AWS.

    Requirements

    • Basic understanding of Kubernetes concepts.
    • An AWS account and IAM credentials.
    • An AWS access key ID and a user’s secret with permissions to edit Route 53 records.
    • An AWS Route 53 zone (can be a subdomain).
    • Terraform v0.13.x installed locally.
    • An SSH key pair for management access.
    • terraform v0.13.x installed .
    • kubectl installed .

    NOTE: The kubectl version used to interact with a Kubernetes cluster needs to be compatible with the version of the Kubernetes control plane version. Ideally, you should install a kubectl binary whose version is identical to the Kubernetes control plane included with a Lokomotive release. However, some degree of version “skew” is tolerated - see the Kubernetes version skew policy document for more information. You can determine the Kubernetes control plane’s version included with a Lokomotive release by looking at the release notes .

    Steps

    Step 1: Install lokoctl

    lokoctl is the command-line interface for managing Lokomotive clusters. You can follow the installer guide to install it locally for your OS.

    Step 2: Create a cluster configuration

    Create a directory for the cluster-related files and navigate to it:

    mkdir lokomotive-demo && cd lokomotive-demo
    

    Create a file named cluster.lokocfg with the following contents:

    cluster "aws" {
      asset_dir            = "./assets"
      cluster_name         = "lokomotive-demo"
      controller_count     = 1
      dns_zone             = "example.com"
      dns_zone_id          = "DNSZONEID"
    
      region               = "us-east-1"
      ssh_pubkeys          = ["ssh-rsa AAAA..."]
    
      worker_pool "pool-1" {
        count         = 2
        ssh_pubkeys   = ["ssh-rsa AAAA..."]
      }
    }
    
    # Needed for ingress to work.
    component "contour" {
      service_type = "NodePort"
    }
    
    # A demo application.
    component "httpbin" {
      ingress_host = "httpbin.lokomotive-demo.example.com"
    }
    

    Replace the parameters above using the following information:

    • dns_zone - a Route 53 zone name. A subdomain will be created under this zone in the following format: <cluster_name>.<zone>
    • dns_zone_id - a Route 53 DNS zone ID found in your AWS console.
    • ssh_pubkeys - A list of strings representing the contents of the public SSH keys which should be authorized on cluster nodes.

    The rest of the parameters may be left as-is. For more information about the configuration options, see the configuration reference .

    Step 3: Deploy the cluster

    NOTE: If you have the AWS CLI installed and configured for an AWS account, you can skip setting the AWS_* variables below. lokoctl follows the standard AWS authentication methods, which means it will use the default AWS CLI profile if no explicit credentials are specified. Similarly, environment variables such as AWS_PROFILE can instruct lokoctl to use a specific AWS CLI profile for AWS authentication.

    Set up your AWS credentials in your shell:

    export AWS_ACCESS_KEY_ID=AKIAIOSFODNN7FAKE
    export AWS_SECRET_ACCESS_KEY=wJalrXUtnFEMI/K7MDENG/bPxRfiCYFAKE
    

    Add a private key corresponding to one of the public keys specified in ssh_pubkeys to your ssh-agent:

    ssh-add ~/.ssh/id_rsa
    ssh-add -L
    

    Deploy the cluster:

    lokoctl cluster apply -v
    

    The deployment process typically takes about 15 minutes. Upon successful completion, an output similar to the following is shown:

    Your configurations are stored in ./assets
    
    Now checking health and readiness of the cluster nodes ...
    
    Node              Ready    Reason          Message
    
    ip-10-0-11-66     True     KubeletReady    kubelet is posting ready status
    ip-10-0-34-253    True     KubeletReady    kubelet is posting ready status
    ip-10-0-92-177    True     KubeletReady    kubelet is posting ready status
    
    Success - cluster is healthy and nodes are ready!
    

    Verification

    Use the generated kubeconfig file to access the cluster:

    export KUBECONFIG=$(pwd)/assets/cluster-assets/auth/kubeconfig
    kubectl get nodes
    

    Sample output:

    NAME             STATUS   ROLES    AGE    VERSION
    ip-10-0-11-66    Ready    <none>   105s   v1.19.4
    ip-10-0-34-253   Ready    <none>   107s   v1.19.4
    ip-10-0-92-177   Ready    <none>   105s   v1.19.4
    

    Verify all pods are ready:

    kubectl get pods -A
    

    Verify you can access httpbin:

    HTTPBIN_HOST=$(kubectl get ing -n httpbin -o jsonpath='{.items[*].spec.rules[0].host}')
    
    curl ${HTTPBIN_HOST}/get
    

    Sample output:

    {
      "args": {},
      "headers": {
        "Accept": "*/*",
        "Content-Length": "0",
        "Host": "httpbin.lokomotive-demo.example.com",
        "User-Agent": "curl/7.70.0",
        "X-Envoy-Expected-Rq-Timeout-Ms": "15000",
        "X-Envoy-External-Address": "49.207.214.243"
      },
      "origin": "49.207.214.243",
      "url": "http://httpbin.lokomotive-demo.example.com/get"
    }
    

    Using the cluster

    At this point, you should have access to a Lokomotive cluster and can use it to deploy applications.

    If you don’t have any Kubernetes experience, you can check out the Kubernetes Basics tutorial.

    NOTE: Lokomotive uses a relatively restrictive Pod Security Policy by default. This policy disallows running containers as root. Refer to the Pod Security Policy documentation for more details. We also deploy a webhook server that disallows usage of default service account’s usage. Refer to the Lokomotive admission webhooks for more information.

    Cleanup

    To destroy the Lokomotive cluster, execute the following command:

    lokoctl cluster destroy --confirm
    

    You can safely delete the working directory created for this quickstart guide if you no longer require it.

    Troubleshooting

    Stuck At Copy Controller Secrets

    If there is an execution error or no progress beyond the output provided below:

    ...
    module.aws-myawscluster.null_resource.copy-controller-secrets: Still creating... (8m30s elapsed)
    module.aws-myawscluster.null_resource.copy-controller-secrets: Still creating... (8m40s elapsed)
    ...
    

    The error probably happens because the ssh_pubkeys provided in the configuration is missing in the ssh-agent.

    In case the deployment process seems to hang at the copy-controller-secrets phase for a long time, check the following:

    • Verify the correct private SSH key was added to ssh-agent.
    • Verify that you can SSH into the created controller node from the machine running lokoctl.

    IAM Permission Issues

    Conclusion

    After walking through this guide, you’ve learned how to set up a Lokomotive cluster on AWS.

    Next steps

    You can now start deploying your workloads on the cluster.

    For more information on installing supported Lokomotive components, you can visit the component configuration references .