New in version 1.1.

From Kubernetes.io:

Kubernetes is an open-source system for automating deployment, operations, and scaling of containerized applications.

Since version 1.1, Mantl ships Kubernetes by default. All you need to do is set the kubeworker_count and kubeworker_type variables in your Terraform configuration (see the example Terraform configurations for where this variable integrates into the workflow.)

kubectl is installed and configured for the default SSH user on the control nodes. Please refer to the Kubernetes getting started documentation for how to use Kubernetes.

To talk to the services launched inside Kubernetes, launch them with the NodePort service type (more on what service types are available), then connect on the assigned port on any Kubernetes worker node. Consul service integration will happen in a future release.

Running kubectl Remotely

If you have a local installation of kubectl, you can run kubectl config to configure access to your Mantl cluster. Here is an example:

kubectl config set-cluster mantl --server=https://control-node/kubeapi --insecure-skip-tls-verify=true
kubectl config set-credentials mantl-basic-auth --username=admin --password=password
kubectl config set-context mantl --cluster=mantl --user=mantl-basic-auth
kubectl config use-context mantl

You can set the value of the cluster and context names (mantl in the above example) as desired. In addtion, make sure you replace the value of control-node and password to values that are applicable for your cluster.

Cloud Provider Integration

Cloud provider integration is enabled by default for AWS and GCE clusters starting in Mantl 1.3. This means that Kubernetes can manage cloud-specific resources such as disk volumes and load balancers. If you wish to disable cloud provider integration, set the variable enable_cloud_provider to false when building your cluster.


If you are planning on destroying your cluster with terraform, you should first use kubectl or the Kubernetes API to delete your Kubernetes-managed resources. Otherwise, it is possible that they will interfere with your ability to successfully terraform destroy your cluster.

DNS Outline

Every node in the cluster hosts etcd and skydns instances. All DNS queries for the .local zone are resolved locally. If a container asks for name in .local domain, the request is routed through dnsmasq to skydns, which accesses data stored in etcd. Updates for container dns names are managed by kube2sky, which acts upon kubeapi events.