--- title: "VMware Tanzu Community Edition Kubernetes Platform in a Homelab" # Title of the blog post. date: 2022-01-12 # Date of post creation. # lastmod: 2022-01-06T09:42:51-06:00 # Date when last modified description: "Gaining familiarity with VMware Tanzu Community Edition by deploying phpIPAM on Kubernetes in my homelab" # Description used for search engine. featured: false # Sets if post is a featured post, making appear on the home page side bar. draft: false # Sets whether to render this page. Draft of true will not be rendered. toc: true # Controls if a table of contents should be generated for first-level links automatically. usePageBundles: true # menu: main # featureImage: "file.png" # Sets featured image on blog post. # featureImageAlt: 'Description of image' # Alternative text for featured image. # featureImageCap: 'This is the featured image.' # Caption (optional). thumbnail: "tanzu_community_edition.png" # Sets thumbnail image appearing inside card on homepage. # shareImage: "share.png" # Designate a separate image for social media sharing. codeLineNumbers: false # Override global value for showing of line numbers within code block. categories: VMware tags: - vmware - linux - kubernetes - docker - containers - tanzu - homelab comments: true # Disable comment if false. --- Back in October, VMware [announced](https://tanzu.vmware.com/content/blog/vmware-tanzu-community-edition-announcement) [Tanzu Community Edition](https://tanzucommunityedition.io/) as way to provide "a full-featured, easy-to-manage Kubernetes platform that’s perfect for users and learners alike." TCE bundles a bunch of open-source components together in a modular, "batteries included but swappable" way: ![Tanzu Community Edition components](tanzu_community_edition.png) I've been meaning to brush up on my Kubernetes skills so I thought deploying and using TCE in my self-contained [homelab](/vmware-home-lab-on-intel-nuc-9/) would be a fun and rewarding learning exercise - and it was! Here's how I did it. ### Planning TCE supports several different deployment scenarios and targets. It can be configured as separate Management and Workload Clusters or as a single integrated Standalone Cluster, and deployed to cloud providers like AWS and Azure, on-premise vSphere, or even a local Docker environment[^yo_dawg]. I'll be using the standard Management + Workload Cluster setup in my on-prem vSphere, so I start by reviewing the [Prepare to Deploy a Cluster to vSphere](https://tanzucommunityedition.io/docs/latest/vsphere/) documentation to get an idea of what I'll need. Looking ahead, part of the installation process creates a local [KIND](https://kind.sigs.k8s.io/) cluster for bootstrapping the Management and Workload clusters. I do most of my home computing (and homelab work) by using the [Linux environment available on my Chromebook](/setting-up-linux-on-a-new-lenovo-chromebook-duet-bonus-arm64-complications/). Unfortunately I know from past experience that KIND will not work within this environment so I'll be using a Debian 10 VM to do the deployment. [^yo_dawg]: Yo dawg, I heard you like containers... #### Networking The Kubernetes node VMs will need to be attached to a network with a DHCP server to assign their addresses, and that network will need to be able to talk to vSphere. My router handles DHCP for the range `192.168.1.101-250` so I'll plan on using that. I'll also need to set aside a few static IPs for this project. These will need to be routable and within the same subnet as the DHCP range, but excluded from that DHCP range. | IP Address | Purpose | | --- | --- | | `192.168.1.60` | Control plane for Management cluster | | `192.168.1.61` | Control plane for Workload cluster | | `192.168.1.64 - 192.168.1.80` | IP range for Workload load balancer | ### Prerequisites Moving on to the [Getting Started](https://tanzucommunityedition.io/docs/latest/getting-started/), I'll need to grab some software before I can actually Get Started. #### Kubernetes control plane image I need to download a VMware OVA which can be used for deploying my Kubernetes nodes from the VMWare Customer Connect portal [here](https://customerconnect.vmware.com/downloads/get-download?downloadGroup=TCE-090)[^register]. There are a few different options available. I'll get the Photon release with the highest Kubernetes version currently available, `photon-3-kube-v1.21.2+vmware.1-tkg.2-12816990095845873721.ova`. Once the file is downloaded, I'll log into my vCenter and use the **Deploy OVF Template** action to deploy a new VM using the OVA. I won't bother booting the machine once deployed but will rename it to `k8s-node` to make it easier to identify later on and then convert it to a template. ![New k8s-node template](k8s-node_template.png) [^register]: Register [here](https://customerconnect.vmware.com/account-registration) if you don't yet have an account. #### Docker I've already got Docker installed on this machine, but if I didn't I would follow the instructions [here](https://docs.docker.com/engine/install/debian/) to get it installed and then follow [these instructions](https://docs.docker.com/engine/install/linux-postinstall/#manage-docker-as-a-non-root-user) to enable management of Docker without root. I also verify that my install is using `cgroup` version 1 as version 2 is not currently supported: ```shell docker info | grep -i cgroup # [tl! .cmd] Cgroup Driver: cgroupfs # [tl! .nocopy:1] Cgroup Version: 1 ``` #### `kubectl` binary Next up, I'll install `kubectl` [as described here](https://kubernetes.io/docs/tasks/tools/install-kubectl-linux/) - though the latest version is currently `1.23` and that won't work with the `1.21` control plane node image I downloaded from VMware (`kubectl` needs to be within one minor version of the control plane). Instead I need to find the latest `1.22` release. I can look at the [releases page on GithHub](https://github.com/kubernetes/kubernetes/releases) to see that the latest release for me is `1.22.5`. With this newfound knowledge I can follow the [Install kubectl binary with curl on Linux](https://kubernetes.io/docs/tasks/tools/install-kubectl-linux/#install-kubectl-binary-with-curl-on-linux) instructions to grab that specific version: ```shell curl -sLO https://dl.k8s.io/release/v1.22.5/bin/linux/amd64/kubectl # [tl! .cmd:1] sudo install -o root -g root -m 0755 kubectl /usr/local/bin/kubectl # [tl! .nocopy:2] [sudo] password for john: kubectl version --client # [tl! .cmd] Client Version: version.Info{Major:"1", Minor:"22", GitVersion:"v1.22.5", # [tl! .nocopy:3] GitCommit:"5c99e2ac2ff9a3c549d9ca665e7bc05a3e18f07e", GitTreeState:"clean", BuildDate:"2021-12-16T08:38:33Z", GoVersion:"go1.16.12", Compiler:"gc", Platform:"linux/amd64"} ``` #### `kind` binary It's not strictly a requirement, but having the `kind` executable available will be handy for troubleshooting during the bootstrap process in case anything goes sideways. It can be installed in basically the same was as `kubectl`: ```shell curl -sLo ./kind https://kind.sigs.k8s.io/dl/v0.11.1/kind-linux-amd64 # [tl! .cmd:2] sudo install -o root -g root -m 0755 kind /usr/local/bin/kind kind version kind v0.11.1 go1.16.5 linux/amd64 # [tl! .nocopy] ``` #### Tanzu CLI The final bit of required software is the Tanzu CLI, which can be downloaded from the [project on GitHub](https://github.com/vmware-tanzu/community-edition/releases). ```shell curl -H "Accept: application/vnd.github.v3.raw" \ # [tl! .cmd] -L https://api.github.com/repos/vmware-tanzu/community-edition/contents/hack/get-tce-release.sh | \ bash -s v0.9.1 linux ``` And then unpack it and run the installer: ```shell tar xf tce-linux-amd64-v0.9.1.tar.gz # [tl! .cmd:2] cd tce-linux-amd64-v0.9.1 ./install.sh ``` I can then verify the installation is working correctly: ```shell tanzu version # [tl! .cmd] version: v0.2.1 # [tl! .nocopy:2] buildDate: 2021-09-29 sha: ceaa474 ``` ### Cluster creation Okay, now it's time for the good stuff - creating some shiny new Tanzu clusters! The Tanzu CLI really does make this very easy to accomplish. #### Management cluster I need to create a Management cluster first and I'd like to do that with the UI, so that's as simple as: ```shell tanzu management-cluster create --ui # [tl! .cmd] ``` I should then be able to access the UI by pointing a web browser at `http://127.0.0.1:8080`... but I'm running this on a VM without a GUI, so I'll need to back up and tell it to bind on `0.0.0.0:8080` so the web installer will be accessible across the network. I can also include `--browser none` so that the installer doesn't bother with trying to launch a browser locally. ```shell tanzu management-cluster create --ui --bind 0.0.0.0:8080 --browser none # [tl! .cmd] # [tl! .nocopy:2] Validating the pre-requisites... Serving kickstart UI at http://[::]:8080 ``` *Now* I can point my local browser to my VM and see the UI: ![The Tanzu Installer UI](installer_ui.png) And then I can click the button at the bottom left to save my eyes[^dark_mode] before selecting the option to deploy on vSphere. ![Configuring the IaaS Provider](installer_iaas_provider.png) I'll plug in the FQDN of my vCenter and provide a username and password to use to connect to it, then hit the **Connect** button. That will prompt me to accept the vCenter's certificate thumbprint, and then I'll be able to select the virtual datacenter that I want to use. Finally, I'll paste in the SSH public key[^gen_key] I'll use for interacting with the cluster. I click **Next** and move on to the Management Cluster Settings. ![Configuring the Management Cluster](installer_management_cluster.png) This is for a lab environment that's fairly memory-constrained, so I'll pick the single-node *Development* setup with a *small* instance type. I'll name the cluster `tce-mgmt` and stick with the default `kube-vip` control plane endpoint provider. I plug in the control plane endpoint IP that I'll use for connecting to the cluster and select the *small* instance type for the worker node type. I don't have an NSX Advanced Load Balancer or any Metadata to configure so I'll skip past those steps and move on to configuring the Resources. ![Configuring Resources](installer_resources.png) Here I pick to place the Tanzu-related resources in a VM folder named `Tanzu`, to store their data on my single host's single datastore, and to deploy to the one-host `physical-cluster` cluster. Now for the Kubernetes Networking Settings: ![Configuring Kubernetes Networking](installer_k8s_networking.png) This bit is actually pretty easy. For Network Name, I select the vSphere network where the `192.168.1.0/24` network I identified earlier lives, `d-Home-Mgmt`. I leave the service and pod CIDR ranges as default. I disable the Identity Management option and then pick the `k8s-node` template I had imported to vSphere earlier. ![Configuring the OS Image](installer_image.png) I skip the Tanzu Mission Control piece (since I'm still waiting on access to [TMC Starter](https://tanzu.vmware.com/tmc-starter)) and click the **Review Configuration** button at the bottom of the screen to review my selections. ![Reviewing the configuration](installer_review.png) See the option at the bottom to copy the CLI command? I'll need to use that since clicking the friendly **Deploy** button doesn't seem to work while connected to the web server remotely. ```shell tanzu management-cluster create \ # [tl! .cmd] --file /home/john/.config/tanzu/tkg/clusterconfigs/dr94t3m2on.yaml -v 6 ``` In fact, I'm going to copy that file into my working directory and give it a more descriptive name so that I can re-use it in the future. ```shell cp ~/.config/tanzu/tkg/clusterconfigs/dr94t3m2on.yaml \ # [tl! .cmd] ~/projects/tanzu-homelab/tce-mgmt.yaml ``` Now I can run the install command: ```shell tanzu management-cluster create --file ./tce-mgmt.yaml -v 6 # [tl! .cmd] ``` After a moment or two of verifying prerequisites, I'm met with a polite offer to enable Tanzu Kubernetes Grid Service in vSphere: ``` vSphere 7.0 Environment Detected. You have connected to a vSphere 7.0 environment which does not have vSphere with Tanzu enabled. vSphere with Tanzu includes an integrated Tanzu Kubernetes Grid Service which turns a vSphere cluster into a platform for running Kubernetes workloads in dedicated resource pools. Configuring Tanzu Kubernetes Grid Service is done through vSphere HTML5 client. Tanzu Kubernetes Grid Service is the preferred way to consume Tanzu Kubernetes Grid in vSphere 7.0 environments. Alternatively you may deploy a non-integrated Tanzu Kubernetes Grid instance on vSphere 7.0. Note: To skip the prompts and directly deploy a non-integrated Tanzu Kubernetes Grid instance on vSphere 7.0, you can set the 'DEPLOY_TKG_ON_VSPHERE7' configuration variable to 'true' Do you want to configure vSphere with Tanzu? [y/N]: n Would you like to deploy a non-integrated Tanzu Kubernetes Grid management cluster on vSphere 7.0? [y/N]: y ``` That's not what I'm after in this case, though, so I'll answer with a `n` and a `y` to confirm that I want the non-integrated TKG deployment. And now I go get coffee as it'll take 10-15 minutes for the deployment to complete. ![Coffee break!](coffee_break.gif) Okay, I'm back - and so is my shell prompt! The deployment completed successfully: ``` Waiting for additional components to be up and running... Waiting for packages to be up and running... Context set for management cluster tce-mgmt as 'tce-mgmt-admin@tce-mgmt'. Management cluster created! You can now create your first workload cluster by running the following: tanzu cluster create [name] -f [file] Some addons might be getting installed! Check their status by running the following: kubectl get apps -A ``` I can run that last command to go ahead and verify that the addon installation has completed: ```shell kubectl get apps -A # [tl! .cmd] NAMESPACE NAME DESCRIPTION SINCE-DEPLOY AGE # [tl! .nocopy:5] tkg-system antrea Reconcile succeeded 26s 6m49s tkg-system metrics-server Reconcile succeeded 36s 6m49s tkg-system tanzu-addons-manager Reconcile succeeded 22s 8m54s tkg-system vsphere-cpi Reconcile succeeded 19s 6m50s tkg-system vsphere-csi Reconcile succeeded 36s 6m50s ``` And I can use the Tanzu CLI to get some other details about the new management cluster: ```shell tanzu management-cluster get tce-mgmt # [tl! .cmd] NAME NAMESPACE STATUS CONTROLPLANE WORKERS KUBERNETES ROLES # [tl! .nocopy:start] tce-mgmt tkg-system running 1/1 1/1 v1.21.2+vmware.1 management Details: NAME READY SEVERITY REASON SINCE MESSAGE /tce-mgmt True 40m ├─ClusterInfrastructure - VSphereCluster/tce-mgmt True 41m ├─ControlPlane - KubeadmControlPlane/tce-mgmt-control-plane True 40m │ └─Machine/tce-mgmt-control-plane-xtdnx True 40m └─Workers └─MachineDeployment/tce-mgmt-md-0 └─Machine/tce-mgmt-md-0-745b858d44-4c9vv True 40m Providers: NAMESPACE NAME TYPE PROVIDERNAME VERSION WATCHNAMESPACE capi-kubeadm-bootstrap-system bootstrap-kubeadm BootstrapProvider kubeadm v0.3.23 capi-kubeadm-control-plane-system control-plane-kubeadm ControlPlaneProvider kubeadm v0.3.23 capi-system cluster-api CoreProvider cluster-api v0.3.23 capv-system infrastructure-vsphere InfrastructureProvider vsphere v0.7.10 # [tl! .nocopy:end] ``` Excellent! Things are looking good so I can move on to create the cluster which will actually run my workloads. [^dark_mode]: Enabling dark mode is probably the most important part of this process. [^gen_key]: If I didn't already have a key pair to use I would generate one with `ssh-keygen -t rsa -b 4096 -C "email@example.com"` and add it to my client with `ssh-add ~/.ssh/id_rsa`. #### Workload cluster I won't use the UI for this but will instead take a copy of my `tce-mgmt.yaml` file and adapt it to suit the workload needs (as described [here](https://tanzucommunityedition.io/docs/latest/workload-clusters/)). ```shell cp tce-mgmt.yaml tce-work.yaml # [tl! .cmd:1] vi tce-work.yaml ``` I only need to change 2 of the parameters in this file: - `CLUSTER_NAME`: from `tce-mgmt` to `tce-work` - `VSPHERE_CONTROL_PLANE_ENDPOINT`: from `192.168.1.60` to `192.168.1.61` I *could* change a few others if I wanted to[^i_wont]: - (Optional) `CLUSTER_PLAN` to change between `dev`/`prod` plans independently - (Optional) `CONTROL_PLANE_MACHINE_COUNT` to deploy an increased number of control plane nodes (must but an odd integer) - (Optional) `WORKER_MACHINE_COUNT` to add worker nodes - (Optional) `NAMESPACE` to deploy the cluster in a specific Kubernetes namespace - (Optional) `OS_NAME` and `OS_VERSION` to use a different machine image for the workload cluster After saving my changes to the `tce-work.yaml` file, I'm ready to deploy the cluster: ```shell tanzu cluster create --file tce-work.yaml # [tl! .cmd] Validating configuration... # [tl! .nocopy:start] Warning: Pinniped configuration not found. Skipping pinniped configuration in workload cluster. Please refer to the documentation to check if you can configure pinniped on workload cluster manually Creating workload cluster 'tce-work'... Waiting for cluster to be initialized... Waiting for cluster nodes to be available... Waiting for addons installation... Waiting for packages to be up and running... Workload cluster 'tce-work' created # [tl! .nocopy:end] ``` Right on! I'll use `tanzu cluster get` to check out the workload cluster: ```shell tanzu cluster get tce-work # [tl! .cmd] NAME NAMESPACE STATUS CONTROLPLANE WORKERS KUBERNETES ROLES # [tl! .nocopy:start] tce-work default running 1/1 1/1 v1.21.2+vmware.1 ℹ Details: NAME READY SEVERITY REASON SINCE MESSAGE /tce-work True 9m31s ├─ClusterInfrastructure - VSphereCluster/tce-work True 10m ├─ControlPlane - KubeadmControlPlane/tce-work-control-plane True 9m31s │ └─Machine/tce-work-control-plane-8km9m True 9m31s └─Workers └─MachineDeployment/tce-work-md-0 └─Machine/tce-work-md-0-687444b744-cck4x True 8m31s # [tl! .nocopy:end] ``` I can also go into vCenter and take a look at the VMs which constitute the two clusters: ![Cluster VMs](clusters_in_vsphere.png) I've highlighted the two Control Plane nodes. They got their IP addresses assigned by DHCP, but [VMware says](https://tanzucommunityedition.io/docs/latest/verify-deployment/#configure-dhcp-reservations-for-the-control-plane-nodes-vsphere-only) that I need to create reservations for them to make sure they don't change. So I'll do just that. ![DHCP reservations on Google Wifi](dhcp_reservations.png) Excellent, I've got a Tanzu management cluster and a Tanzu workload cluster. What now? [^i_wont]: I'm not going to, but I totally could. ### Working with Tanzu If I run `kubectl get nodes` right now, I'll only get information about the management cluster: ```shell kubectl get nodes # [tl! .cmd] NAME STATUS ROLES AGE VERSION # [tl! .nocopy:2] tce-mgmt-control-plane-xtdnx Ready control-plane,master 18h v1.21.2+vmware.1 tce-mgmt-md-0-745b858d44-4c9vv Ready 17h v1.21.2+vmware.1 ``` #### Setting the right context To be able to deploy stuff to the workload cluster, I need to tell `kubectl` how to talk to it. And to do that, I'll first need to use `tanzu` to capture the cluster's kubeconfig: ```shell tanzu cluster kubeconfig get tce-work --admin # [tl! .cmd] Credentials of cluster 'tce-work' have been saved # [tl! .nocopy:1] You can now access the cluster by running 'kubectl config use-context tce-work-admin@tce-work' ``` I can now run `kubectl config get-contexts` and see that I have access to contexts on both management and workload clusters: ```shell kubectl config get-contexts # [tl! .cmd] CURRENT NAME CLUSTER AUTHINFO NAMESPACE # [tl! .nocopy:2] * tce-mgmt-admin@tce-mgmt tce-mgmt tce-mgmt-admin tce-work-admin@tce-work tce-work tce-work-admin ``` And I can switch to the `tce-work` cluster like so: ```shell kubectl config use-context tce-work-admin@tce-work # [tl! .cmd] Switched to context "tce-work-admin@tce-work". # [tl! .nocopy] kubectl get nodes # [tl! .cmd] NAME STATUS ROLES AGE VERSION # [tl! .nocopy:2] tce-work-control-plane-8km9m Ready control-plane,master 17h v1.21.2+vmware.1 tce-work-md-0-687444b744-cck4x Ready 17h v1.21.2+vmware.1 ``` There they are! #### Deploying the `yelb` demo app Before I move on to deploying actually *useful* workloads, I'll start with deploying a quick demo application as described in William Lam's post on [Interesting Kubernetes application demos](https://williamlam.com/2020/06/interesting-kubernetes-application-demos.html). `yelb` is a web app which consists of a UI front end, application server, database server, and Redis caching service so it's a great little demo to make sure Kubernetes is working correctly. I can check out the sample deployment that William put together [here](https://github.com/lamw/vmware-k8s-app-demo/blob/master/yelb.yaml), and then deploy it with: ```shell kubectl create ns yelb # [tl! .cmd] namespace/yelb created # [tl! .nocopy:1] kubectl apply -f https://raw.githubusercontent.com/lamw/vmware-k8s-app-demo/master/yelb.yaml # [tl! .cmd] service/redis-server created # [tl! .nocopy:start] service/yelb-db created service/yelb-appserver created service/yelb-ui created deployment.apps/yelb-ui created deployment.apps/redis-server created deployment.apps/yelb-db created deployment.apps/yelb-appserver created # [tl! .nocopy:end] kubectl -n yelb get pods # [tl! .cmd] NAME READY STATUS RESTARTS AGE # [tl! .nocopy:4] redis-server-74556bbcb7-r9jqc 1/1 Running 0 10s yelb-appserver-d584bb889-2jspg 1/1 Running 0 10s yelb-db-694586cd78-wb8tt 1/1 Running 0 10s yelb-ui-8f54fd88c-k2dw9 1/1 Running 0 10s ``` Once the app is running, I can point my web browser at it to see it in action. But what IP do I use? ```shell kubectl -n yelb get svc/yelb-ui # [tl! .cmd] NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE # [tl! .nocopy:1] yelb-ui NodePort 100.71.228.116 80:30001/TCP 84s ``` This demo is using a `NodePort` type service to expose the front end, which means it will be accessible on port `30001` on the node it's running on. I can find that IP by: ```shell kubectl -n yelb describe pod $(kubectl -n yelb get pods | grep yelb-ui | awk '{print $1}') | grep "Node:" # [tl! .cmd] Node: tce-work-md-0-687444b744-cck4x/192.168.1.145 # [tl! .nocopy] ``` So I can point my browser at `http://192.168.1.145:30001` and see the demo: ![yelb demo page](yelb_nodeport_demo.png) After marveling at my own magnificence[^magnificence] for a few minutes, I'm ready to move on to something more interesting - but first, I'll just delete the `yelb` namespace to clean up the work I just did: ```shell kubectl delete ns yelb # [tl! .cmd] namespace "yelb" deleted # [tl! .nocopy] ``` Now let's move on and try to deploy `yelb` behind a `LoadBalancer` service so it will get its own IP. William has a [deployment spec](https://github.com/lamw/vmware-k8s-app-demo/blob/master/yelb-lb.yaml) for that too. ```shell kubectl create ns yelb # [tl! .cmd] namespace/yelb created # [tl! .nocopy:1] kubectl apply -f https://raw.githubusercontent.com/lamw/vmware-k8s-app-demo/master/yelb-lb.yaml # [tl! .cmd] service/redis-server created # [tl! .nocopy:8] service/yelb-db created service/yelb-appserver created service/yelb-ui created deployment.apps/yelb-ui created deployment.apps/redis-server created deployment.apps/yelb-db created deployment.apps/yelb-appserver created kubectl -n yelb get pods # [tl! .cmd] NAME READY STATUS RESTARTS AGE # [tl! .nocopy:4] redis-server-74556bbcb7-q6l62 1/1 Running 0 7s yelb-appserver-d584bb889-p5qgd 1/1 Running 0 7s yelb-db-694586cd78-hjtn4 1/1 Running 0 7s yelb-ui-8f54fd88c-pm9qw 1/1 Running 0 7s ``` And I can take a look at that service... ```shell kubectl -n yelb get svc/yelb-ui # [tl! .cmd] NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE # [tl! .nocopy:1] yelb-ui LoadBalancer 100.67.177.185 80:32339/TCP 15s ``` Wait a minute. That external IP is *still* ``. What gives? Oh yeah I need to actually deploy and configure a load balancer before I can balance anything. That's up next. [^magnificence]: Mr. Anderson. #### Deploying `kube-vip` as a load balancer Fortunately, William Lam [wrote up some tips](https://williamlam.com/2021/10/quick-tip-install-kube-vip-as-service-load-balancer-with-tanzu-community-edition-tce.html) for handling that too. It's [based on work by Scott Rosenberg](https://github.com/vrabbi/tkgm-customizations). The quick-and-dirty steps needed to make this work are: ```shell git clone https://github.com/vrabbi/tkgm-customizations.git # [tl! .cmd:3] cd tkgm-customizations/carvel-packages/kube-vip-package kubectl apply -n tanzu-package-repo-global -f metadata.yml kubectl apply -n tanzu-package-repo-global -f package.yaml cat << EOF > values.yaml # [tl! .cmd] vip_range: 192.168.1.64-192.168.1.80 EOF tanzu package install kubevip -p kubevip.terasky.com -v 0.3.9 -f values.yaml # [tl! .cmd] ``` Now I can check out the `yelb-ui` service again: ```shell kubectl -n yelb get svc/yelb-ui # [tl!.cmd] NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE # [tl! .nocopy:1] yelb-ui LoadBalancer 100.67.177.185 192.168.1.65 80:32339/TCP 4h35m ``` And it's got an IP! I can point my browser to `http://192.168.1.65` now and see: ![Successful LoadBalancer test!](yelb_loadbalancer_demo.png) I'll keep the `kube-vip` load balancer since it'll come in handy, but I have no further use for `yelb`: ```shell kubectl delete ns yelb # [tl! .cmd] namespace "yelb" deleted # [tl! .nocopy] ``` #### Persistent Volume Claims, Storage Classes, and Storage Policies At some point, I'm going to want to make sure that data from my Tanzu workloads stick around persistently - and for that, I'll need to [define some storage stuff](https://tanzucommunityedition.io/docs/latest/vsphere-cns/). First up, I'll add a new tag called `tkg-storage-local` to the `nuchost-local` vSphere datastore that I want to use for storing Tanzu volumes: ![Tag (and corresponding category) applied ](storage_tag.png) Then I create a new vSphere Storage Policy called `tkg-storage-policy` which states that data covered by the policy should be placed on the datastore(s) tagged with `tkg-storage-local`: ![My Tanzu storage policy](storage_policy.png) So that's the vSphere side of things sorted; now to map that back to the Kubernetes side. For that, I'll need to define a Storage Class tied to the vSphere Storage profile so I drop these details into a new file called `vsphere-sc.yaml`: ```yaml # torchlight! {"lineNumbers": true} kind: StorageClass apiVersion: storage.k8s.io/v1 metadata: name: vsphere provisioner: csi.vsphere.vmware.com parameters: storagePolicyName: tkg-storage-policy ``` And then apply it with : ```shell kubectl apply -f vsphere-sc.yaml # [tl! .cmd] storageclass.storage.k8s.io/vsphere created # [tl! .nocopy] ``` I can test that I can create a Persistent Volume Claim against the new `vsphere` Storage Class by putting this in a new file called `vsphere-pvc.yaml`: ```yaml # torchlight! {"lineNumbers": true} apiVersion: v1 kind: PersistentVolumeClaim metadata: labels: name: vsphere-demo-1 name: vsphere-demo-1 spec: accessModes: - ReadWriteOnce storageClassName: vsphere resources: requests: storage: 5Gi ``` And applying it: ```shell kubectl apply -f demo-pvc.yaml # [tl! .cmd] persistentvolumeclaim/vsphere-demo-1 created # [tl! .nocopy] ``` I can see the new claim, and confirm that its status is `Bound`: ```shell kubectl get pvc # [tl! .cmd] NAME STATUS VOLUME CAPACITY ACCESS MODES STORAGECLASS AGE # [tl! .nocopy:1] vsphere-demo-1 Bound pvc-36cc7c01-a1b3-4c1c-ba0d-dff3fd47f93b 5Gi RWO vsphere 4m25s ``` And for bonus points, I can see that the container volume was created on the vSphere side: ![Container Volume in vSphere](container_volume_in_vsphere.png) So that's storage sorted. I'll clean up my test volume before moving on: ```shell kubectl delete -f demo-pvc.yaml # [tl! .cmd] persistentvolumeclaim "vsphere-demo-1" deleted # [tl! .nocopy] ``` ### A real workload - phpIPAM Demos are all well and good, but how about a real-world deployment to tie it all together? I've been using a [phpIPAM instance for assigning static IP addresses for my vRealize Automation deployments](/integrating-phpipam-with-vrealize-automation-8/), but have *only* been using it to monitor IP usage within the network ranges to which vRA will provision machines. I recently decided that I'd like to expand phpIPAM's scope so it can keep an eye on *all* the network ranges within the environment. That's not a big ask in [my little self-contained homelab](/vmware-home-lab-on-intel-nuc-9/), but having a single system scanning all the ranges of a large production network probably wouldn't scale too well. Fortunately the phpIPAM project provides a [remote scanning agent](https://github.com/phpipam/phpipam-agent) which can be used for keeping an eye on networks and reporting back to the main phpIPAM server. With this, I could deploy an agent to each region (or multiple agents to a region!) and divide up the network into chunks that each agent would be responsible for scanning. But that's a pretty lightweight task for a single server to manage, and who wants to deal with configuring multiple instances of the same thing? Not this guy. So I set to work exploring some containerization options, and I found [phpipam-docker](https://github.com/phpipam-docker/phpipam-docker). That would easily replicate my existing setup in a trio of containers (one for the web front-end, one for the database back-end, and one with `cron` jobs to run scans at regular intervals)... but doesn't provide a remote scan capability. I also found a [dockerized phpipam-agent](https://github.com/pierrecdn/phpipam-agent), but this one didn't quite meet my needs. It did provide me a base to work off of though so a few days of [tinkering](https://github.com/jbowdre/phpipam-agent-docker) resulted in me publishing my first [Docker image](https://github.com/jbowdre/phpipam-agent-docker/pkgs/container/phpipam-agent). I've still some work to do before this application stack is fully ready for production but it's at a point where I think it's worth doing a test deploy. To start, I'll create a new namespace to keep things tidy: ```shell kubectl create ns ipam # [tl! .cmd] namespace/ipam created # [tl! .nocopy] ``` I'm going to wind up with four pods: - `phpipam-db` for the database back-end - `phpipam-www` for the web front-end - `phpipam-cron` for the local cron jobs, which will be largely but not completely[^dns_scans] replaced by: - `phpipam-agent` for my remote scan agent I'll use each container's original `docker-compose` configuration and adapt that into something I can deploy on Kubernetes. [^dns_scans]: My `phpipam-agent` image won't (yet?) do the DNS lookups that `phpipam-cron` can. #### phpipam-db The phpIPAM database will live inside a MariaDB container. Here's the relevant bit from `docker-compose`: ```yaml # torchlight! {"lineNumbers": true} services: phpipam-db: image: mariadb:latest ports: - "3306:3306" environment: - MYSQL_ROOT_PASSWORD=VMware1!VMWare1! volumes: - phpipam-db-data:/var/lib/mysql ``` So it will need a `Service` exposing the container's port `3306` so that other pods can connect to the database. For my immediate demo, using `type: ClusterIP` will be sufficient since all the connections will be coming from within the cluster. When I do this for real, it will need to be `type: LoadBalancer` so that the agent running on a different cluster can connect. And it will need a `PersistentVolumeClaim` so it can store the database data at `/var/lib/mysql`. It will also get passed an environment variable to set the initial `root` password on the database instance (which will be used later during the phpIPAM install to create the initial `phpipam` database). It might look like this on the Kubernetes side: ```yaml # torchlight! {"lineNumbers": true} # phpipam-db.yaml apiVersion: v1 kind: Service metadata: name: phpipam-db labels: app: phpipam-db namespace: ipam spec: type: ClusterIP ports: - name: mysql port: 3306 protocol: TCP targetPort: 3306 selector: app: phpipam-db --- apiVersion: v1 kind: PersistentVolumeClaim metadata: labels: name: phpipam-db name: phpipam-db-pvc namespace: ipam spec: accessModes: - ReadWriteOnce storageClassName: vsphere resources: requests: storage: 5Gi --- apiVersion: apps/v1 kind: Deployment metadata: name: phpipam-db namespace: ipam spec: selector: matchLabels: app: phpipam-db replicas: 1 template: metadata: labels: app: phpipam-db spec: containers: - name: phpipam-db image: mariadb:latest env: - name: MYSQL_ROOT_PASSWORD value: "VMware1!VMware1!" ports: - name: mysql containerPort: 3306 volumeMounts: - name: phpipam-db-vol mountPath: /var/lib/mysql volumes: - name: phpipam-db-vol persistentVolumeClaim: claimName: phpipam-db-pvc ``` Moving on: #### phpipam-www This is the `docker-compose` excerpt for the web component: ```yaml # torchlight! {"lineNumbers": true} services: phpipam-web: image: phpipam/phpipam-www:1.5x ports: - "80:80" environment: - TZ=UTC - IPAM_DATABASE_HOST=phpipam-db - IPAM_DATABASE_PASS=VMware1! - IPAM_DATABASE_WEBHOST=% volumes: - phpipam-logo:/phpipam/css/images/logo ``` Based on that, I can see that my `phpipam-www` pod will need a container running the `phpipam/phpipam-www:1.5x` image, a `Service` of type `LoadBalancer` to expose the web interface on port `80`, a `PersistentVolumeClaim` mounted to `/phpipam/css/images/logo`, and some environment variables passed in to configure the thing. Note that the `IPAM_DATABASE_PASS` variable defines the password used for the `phpipam` user on the database (not the `root` user referenced earlier), and the `IPAM_DATABASE_WEBHOST=%` variable will define which hosts that `phpipam` database user will be able to connect from; setting it to `%` will make sure that my remote agent can connect to the database even if I don't know where the agent will be running. Here's how I'd adapt that into a structure that Kubernetes will understand: ```yaml # torchlight! {"lineNumbers": true} # phpipam-www.yaml apiVersion: v1 kind: Service metadata: name: phpipam-www labels: app: phpipam-www namespace: ipam spec: type: LoadBalancer ports: - name: http port: 80 protocol: TCP targetPort: 80 selector: app: phpipam-www --- apiVersion: v1 kind: PersistentVolumeClaim metadata: labels: name: phpipam-www name: phpipam-www-pvc namespace: ipam spec: accessModes: - ReadWriteOnce storageClassName: vsphere resources: requests: storage: 100Mi --- apiVersion: apps/v1 kind: Deployment metadata: name: phpipam-www namespace: ipam spec: selector: matchLabels: app: phpipam-www replicas: 1 template: metadata: labels: app: phpipam-www spec: containers: # [tl! focus:2] - name: phpipam-www image: phpipam/phpipam-www:1.5x env: - name: TZ value: "UTC" - name: IPAM_DATABASE_HOST value: "phpipam-db" - name: IPAM_DATABASE_PASS value: "VMware1!" - name: IPAM_DATABASE_WEBHOST value: "%" ports: - containerPort: 80 volumeMounts: - name: phpipam-www-vol mountPath: /phpipam/css/images/logo volumes: - name: phpipam-www-vol persistentVolumeClaim: claimName: phpipam-www-pvc ``` #### phpipam-cron This container has a pretty simple configuration in `docker-compose`: ```yaml # torchlight! {"lineNumbers": true} services: phpipam-cron: image: phpipam/phpipam-cron:1.5x environment: - TZ=UTC - IPAM_DATABASE_HOST=phpipam-db - IPAM_DATABASE_PASS=VMware1! - SCAN_INTERVAL=1h ``` No exposed ports, no need for persistence - just a base image and a few variables to tell it how to connect to the database and how often to run the scans: ```yaml # torchlight! {"lineNumbers": true} # phpipam-cron.yaml apiVersion: apps/v1 kind: Deployment metadata: name: phpipam-cron namespace: ipam spec: selector: matchLabels: app: phpipam-cron replicas: 1 template: metadata: labels: app: phpipam-cron spec: containers: - name: phpipam-cron image: phpipam/phpipam-cron:1.5x env: - name: IPAM_DATABASE_HOST value: "phpipam-db" - name: IPAM_DATABASE_PASS value: "VMWare1!" - name: SCAN_INTERVAL value: "1h" - name: TZ value: "UTC" ``` #### phpipam-agent And finally, my remote scan agent. Here's the `docker-compose`: ```yaml # torchlight! {"lineNumbers": true} services: phpipam-agent: container_name: phpipam-agent restart: unless-stopped image: ghcr.io/jbowdre/phpipam-agent:latest environment: - IPAM_DATABASE_HOST=phpipam-db - IPAM_DATABASE_NAME=phpipam - IPAM_DATABASE_USER=phpipam - IPAM_DATABASE_PASS=VMware1! - IPAM_DATABASE_PORT=3306 - IPAM_AGENT_KEY= - IPAM_SCAN_INTERVAL=5m - IPAM_RESET_AUTODISCOVER=true - IPAM_REMOVE_DHCP=true - TZ=UTC ``` It's got a few additional variables to make it extra-configurable, but still no need for persistence or network exposure. That `IPAM_AGENT_KEY` variable will need to get populated the appropriate key generated within the new phpIPAM deployment, but we can deal with that later. For now, here's how I'd tell Kubernetes about it: ```yaml # torchlight! {"lineNumbers": true} # phpipam-agent.yaml apiVersion: apps/v1 kind: Deployment metadata: name: phpipam-agent namespace: ipam spec: selector: matchLabels: app: phpipam-agent replicas: 1 template: metadata: labels: app: phpipam-agent spec: containers: - name: phpipam-agent image: ghcr.io/jbowdre/phpipam-agent:latest env: - name: IPAM_DATABASE_HOST value: "phpipam-db" - name: IPAM_DATABASE_NAME value: "phpipam" - name: IPAM_DATABASE_USER value: "phpipam" - name: IPAM_DATABASE_PASS value: "VMware1!" - name: IPAM_DATABASE_PORT value: "3306" - name: IPAM_AGENT_KEY value: "" - name: IPAM_SCAN_INTERVAL value: "5m" - name: IPAM_RESET_AUTODISCOVER value: "true" - name: IPAM_REMOVE_DHCP value: "true" - name: TZ value: "UTC" ``` #### Deployment and configuration of phpIPAM I can now go ahead and start deploying these containers, starting with the database one (upon which all the others rely): ```shell kubectl apply -f phpipam-db.yaml # [tl! .cmd] service/phpipam-db created # [tl! .nocopy:2] persistentvolumeclaim/phpipam-db-pvc created deployment.apps/phpipam-db created ``` And the web server: ```shell kubectl apply -f phpipam-www.yaml # [tl! .cmd] service/phpipam-www created # [tl! .nocopy:2] persistentvolumeclaim/phpipam-www-pvc created deployment.apps/phpipam-www created ``` And the cron runner: ```shell kubectl apply -f phpipam-cron.yaml # [tl! .cmd] deployment.apps/phpipam-cron created # [tl! .nocopy] ``` I'll hold off on the agent container for now since I'll need to adjust the configuration slightly after getting phpIPAM set up, but I will go ahead and check out my work so far: ```shell kubectl -n ipam get all # [tl! .cmd] NAME READY STATUS RESTARTS AGE # [tl! .nocopy:start] pod/phpipam-cron-6c994897c4-6rsnp 1/1 Running 0 4m30s pod/phpipam-db-5f4c47d4b9-sb5bd 1/1 Running 0 16m pod/phpipam-www-769c95c68d-94klg 1/1 Running 0 5m59s NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE service/phpipam-db ClusterIP 100.66.194.69 3306/TCP 16m service/phpipam-www LoadBalancer 100.65.232.238 192.168.1.64 80:31400/TCP 5m59s NAME READY UP-TO-DATE AVAILABLE AGE deployment.apps/phpipam-cron 1/1 1 1 4m30s deployment.apps/phpipam-db 1/1 1 1 16m deployment.apps/phpipam-www 1/1 1 1 5m59s NAME DESIRED CURRENT READY AGE replicaset.apps/phpipam-cron-6c994897c4 1 1 1 4m30s replicaset.apps/phpipam-db-5f4c47d4b9 1 1 1 16m replicaset.apps/phpipam-www-769c95c68d 1 1 1 5m59s # [tl! .nocopy:end] ``` And I can point my browser to the `EXTERNAL-IP` associated with the `phpipam-www` service to see the initial setup page: ![phpIPAM installation page](phpipam_install_page.png) I'll click the **New phpipam installation** option to proceed to the next step: ![Database initialization options](phpipam_database_install_options.png) I'm all for easy so I'll opt for **Automatic database installation**, which will prompt me for the credentials of an account with rights to create a new database within the MariaDB instance. I'll enter `root` and the password I used for the `MYSQL_ROOT_PASSWORD` variable above: ![Automatic database install](phpipam_automatic_database_install.png) I click the **Install database** button and I'm then met with a happy success message saying that the `phpipam` database was successfully created. And that eventually gets me to the post-install screen, where I set an admin password and proceed to log in: ![We made it to the post-install!](phpipam_post_install.png) To create a new scan agent, I go to **Menu > Administration > Server management > Scan agents**. ![Scan agents screen](scan_agents.png) And click the button to create a new one: ![Creating a new agent](create_new_agent.png) I'll copy the agent code and plug it into my `phpipam-agent.yaml` file: ```yaml - name: IPAM_AGENT_KEY value: "4DC5GLo-F_35cy7BEPnGn7HivtjP_o-v" ``` And then deploy that: ```shell kubectl apply -f phpipam-agent.yaml # [tl! .cmd] deployment.apps/phpipam-agent created # [tl! .nocopy] ``` The scan agent isn't going to do anything until it's assigned to a subnet though, so now I head to **Administration > IP related management > Sections**. phpIPAM comes with a few default sections and ranges and such defined so I'll delete those and create a new one that I'll call `Lab`. ![Section management](section_management.png) Now I can create a new subnet within the `Lab` section by clicking the **Subnets** menu, selecting the `Lab` section, and clicking **+ Add subnet**. ![Empty subnets menu](subnets_empty.png) I'll define the new subnet as `192.168.1.0/24`. Once I enable the option to *Check hosts status*, I'll then be able to specify my new `remote-agent` as the scanner for this subnet. ![Creating a new subnet](creating_new_subnet.png) ![A new (but empty) subnet](new_subnet_pre_scan.png) It shows the scanner associated with the subnet, but no data yet. I'll need to wait a few minutes for the first scan to kick off (at the five-minute interval I defined in the configuration). ![GIF which says 'Five Minutes Later!'](five_minutes.gif) ![Newly discovered IPs!](newly-discovered_IPs.png) Woah, it actually works! ### Conclusion I still need to do more work to the containerized phpIPAM stack ready for production, but I'm feeling pretty good for having deployed a functional demo of it at this point! And working on this was a nice excuse to get a bit more familiar with Tanzu Community Edition specifically, Kubernetes in general, and Docker (I learned a ton while assembling the `phpipam-agent` image!). I find I always learn more about a new-to-me technology when I have an actual project to do rather than just going through the motions of a lab exercise. Maybe my notes will be useful to you, too.