WilliamLam.com

  • About
    • About
    • Privacy
  • VMware Cloud Foundation
  • VKS
  • Homelab
    • Resources
    • Nested Virtualization
  • VMware Nostalgia
  • Apple

Deploy Harbor in an Air-Gapped environment for Tanzu Kubernetes Grid (TKG)

04.24.2020 by William Lam // 1 Comment

When using Tanzu Kubernetes Grid (TKG) and the new TKG CLI, outbound internet connectivity is required as part of the initial setup on the machine running TKG CLI but also on the TKG Management Cluster which is automatically stood up as part of the deployment. For demo and testing purposes, this is usually not a problem but for anyone looking to run this in a Production or datacenter environment, direct internet access is generally not available.

TKG does support air-gapped environments today by requiring a private container registry that has been configured with all the required containers. Once your registry has been setup, you will also need to update the TKG YAML manifest files to specify your private registry as by default, it will point to registry.tkg.vmware.run. You can use any container registry that is supported with Kubernetes including the popular Harbor solution. One thing to note is that your private registry must have a proper signed SSL certificate, custom CA certificates or self-signed certificates are not officially supported today with TKG.

Since I recently had to set this up for a project I am working on, which I hope to talk about in a future blog post, I thought it would be useful to share the instructions on how to setup and configure Harbor to be used in-conjunction with TKG as well as any other solution that requires a container registry running in your own environment. In my deployment, I will be using Let's Encrypt for generating the required SSL certificate, but you can use any existing service for performing this operation. I will also be installing Harbor on Photon OS, but you can use any operating system of your choice that Harbor is supported on.


Pre-Requisites 

  • Access to a public DNS domain which you have ownership of (e.g. adding new records)
  • Access to your internal DNS server to add a custom DNS zone lookup entry (e.g. registry.<yourdomain>.com)

[Read more...]

Categories // Docker, Kubernetes, VMware Tanzu, vSphere Tags // Kubernetes, Tanzu Kubernetes Grid, TKG, TKG CLI

Automated vSphere 7 and vSphere with Kubernetes Lab Deployment Script

04.13.2020 by William Lam // 117 Comments

I know many of you have been asking me about my vSphere with Kubernetes automation script which I had been sharing snippets of on Twitter. For the past couple of weeks, I have been hard at work making the required changes between the vSphere 7 Beta and GA workflows, some additional testing and of course documentation. Hopefully the wait was worth it (I think it is) and if you enjoy the script or have benefited, please consider adding 🌟to the Github repo to show your support! Thanks and enjoy

Had to make some updates to one of my vGhetto Automated Lab Deployment Scripts

💥44min to automate all required #vSphere7 infrastructure! 🤛🎤🥳

1 x VCSA 7.0
3 x ESXi + vSAN 7.0
1 x NSX-T 3.0 UA
1 x NSX-T Edge

Need to clean up #ProjectPacific wording but its working great! pic.twitter.com/ZInPgVgbGS

— William Lam (@lamw.bsky.social | @*protected email*) (@lamw) April 4, 2020

The Github repository:

  • https://github.com/lamw/vghetto-vsphere-with-kubernetes-external-nsxt-automated-lab-deployment

Before getting started, please carefully read through the requirements section along with the complete sample end-to-end execution if you are new to vSphere with Kubernetes. You will need to have a VMware Cloud Foundation (VCF) 4.0 license before you can get started and specifically an NSX-T Advance license which is one of the required parameters within the script. If you do not have access to a VCF 4 license, I strongly recommend taking part in the recent VMUG Advantage Homelab Group Buy effort which I had started to easily get access to the latest VMware releases along with a nice 15% discount!

The script supports deploying both a standard vSphere 7 environment with just VCSA, ESXi and vSAN as well as the complete solution which includes NSX-T to support vSphere with Kubernetes. For more details, please refer to the FAQ.

Categories // Automation, Kubernetes, Nested Virtualization, NSX, VMware Tanzu, VSAN, vSphere, vSphere 7.0 Tags // Kubernetes, NSX-T, Project Pacific, VMware Cloud Foundation, vSphere 7.0, vSphere with Kubernetes

Integrating vCenter Event Broker Appliance (VEBA) with vRealize Orchestrator 

03.19.2020 by William Lam // Leave a Comment

VMware vRealize Orchestrator (vRO) is still used by many of our customers to automate various infrastructure tasks whether that is running it in standalone mode connected to a vCenter Server(s) or powering the extensibility of vRealize Automation. For customers who have already built an extensive collection of vRO workflows, it certainly makes sense that they would like to reuse the Automation they have already created.

For customers that want to easily subscribe to specific vCenter Server event(s) and trigger specific Automation workflows, the vCenter Event Broker Appliance (VEBA) Fling has been a very popular solution. Customers no longer have to write or manage the code to retrieve specific event(s) from vCenter Server, instead they can focus on the really interesting part of the workflow which is automation or "business logic" that does something specific with a given vCenter Server event. Some common examples including notification using Slack or SMS, vSphere automation such as applying a vSphere Tag or hardening a VM configuration to alerting using PagerDuty to ticket creation in ServiceNow or Salesforce.com, the list goes on and on.

The great news is that vRO customers can also benefit from VEBA while leveraging their existing vRO workflows. VEBA can integrate with any vRO workflow by simply having a function, written using any language of your choice, that calls into the vRO's REST API to trigger a specific vRO workflow. To demonstrate this integration, I have built VEBA function using PowerShell which executes a vSphere Tagging vRO workflow which can be found here and below is a video on how this works.

Categories // Automation, vRealize Suite, vSphere Tags // VEBA, VMware Event Broker Appliance, vrealize orchestrator, vRO

  • « Previous Page
  • 1
  • …
  • 37
  • 38
  • 39
  • 40
  • 41
  • …
  • 109
  • Next Page »

Search

Thank Author

Author

William is Distinguished Platform Engineering Architect in the VMware Cloud Foundation (VCF) Division at Broadcom. His primary focus is helping customers and partners build, run and operate a modern Private Cloud using the VMware Cloud Foundation (VCF) platform.

Connect

  • Bluesky
  • Email
  • GitHub
  • LinkedIn
  • Mastodon
  • Reddit
  • RSS
  • Twitter
  • Vimeo

Recent

  • VMUG Connect 2025 - Minimal VMware Cloud Foundation (VCF) 5.x in a Box  05/15/2025
  • Programmatically accessing the Broadcom Compatibility Guide (BCG) 05/06/2025
  • Quick Tip - Validating Broadcom Download Token  05/01/2025
  • Supported chipsets for the USB Network Native Driver for ESXi Fling 04/23/2025
  • vCenter Identity Federation with Authelia 04/16/2025

Advertisment

Privacy & Cookies: This site uses cookies. By continuing to use this website, you agree to their use.
To find out more, including how to control cookies, see here: Cookie Policy

Copyright WilliamLam.com © 2025

 

Loading Comments...