VMware Cloud on AWS

Running Microsoft SQL Server Big Data Clusters in Tanzu Kubernetes Grid on VMware Cloud on AWS – Part II

In Part I, we have already discussed running SQL Server Big Data Cluster that is managed by Tanzu Kubernetes Grid (TKGm) on VMware Cloud on AWS. In this blog post, we’ll continue exploring how to run SQL BDC in the Tanzu Kubernetes Grid Service (TKGs) that is managed by Workload Management (WCP) on VMware Cloud on AWS (Tanzu Kubernetes Grid with VMware Cloud on AWS).

How to Activate Tanzu Kubernetes Grid with VMware Cloud on AWS

The activation of Tanzu Kubernetes Grid in VMware Cloud on AWS is a per-cluster workflow. You may see the activation/deactivation option in the SDDC card.

You may refer to Activate Tanzu Kubernetes Grid in an SDDC Cluster documentation for detailed requirements and deployment steps. Several CIDR blocks for Tanzu workload control plane need to be defined during this process, as shown in the following picture. The “Validate and Proceed” button will help verify whether the CIDR blocks are ready to create the Supervisor cluster as the workload control plane.

Click the “Activate Tanzu Kubernetes Grid” button and a workflow will be initiated at the backend. The process may take 10-30 minutes to complete.

Create and Configure vSphere namespaces

Before you deploy the SQL BDC workloads, you need to create a vSphere namespace and define the necessary resources that are required to run the container workloads.

  1. Firstly, let’s create a namespace called “sqlbdc”.

2. Assign the namespace permission to “[email protected]” account.

3. Assign a storage policy to the namespace to allow dynamic provisioning of persistent volumes required by SQL BDC. We created a storage policy called “sqlbdc-storage-policy”.

4. Associate the namespace with a content library for ova images of Tanzu Kubernetes releases.

5. Add predefined or customized VM class to the namespace.

6. Finally the “sqlbdc” namespace may look like this:

Prepare a TKGs jumpbox as kubectl client to manage the namespace and workload

The workloads in the vSphere namespaces are managed through kubectl command line interface. To manage it in VMware Cloud on AWS, it is recommended to prepare a jumpbox VM as a kubectl client.

We deployed an Ubuntu 20.04 VM as a jumpbox and enabled public Internet access. You may refer to Configure the Bootstrap machine in VMC on AWS in the Part I blog post.

SSH to the jumpbox and download the kubectl vsphere tool.

Now let’s login to the “sqlbdc” namespace.

Make sure we are under the “sqlbdc” namespace context

Deploy a Tanzu Kubernetes Cluster using YAML file

Now we can proceed to deploy a Tanzu Kubernetes cluster as a workload cluster for SQL BDC. Verify the storage class:

Tanzu Kubernetes releases:

Create a sqlbdc-tkc.yaml file to provision the Tanzu Kubernetes cluster.

sqlbdc-tkc.yaml

Create the TKC and check the cluster status using the following command. The cluster should be up and running within several minutes.

Configure the Tanzu Kubernetes Cluster

To deploy SQL BDC workloads in Tanzu Kubernetes cluster, first login to the cluster using the kubectl vsphere command

Before deploying any workload in the TKC, create role bindings for pod security policy. For more details, refer to Example Role Bindings for Pod Security Policy

Install the following tools that are required by SQL BDC:

  1. Azure Data CLI – used to create and manage SQL BDC cluster
  2. Azure Data Studio – used to generate deployment profile of SQL BDC cluster

The SQL BDC deployment process requires network access to the Tanzu Kubernetes cluster nodes to check the service endpoint status, e.g. the SQL BDC control plane endpoint at port 30080. By default the Tanzu Kubernetes cluster has firewall rules enabled that will block the BDC endpoint port. You’ll probably get stuck halfway in the deployment process as shown below:

If you login to the cluster using azdata, it will report the following error message:

Here’s a workaround for a successful deployment of SQL BDC on VMware Cloud on AWS.

1. Add a network adapter on the TKGs jumpbox and connect to the same portgroup of the TKC cluster, e.g., “vnet-domain-c55:dff3902e-a4df-4a59-a496-5c93d27b9187-sqlbdc-sqlbdc-tkc-vnet-0”

2. Get the network subsets of the TKC Segment from NSX manager

3. Assign a static IP address that are within the same network range of the TKC node.

4. Follow the steps described in SSH to Tanzu Kubernetes Cluster Nodes as the System User Using a Password. SSH to each of the TKC node and add a firewall rule to allow the 30080 port.

Deploy SQL Server Big Data Clusters in Tanzu Kubernetes cluster

Now you can deploy the SQL BDC in Tanzu Kubernetes Cluster on VMware Cloud on AWS. The BDC cluster is deployed with Azure Data CLI and the deployment profile can be generated in Azure Data Studio.

Get the running pods of SQL BDC.

Get the SQL BDC endpoints as shown in the following picture. Besides the 30080 port we enabled previously to allow successful deployment of SQL BDC on VMware Cloud on AWS environment, you also need to enable the rest of the endpoint port (e.g. 30043 for Spark) on each of the TKC node to allow access.