Before you can deploy universes to Kubernetes using YugabyteDB Anywhere (YBA), you must create a provider configuration.

Prerequisites

To deploy YugabyteDB universes on Kubernetes, you need to provide your Kubernetes provider credentials. YBA uses those credentials to automatically provision and de-provision the pods that run YugabyteDB.

Before you create a Kubernetes provider, perform the following:

  • Create a yugabyte-platform-universe-management service account.
  • Create a kubeconfig file of the service account you created to configure access to the Kubernetes cluster.

Refer to To deploy nodes.

Configure Kubernetes

Navigate to Integrations > Infrastructure > Managed Kubernetes Service to see a list of all currently configured Kubernetes providers.

View and edit providers

To view a provider, select it in the list of Managed Kubernetes Service Configs to display the Overview.

To edit the provider, select Config Details, make changes, and click Apply Changes. For more information, refer to Provider settings. Note that, depending on whether the provider has been used to create a universe, you can only edit a subset of options.

To view the universes created using the provider, select Universes.

To delete the provider, click Actions and choose Delete Configuration. You can only delete providers that are not in use by a universe.

Create a provider

To create a Kubernetes provider:

  1. Click Create Config to open the Create Kubernetes Provider Configuration page.

    Create Kubernetes provider

  2. Enter the provider details. Refer to Provider settings.

    To fill the provider configuration values using the configuration of the same Kubernetes cluster that your instance of YBA is installed on, click Autofill local cluster config.

    Auto-fill Kubernetes provider

  3. Click Validate and Save Configuration when you are done and wait for the configuration to validate and complete.

    If you want to save your progress, you can skip validation by choosing the Ignore and save provider configuration anyway option, which saves the provider configuration without validating. Note that you may not be able to create universes using an incomplete or unvalidated provider.

Provider settings

Provider Name

Enter a Provider name. The Provider name is an internal tag used for organizing cloud providers.

Cloud Info

Choose the Kubernetes Provider Type.

In the Image Registry field, specify from where to pull the YugabyteDB image. Accept the default setting, unless you are hosting the registry, in which case refer to steps described in Pull and push YugabyteDB Docker images to private container registry.

Use Pull Secret to upload the pull secret to download the image of the Enterprise YugabyteDB that is in a private repository. Your Yugabyte sales representative should have provided this secret.

Use Kube Config to upload the kubeconfig file. If specified, this configuration file is used for all availability zones in all regions.

Alternately, you can define separate kubeconfig files for each zone when defining the regions. See Configure region and zones.

Configure region and zones

Continue configuring your Kubernetes provider by clicking Add region and completing the Add new region dialog as follows:

  1. Use the Region field to select the region.

  2. Click Add Zone and complete the corresponding portion of the dialog. Notice that there are might be multiple zones.

  3. Use the Zone Code field to select a zone label that should match the value of failure domain zone label on the nodes. For example, topology.kubernetes.io/zone would place the pods in that zone.

  4. Optionally, use Kube Config to upload the kubeconfig file. If this file is available at the provider level, you are not required to supply it.

  5. Optionally, use the Storage Classes field to enter a comma-delimited value. If you do not specify this value, it would default to standard. You need to ensure that this storage class exists in your Kubernetes cluster. Refer to Hardware requirements for pods.

  6. Optionally, use the Kube Pod Address Template field to enter the pod address template.

  7. Optionally, in the Kube Domain field, provide the DNS domain name used in the Kubernetes cluster.

  8. Use the Kube Namespace field to specify the namespace. If the provided service account has the Cluster Admin permissions, you are not required to complete this field. The service account used in the provided kubeconfig file should have access to this namespace.

  9. Complete the Overrides field using one of the provided options. If you do not specify anything, YBA uses defaults specified inside the Helm chart. For additional information, see Open source Kubernetes.

  10. If you are using Kubernetes cert-manager to manage TLS certificates, specify the issuer type and enter the issuer name. For more information, refer to Enable encryption in transit.

If required, add a new zone by clicking Add Zone, as your configuration may have multiple zones.

Click Add Region when you are done.

Overrides

The following overrides are available:

  • Overrides to add service-level annotations:

    serviceEndpoints:
      - name: "yb-master-service"
        type: "LoadBalancer"
        annotations:
          service.beta.kubernetes.io/aws-load-balancer-internal: "0.0.0.0/0"
        app: "yb-master"
        ports:
          ui: "7000"
    
      - name: "yb-tserver-service"
        type: "LoadBalancer"
        annotations:
          service.beta.kubernetes.io/aws-load-balancer-internal: "0.0.0.0/0"
        app: "yb-tserver"
        ports:
          ycql-port: "9042"
          ysql-port: "5433"
    
  • Overrides to disable LoadBalancer:

    enableLoadBalancer: False
    
  • Overrides to change the cluster domain name:

    domainName: my.cluster
    

    YugabyteDB servers and other components communicate with each other using the Kubernetes Fully Qualified Domain Names (FQDN). The default domain is cluster.local.

  • Overrides to add annotations at StatefulSet level:

    networkAnnotation:
      annotation1: 'foo'
      annotation2: 'bar'
    
  • Overrides to add custom resource allocation for YB-Master and YB-TServer pods:

    resource:
      master:
        requests:
          cpu: 2
          memory: 2Gi
        limits:
          cpu: 2
          memory: 2Gi
      tserver:
        requests:
          cpu: 2
          memory: 4Gi
        limits:
          cpu: 2
          memory: 4Gi
    

    This overrides instance types selected in the YBA universe creation flow.

  • Overrides to enable Istio compatibility:

    istioCompatibility:
      enabled: true
    

    This is required when Istio is used with Kubernetes.

  • Overrides to publish node IP as the server broadcast address.

    By default, YB-Master and YB-TServer pod fully-qualified domain names (FQDN) are used in the cluster as the server broadcast address. To publish the IP addresses of the nodes on which YB-TServer pods are deployed, add the following YAML to each zone override configuration:

    tserver:
      extraEnv:
      - name: NODE_IP
        valueFrom:
          fieldRef:
            fieldPath: status.hostIP
      serverBroadcastAddress: "$(NODE_IP)"
      affinity:
        podAntiAffinity:
          requiredDuringSchedulingIgnoredDuringExecution:
          - labelSelector:
              matchExpressions:
              - key: app
                operator: In
                values:
                - "yb-tserver"
            topologyKey: kubernetes.io/hostname
    
    # Required to esure that the Kubernetes FQDNs are used for
    # internal communication between the nodes and node-to-node
    # TLS certificates are validated correctly
    
    gflags:
      master:
        use_private_ip: cloud
      tserver:
        use_private_ip: cloud
    
    serviceEndpoints:
      - name: "yb-master-ui"
        type: LoadBalancer
        app: "yb-master"
        ports:
          http-ui: "7000"
    
      - name: "yb-tserver-service"
        type: NodePort
        externalTrafficPolicy: "Local"
        app: "yb-tserver"
        ports:
          tcp-yql-port: "9042"
          tcp-ysql-port: "5433"
    
  • Overrides to run YugabyteDB as a non-root user:

    podSecurityContext:
      enabled: true
      ## Set to false to stop the non-root user validation
      runAsNonRoot: true
      fsGroup: 10001
      runAsUser: 10001
      runAsGroup: 10001
    

    Note that you cannot change users during the Helm upgrades.

  • Overrides to add tolerations in YB-Master and YB-TServer pods:

    ## Consider the node has the following taint:
    ## kubectl taint nodes node1 dedicated=experimental:NoSchedule-
    
    master:
      tolerations:
      - key: dedicated
        operator: Equal
        value: experimental
        effect: NoSchedule
    
    tserver:
      tolerations: []
    

    Tolerations work in combination with taints: Taints are applied on nodes and Tolerations are applied to pods. Taints and tolerations ensure that pods do not schedule onto inappropriate nodes. You need to set nodeSelector to schedule YugabyteDB pods onto specific nodes, and then use taints and tolerations to prevent other pods from getting scheduled on the dedicated nodes, if required. For more information, see tolerations and Toleration API.

  • Overrides to use nodeSelector to schedule YB-Master and YB-TServer pods on dedicated nodes:

    ## To schedule a pod on a node that has a topology.kubernetes.io/zone=asia-south2-a label
    
    nodeSelector:
      topology.kubernetes.io/zone: asia-south2-a
    

    For more information, see nodeSelector and Kubernetes: Node Selector.

  • Overrides to add affinity in YB-Master and YB-TServer pods:

    ## To prevent scheduling of multiple master pods on single kubernetes node
    
    master:
      affinity:
        podAntiAffinity:
          requiredDuringSchedulingIgnoredDuringExecution:
          - labelSelector:
              matchExpressions:
              - key: app
                operator: In
                values:
                - "yb-master"
            topologyKey: kubernetes.io/hostname
    
    tserver:
      affinity: {}
    

    affinity allows the Kubernetes scheduler to place a pod on a set of nodes or a pod relative to the placement of other pods. You can use nodeAffinity rules to control pod placements on a set of nodes. In contrast, podAffinity or podAntiAffinity rules provide the ability to control pod placements relative to other pods. For more information, see Affinity API.

  • Overrides to add annotations to YB-Master and YB-TServer pods:

    master:
      podAnnotations:
        application: "yugabytedb"
    
    tserver:
      podAnnotations: {}
    

    The Kubernetes annotations can attach arbitrary metadata to objects. For more information, see Annotations.

  • Overrides to add labels to YB-Master and YB-TServer pods:

    master:
      podLabels:
        environment: production
        app: yugabytedb
        prometheus.io/scrape: true
    
    tserver:
      podLabels: {}
    

    The Kubernetes labels are key-value pairs attached to objects. The labels are used to specify identifying attributes of objects that are meaningful and relevant to you. For more information, see Labels.

  • Preflight check overrides, such as DNS address resolution, disk IO, available port, ulimit:

    ## Default values
    preflight:
      ## Set to true to skip disk IO check, DNS address resolution, port bind checks
      skipAll: false
    
      ## Set to true to skip port bind checks
      skipBind: false
    
      ## Set to true to skip ulimit verification
      ## SkipAll has higher priority
      skipUlimit: false
    

    For more information, see Helm chart: Prerequisites.

  • Overrides to use a secret for LDAP authentication. Refer to Create secrets for Kubernetes.

Configure Kubernetes multi-cluster environment

If you plan to create multi-region YugabyteDB universes, you can set up Multi-Cluster Services (MCS) across your Kubernetes clusters. This section covers implementation specific details for setting up MCS on various cloud providers and service mesh tools.

YBA support for MCS is in Early Access

The Kubernetes MCS API is currently in alpha, though there are various implementations of MCS which are considered to be stable. To know more, see API versioning in the Kubernetes documentation.

MCS support in YBA is currently in Early Access. Keep in mind following caveats:

  • Universe metrics may not display correct metrics for all the pods.
  • xCluster replication needs an additional manual step to work on OpenShift MCS.
  • Yugabyte Support assistance is needed for upgrades.

Prepare Kubernetes clusters for GKE MCS

GKE MCS allows clusters to be combined as a fleet on Google Cloud. These fleet clusters can export services, which enables you to do cross-cluster communication. For more information, see Multi-cluster Services in the Google Cloud documentation.

To enable MCS on your GKE clusters, see Configuring multi-cluster Services. Note down the unique membership name of each cluster in the fleet, it will be used during the cloud provider setup in YBA.

Prepare OpenShift clusters for MCS

Red Hat OpenShift Container Platform uses the Advanced Cluster Management for Kubernetes (RHACM) and its Submariner add-on to enable MCS. At a very high level this involves following steps:

  1. Create a management cluster and install RHACM on it. For details, see Installing Red Hat Advanced Cluster Management for Kubernetes in the Red Hat documentation.
  2. Provision the OpenShift clusters which will be connected together. Ensure that the CIDRs mentioned in the cluster configuration file at networking.clusterNetwork, networking.serviceNetwork, and networking.machineNetwork are non-overlapping across the multiple clusters. You can find more details about these options in provider-specific sections under the OpenShift Container Platform installation overview (look for sections named "Installing a cluster on [provider name] with customizations").
  3. Import the clusters into RHACM as a cluster set, and install the Submariner add-on on them. For more information, see Configuring Submariner.
  4. Note down the cluster names from the cluster set, as these will be used during the cloud provider setup in YBA.

Prepare Kubernetes clusters for Istio multicluster

An Istio service mesh can span multiple clusters, which allows you to configure MCS. It supports different topologies and network configurations. To install an Istio mesh across multiple Kubernetes clusters, see Install Multicluster in the Istio documentation.

The Istio configuration for each cluster should have following options:

apiVersion: install.istio.io/v1alpha1
kind: IstioOperator
spec:
  meshConfig:
    defaultConfig:
      proxyMetadata:
        ISTIO_META_DNS_CAPTURE: "true"
        ISTIO_META_DNS_AUTO_ALLOCATE: "true"
  # rest of the configuration…

Refer to Multi-Region YugabyteDB Deployments on Kubernetes with Istio for a step-by-step guide and an explanation of the options being used.

Configure the cloud provider for MCS

After you have the cluster set up, follow the instructions in Configure the Kubernetes cloud provider, and refer to this section for region and zone configuration required for multi-region universes.

Configure region and zone for GKE MCS

Follow the steps in Configure region and zones and set values for all the zones from your Kubernetes clusters connected via GKE MCS as follows:

  1. Specify fields such as Region, Zone, and so on as you would normally.

  2. Set the Cluster DNS Domain to clusterset.local.

  3. Upload the correct Kube Config of the cluster.

  4. Set the Pod Address Template to {pod_name}.<cluster membership name>.{service_name}.{namespace}.svc.{cluster_domain}, where the <cluster membership name> is the membership name of the Kubernetes cluster set during the fleet setup.

  5. Set the Overrides as follows:

    multicluster:
      createServiceExports: true
      kubernetesClusterId: "<cluster membership name>"
      mcsApiVersion: "net.gke.io/v1"
    

For example, if your cluster membership name is yb-asia-south1, then the Add new region screen would look as follows:

Add new region screen of YBA with GKE MCS

Configure region and zones for OpenShift MCS

Follow the instructions in Configure the OpenShift cloud provider and Create a provider. For all the zones from your OpenShift clusters connected via MCS (Submariner), add a region as follows:

  1. Specify fields such as Region, Zone, and so on as you would normally.

  2. Set the Cluster DNS Domain to clusterset.local.

  3. Upload the correct Kube Config of the cluster.

  4. Set the Pod Address Template to {pod_name}.<cluster name>.{service_name}.{namespace}.svc.{cluster_domain}, where the <cluster name> is the name of the OpenShift cluster set during the cluster set creation.

  5. Set the Overrides as follows:

    multicluster:
      createServiceExports: true
      kubernetesClusterId: "<cluster name>"
    

For example, if your cluster name is yb-asia-south1, then the values will be as follows:

  • Pod Address Template {pod_name}.yb-asia-south1.{service_name}.{namespace}.svc.{cluster_domain}

  • Overrides

    multicluster:
      createServiceExports: true
      kubernetesClusterId: "yb-asia-south1"
    

Configure region and zones for Istio

Follow the steps in Configure region and zones and set values for all the zones from your Kubernetes clusters connected via Istio as follows.

  1. Specify fields such as Region, Zone, and so on as you would normally.

  2. Upload the correct Kube Config of the cluster.

  3. Set the Pod Address Template to {pod_name}.{namespace}.svc.{cluster_domain}.

  4. Set the Overrides as follows:

    istioCompatibility:
      enabled: true
    multicluster:
      createServicePerPod: true