6 Creating an Order Balancer Cloud Native Instance
This chapter describes how to create an Order Balancer cloud native instance in your cloud environment using the operational scripts and the base Order Balancer configuration provided in the Order Balancer cloud native toolkit. You can create an Order Balancer instance quickly to become familiar with the process, explore the configuration, and structure your own project. You can create multiple Order Balancer instances by using the same Order Balancer image. This procedure is intended to validate that you are able to create an Order Balancer instance in your environment.
Before you create an Order Balancer instance, you must do the following:
- Download the ASAP cloud native tar file and extract the
ob-cntk.zip
file. For more information about downloading the Order Balancer cloud native toolkit, see "Downloading the ASAP Cloud Native Artifacts". - Install the Traefik container images
Installing the Order Balancer Artifacts and the Toolkit
Build container image for the Order Balancer application using the Order Balancer cloud native Image Builder.
You must create a private repository for this image, ensuring that all nodes in the cluster have access to the repository. See "About Container Image Management" for more details.
Copy the Order Balancer cloud native toolkit that is the
ob-cntk.zip
file to one of the nodes in the Kubernetes cluster and
do the following:
- On Oracle Linux: Where Kubernetes is hosted on Oracle Linux, download and extract the tar archive to each host that has connectivity to the Kubernetes cluster.
- On OKE: For an environment where Kubernetes is running in OKE, extract the contents of the tar archive on each OKE client host. The OKE client host is the bastion host/s that is set up to communicate with the OKE cluster.
$ export OB_CNTK=ob_cntk_path
In the previous example, ob_cntk_path
is the
installation directory of the Order Balancer cloud native toolkit.
Installing the Traefik Container Image
To leverage the Order Balancer cloud native samples that integrate with Traefik, the Kubernetes environment must have the Traefik ingress controller installed and configured.
If you are working in an environment where the Kubernetes cluster is shared,
confirm whether Traefik has already been installed and configured for Order Balancer
cloud native. If Traefik is already installed and configured, set your TRAEFIK_NS
environment variable to the appropriate name
space.
The instance of Traefik that you installed to validate your cloud environment must be removed as it does not leverage the Order Balancer cloud native samples. Ensure that you have removed this installation in addition to purging the Helm release. Check that any roles and rolebindings created by Traefik are removed. There could be a clusterrole and clusterrolebinding called "traefik-operator". There could also be a role and rolebinding called "traefik-operator" in the $TRAEFIK_NS name space. Delete all of these before you set up Traefik.
To download and install the Traefik container image:
- Ensure that Podman in your Kubernetes cluster can pull images from the container registry. See ASAP Compatibility Matrix for the required and supported versions of the Traefik image.
- Run the following command to create a name space ensuring that it does
not already exist:
Note:
You might want to add the traefik name space to the environment setup such as.bashrc
.kubectl get namespaces export TRAEFIK_NS=traefik kubectl create namespace $TRAEFIK_NS
- Run the following commands to install Traefik using the
$OB_CNTK/samples/charts/traefik/values.yaml
file in the samples:Note:
Setkubernetes.namespaces
and the chart version specifically using command-line.helm repo add traefik repo_link helm install traefik-operator traefik/traefik \ --namespace $TRAEFIK_NS \ --values $OB_CNTK/samples/charts/traefik/values.yaml \ --set "kubernetes.namespaces={$TRAEFIK_NS}" --version 26.0.0
In the example, repo_link is https://helm.traefik.io/traefik or https://traefik.github.io/charts based on the helm chart version. For more details, see: https://github.com/traefik/traefik-helm-chart
After the installation, Traefik monitors the name spaces listed in its
kubernetes.namespaces
field for Ingress objects. The
scripts in the toolkit manage this name space list as part of creating and tearing down
Order Balancer cloud native projects.
When the values.yaml
Traefik sample in the Order Balancer
cloud native toolkit is used as is, Traefik is exposed to the network outside of the
Kubernetes cluster through port 30305. To use a different port, edit the YAML file
before installing Traefik. Traefik metrics are also available for Prometheus to scrape
from the standard annotations.
Traefik function can be viewed using the Traefik dashboard. Create the
Traefik dashboard by running the instructions provided in the
$OB_CNTK/samples/charts/traefik/traefik-dashboard.yaml
file. To
access this dashboard, the URL is: http://traefik.asap.org
. This is if
you use the values.yaml
file provided with the Order Balancer cloud
native toolkit; it is possible to change the hostname as well as the port to your
desired values.
Creating an Order Balancer Instance
This section describes how to create an Order Balancer instance.
Setting Environment Variables
- Path to your private specification repository
- Traefik name space
To set the environment variables:
- Set the
TRAEFIK_NS
variable for Traefik name space as follows:$ export TRAEFIK_NS=Treafik Namespace
Creating Secrets
You must store sensitive data and credential information in the form of Kubernetes Secrets that the scripts and Helm charts in the toolkit consume. Managing secrets is out of the scope of the toolkit and must be implemented while adhering to your organization's corporate policies. Additionally, ASAP cloud native does not establish password policies.
- imagepull-secret: If the private registry or repository is password protected, create this secret.
- tls-secret: If the traefik ingress is ssl-enabled, create this secret. For more information about creating tls-secret, see "Setting Up ASAP Cloud Native for Incoming Access."
To create imagepull-secret:
- Run the following
command:
podman login
- Enter the credentials or access token.
The login process creates or updates the
config.json
file that contains an authorization token. - Run the following command to run the
secret:
kubectl create secret generic asap-imagepull -n namespace --from-file=.podmanconfigjson=$HOME/.podman/config.json --type=kubernetes.io/podmanconfigjson
Registering the Namespace
After you set the environment variables, register the name space. To register the name space, run the following command:
$OB_CNTK/scripts/register-namespace.sh -p sr -t targets
# For example, $OB_CNTK/scripts/register-namespace.sh -p sr -t traefik
Where -p
is the namespace where Order Balancer is being
deployed.
Note:
traefik
is the name of
the targets for registration of the namespace sr. The script uses
TRAEFIK_NS
to find these targets. Do not provide the
traefik
target if you are not using Traefik.
Creating an Order Balancer Instance
This procedure describes how to create an Order Balancer instance in your environment using the scripts that are provided with the toolkit.
- Update the
$OB_CNTK/charts/asap/values.yaml
file with the following values:replicaCount: 1 image: repository: obcn pullPolicy: IfNotPresent tag: "7.4.1.0.0" imagePullSecrets: - name: asap-imagepull obEnv: envid: ob96 port: 7501 host: obhost domain: ob persistence: enabled: true nameOverride: "" fullnameOverride: "" serviceAccount: # Specifies whether a service account should be created create: true # Annotations to add to the service account annotations: {} # The name of the service account to use. # If not set and create is true, a name is generated using the fullname template name: "" podAnnotations: {} podSecurityContext: {} # fsGroup: 2000 securityContext: {} # capabilities: # drop: # - ALL # readOnlyRootFilesystem: true # runAsNonRoot: true # runAsUser: 1000 servicechannel: name: channelport type: ClusterIP port: 7502 service: name: adminport type: ClusterIP port: 7501 ingress: type: TRAEFIK enabled: true sslIncoming: false adminsslhostname: adminobhostssl.asap.org adminhostname: adminobhost.asap.org secretName: project-instance-obcn-tls-cert hosts: - host: adminobhost.asap.org paths: [] tls: [] # - secretName: chart-example-tls # hosts: # - chart-example.local resources: {} autoscaling: enabled: false nodeSelector: {} tolerations: [] # To avoid co-locating two OB replicas/pods on the same node remove {} and # uncomment below lines. Replace <OB-ENV-ID> with OB Env ID affinity: {} # podAntiAffinity: # requiredDuringSchedulingIgnoredDuringExecution: # - labelSelector: # matchExpressions: # - key: "app" # operator: In # values: # - obEnv.envid # topologyKey: "kubernetes.io/hostname"
Where:replicaCount
is the number of Order Balancer instances to be created.repository
is the repository name of the configured container registry.tag
is the version name that is used when you create a final Order Balancer image from the container.name
is the name of the imagepull-secret if it is configured. For more information about imagepull-secret, see "Creating Secrets".envid
is the unique environment ID of the instance. This ID should include only the lower-case alphanumeric characters. For example, asapinstance1.port
is the port number of the WebLogic Server where Order Balancer is deployed.host
is the hostname of the podman container.domain
is the Order Balancer WebLogic domain name as configured inob.properties
while creating image.servicechannel.port
is the channel port when you create a channel in the WebLogic domain.service.port
is the admin port of the WebLogic Server.type
is the ingress controller type. This type can be TRAEFIK or GENERIC or OTHER.enabled
is the status of the ingress controller whether it is enabled or not. The value is true or false. By default, this is set to true.sslIncoming
is the status of the SSL/TLS configuration on incoming connections whether it is enabled or not. The configuration value is true or false. By default, this is set to false. If you want to set the value to true, create keys, certificate, and secret by following the instructions in the "Setting Up ASAP Cloud Native for Incoming Access" section.adminsslhostname
is the host name of the https access.adminhostname
is the host name of the http access.secretName
is the secret name of the certificate created for SSL/TLS. For more information about creating keys and secret name, see "Setting Up ASAP Cloud Native for Incoming Access."podAntiAffinity
can be uncommented to avoid co-locating two Order Balancer pods on the same node.requiredDuringSchedulingIgnoredDuringExecution
tells the Kubernetes Scheduler that it should never co-locate two Pods which have app label asenvid
in the domain defined by thetopologyKey.
topologyKey
with the value kubernetes.io/hostname indicates that the domain is an individual node.
Note:
adminsslhostname
andsecretName
are applicable only ifsslIncoming
is set to true. - Create PV and PVC for the Order Balancer instance. The pv path is
used to store the wallet and logs of the ASAP instance added to the Order
Balancer. The sample files are available in the
ob_cntk.zip
at$OB_CNTK/samples/nfs/
file. Use thehostPath
to access the specified path on the local nodes. Usenfs
to access specified path on all the nodes.Note:
Creating PV and PVC is a mandatory step for the Order Balancer instance.# This is pv.yaml apiVersion: v1 kind: PersistentVolume metadata: name: <project>-<obEnv.envid>-nfs-pv labels: type: local spec: storageClassName: wallet capacity: storage: 10Gi accessModes: - ReadWriteMany # Valid values are Retain, Delete or Recycle persistentVolumeReclaimPolicy: Retain # hostPath: nfs: server: <server> path: <path> # This is pvc.yaml apiVersion: v1 kind: PersistentVolumeClaim metadata: name: <obEnv.envid>-pvc namespace: <project> spec: storageClassName: wallet accessModes: - ReadWriteMany resources: requests: storage: 10Gi
Where:<project>
is the namespace. In the above example, value is sr.<obEnv.envid>
is the environment ID provided in the$OB_CNTK/samples/charts/traefik/values.yaml
file. In the above example, value is ob96.
The below paths from both Order Balancer pods are mounted on to the PV path as:
For wallet files:
/u01/oracle/user_projects/domains/<domainName>/oracle_communications/asap
toPV_mount_path/asap
For log files of each Order Balancer pod:
/u01/oracle/user_projects/domains/<domainName>/oracle_communications/logs
toPV_mount_path/OB_pod_name/logs
For example,
/mnt/OB96-1/logs and /mnt/OB96-0/logs
- Run the following command to create pv and
pvc:
Kubectl apply -f pv.yaml Kubectl apply -f pvc.yaml -n sr
- Verify that whether pv and pvc are created successfully or not by
running the following
command:
kubectl get pv kubectl get pvc -n sr
- Run the following command to create the Order Balancer
instance:
$OB_CNTK/scripts/create-instance.sh -p sr -i quick
The
create-instance.sh
script uses the Helm chart located in thecharts/ob
directory to deploy the Order Balancer image, service, and ingress controller for your instance. If the script fails, see "Troubleshooting Issues with the Scripts" before you make additional attempts. - Validate the important input details such as Image name and tag,
specification files used (Values Applied), hostname, and port for ingress
routing:
Calling helm lint ==> Linting ../charts/ob [INFO] Chart.yaml: icon is recommended 1 chart(s) linted, 0 chart(s) failed Project Namespace : sr Instance Fullname : sr-quick NAME : sr-quick LB_PORT : 30305 LAST DEPLOYED : Timestamp NAMESPACE : sr STATUS : deployed REVISION : 1 TEST SUITE : n/a
If PV/PVC is not configured before, the Pod will be in the pending state as shown below:Kubectl describe pod <pod-name> -n sr Events: Type Reason Age From Message ---- ------ ---- ---- ------- Warning FailedScheduling 2m31s default-scheduler persistentvolumeclaim "sr-pvc" not found Warning FailedScheduling 2m31s default-scheduler persistentvolumeclaim "sr-pvc" not found
- When the Pod state shows READY 1/1, your Order Balancer instance is up and running.
The base hostname required to access this instance using HTTP is adminobhostnonssl.asap.org
. See "Planning and Validating Your Cloud Environment" for details about hostname resolution.
The create-instance
script prints out the following valuable
information that you can use when you work with your Order Balancer domain:
- The URL for accessing the WebLogic UI, which is provided through the
ingress controller at host:
http://adminobhostnonssl.asap.org:30305/console
.
Validating the Order Balancer Instance
After creating an instance, you can validate it by accessing the WebLogic Server console.
Run the following command to display the Pod details of the Order Balancer instances that you have created:
$ kubectl get all -n OB_namespace
NAME READY STATUS RESTARTS AGE
pod/ob96-0 1/1 Running 0 24h
pod/ob96-1 1/1 Running 0 24h
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
service/ob96-hlservice ClusterIP None <none> 6723/TCP,6724/TCP 3d18h
service/ob96-service ClusterIP 10.106.240.147 <none> 6723/TCP,6724/TCP 3d18h
NAME READY AGE
statefulset.apps/ob96 2/2 3d18h
Note:
After the Order Balancer instances are created, it may take a few minutes to start Order Balancer servers.To access WebLogic Administration Console outside the cluster, enter the following URL in the browser:
http://adminhostnonssl.asap.org:nonSSL_Traefik_Port/console
https://adminhostssl.asap.org:SSL_Traefik_port/console
The nonSSL_Traefik_Port and SSL_Traefik_port are mentioned in
$ASAP_CNTK/samples/charts/traefik/values.yaml
.
The system prompts for the user name and password. Enter the WebLogic domain user name and password.
hostname
and
master_ip
address on the machine where the URL is getting
accessed.
Note:
The hosts file is located in/etc/hosts
on Linux and MacOS machines and in
C:\Windows\System32\drivers\etc\hosts
on Windows
machines.
ip_address adminhostnonssl.asap.org
Scaling the Order Balancer Instance
You can create multiple instances of Order Balancer by setting the
replicaCount
accordingly in the
values.yaml
file. The Order Balancer helm chart templates create
multiple instances as StatefulSet pods. The unique pod names are automatically
configured by suffixing an ordinal number with the StatefulSet Name. For example, if the
StatefulSet Name is OBapp, the pod names are OBapp-0, OBapp-1 and so on. The StatefulSet
policy ensures that only one pod with the given name is run at a given point of
time.
The number of Order Balancer instances can be scaled up or down by modifying the replicas count in StatefulSet:
kubectl scale statefulsets stateful_set_name --replicas=replica_count
Where
stateful_set_name
is the name of the StatefulSet. In
this example, it is OBapp.
replica_count
is the number of the Order Balancer
instances to be created.
Deleting and Recreating Your Order Balancer Instance
Deleting Your Order Balancer Instance
$OB_CNTK/scripts/delete-instance.sh -p sr -i quick
Re-creating Your Order Balancer Instance
When you delete an Order Balancer instance, the database state for that instance still remains unaffected. You can re-create multiple Order Balancer instance by using the same image.
$OB_CNTK/scripts/create-instance.sh -p sr -i quick
Cleaning Up the Environment
- Delete the
instance:
$OB_CNTK/scripts/delete-instance.sh -p sr -i quick
- Delete the name space, which in turn deletes the Kubernetes name
space and the
secrets:
$OB_CNTK/scripts/unregister-namespace.sh -p sr -d -t target
Note:
traefik
is the name of the target for registration of the name space. The script uses TRAEFIK_NS to find this target. Do not provide the "traefik" target if you are not using Traefik. - Delete the Order Balancer database users.
- Delete pv and pvc using the following
commands:
Kubectl delete pv <pv-name> Kubectl delete pvc <pvc-name> -n sr
Troubleshooting Issues with the Scripts
This section provides information about troubleshooting some issues that you may come across when running the scripts.
kubectl describe pod name -n sr
Cleanup Failed Instance
When a create-instance script fails, you must clean up the instance before making another attempt at instance creation.
Note:
Do not retry running thecreate-instance.sh
script or the
upgrade-instance.sh
script immediately to fix any errors, as they
would return errors. The upgrade-instance.sh
script may work but
re-running it does not complete the operation.
$OB_CNTK/scripts/delete-instance.sh -p sr -i quick
Recreating an Instance
create-instance.sh
script as this will fail. Instead, perform the
cleanup activities and then run the following
command:$OB_CNTK/scripts/create-instance.sh -p sr -i quick
Next Steps
The Order Balancer instance is ready to add the ASAP cloud native instance.
For more information about managing ASAP instances, see "Setting Up ASAP for High Availability" in ASAP System Administrator's Guide.