17.3 Known Issues

This section contains information about known issues.

Domain Creation Failure With WLST

The instructions in this section relate to problems creating Oracle Identity Governance (OIG) domains using WLST. See, Creating OIG Domains Using WLST Offline Scripts.

If the OIG domain creation fails, run the following to diagnose the issue:
kubectl logs <domain_job> -n <domain_namespace>
For example:
kubectl logs governancedomain-create-fmw-infra-sample-domain-job-c6vfb -n oigns
Also run:
kubectl describe pod <domain_job> -n <domain_namespace>
For example:
kubectl describe pod governancedomain-create-fmw-infra-sample-domain-job-c6vfb -n oigns
Using the output you should be able to diagnose the problem and resolve the issue.
If any of the above commands return the following error:
Failed to start container "create-fmw-infra-sample-domain-job": Error response from daemon: error while creating mount source path
'/nfs_volumes/oig/governancedomainpv ': mkdir /nfs_volumes/oig/governancedomainpv : permission denied
Then there is a permissions error on the directory for the PV and PVC and the following should be checked:
  • The directory has 777 permissions: chmod -R 777 <persistent_volume>/governancedomainpv.
  • If it does have the permissions, check if an oracle user exists and the uid is 1000 and gid is 0.

    Create the oracle user if it doesn’t exist and set the uid to 1000 and gid to 0.

  • Edit the $WORKDIR/kubernetes/create-weblogic-domain-pv-pvc/create-pv-pvc-inputs.yaml and add a slash to the end of the directory for the weblogicDomainStoragePath parameter:
    weblogicDomainStoragePath: /nfs_volumes/oig/governancedomainpv/
    
Once you have diagnosed the problem, clean down the failed domain creation by following: Then follow the instructions again in Creating OIG Domains Using WLST Offline Scripts

Domain Creation Failure With WDT Models

The instructions in this section relate to problems creating OIG domains using WDT models. See, Creating OIG Domains Using WDT Models.

If the domain creation fails while creating domain resources using the domain.yaml file, run the following steps to diagnose the issue:
  1. Check the domain events, by running the following command:
    kubectl describe domain <domain name> -n <domain_namespace>
    
    For example:
    kubectl describe domain governancedomain -n oigns
    
    Using the output, you should be able to diagnose the problem and resolve the issue.
  2. If the introspect job fails due to validation errors, then you can recreate the domain resources using the commands:
    kubectl delete -f domain.yaml
    kubectl create -f domain.yaml
    
  3. If the domain creation fails because of database issues:
    1. Create a helper pod:
       kubectl run --image=container-registry.oracle.com/middleware/oig_cpu:14.1.2.1.0-jdk17-ol8-<YYMMDD> --image-pull-policy="IfNotPresent" --overrides='{"apiVersion": "v1","spec":{"imagePullSecrets": [{"name": "orclcred"}]}}' helper -n oigns -- sleep infinity
      
    2. Once you have diagnosed the problem, clean down the failed domain creation by following:
    3. Execute the steps in Creating OIG Domains Using WDT Models again.

    Note:

    You might need to recreate the domain creation image depending upon the errors. Domain creation logs are stored in <persistent_volume>/domains/wdt-logs.
  4. If there is any issues bringing up the administration server, or OIG managed server pods, you can run the following to check the logs:
    kubectl logs <pod> -n <domain_namespace>
    For example:
    kubectl logs governancedomain-adminserver -n oigns
    If the above does not give any information you can also run:
    kubectl describe pod <pod> -n <domain_namespace>
    For example:
    kubectl describe pod governancedomain-adminserver -n oigns
For more details related to debugging issues, refer to Domain Debugging.

Pods Restarting Due to LivenessProbe

If the server pods keep restarting due to livenessProbe or readinessProbe failure, then make the following changes in the oim-cluster and soa-cluster respectively.

If the restart is due to resource (CPU in this case) limit issue, then the CPU parameter need to be adjusted as needed:
kubectl edit cluster <cluster> -n <domain_namespace>
For example:
kubectl edit cluster governancedomain-oim-cluster -n oigns
In the edit session change the CPU parameter as follows:
spec:
  clusterName: oim_cluster
  replicas: 1
  serverPod:
    env:
    - name: USER_MEM_ARGS
      value: '-Djava.security.egd=file:/dev/./urandom -Xms4096m -Xmx10240m -Dweblogic.rjvm.allowUnknownHost=true '
    livenessProbe:
      failureThreshold: 3
      periodSeconds: 60
      timeoutSeconds: 60
    readinessProbe:
      failureThreshold: 3
      periodSeconds: 60
      timeoutSeconds: 60
    resources:
      limits:
        cpu: 1700m
        memory: 6Gi
      requests:
        cpu: 500m
        memory: 4Gi

Patch Domain Failures

The instructions in this section relate to problems patching a deployment with a new image as per Patching a Container Image.

If the OIG domain patching fails when running patch_oig_domain.sh, run the following to diagnose the issue:
kubectl describe domain <domain name> -n <domain_namespace>
For example:
kubectl describe domain governancedomain -n oigns

Using the output you should be able to diagnose the problem and resolve the issue.

If the domain is already patched successfully and the script failed at the last step of waiting for pods to come up with the new image, then you do not need to rerun the script again after issue resolution. The pods will come up automatically once you resolve the underlying issue.

If the script is stuck at the following message for a long time:
[INFO] Waiting for weblogic pods to be ready..This may take several minutes, do not close the window. Check log /scratch/OIGK8Slatest/fmw-kubernetes/OracleIdentityGovernance/kubernetes/domain-lifecycle/log/oim_patch_log-<DATE>/monitor_weblogic_pods.log for progress
Run the following command to diagnose the issue:
kubectl get pods -n <domain_namespace>
For example:
kubectl get pods -n oigns
Run the following to check the logs of the Admininstration Server, SOA server or OIM server pods, as there may be an issue that is not allowing the domain pods to start properly:
kubectl logs <pod> -n oigns
If the above does not glean any information you can also run:
kubectl describe pod <pod> -n oigns
Further diagnostic logs can also be found under the $WORKDIR/kubernetes/domain-lifecycle.

Once any issue is resolved the pods will come up automatically without the need to rerun the script.