-
Bug
-
Resolution: Done
-
Highest
-
Dublin Release
-
None
-
Integration-OOM-Staging-Daily in Windriver lab
After master branch ONAP is deployed, Policy fails healthcheck with timeout. Redeploying Policy doesn't help.
------------------------------------------------------------------------------ Basic Policy Health Check | FAIL | Test timeout 10 seconds exceeded. ------------------------------------------------------------------------------
It appears drools pod doesn't come up:
root@staging-rancher:~# kubectl -n onap get pod |grep policy dev-policy-brmsgw-7bb48fb5f5-8tjsj 1/1 Running 0 38m dev-policy-drools-0 0/1 CrashLoopBackOff 7 38m dev-policy-nexus-7d5f84d8b6-7kk4s 1/1 Running 0 38m dev-policy-pap-77cd6f88bf-v42rk 2/2 Running 0 38m dev-policy-pdp-0 2/2 Running 0 38m dev-policy-policy-apex-pdp-0 1/1 Running 0 38m dev-policy-policy-distribution-77496d7888-j56zl 1/1 Running 0 38m dev-policy-policydb-74d46689cb-5llwj 1/1 Running 0 38m
Pod describe shows:
root@staging-rancher:~# kubectl -n onap describe pod dev-policy-drools-0 Name: dev-policy-drools-0 Namespace: onap Node: staging-k8s-03/10.0.0.11 Start Time: Tue, 05 Feb 2019 22:06:00 +0000 Labels: app=drools controller-revision-hash=dev-policy-drools-6bc59d4b85 release=dev-policy statefulset.kubernetes.io/pod-name=dev-policy-drools-0 Annotations: <none> Status: Running IP: 10.42.158.224 Controlled By: StatefulSet/dev-policy-drools Init Containers: drools-readiness: Container ID: docker://a98773e9bac1cd5b2dc54fa5751275051a58406bdc6e4c682403b81a845bdc21 Image: oomk8s/readiness-check:2.0.0 Image ID: docker-pullable://oomk8s/readiness-check@sha256:7daa08b81954360a1111d03364febcb3dcfeb723bcc12ce3eb3ed3e53f2323ed Port: <none> Host Port: <none> Command: /root/ready.py Args: --container-name policydb --container-name nexus State: Terminated Reason: Completed Exit Code: 0 Started: Tue, 05 Feb 2019 22:06:24 +0000 Finished: Tue, 05 Feb 2019 22:09:41 +0000 Ready: True Restart Count: 0 Environment: NAMESPACE: onap (v1:metadata.namespace) Mounts: /var/run/secrets/kubernetes.io/serviceaccount from default-token-x4r4h (ro) Containers: drools: Container ID: docker://238ed84ba9c2dd0cfb92b735b2deda21029176f3c40e255ccae6eb6bafcb2a6c Image: 10.12.5.2:5000/onap/policy-drools:1.3-SNAPSHOT-latest Image ID: docker-pullable://10.12.5.2:5000/onap/policy-drools@sha256:66684af3027434af76bae0f0f746fe972980f9b314eddb869926b37dc61f5ba3 Ports: 6969/TCP, 9696/TCP Host Ports: 0/TCP, 0/TCP Command: /bin/bash -c ./do-start.sh State: Waiting Reason: CrashLoopBackOff Last State: Terminated Reason: Error Exit Code: 137 Started: Tue, 05 Feb 2019 22:51:20 +0000 Finished: Tue, 05 Feb 2019 22:55:20 +0000 Ready: False Restart Count: 9 Liveness: tcp-socket :6969 delay=180s timeout=1s period=10s #success=1 #failure=3 Readiness: tcp-socket :6969 delay=60s timeout=1s period=10s #success=1 #failure=3 Environment: REPLICAS: 1 Mounts: /etc/localtime from localtime (ro) /tmp/logback.xml from policy-logback (rw) /tmp/policy-install/config/apps-install.sh from drools-config (rw) /tmp/policy-install/config/base.conf from drools-config (rw) /tmp/policy-install/config/drools-preinstall.sh from drools-config (rw) /tmp/policy-install/config/drools-tweaks.sh from drools-config (rw) /tmp/policy-install/config/feature-healthcheck.conf from drools-secret (rw) /tmp/policy-install/config/feature-pooling-dmaap.conf from drools-config (rw) /tmp/policy-install/config/policy-management.conf from drools-config (rw) /usr/share/maven/conf/settings.xml from drools-settingsxml (rw) /var/log/onap from policy-logs (rw) /var/run/secrets/kubernetes.io/serviceaccount from default-token-x4r4h (ro) Conditions: Type Status Initialized True Ready False ContainersReady False PodScheduled True Volumes: localtime: Type: HostPath (bare host directory volume) Path: /etc/localtime HostPathType: filebeat-conf: Type: ConfigMap (a volume populated by a ConfigMap) Name: dev-policy-filebeat-configmap Optional: false policy-logs: Type: EmptyDir (a temporary directory that shares a pod's lifetime) Medium: policy-data-filebeat: Type: EmptyDir (a temporary directory that shares a pod's lifetime) Medium: policy-logback: Type: ConfigMap (a volume populated by a ConfigMap) Name: dev-policy-drools-log-configmap Optional: false drools-settingsxml: Type: ConfigMap (a volume populated by a ConfigMap) Name: dev-policy-drools-settings-configmap Optional: false drools-config: Type: ConfigMap (a volume populated by a ConfigMap) Name: dev-policy-drools-configmap Optional: false drools-secret: Type: Secret (a volume populated by a Secret) SecretName: dev-policy-drools-secret Optional: false default-token-x4r4h: Type: Secret (a volume populated by a Secret) SecretName: default-token-x4r4h Optional: false QoS Class: BestEffort Node-Selectors: <none> Tolerations: node.kubernetes.io/not-ready:NoExecute for 300s node.kubernetes.io/unreachable:NoExecute for 300s Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal Scheduled 51m default-scheduler Successfully assigned onap/dev-policy-drools-0 to staging-k8s-03 Normal Pulled 51m kubelet, staging-k8s-03 Container image "oomk8s/readiness-check:2.0.0" already present on machine Normal Created 51m kubelet, staging-k8s-03 Created container Normal Started 51m kubelet, staging-k8s-03 Started container Normal Pulling 47m kubelet, staging-k8s-03 pulling image "10.12.5.2:5000/onap/policy-drools:1.3-SNAPSHOT-latest" Normal Pulled 47m kubelet, staging-k8s-03 Successfully pulled image "10.12.5.2:5000/onap/policy-drools:1.3-SNAPSHOT-latest" Normal Created 47m kubelet, staging-k8s-03 Created container Normal Started 47m kubelet, staging-k8s-03 Started container Warning Unhealthy 16m (x22 over 44m) kubelet, staging-k8s-03 Liveness probe failed: dial tcp 10.42.158.224:6969: connect: connection refused Warning Unhealthy 5m (x163 over 46m) kubelet, staging-k8s-03 Readiness probe failed: dial tcp 10.42.158.224:6969: connect: connection refused Warning BackOff 1m (x31 over 15m) kubelet, staging-k8s-03 Back-off restarting failed container
drools pod console log is attached.