issues with installer digital_oi_installer

book

Article ID: 144687

calendar_today

Updated On:

Products

DX Operational Intelligence

Issue/Introduction

After execution digital_oi_installer_1.3.2_16.bin everything finished fine.

However not all PVs are in Bound state, doiproject-hadoop-data-0-pv has state Available.

Created Storage manually for that PV in Opehshift Console and state changed to Bound. 

However after DOI project was deployment few pods are not in running status:

 - Clientnodemanager is on stuck state,

 - Jobmanager failed to deploy. 


Messages in the install log (digital_oi_installer.log):

creating storage class doi-usage
Error from server (AlreadyExists): error when creating "docs/ca-digital-oi-storageclass.yaml": storageclasses.storage.k8s.io "doi-usage" already exists
Start create pv in cluster admin
mv: cannot stat ‘../doiproject/pv-docs/hadoop-data-1-pv.yaml’: No such file or directory
mv: cannot stat ‘../doiproject/pv-docs/ngtas-backup-pv.yaml’: No such file or directory

Error from server (AlreadyExists): error when creating "docs/kubestatemetrics/kube-state-metrics-cluster-role-binding.yaml": clusterrolebindings.rbac.authorization.k8s.io "kube-state-metrics" already exists
Error from server (AlreadyExists): error when creating "docs/kubestatemetrics/kube-state-metrics-cluster-role.yaml": clusterroles.rbac.authorization.k8s.io "kube-state-metrics" already exists

Installer log (CA_digital_oi_1.3.2_Install_01_20_2020_01_10_49.log) seems to say everything was a success.

Environment

Release : 1.3.2

Component : CA DOI AO PLATFORM COMPONENTS

Resolution

To fix this issue, please follow these steps:

oc delete pv acnproj-hadoop-data-0-pv


Create a yaml file "filename.yaml" with the following contents (Adjust the values according to your environments):


apiVersion: v1
kind: PersistentVolume
metadata:
  name: doiproject-hadoop-data-0-pv
  labels:
    doiPVName: doiproject-hadoop-data-pv
spec:
  storageClassName: doi-usage
  capacity:
    storage: 2Gi
  accessModes:
    - ReadWriteMany
  persistentVolumeReclaimPolicy: Retain
  nfs:
    path: /doi/nfs/hadoop-data1
    server: <IP>


run oc create -f  "filename.yaml" 



With above steps, it should resolve this issue.