Unsolved
1 Rookie
•
11 Posts
0
138
April 16th, 2025 14:25
InsightIQ 'simple' upgrade from 5.1 to 5.2 fails pre-check "There are 1 unhealthy pods."
When running the upgrade script ./upgrade-iiq.sh on an InsightIQ 5.1 'simple' (VMware OVA) it fails with:
[2025-04-15 12:12:19] [INFO] All nodes are in a healthy state.[2025-04-15 12:12:19] [INFO] There are 1 unhealthy pods.[2025-04-15 12:12:19] [INFO] Some pods are not in a healthy state. Exiting..[2025-04-15 12:12:19] [ERROR] [PRECHECK] Failed to run precheck, please run showupg --log to see more detailed logs.[2025-04-15 12:12:19] [INFO] [PRECHECK] Pods were not scaled down. No scaling up required.
We know nothing about Kubernetes, and have little desire to.
The command :
kubectl get pods -n atlantic
which is suggested in the documentation shows a list of objects that are all 'running' except for one :
clustermanagement-nfs-d75d6f57b-k74gc 0/2 Init:0/1 0 24h
Running the script again (as suggested) does no better. The credentials are good, diskspace is good. We just upgraded our one PowerScale to 9.7.1.7 and we knew that 5.1 was not compatible (and indeed monitoring is broken). So we would rather like to be able to upgrade to 5.2.
We would raise an SR but it seems that InsightIQ only has telephone support.
Has anyone any suggestions where to look for 'unhealthy pods'?
Thanks


DELL-Josh Cr
Moderator
•
9.5K Posts
0
April 16th, 2025 20:33
Hi,
Thanks for your question.
What does showupg –log show?
You can try recreating that pod.
1.Login/SSH to InsightIQ primary-node
2. Edit the clustermanagement deployment using the below command:
kubectl edit deployment clustermanagement-nfs -n atlantic
3. deployment.yaml file will open in default editor
4. Remove the mount "cms-nfs-vol" from containers and volume "cms-nfs-vol" from volumes
5. After removing sections mentioned in step 4, Save the deployment file ':wq!'
6. clustermanagement-nfs will restart automatically, It will delete the existing POD and create a new one.
7. Check the pod status. New clustermanagement-nfs pod will be spawned without any older NFS mount info:
kubectl get pods -n atlantic | grep clustermanagement-nfs
8. At this point the customer should be able to remove the NFS export from their cluster and the clustermanagement-nfs pod should not go into an init 0/1 state anymore.
Let us know if you have any additional questions.
WilliamDLB
1 Rookie
•
11 Posts
1
April 17th, 2025 10:13
@DELL-Josh Cr
That is encouraging. I followed the above procedure and afterwards for a while it showed this:
but then the 'Terminating' one went away. All pods were now in a 'running' state, and re-running the upgrade has completed successfully:
Very many thanks as I could never have worked that out myself.
DELL-Josh Cr
Moderator
•
9.5K Posts
0
April 17th, 2025 13:23
Great to hear.