I was discussing with Frank and Wendy that I saw a similar issue when running a sanity test on an AIO-SX.
I observed a test failure that reported the following:
CLI 'openstack --os-username 'admin' --os-password 'Li69nux*' --os-project-name admin --os-auth-url http://keystone.openstack.svc.cluster.local/v3 --os-user-domain-name Default --os-project-domain-name Default --os-identity-api-version 3 --os-interface internal --os-region-name RegionOne server show 1820414d-cf2d-46c1-877c-3a98a1b18be3' failed to execute.
Output: No server with a name or ID of '1820414d-cf2d-46c1-877c-3a98a1b18be3' exists.
In this scenario a reboot -f was perform on the host that contained 5 VMs. During the period after rebooting the Kubernetes OpenStack pods are starting and reaching a steady state condition. Also the 5 VMs are being restarted.
I suspect that in my scenario we were potentially experiencing some resource constraints on the system probably CPU and/or disk which resulted in inconsistent API results.
I suspect that this types of LP fall into the performance tuning pile of LPs. Resolution of these more than likely need to be done on H/W with additional instrumentation to identify bottlenecks: top, schedtop, iostat, etc...
I was discussing with Frank and Wendy that I saw a similar issue when running a sanity test on an AIO-SX.
I observed a test failure that reported the following:
CLI 'openstack --os-username 'admin' --os-password 'Li69nux*' --os-project-name admin --os-auth-url http:// keystone. openstack. svc.cluster. local/v3 --os-user- domain- name Default --os-project- domain- name Default --os-identity- api-version 3 --os-interface internal --os-region-name RegionOne server show 1820414d- cf2d-46c1- 877c-3a98a1b18b e3' failed to execute.
Output: No server with a name or ID of '1820414d- cf2d-46c1- 877c-3a98a1b18b e3' exists.
In this scenario a reboot -f was perform on the host that contained 5 VMs. During the period after rebooting the Kubernetes OpenStack pods are starting and reaching a steady state condition. Also the 5 VMs are being restarted.
I suspect that in my scenario we were potentially experiencing some resource constraints on the system probably CPU and/or disk which resulted in inconsistent API results.
I suspect that this types of LP fall into the performance tuning pile of LPs. Resolution of these more than likely need to be done on H/W with additional instrumentation to identify bottlenecks: top, schedtop, iostat, etc...