2021-12-07T21:44:52.687682317Z stderr F F1207 21:44:52.687211 1 main.go:80] Error getting server version: the server has asked for the client to provide credentials
Alarms
-
Test Activity
Regression
Workaround
-
More info:
The problem might be that certificate is missing controller-0's cluster host IP in SANs
Brief Description
platform-integ-apps fails to reach applied state after SX system is migrated to DX system
Severity
major
Steps to Reproduce
1)install subcloud as SX subcloud- overrides. yaml" on centralcontroller
2)Create the following file "migrate-
[sysadmin@ controller- 0 ~(keystone_admin)]$ cat migrate- subcloud- overrides. yaml ssh_pass" : ******, oam_node_ 0_address" : "2620:10A: A001:A103: :218", oam_node_ 1_address" : "2620:10A: A001:A103: :42", controller- 0 ~(keystone_admin)]$
---
{
"ansible_
"external_
"external_
}
[sysadmin@
3)Run migrate_ sx_to_dx. yml on central cloud and verify that there are not errors in the output
4)When the subcloud is online, managed, and login to subcloud and verify that sx is converted to duplex
[sysadmin@ controller- 0 ~(keystone_admin)]$ system show ------- ------- ----+-- ------- ------- ------- ------- ------- -+ ------- ------- ----+-- ------- ------- ------- ------- ------- -+ 05T18:36: 08.995159+ 00:00 | cloud_role | subcloud | project_ name | services | 07T19:47: 27.308407+ 00:00 | c387-4104- 9d75-e6948454ff e3 | ------- ------- ----+-- ------- ------- ------- ------- ------- -+ controller- 0 ~(keystone_admin)]$
+------
| Property | Value |
+------
| contact | None |
| created_at | 2021-12-
| description | None |
| distributed_
| https_enabled | True |
| latitude | None |
| location | None |
| longitude | None |
| name | dc-subcloud12 |
| region_name | subcloud12 |
| sdn_enabled | False |
| security_feature | spectre_meltdown_v1 |
| service_
| shared_services | [] |
| software_version | 21.12 |
| system_mode | duplex |
| system_type | All-in-one |
| timezone | UTC |
| updated_at | 2021-12-
| uuid | 0c86a371-
| vswitch_type | none |
+------
[sysadmin@
5)But the platform-integ-apps failed to reach applied state and it stuck at 25% forever
[sysadmin@ controller- 0 ~(keystone_admin)]$ system application-list ------- ------- ------+ ------- --+---- ------- ------- ------- ------- ---+--- ------- ------- ------- ------- ------- --+---- ------+ ------- ------- ------- ------- ------- ----+ ------- ------- ------+ ------- --+---- ------- ------- ------- ------- ---+--- ------- ------- ------- ------- ------- --+---- ------+ ------- ------- ------- ------- ------- ----+ manifest | certmanager- manifest. yaml | applied | completed | controller | 1.1-17 | nginx-ingress- controller- manifest | nginx_ingress_ controller_ manifest. yaml | applied | completed | integration- manifest | manifest.yaml | applying | processing chart: stx-rbd-provisioner | ------- ------- ------+ ------- --+---- ------- ------- ------- ------- ---+--- ------- ------- ------- ------- ------- --+---- ------+ ------- ------- ------- ------- ------- ----+
+------
| application | version | manifest name | manifest file | status | progress |
+------
| cert-manager | 1.0-25 | cert-manager-
| nginx-ingress-
| oidc-auth-apps | 1.0-59 | oidc-auth-manifest | manifest.yaml | applied | completed |
| platform-integ-apps | 1.0-42 | platform-
| | | | | | , overall completion: 25.0% |
| | | | | | |
| rook-ceph-apps | 1.0-13 | rook-ceph-manifest | manifest.yaml | uploaded | completed |
| vault | 1.0-22 | vault-manifest | vault_manifest.yaml | applied | completed |
+------
when tried to apply again, it failed
Expected Behavior
after migration the app should be applied properly
Actual Behavior
platform-integ-apps fails to reach applied state
Reproducibility
100%
System Configuration
SX subcloud. Seen on AIO-SX standalone too
Branch/Pull Time/Commit
21.12
Last Pass
21.05
Timestamp/Logs
/var/log/ armada/ platform- integ-apps- apply_2021- 12-07-21- 12-42.log
2021-12-07 21:42:51.198 178 ERROR armada. handlers. tiller [-] [chart= kube-system- rbd-provisioner ]: Error while installing release stx-rbd- provisioner: grpc._channel. _InactiveRpcErr or: <_InactiveRpcError of RPC that terminated with: error_string = "{"created" :"@1638913371. 198181944" ,"description" :"Error received from peer ipv4:127. 0.0.1:24134" ,"file" :"src/core/ lib/surface/ call.cc" ,"file_ line":1067, "grpc_message" :"release stx-rbd-provisioner failed: timed out waiting for the condition" ,"grpc_ status" :2}" handlers. tiller Traceback (most recent call last): handlers. tiller File "/usr/local/ lib/python3. 6/dist- packages/ armada/ handlers/ tiller. py", line 465, in install_release handlers. tiller metadata= self.metadata) handlers. tiller File "/usr/local/ lib/python3. 6/dist- packages/ grpc/_channel. py", line 923, in __call__ handlers. tiller return _end_unary_ response_ blocking( state, call, False, None) handlers. tiller File "/usr/local/ lib/python3. 6/dist- packages/ grpc/_channel. py", line 826, in _end_unary_ response_ blocking handlers. tiller raise _InactiveRpcErr or(state) handlers. tiller grpc._channel. _InactiveRpcErr or: <_InactiveRpcError of RPC that terminated with: handlers. tiller >--status = StatusCode.UNKNOWN handlers. tiller >--details = "release stx-rbd-provisioner failed: timed out waiting for the condition" handlers. tiller >--debug_ error_string = "{"created" :"@1638913371. 198181944" ,"description" :"Error received from peer ipv4:127. 0.0.1:24134" ,"file" :"src/core/ lib/surface/ call.cc" ,"file_ line":1067, "grpc_message" :"release stx-rbd-provisioner failed: timed out waiting for the condition" ,"grpc_ status" :2}" handlers. tiller > handlers. tiller ^[[00m handlers. tiller [-] [chart= kube-system- rbd-provisioner ]: Helm getting release status for release= stx-rbd- provisioner, version=0 get_release_status /usr/local/ lib/python3. 6/dist- packages/ armada/ handlers/ tiller. py:531^ [[00m
>---status = StatusCode.UNKNOWN
>---details = "release stx-rbd-provisioner failed: timed out waiting for the condition"
>---debug_
>
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.198 178 ERROR armada.
2021-12-07 21:42:51.199 178 DEBUG armada.
/var/logs/ pods/kube- system_ rbd-provisioner -759dfb8b6b- cfbnf_fedb37bb- ca78-4e48- 93b2-9d14d98327 da/rbd- provisioner/ 26.log
2021-12- 07T21:44: 52.687682317Z stderr F F1207 21:44:52.687211 1 main.go:80] Error getting server version: the server has asked for the client to provide credentials
Alarms
-
Test Activity
Regression
Workaround
-
More info:
The problem might be that certificate is missing controller-0's cluster host IP in SANs