Connection timed out
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
Mirantis OpenStack |
Invalid
|
Medium
|
Pavel |
Bug Description
Detailed bug description:
Problem discovered on MOS 9.0 ISO #495 RC2
Problem was discovered during execution of NovaServers.
Nova log file atached to this bug.
Nova api log of this issue:
2016-07-07 13:09:44.688 19298 DEBUG keystoneauth.
2016-07-07 13:09:48.363 19400 DEBUG nova.metadata.
2016-07-07 13:09:48.366 19400 INFO nova.metadata.
2016-07-07 13:09:48.811 19311 DEBUG nova.osapi_
2016-07-07 13:09:48.813 19311 DEBUG nova.api.
2016-07-07 13:09:48.814 19311 INFO nova.osapi_
2016-07-07 13:09:56.098 19271 DEBUG oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.098 19271 ERROR oslo.messaging.
2016-07-07 13:09:56.100 19271 ERROR oslo.messaging.
Steps to reproduce:
Start rally test NovaServers.
Expected results:
Test passed. Server flavor was chenged.
Actual result:
Test failed.
Reproducibility:
Iterations: 97, Failures: 6
Workaround:
n/a
Impact:
unknown
Description of the environment:
* 10 baremetal nodes:
- CPU: 12 x 2.10 GHz
- Disks: 2 drives (SSD - 80 GB, HDD - 931.5 GB), 1006.0 GB total
- Memory: 2 x 16.0 GB, 32.0 GB total
- NUMA topology: 1 NUMA node
* Node roles:
- 3 controllers (1 was is offline because of disk problems)
- 7 computes
* Details:
- OS on controllers: Mitaka on Ubuntu 14.04
- OS on computes: RHEL
- Compute: KVM
- Neutron with VLAN segmentation
Additional information:
Diagnostic snapshot download link: http://
affects: | nova → mos |
Changed in mos: | |
importance: | Undecided → Medium |
milestone: | none → 9.0 |
Changed in mos: | |
assignee: | nobody → MOS Oslo (mos-oslo) |
milestone: | 9.0 → 9.1 |
status: | New → Confirmed |
Pavel, we need an env with repro to investigate the issue.
tl;dr
Logs from var/log/ remote/ node-17. domain. tld/rabbitmq. log.3.gz (http:// paste.openstack .org/show/ 532892/) show that at 2016-07-07T12:51:41 node-17 lost contact with both its peers (node-11 and node-19) and was shut down. Later, at 2016-07-07T13:09:40 node-19 also lost connection with node-11 and also was shut down (rabbitmq log from node-19 http:// paste.openstack .org/show/ 532893/). Eventually nodes recovered and joined back to node-11.
In both cases it is not clear what was the root cause, but strange entries from var/log/ remote/ node-19. domain. tld/lrmd. log.4.gz like 07T13:11: 11.951448+ 00:00 notice: notice: vip__vrouter_ monitor_ 5000:8313: stderr [ cat: /sys/class/ net/br- mgmt/carrier: Invalid argument] 07T13:14: 18.256347+ 00:00 warning: warning: Notification of client crmd/d5013cf8- 70df-49fb- 8dec-0842aa92db e3 failed
2016-07-
2016-07-
indicate that something was not right with the system and most probably it was either overloaded or network bandwidth was exhausted. Since the environment is small and the test executed seems to be rather light for controllers, the issue looks strange and needs to be investigated on a live environment.