Activity log for bug #1567668

Date Who What changed Old value New value Message
2016-04-07 21:30:23 Assaf Muller bug added bug
2016-04-07 22:56:39 Assaf Muller neutron: status New Confirmed
2016-04-08 09:42:46 OpenStack Infra neutron: status Confirmed In Progress
2016-04-08 09:42:46 OpenStack Infra neutron: assignee Jakub Libosvar (libosvar)
2016-04-11 15:05:58 Assaf Muller description Here's an example: http://logs.openstack.org/13/302913/1/check/gate-neutron-dsvm-functional/91dd537/console.html Logstash query: build_name:"gate-neutron-dsvm-functional" AND build_status:"FAILURE" AND message:"Killed timeout -s 9" 45 hits in the last 7 days. Ihar and I checked the timing, and it started happening as we merged: https://review.openstack.org/#/c/298056/ There's a few problems here: 1) It appears like a test is freezing up. We have a per-test timeout defined. The timeout is defined by OS_TEST_TIMEOUT in tox.ini, and is enforced via a fixtures.Timeout fixture set up in the oslotest base class. It looks like that timeout doesn't always work. 2) When the global 2 hours job timeout is hit, it doesn't perform post-tests tasks such as copying over log files, which makes these problems a lot harder to troubleshoot. 3) And of course, there is some sort of issue with likely https://review.openstack.org/#/c/298056/. We can fix via a revert, which will increase the failure rate of fullstack. Since I've been unable to reproduce this issue locally, I'd like to hold off on a revert and try to get some more information by tackling some combination of problems 1 and 2, and then adding more logging to figure it out. Here's an example: http://logs.openstack.org/13/302913/1/check/gate-neutron-dsvm-functional/91dd537/console.html Logstash query: build_name:"gate-neutron-dsvm-functional" AND build_status:"FAILURE" AND message:"Killed timeout -s 9" 45 hits in the last 7 days. Ihar and I checked the timing, and it started happening as we merged: https://review.openstack.org/#/c/298056/ (EDIT: After some investigating, this doesn't look like the root cause). There's a few problems here: 1) It appears like a test is freezing up. We have a per-test timeout defined. The timeout is defined by OS_TEST_TIMEOUT in tox.ini, and is enforced via a fixtures.Timeout fixture set up in the oslotest base class. It looks like that timeout doesn't always work. 2) When the global 2 hours job timeout is hit, it doesn't perform post-tests tasks such as copying over log files, which makes these problems a lot harder to troubleshoot. 3) And of course, there is some sort of issue with likely https://review.openstack.org/#/c/298056/. We can fix via a revert, which will increase the failure rate of fullstack. Since I've been unable to reproduce this issue locally, I'd like to hold off on a revert and try to get some more information by tackling some combination of problems 1 and 2, and then adding more logging to figure it out.
2016-05-09 15:23:16 Dr. Jens Harbott bug added subscriber Dr. Jens Rosenboom
2016-05-13 16:18:23 Brian Haley bug added subscriber Brian Haley
2016-05-24 14:18:54 Ihar Hrachyshka neutron: milestone newton-1
2016-05-25 19:18:51 OpenStack Infra neutron: assignee Jakub Libosvar (libosvar) Henry Gessau (gessau)
2016-05-26 09:54:22 OpenStack Infra neutron: assignee Henry Gessau (gessau) Jakub Libosvar (libosvar)
2016-05-28 03:47:49 OpenStack Infra neutron: status In Progress Fix Released
2016-06-30 13:36:59 Ihar Hrachyshka tags functional-tests gate-failure functional-tests gate-failure neutron-proactive-backport-potential
2016-08-03 03:12:26 OpenStack Infra tags functional-tests gate-failure neutron-proactive-backport-potential functional-tests gate-failure in-stable-mitaka neutron-proactive-backport-potential
2016-10-07 15:47:11 Ihar Hrachyshka tags functional-tests gate-failure in-stable-mitaka neutron-proactive-backport-potential functional-tests gate-failure in-stable-mitaka