FAIL: tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance -- nova VM fails to go to running state when using neutron

Bug #1233923 reported by Davanum Srinivas (DIMS)
6
This bug affects 1 person
Affects Status Importance Assigned to Milestone
OpenStack Compute (nova)
Invalid
Undecided
Unassigned
neutron
Expired
Undecided
Unassigned

Bug Description

http://logs.openstack.org/10/49210/2/check/check-tempest-devstack-vm-neutron/1a9306e/logs/screen-n-cpu.txt.gz

2013-10-01 23:57:11.785 2937 INFO nova.compute.manager [-] [instance: 771f75a5-a649-4946-b495-a57558915b03] During sync_power_state the instance has a pending task. Skip.
Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/eventlet/queue.py", line 107, in switch
    self.greenlet.switch(value)
  File "/usr/local/lib/python2.7/dist-packages/eventlet/greenthread.py", line 194, in main
    result = function(*args, **kwargs)
  File "/opt/stack/new/nova/nova/virt/libvirt/driver.py", line 447, in _dispatch_thread
    self._dispatch_events()
  File "/opt/stack/new/nova/nova/virt/libvirt/driver.py", line 507, in _dispatch_events
    assert _c
AssertionError

http://logs.openstack.org/10/49210/2/check/check-tempest-devstack-vm-neutron/1a9306e/testr_results.html.gz

ft71.3: tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance[gate,smoke]_StringException: Empty attachments:
  stderr
  stdout

pythonlogging:'': {{{2013-10-01 23:57:07,853 state: pending}}}

Traceback (most recent call last):
  File "tempest/thirdparty/boto/test_ec2_instance_run.py", line 150, in test_run_stop_terminate_instance
    self.assertInstanceStateWait(instance, "running")
  File "tempest/thirdparty/boto/test.py", line 356, in assertInstanceStateWait
    state = self.waitInstanceState(lfunction, wait_for)
  File "tempest/thirdparty/boto/test.py", line 341, in waitInstanceState
    self.valid_instance_state)
  File "tempest/thirdparty/boto/test.py", line 331, in state_wait_gone
    state = state_wait(lfunction, final_set, valid_set)
  File "tempest/thirdparty/boto/utils/wait.py", line 57, in state_wait
    (dtime, final_set, status))
AssertionError: State change timeout exceeded!(400s) While waitingfor set(['running', '_GONE']) at "pending"

More importantly, this failure does NOT have "Cannot understand JSON" and is thus not the same as bug : https://bugs.launchpad.net/tempest/+bug/1223161

Revision history for this message
Joe Gordon (jogo) wrote :
Revision history for this message
Joe Gordon (jogo) wrote :

I dug into this yesterday, and it appears that nova-compute is missing its 'Got sempahore / lockl 'update_usage" call so the DB doesn't know the instance is booted. Because this only happens with neutron there is probably some strange interaction, such as a synchronous call to neutron that hangs for some reason.

http://logs.openstack.org/98/49198/1/check/check-tempest-devstack-vm-neutron/e3cd6d8/logs/screen-n-cpu.txt.gz#_2013-10-01_16_57_22_043

summary: FAIL:
tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance
- -- AssertionError
+ -- nova VM fails to go to running state when using neutron
Revision history for this message
Joe Gordon (jogo) wrote :

18:08 < jog0> http://logs.openstack.org/67/49367/1/check/check-tempest-devstack-vm-neutron/f333c6f/logs/screen-q-agt.txt.gz?
18:08 < jog0> search for '4a5a7f6e78ab' that is the failed instance-id
18:08 < jog0> I See it 843 times!
18:08 < jog0> is that right?
18:10 < markmcclain> yeah.. that is the result of polling the vswitch
18:10 < jog0> for a passing copy of q-agt I only see 'vm-uuid' under 300 times
18:11 < jog0> it looks like the vswitch is polled until the end of time

Changed in neutron:
status: New → Confirmed
Joe Gordon (jogo)
tags: added: havana-rc-potential
Thierry Carrez (ttx)
tags: added: havana-backport-potential
removed: havana-rc-potential
Revision history for this message
Russell Bryant (russellb) wrote :

The logstash query is no longer finding any hits on this one.

Changed in nova:
status: New → Invalid
Revision history for this message
Cedric Brandily (cbrandily) wrote :

This bug is > 365 days without activity. We are unsetting assignee and milestone and setting status to Incomplete in order to allow its expiry in 60 days.

If the bug is still valid, then update the bug status.

Changed in neutron:
status: Confirmed → Incomplete
Alan Pevec (apevec)
tags: removed: havana-backport-potential
Revision history for this message
Launchpad Janitor (janitor) wrote :

[Expired for neutron because there has been no activity for 60 days.]

Changed in neutron:
status: Incomplete → Expired
To post a comment you must log in.
This report contains Public information  
Everyone can see this information.

Other bug subscribers

Remote bug watches

Bug watches keep track of this bug in other bug trackers.