Activity log for bug #1412578

Date Who What changed Old value New value Message
2015-01-19 21:57:45 Charles Wang bug added bug
2015-01-19 22:07:46 Charles Wang description Sometimes async_task_executor gets stuck and is unable to processed tasks queued in rabbitMQ. Restarting the async_task_executor service fixes it. This could be oslo messaging issue, or cassandra driver issue. The message got stuck in the queue. The following log may be related. ERROR oslo.messaging.rpc.dispatcher [-] Exception during message handling: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 134, in _dispatch_and_reply incoming.message)) File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 177, in _dispatch return self._do_dispatch(endpoint, method, ctxt, args) File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 123, in _do_dispatch result = getattr(endpoint, method)(ctxt, **new_args) File "/usr/bin/magnetodb-async-task-executor", line 120, in create self._table_info_repo.update(context, table_info, ["status"]) File "/usr/lib/python2.7/dist-packages/magnetodb/storage/table_info_repo/cassandra_impl.py", line 174, in update "".join(query_builder), consistent=True File "/usr/lib/python2.7/dist-packages/magnetodb/common/cassandra/cluster_handler.py", line 164, in execute_query raise ex NoHostAvailable: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Sometimes async_task_executor gets stuck and is unable to processed tasks queued in rabbitMQ. Restarting the async_task_executor service fixes it. This could be oslo messaging issue, or cassandra driver issue. The message got stuck in the queue. The following log may be related. ERROR oslo.messaging.rpc.dispatcher [-] Exception during message handling: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Traceback (most recent call last):   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 134, in _dispatch_and_reply     incoming.message))   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 177, in _dispatch     return self._do_dispatch(endpoint, method, ctxt, args)   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 123, in _do_dispatch     result = getattr(endpoint, method)(ctxt, **new_args)   File "/usr/bin/magnetodb-async-task-executor", line 120, in create     self._table_info_repo.update(context, table_info, ["status"])   File "/usr/lib/python2.7/dist-packages/magnetodb/storage/table_info_repo/cassandra_impl.py", line 174, in update     "".join(query_builder), consistent=True   File "/usr/lib/python2.7/dist-packages/magnetodb/common/cassandra/cluster_handler.py", line 164, in execute_query     raise ex NoHostAvailable: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)})
2015-01-19 22:13:44 Charles Wang description Sometimes async_task_executor gets stuck and is unable to processed tasks queued in rabbitMQ. Restarting the async_task_executor service fixes it. This could be oslo messaging issue, or cassandra driver issue. The message got stuck in the queue. The following log may be related. ERROR oslo.messaging.rpc.dispatcher [-] Exception during message handling: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Traceback (most recent call last):   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 134, in _dispatch_and_reply     incoming.message))   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 177, in _dispatch     return self._do_dispatch(endpoint, method, ctxt, args)   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 123, in _do_dispatch     result = getattr(endpoint, method)(ctxt, **new_args)   File "/usr/bin/magnetodb-async-task-executor", line 120, in create     self._table_info_repo.update(context, table_info, ["status"])   File "/usr/lib/python2.7/dist-packages/magnetodb/storage/table_info_repo/cassandra_impl.py", line 174, in update     "".join(query_builder), consistent=True   File "/usr/lib/python2.7/dist-packages/magnetodb/common/cassandra/cluster_handler.py", line 164, in execute_query     raise ex NoHostAvailable: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Sometimes async_task_executor gets stuck and is unable to processed tasks queued in rabbitMQ. Restarting the async_task_executor service fixes it. This could be oslo messaging issue, or cassandra driver issue. The async_task_executor code is not handling exception caused by not being able to connect to C* cluster when checking table status in table_info. The exception falls back to oslo messaging rpc dispatcher. The message gets stuck in the queue. The following log may be related. ERROR oslo.messaging.rpc.dispatcher [-] Exception during message handling: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Traceback (most recent call last):   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 134, in _dispatch_and_reply     incoming.message))   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 177, in _dispatch     return self._do_dispatch(endpoint, method, ctxt, args)   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 123, in _do_dispatch     result = getattr(endpoint, method)(ctxt, **new_args)   File "/usr/bin/magnetodb-async-task-executor", line 120, in create     self._table_info_repo.update(context, table_info, ["status"])   File "/usr/lib/python2.7/dist-packages/magnetodb/storage/table_info_repo/cassandra_impl.py", line 174, in update     "".join(query_builder), consistent=True   File "/usr/lib/python2.7/dist-packages/magnetodb/common/cassandra/cluster_handler.py", line 164, in execute_query     raise ex NoHostAvailable: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)})
2015-01-19 22:16:07 Charles Wang description Sometimes async_task_executor gets stuck and is unable to processed tasks queued in rabbitMQ. Restarting the async_task_executor service fixes it. This could be oslo messaging issue, or cassandra driver issue. The async_task_executor code is not handling exception caused by not being able to connect to C* cluster when checking table status in table_info. The exception falls back to oslo messaging rpc dispatcher. The message gets stuck in the queue. The following log may be related. ERROR oslo.messaging.rpc.dispatcher [-] Exception during message handling: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Traceback (most recent call last):   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 134, in _dispatch_and_reply     incoming.message))   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 177, in _dispatch     return self._do_dispatch(endpoint, method, ctxt, args)   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 123, in _do_dispatch     result = getattr(endpoint, method)(ctxt, **new_args)   File "/usr/bin/magnetodb-async-task-executor", line 120, in create     self._table_info_repo.update(context, table_info, ["status"])   File "/usr/lib/python2.7/dist-packages/magnetodb/storage/table_info_repo/cassandra_impl.py", line 174, in update     "".join(query_builder), consistent=True   File "/usr/lib/python2.7/dist-packages/magnetodb/common/cassandra/cluster_handler.py", line 164, in execute_query     raise ex NoHostAvailable: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Sometimes async_task_executor gets stuck and is unable to processed tasks queued in rabbitMQ. Restarting the async_task_executor service fixes it. The async_task_executor code is not handling exception caused by not being able to connect to C* cluster when checking table status in table_info. The exception falls back to oslo messaging rpc dispatcher. The message gets stuck in the queue. Failed tasks should be requeud, or at least all exceptions should handled to report it and return, instead of falling back to oslo messaging rpc dispatcher. The following log may be related. ERROR oslo.messaging.rpc.dispatcher [-] Exception during message handling: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Traceback (most recent call last):   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 134, in _dispatch_and_reply     incoming.message))   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 177, in _dispatch     return self._do_dispatch(endpoint, method, ctxt, args)   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 123, in _do_dispatch     result = getattr(endpoint, method)(ctxt, **new_args)   File "/usr/bin/magnetodb-async-task-executor", line 120, in create     self._table_info_repo.update(context, table_info, ["status"])   File "/usr/lib/python2.7/dist-packages/magnetodb/storage/table_info_repo/cassandra_impl.py", line 174, in update     "".join(query_builder), consistent=True   File "/usr/lib/python2.7/dist-packages/magnetodb/common/cassandra/cluster_handler.py", line 164, in execute_query     raise ex NoHostAvailable: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)})
2015-01-19 22:19:53 Charles Wang description Sometimes async_task_executor gets stuck and is unable to processed tasks queued in rabbitMQ. Restarting the async_task_executor service fixes it. The async_task_executor code is not handling exception caused by not being able to connect to C* cluster when checking table status in table_info. The exception falls back to oslo messaging rpc dispatcher. The message gets stuck in the queue. Failed tasks should be requeud, or at least all exceptions should handled to report it and return, instead of falling back to oslo messaging rpc dispatcher. The following log may be related. ERROR oslo.messaging.rpc.dispatcher [-] Exception during message handling: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Traceback (most recent call last):   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 134, in _dispatch_and_reply     incoming.message))   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 177, in _dispatch     return self._do_dispatch(endpoint, method, ctxt, args)   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 123, in _do_dispatch     result = getattr(endpoint, method)(ctxt, **new_args)   File "/usr/bin/magnetodb-async-task-executor", line 120, in create     self._table_info_repo.update(context, table_info, ["status"])   File "/usr/lib/python2.7/dist-packages/magnetodb/storage/table_info_repo/cassandra_impl.py", line 174, in update     "".join(query_builder), consistent=True   File "/usr/lib/python2.7/dist-packages/magnetodb/common/cassandra/cluster_handler.py", line 164, in execute_query     raise ex NoHostAvailable: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Sometimes async_task_executor gets stuck and is unable to processed tasks queued in rabbitMQ. Restarting the async_task_executor service fixes it. The async_task_executor code is not handling exception caused by not being able to connect to C* cluster when checking table status in table_info. The exception falls back to oslo messaging rpc dispatcher. The message gets stuck in the queue. Failed tasks should be requeud, or at least all exceptions should handled to report it and return, instead of falling back to oslo messaging rpc dispatcher. Oslo messaginf seems unable to re-dispatch the message and blocks all other messages in the queue. The following log may be related. ERROR oslo.messaging.rpc.dispatcher [-] Exception during message handling: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Traceback (most recent call last):   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 134, in _dispatch_and_reply     incoming.message))   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 177, in _dispatch     return self._do_dispatch(endpoint, method, ctxt, args)   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 123, in _do_dispatch     result = getattr(endpoint, method)(ctxt, **new_args)   File "/usr/bin/magnetodb-async-task-executor", line 120, in create     self._table_info_repo.update(context, table_info, ["status"])   File "/usr/lib/python2.7/dist-packages/magnetodb/storage/table_info_repo/cassandra_impl.py", line 174, in update     "".join(query_builder), consistent=True   File "/usr/lib/python2.7/dist-packages/magnetodb/common/cassandra/cluster_handler.py", line 164, in execute_query     raise ex NoHostAvailable: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)})
2015-01-19 22:38:12 Charles Wang description Sometimes async_task_executor gets stuck and is unable to processed tasks queued in rabbitMQ. Restarting the async_task_executor service fixes it. The async_task_executor code is not handling exception caused by not being able to connect to C* cluster when checking table status in table_info. The exception falls back to oslo messaging rpc dispatcher. The message gets stuck in the queue. Failed tasks should be requeud, or at least all exceptions should handled to report it and return, instead of falling back to oslo messaging rpc dispatcher. Oslo messaginf seems unable to re-dispatch the message and blocks all other messages in the queue. The following log may be related. ERROR oslo.messaging.rpc.dispatcher [-] Exception during message handling: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Traceback (most recent call last):   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 134, in _dispatch_and_reply     incoming.message))   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 177, in _dispatch     return self._do_dispatch(endpoint, method, ctxt, args)   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 123, in _do_dispatch     result = getattr(endpoint, method)(ctxt, **new_args)   File "/usr/bin/magnetodb-async-task-executor", line 120, in create     self._table_info_repo.update(context, table_info, ["status"])   File "/usr/lib/python2.7/dist-packages/magnetodb/storage/table_info_repo/cassandra_impl.py", line 174, in update     "".join(query_builder), consistent=True   File "/usr/lib/python2.7/dist-packages/magnetodb/common/cassandra/cluster_handler.py", line 164, in execute_query     raise ex NoHostAvailable: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Sometimes async_task_executor gets stuck and is unable to process tasks queued in rabbitMQ. Restarting the async_task_executor service fixes it. The async_task_executor code is not handling exception caused by not being able to connect to C* cluster when checking table status in table_info. The exception falls back to oslo messaging rpc dispatcher. The message gets stuck in the queue. Failed tasks should be requeud, or at least all exceptions should handled, instead of falling back to oslo messaging rpc dispatcher. Oslo messaginf seems unable to re-dispatch the message and blocks all other messages in the queue. The following log may be related. ERROR oslo.messaging.rpc.dispatcher [-] Exception during message handling: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)}) Traceback (most recent call last):   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 134, in _dispatch_and_reply     incoming.message))   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 177, in _dispatch     return self._do_dispatch(endpoint, method, ctxt, args)   File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/dispatcher.py", line 123, in _do_dispatch     result = getattr(endpoint, method)(ctxt, **new_args)   File "/usr/bin/magnetodb-async-task-executor", line 120, in create     self._table_info_repo.update(context, table_info, ["status"])   File "/usr/lib/python2.7/dist-packages/magnetodb/storage/table_info_repo/cassandra_impl.py", line 174, in update     "".join(query_builder), consistent=True   File "/usr/lib/python2.7/dist-packages/magnetodb/common/cassandra/cluster_handler.py", line 164, in execute_query     raise ex NoHostAvailable: ('Unable to complete the operation against any hosts', {<Host: 192.168.19.241 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.240 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.243 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.242 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.238 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.239 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.234 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.235 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.236 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.237 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.232 datacenter1>: ConnectionException('Pool is shutdown',), <Host: 192.168.19.233 datacenter1>: ConnectionException('Pool is shutdown',)})
2015-01-23 13:42:13 Ilya Sviridov bug added subscriber Illia Khudoshyn
2015-01-23 13:44:38 Ilya Sviridov magnetodb: status New Triaged
2015-01-23 13:44:44 Ilya Sviridov magnetodb: importance Undecided High
2015-01-23 13:44:51 Ilya Sviridov tags ux
2015-01-23 13:51:10 Ilya Sviridov magnetodb: milestone kilo-2
2015-02-05 16:09:52 Ilya Sviridov magnetodb: milestone kilo-2 kilo-3
2015-03-20 21:09:47 Ilya Sviridov magnetodb: milestone kilo-3 l-1
2015-03-20 21:09:49 Ilya Sviridov magnetodb: milestone l-1 2015.1
2015-04-29 12:38:36 Andrii Ostapenko magnetodb: milestone 2015.1 l-1