status-history-pruner fails under load
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
Canonical Juju |
Fix Released
|
High
|
Christian Muirhead |
Bug Description
2017-06-07 17:23:42 ERROR juju.worker.
This seemed to be accompanied with a mongotop entry of:
juju.
Note also, because of bugs like https:/
We need to make sure that our various pruning tasks aren't trying to take too large of a bite at a time, and can make forward progress when its most important.
Changed in juju: | |
assignee: | nobody → Christian Muirhead (2-xtian) |
status: | Triaged → In Progress |
tags: | added: adrastea |
Changed in juju: | |
milestone: | 2.2.1 → 2.2.2 |
Changed in juju: | |
status: | In Progress → Fix Committed |
Changed in juju: | |
status: | Fix Committed → Fix Released |
It may be that we're trying to delete too much stuff at one time.
IIRC from txn pruning, it can take ~1min to delete 100,000 (200k?) txn records from juju.txns. On the current database, they are dealing with 18M transactions.
I tried to run a query to delete 17M of them.
At 100k/min that would be 2.8hrs to complete.
Maybe we just need to break it up into smaller chunks at a time?
In the txn pruner we went with something like 1000 records at a time, and then dump how far we've gotten every 15s.