Comment 47 for bug 1910866

Revision history for this message
faattori (fatordee) wrote :

I have just encountered this bug with Ubuntu 23.04 with kernel 6.2.0-20-generic. System is using default settings.

NVME is a Samsung SSD 970 EVO Plus 2TB with latest 4B2QEXM7 firmware available since it left factory.
Motherboard is Asus TUF GAMING X670E-PLUS WIFI with firmware 1410.

My issue happens only after an extended period of time, more than a week +-day or two.

System turns to read-only and the last thing in journalctl -f I see this:

touko 27 03:21:01 cereza kernel: nvme nvme0: I/O 657 (I/O Cmd) QID 14 timeout, aborting
touko 27 03:21:01 cereza kernel: nvme nvme0: Abort status: 0x0
touko 27 03:21:31 cereza kernel: nvme nvme0: I/O 657 (I/O Cmd) QID 14 timeout, aborting
touko 27 03:21:31 cereza kernel: nvme nvme0: Abort status: 0x0
touko 27 03:21:35 cereza kernel: nvme nvme0: I/O 12 QID 0 timeout, reset controller

I have now enabled nvme_core.default_ps_max_latency_us=1200 to see if the issue appears again since this should disable the lowest power state of the drive according to smartctl:

Supported Power States
St Op Max Active Idle RL RT WL WT Ent_Lat Ex_Lat
 0 + 7.59W - - 0 0 0 0 0 0
 1 + 7.59W - - 1 1 1 1 0 200
 2 + 7.59W - - 2 2 2 2 0 1000
 3 - 0.0500W - - 3 3 3 3 2000 1200
 4 - 0.0050W - - 4 4 4 4 500 9500