Bug 112331 - [CI][SHARDS]igt@gem_eio@suspend - dmesg-warn - WARNING: CPU: 4 PID: 33 at drivers/pci/msi.c:1278 pci_irq_vector+0x66/0x70
Summary: [CI][SHARDS]igt@gem_eio@suspend - dmesg-warn - WARNING: CPU: 4 PID: 33 at dri...
Status: RESOLVED MOVED
Alias: None
Product: DRI
Classification: Unclassified
Component: DRM/Intel (show other bugs)
Version: DRI git
Hardware: Other All
: not set not set
Assignee: Intel GFX Bugs mailing list
QA Contact: Intel GFX Bugs mailing list
URL:
Whiteboard:
Keywords:
Depends on:
Blocks:
 
Reported: 2019-11-19 07:22 UTC by Lakshmi
Modified: 2019-11-29 19:50 UTC (History)
1 user (show)

See Also:
i915 platform: TGL
i915 features:


Attachments

Description Lakshmi 2019-11-19 07:22:31 UTC
https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_7364/shard-tglb8/igt@gem_eio@suspend.html
4> [171.254317] nvme nvme0: I/O 859 QID 5 timeout, aborting
<4> [171.254728] nvme nvme0: I/O 860 QID 5 timeout, aborting
<4> [171.254771] nvme nvme0: I/O 861 QID 5 timeout, aborting
<4> [171.254784] nvme nvme0: I/O 862 QID 5 timeout, aborting
<4> [171.254798] nvme nvme0: I/O 863 QID 5 timeout, aborting
<4> [201.461847] nvme nvme0: I/O 18 QID 0 timeout, reset controller
<4> [201.476970] ------------[ cut here ]------------
<4> [201.476977] WARNING: CPU: 4 PID: 33 at drivers/pci/msi.c:1278 pci_irq_vector+0x66/0x70
<4> [201.476979] Modules linked in: vgem snd_hda_codec_hdmi i915 mei_hdcp x86_pkg_temp_thermal coretemp crct10dif_pclmul crc32_pclmul snd_hda_intel snd_intel_dspcfg snd_hda_codec snd_hwdep snd_hda_core ghash_clmulni_intel cdc_ether usbnet mii snd_pcm mei_me mei intel_lpss_pci prime_numbers
<4> [201.476991] CPU: 4 PID: 33 Comm: kworker/4:0H Tainted: G     U            5.4.0-rc8-CI-CI_DRM_7364+ #1
<4> [201.476993] Hardware name: Intel Corporation Tiger Lake Client Platform/TigerLake U DDR4 SODIMM RVP, BIOS TGLSFWI1.R00.2321.A08.1909162051 09/16/2019
<4> [201.476996] Workqueue: kblockd blk_mq_timeout_work
<4> [201.476998] RIP: 0010:pci_irq_vector+0x66/0x70
<4> [201.476999] Code: c3 a8 10 75 0d 85 f6 75 21 8b 87 4c 05 00 00 01 f0 c3 48 8b 87 58 04 00 00 39 70 14 77 eb 0f 0b b8 ea ff ff ff c3 8b 42 10 c3 <0f> 0b b8 ea ff ff ff c3 66 90 48 83 ec 10 0f b6 77 59 65 48 8b 04
<4> [201.477001] RSP: 0018:ffffc900001ffc88 EFLAGS: 00010206
<4> [201.477002] RAX: 0000000000000000 RBX: ffff88849ddca000 RCX: 00000000fffe7a0f
<4> [201.477003] RDX: 0000000000000000 RSI: 0000000000000005 RDI: ffff88849c31c000
<4> [201.477004] RBP: ffff888499c30d80 R08: 00000000fffe7a0f R09: 00000000fffffffe
<4> [201.477005] R10: 00000000b534f170 R11: 00000000550975ef R12: 000000000000035b
<4> [201.477006] R13: ffff88849c31c000 R14: ffff88849a40c780 R15: 000000000000001b
<4> [201.477007] FS:  0000000000000000(0000) GS:ffff8884a0600000(0000) knlGS:0000000000000000
<4> [201.477008] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
<4> [201.477010] CR2: 0000561102b08fd8 CR3: 0000000005210001 CR4: 0000000000760ee0
<4> [201.477011] PKRU: 55555554
<4> [201.477011] Call Trace:
<4> [201.477016]  nvme_poll_irqdisable+0x30e/0x440
<4> [201.477020]  ? __lock_acquire+0x460/0x15d0
<4> [201.477023]  nvme_timeout+0xf9/0x310
<4> [201.477027]  ? find_held_lock+0x2d/0x90
<4> [201.477031]  blk_mq_check_expired+0xe6/0x140
<4> [201.477034]  blk_mq_queue_tag_busy_iter+0x230/0x410
<4> [201.477036]  ? __blk_mq_requeue_request+0x180/0x180
<4> [201.477038]  ? __blk_mq_requeue_request+0x180/0x180
<4> [201.477042]  blk_mq_timeout_work+0x105/0x250
<4> [201.477046]  process_one_work+0x26a/0x620
<4> [201.477050]  worker_thread+0x37/0x380
<4> [201.477053]  ? process_one_work+0x620/0x620
<4> [201.477056]  kthread+0x119/0x130
<4> [201.477058]  ? kthread_park+0x80/0x80
<4> [201.477063]  ret_from_fork+0x24/0x50
<4> [201.477068] irq event stamp: 134
<4> [201.477070] hardirqs last  enabled at (133): [<ffffffff819ea214>] _raw_spin_unlock_irq+0x24/0x50
<4> [201.477073] hardirqs last disabled at (134): [<ffffffff81001c0a>] trace_hardirqs_off_thunk+0x1a/0x20
<4> [201.477076] softirqs last  enabled at (0): [<ffffffff810acec8>] copy_process+0x538/0x1d80
<4> [201.477078] softirqs last disabled at (0): [<0000000000000000>] 0x0
<4> [201.477079] ---[ end trace 154d1381cb21ebc8 ]---
<4> [201.477086] nvme nvme0: I/O 859 QID 5 timeout, reset controller
<4> [201.478700] nvme nvme0: Abort status: 0x370
<4> [201.478706] nvme nvme0: Abort status: 0x370
<4> [201.478713] nvme nvme0: Abort status: 0x370
<4> [201.478714] nvme nvme0: Abort status: 0x370
<4> [201.478716] nvme nvme0: Abort status: 0x370
<3> [201.478821] PM: dpm_run_callback(): pci_pm_resume+0x0/0xa0 returns -16
<3> [201.478829] PM: Device 0000:01:00.0 failed to resume async: error -16
<6> [201.511499] OOM killer enabled.
<6> [201.511501] Restarting tasks ... 
<6> [201.520577] mei_hdcp mei::b638ab7e-94e2-4ea2-a552-d1c54b627f04:01: bound 0000:00:02.0 (ops i915_hdcp_component_ops [i915])
<4> [201.524508] done.
Comment 1 CI Bug Log 2019-11-19 07:23:30 UTC
The CI Bug Log issue associated to this bug has been updated.

### New filters associated

* TGL: igt@gem_eio@suspend - dmesg-warn - WARNING: CPU: 4 PID: 33 at drivers/pci/msi.c:1278 pci_irq_vector+0x66/0x70
  - https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_7364/shard-tglb8/igt@gem_eio@suspend.html
Comment 2 Chris Wilson 2019-11-19 10:03:44 UTC
And there we finally see the tgl S3 death in its full glory! After nvme dies, ext4 starts to panic.
Comment 3 Martin Peres 2019-11-29 19:50:23 UTC
-- GitLab Migration Automatic Message --

This bug has been migrated to freedesktop.org's GitLab instance and has been closed from further activity.

You can subscribe and participate further through the new bug through this link to our GitLab instance: https://gitlab.freedesktop.org/drm/intel/issues/620.


Use of freedesktop.org services, including Bugzilla, is subject to our Code of Conduct. How we collect and use information is described in our Privacy Policy.