Bug 112168 - [CI][RESUME]igt@kms_* - timeout - Received signal SIGQUIT. Stack trace: #0 [fatal_sig_handler+0xd6]
Summary: [CI][RESUME]igt@kms_* - timeout - Received signal SIGQUIT. Stack trace: #0 [f...
Status: RESOLVED MOVED
Alias: None
Product: DRI
Classification: Unclassified
Component: DRM/Intel (show other bugs)
Version: DRI git
Hardware: Other All
: not set not set
Assignee: Intel GFX Bugs mailing list
QA Contact: Intel GFX Bugs mailing list
URL:
Whiteboard:
Keywords:
Depends on:
Blocks:
 
Reported: 2019-10-29 11:17 UTC by Lakshmi
Modified: 2019-11-29 19:44 UTC (History)
1 user (show)

See Also:
i915 platform: TGL
i915 features:


Attachments

Description Lakshmi 2019-10-29 11:17:29 UTC
https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_7167/shard-tglb7/igt@kms_color@pipe-b-ctm-red-to-blue.html
	
Received signal SIGQUIT.
Stack trace: 
 #0 [fatal_sig_handler+0xd6]
 #1 [killpg+0x40]
 #2 [__write+0x12]
 #3 [writeN+0x2d]
 #4 [igt_sysfs_vprintf+0xf4]
 #5 [igt_sysfs_printf+0x8f]
 #6 [igt_drop_caches_set+0x2a]
 #7 [drm_open_driver+0x60]
 #8 [drm_open_driver_master+0x6]
 #9 [__real_main1210+0x233]
 #10 [main+0x27]
 #11 [__libc_start_main+0xe7]
 #12 [_start+0x2a]

This is split from Bug 112126 and specific to display tests.
Comment 1 CI Bug Log 2019-10-29 11:18:33 UTC
The CI Bug Log issue associated to this bug has been updated.

### New filters associated

* TGL: igt@kms_* - timeout - Received signal SIGQUIT. Stack trace: #0 [fatal_sig_handler+0xd6]
  (No new failures associated)
Comment 2 Uma Shankar 2019-10-30 10:31:02 UTC
I tried to check from display perspective. This looks to be an after affect of some locking issues caused by tests executed earlier:
Showing all locks held in the system:
 <4>[ 735.147923] 4 locks held by dmesg/1005:
 <4>[ 735.147925] 3 locks held by igt_runner/1022:
 <4>[ 735.147927] #0: ffff88849ead9408 (sb_writers#4)
{.+.+}, at: vfs_write+0x1a4/0x1d0
 <4>[ 735.147933] #1: ffffffff82247a20 (rcu_read_lock){....}, at: __handle_sysrq+0x0/0x220
 <4>[ 735.147936] #2: ffffffff82247a20 (rcu_read_lock){....}, at: debug_show_all_locks+0xe/0x1a0
 <4>[ 735.147939] 4 locks held by gem_linear_blit/1049:
 <4>[ 735.147942] #0: ffff88849e2d9408 (sb_writers#14){.+.+} 
, at: vfs_write+0x1a4/0x1d0
 <4>[ 735.147948] #1: ffff888491ada9b0 (&attr->mutex){..}, at: simple_attr_write+0x36/0xd0
 <4>[ 735.147956] #2: ffff88849b420e70 (&timeline->mutex){..}, at: intel_gt_retire_requests_timeout+0xe2/0x540 [i915]
 <4>[ 735.147994] #3: ffff888481b5c570 (&gt->reset.mutex){..}, at: i915_request_wait+0xc9/0x880 [i915]

This is causing the kms_color to wait more than the timeout and causing these SIGQUIT. 

From the logs below tests seem to have triggered this:
[IGT] gem_ctx_engines
[IGT] prime_busy
[IGT] i915_pm_rps
[IGT] gem_linear_blits
Comment 3 Martin Peres 2019-11-29 19:44:53 UTC
-- GitLab Migration Automatic Message --

This bug has been migrated to freedesktop.org's GitLab instance and has been closed from further activity.

You can subscribe and participate further through the new bug through this link to our GitLab instance: https://gitlab.freedesktop.org/drm/intel/issues/561.


Use of freedesktop.org services, including Bugzilla, is subject to our Code of Conduct. How we collect and use information is described in our Privacy Policy.