Displaying 20 results from an estimated 53 matches for "0x750".
Did you mean:
0x50
2017 Jan 24
1
[PATCH 2/2] drm/nouveau: Queue hpd_work on (runtime) resume
...m_callback+0x24/0x80
[ 246.899695] [<ffffffff8c4ced30>] ? pci_pm_runtime_resume+0xa0/0xa0
[ 246.899698] [<ffffffff8c5fffee>] rpm_suspend+0x11e/0x6f0
[ 246.899701] [<ffffffff8c60149b>] pm_runtime_work+0x7b/0xc0
[ 246.899707] [<ffffffff8c0afe58>] process_one_work+0x1f8/0x750
[ 246.899710] [<ffffffff8c0afdd9>] ? process_one_work+0x179/0x750
[ 246.899713] [<ffffffff8c0b03fb>] worker_thread+0x4b/0x4f0
[ 246.899717] [<ffffffff8c0bf8fc>] ? preempt_count_sub+0x4c/0x80
[ 246.899720] [<ffffffff8c0b03b0>] ? process_one_work+0x750/0x750
[ 246.899...
2020 Jan 09
1
[BUG] nouveau lockdep splat
...lock+0x134/0xc70
[ 98.459526] nouveau_svmm_invalidate_range_start+0x71/0x110 [nouveau]
[ 98.466593] __mmu_notifier_invalidate_range_start+0x25c/0x320
[ 98.473031] unmap_vmas+0x10c/0x200
[ 98.477130] unmap_region+0x1a4/0x240
[ 98.481410] __do_munmap+0x3e0/0x750
[ 98.485535] __vm_munmap+0xbc/0x130
[ 98.489599] __x64_sys_munmap+0x3c/0x50
[ 98.493951] do_syscall_64+0x68/0x280
[ 98.498162] entry_SYSCALL_64_after_hwframe+0x49/0xbe
[ 98.503778]
[ 98.503778] -> #2 (mmu_notifier_invalidate_range_start){+.+.}:
[ 98.511...
2019 Aug 06
2
Xorg indefinitely hangs in kernelspace
...secs"
disables this message.
[354073.738332] Xorg D 0 920 854 0x00404004
[354073.738334] Call Trace:
[354073.738340] __schedule+0x2ba/0x650
[354073.738342] schedule+0x2d/0x90
[354073.738343] schedule_preempt_disabled+0xe/0x10
[354073.738345] __ww_mutex_lock.isra.11+0x3e0/0x750
[354073.738346] __ww_mutex_lock_slowpath+0x16/0x20
[354073.738347] ww_mutex_lock+0x34/0x50
[354073.738352] ttm_eu_reserve_buffers+0x1f9/0x2e0 [ttm]
[354073.738356] qxl_release_reserve_list+0x67/0x150 [qxl]
[354073.738358] ? qxl_bo_pin+0xaa/0x190 [qxl]
[354073.738359] qxl_cursor_atomic_update+...
2019 Aug 06
2
Xorg indefinitely hangs in kernelspace
...secs"
disables this message.
[354073.738332] Xorg D 0 920 854 0x00404004
[354073.738334] Call Trace:
[354073.738340] __schedule+0x2ba/0x650
[354073.738342] schedule+0x2d/0x90
[354073.738343] schedule_preempt_disabled+0xe/0x10
[354073.738345] __ww_mutex_lock.isra.11+0x3e0/0x750
[354073.738346] __ww_mutex_lock_slowpath+0x16/0x20
[354073.738347] ww_mutex_lock+0x34/0x50
[354073.738352] ttm_eu_reserve_buffers+0x1f9/0x2e0 [ttm]
[354073.738356] qxl_release_reserve_list+0x67/0x150 [qxl]
[354073.738358] ? qxl_bo_pin+0xaa/0x190 [qxl]
[354073.738359] qxl_cursor_atomic_update+...
2016 Nov 21
2
[PATCH 1/2] drm/nouveau: Rename acpi_work to hpd_work
We need to call drm_helper_hpd_irq_event() on resume to properly detect
monitor connection / disconnection on some laptops. For runtime-resume
(which gets called on resume from normal suspend too) we must call
drm_helper_hpd_irq_event() from a workqueue to avoid a deadlock.
Rename acpi_work to hpd_work, and move it out of the #ifdef CONFIG_ACPI
blocks to make it suitable for generic work.
2018 Feb 23
2
v4.16-rc2: virtio-block + ext4 lockdep splats / sleeping from invalid context
...eously hit a
number of splats in the block layer:
* inconsistent {HARDIRQ-ON-W} -> {IN-HARDIRQ-R} usage in
jbd2_trans_will_send_data_barrier
* BUG: sleeping function called from invalid context at mm/mempool.c:320
* WARNING: CPU: 0 PID: 0 at block/blk.h:297 generic_make_request_checks+0x670/0x750
... I've included the full splats at the end of the mail.
These all happen in the context of the virtio block IRQ handler, so I
wonder if this calls something that doesn't expect to be called from IRQ
context. Is it valid to call blk_mq_complete_request() or
blk_mq_end_request() from an I...
2018 Feb 23
2
v4.16-rc2: virtio-block + ext4 lockdep splats / sleeping from invalid context
...eously hit a
number of splats in the block layer:
* inconsistent {HARDIRQ-ON-W} -> {IN-HARDIRQ-R} usage in
jbd2_trans_will_send_data_barrier
* BUG: sleeping function called from invalid context at mm/mempool.c:320
* WARNING: CPU: 0 PID: 0 at block/blk.h:297 generic_make_request_checks+0x670/0x750
... I've included the full splats at the end of the mail.
These all happen in the context of the virtio block IRQ handler, so I
wonder if this calls something that doesn't expect to be called from IRQ
context. Is it valid to call blk_mq_complete_request() or
blk_mq_end_request() from an I...
2018 Feb 26
0
v4.16-rc2: virtio-block + ext4 lockdep splats / sleeping from invalid context
...e block layer:
>
> * inconsistent {HARDIRQ-ON-W} -> {IN-HARDIRQ-R} usage in
> jbd2_trans_will_send_data_barrier
>
> * BUG: sleeping function called from invalid context at mm/mempool.c:320
>
> * WARNING: CPU: 0 PID: 0 at block/blk.h:297 generic_make_request_checks+0x670/0x750
>
> ... I've included the full splats at the end of the mail.
>
> These all happen in the context of the virtio block IRQ handler, so I
> wonder if this calls something that doesn't expect to be called from IRQ
> context. Is it valid to call blk_mq_complete_request() or...
2019 Apr 30
2
Xorg hangs in kernelspace with qxl
...Not tainted 5.0.0-13-generic #14-Ubuntu
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
Xorg D 0 879 790 0x00400004
Call Trace:
__schedule+0x2d0/0x840
schedule+0x2c/0x70
schedule_preempt_disabled+0xe/0x10
__ww_mutex_lock.isra.11+0x3e0/0x750
__ww_mutex_lock_slowpath+0x16/0x20
ww_mutex_lock+0x34/0x50
ttm_eu_reserve_buffers+0x1f9/0x2e0 [ttm]
qxl_release_reserve_list+0x67/0x150 [qxl]
? qxl_bo_pin+0x11d/0x200 [qxl]
qxl_cursor_atomic_update+0x1b0/0x2e0 [qxl]
drm_atomic_helper_commit_planes+0xb9/0x220 [drm_kms_helper]
drm_atomic_help...
2019 Apr 30
2
Xorg hangs in kernelspace with qxl
...Not tainted 5.0.0-13-generic #14-Ubuntu
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
Xorg D 0 879 790 0x00400004
Call Trace:
__schedule+0x2d0/0x840
schedule+0x2c/0x70
schedule_preempt_disabled+0xe/0x10
__ww_mutex_lock.isra.11+0x3e0/0x750
__ww_mutex_lock_slowpath+0x16/0x20
ww_mutex_lock+0x34/0x50
ttm_eu_reserve_buffers+0x1f9/0x2e0 [ttm]
qxl_release_reserve_list+0x67/0x150 [qxl]
? qxl_bo_pin+0x11d/0x200 [qxl]
qxl_cursor_atomic_update+0x1b0/0x2e0 [qxl]
drm_atomic_helper_commit_planes+0xb9/0x220 [drm_kms_helper]
drm_atomic_help...
2019 Sep 06
4
Xorg indefinitely hangs in kernelspace
...ge.
> [354073.738332] Xorg D 0 920 854 0x00404004
> [354073.738334] Call Trace:
> [354073.738340] __schedule+0x2ba/0x650
> [354073.738342] schedule+0x2d/0x90
> [354073.738343] schedule_preempt_disabled+0xe/0x10
> [354073.738345] __ww_mutex_lock.isra.11+0x3e0/0x750
> [354073.738346] __ww_mutex_lock_slowpath+0x16/0x20
> [354073.738347] ww_mutex_lock+0x34/0x50
> [354073.738352] ttm_eu_reserve_buffers+0x1f9/0x2e0 [ttm]
> [354073.738356] qxl_release_reserve_list+0x67/0x150 [qxl]
> [354073.738358] ? qxl_bo_pin+0xaa/0x190 [qxl]
> [354073.7383...
2019 Sep 06
4
Xorg indefinitely hangs in kernelspace
...ge.
> [354073.738332] Xorg D 0 920 854 0x00404004
> [354073.738334] Call Trace:
> [354073.738340] __schedule+0x2ba/0x650
> [354073.738342] schedule+0x2d/0x90
> [354073.738343] schedule_preempt_disabled+0xe/0x10
> [354073.738345] __ww_mutex_lock.isra.11+0x3e0/0x750
> [354073.738346] __ww_mutex_lock_slowpath+0x16/0x20
> [354073.738347] ww_mutex_lock+0x34/0x50
> [354073.738352] ttm_eu_reserve_buffers+0x1f9/0x2e0 [ttm]
> [354073.738356] qxl_release_reserve_list+0x67/0x150 [qxl]
> [354073.738358] ? qxl_bo_pin+0xaa/0x190 [qxl]
> [354073.7383...
2016 Jun 30
6
[PATCH] backlight: Avoid double fbcon backlight handling
...] [<ffffffff8154e611>] do_bind_con_driver+0x1c1/0x3a0
[ 18.984143] [<ffffffff8154eaf6>] do_take_over_console+0x116/0x180
[ 18.984145] [<ffffffff814bd3a7>] do_fbcon_takeover+0x57/0xb0
[ 18.984147] [<ffffffff814c1e48>] fbcon_event_notify+0x658/0x750
[ 18.984150] [<ffffffff810abcae>] notifier_call_chain+0x3e/0xb0
[ 18.984152] [<ffffffff810ac1ad>] __blocking_notifier_call_chain+0x4d/0x70
[ 18.984154] [<ffffffff810ac1e6>] blocking_notifier_call_chain+0x16/0x20
[ 18.984156] [<ffffffff814c748...
2009 Dec 04
2
[LLVMdev] linking a parser bitcode
...here the sj/lj stuff is coming from. Does this mean that the LLVM libraries we're using are broken?
Type.cpp
..\..\..\..\llvm\lib/libLLVMCore.a(Type.cpp.obj):Type.cpp.text+0x722): undefined reference to `__gxx_personality_sj0'
..\..\..\..\llvm\lib/libLLVMCore.a(Type.cpp.obj):Type.cpp.text+0x750): undefined reference to `_Unwind_SjLj_Register'
..\..\..\..\llvm\lib/libLLVMCore.a(Type.cpp.obj):Type.cpp.text+0x848): undefined reference to `_Unwind_SjLj_Resume'
..\..\..\..\llvm\lib/libLLVMCore.a(Type.cpp.obj):Type.cpp.text+0xa31): undefined reference to `_Unwind_SjLj_Resume'
Thank...
2018 Aug 05
2
[PATCH net-next 0/6] virtio_net: Add ethtool stat items
...44] R13: 0000000000000000 R14: 00007ffe83f38728 R15:
00007ffe83f37fd8
[ 46.168778] Allocated by task 499:
[ 46.168784] kasan_kmalloc+0xa0/0xd0
[ 46.168789] __kmalloc+0x191/0x3a0
[ 46.168795] mpi_powm+0x956/0x2360
[ 46.168801] rsa_enc+0x1f0/0x3a0
[ 46.168806] pkcs1pad_verify+0x4c4/0x750
[ 46.168815] public_key_verify_signature+0x58b/0xac0
[ 46.168821] pkcs7_validate_trust+0x3bd/0x710
[ 46.168830] verify_pkcs7_signature+0xe8/0x1b0
[ 46.168837] mod_verify_sig+0x1d4/0x2a0
[ 46.168842] load_module+0x1689/0x6590
[ 46.168847] __do_sys_finit_module+0x192/0x1c0
[ 46.16...
2018 Aug 05
2
[PATCH net-next 0/6] virtio_net: Add ethtool stat items
...44] R13: 0000000000000000 R14: 00007ffe83f38728 R15:
00007ffe83f37fd8
[ 46.168778] Allocated by task 499:
[ 46.168784] kasan_kmalloc+0xa0/0xd0
[ 46.168789] __kmalloc+0x191/0x3a0
[ 46.168795] mpi_powm+0x956/0x2360
[ 46.168801] rsa_enc+0x1f0/0x3a0
[ 46.168806] pkcs1pad_verify+0x4c4/0x750
[ 46.168815] public_key_verify_signature+0x58b/0xac0
[ 46.168821] pkcs7_validate_trust+0x3bd/0x710
[ 46.168830] verify_pkcs7_signature+0xe8/0x1b0
[ 46.168837] mod_verify_sig+0x1d4/0x2a0
[ 46.168842] load_module+0x1689/0x6590
[ 46.168847] __do_sys_finit_module+0x192/0x1c0
[ 46.16...
2019 Sep 06
0
[Spice-devel] Xorg indefinitely hangs in kernelspace
...Xorg D 0 920 854 0x00404004
> > [354073.738334] Call Trace:
> > [354073.738340] __schedule+0x2ba/0x650
> > [354073.738342] schedule+0x2d/0x90
> > [354073.738343] schedule_preempt_disabled+0xe/0x10
> > [354073.738345] __ww_mutex_lock.isra.11+0x3e0/0x750
> > [354073.738346] __ww_mutex_lock_slowpath+0x16/0x20
> > [354073.738347] ww_mutex_lock+0x34/0x50
> > [354073.738352] ttm_eu_reserve_buffers+0x1f9/0x2e0 [ttm]
> > [354073.738356] qxl_release_reserve_list+0x67/0x150 [qxl]
> > [354073.738358] ? qxl_bo_pin+0xaa/0x19...
2018 Jan 10
1
soft lockup after set multicast_router of bridge and it's port to 2
...le_frame_finish+0x0/0x2a0 [bridge]
[<ffffffff814736b6>] ? nf_hook_slow+0x76/0x120
[<ffffffffa04f48f0>] ? br_handle_frame_finish+0x0/0x2a0 [bridge]
[<ffffffffa04f4d1c>] ? br_handle_frame+0x18c/0x250 [bridge]
[<ffffffff81445709>] ? __netif_receive_skb+0x529/0x750
[<ffffffff814397da>] ? __alloc_skb+0x7a/0x180
[<ffffffff814492f8>] ? netif_receive_skb+0x58/0x60
[<ffffffff81449400>] ? napi_skb_finish+0x50/0x70
[<ffffffff8144ab79>] ? napi_gro_receive+0x39/0x50
[<ffffffffa016887f>] ? bnx2x_rx_int+0x83f/0x1630...
2011 May 05
12
Having parent transid verify failed
...el: [13560.752108] [<ffffffff813b0cf9>] ?
mutex_unlock+0x9/0x10
May 5 14:15:14 mail kernel: [13560.752115] [<ffffffffa087e9f4>] ?
btrfs_run_ordered_operations+0x1f4/0x210 [btrfs]
May 5 14:15:14 mail kernel: [13560.752122] [<ffffffffa0860fa3>]
btrfs_commit_transaction+0x263/0x750 [btrfs]
May 5 14:15:14 mail kernel: [13560.752126] [<ffffffff81079ff0>] ?
autoremove_wake_function+0x0/0x40
May 5 14:15:14 mail kernel: [13560.752131] [<ffffffffa085a9bd>]
transaction_kthread+0x26d/0x290 [btrfs]
May 5 14:15:14 mail kernel: [13560.752137] [<ffffffffa085a750>...
2018 Feb 26
2
v4.16-rc2: virtio-block + ext4 lockdep splats / sleeping from invalid context
...* inconsistent {HARDIRQ-ON-W} -> {IN-HARDIRQ-R} usage in
> > jbd2_trans_will_send_data_barrier
> >
> > * BUG: sleeping function called from invalid context at mm/mempool.c:320
> >
> > * WARNING: CPU: 0 PID: 0 at block/blk.h:297 generic_make_request_checks+0x670/0x750
> >
> > ... I've included the full splats at the end of the mail.
> >
> > These all happen in the context of the virtio block IRQ handler, so I
> > wonder if this calls something that doesn't expect to be called from IRQ
> > context. Is it valid to call...