LKML Archive on lore.kernel.org
help / color / mirror / Atom feed
* 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000
@ 2019-05-12  8:27 Justin Piszcz
  2019-05-12  8:34 ` Justin Piszcz
  2019-05-20 11:56 ` Mel Gorman
  0 siblings, 2 replies; 12+ messages in thread
From: Justin Piszcz @ 2019-05-12  8:27 UTC (permalink / raw)
  To: LKML

Hello,

I've turned off zram/zswap and I am still seeing the following during
periods of heavy I/O, I am returning to 5.0.xx in the meantime.

Kernel: 5.1.1
Arch: x86_64
Dist: Debian x86_64

[29967.019411] BUG: unable to handle kernel paging request at ffffea0002030000
[29967.019414] #PF error: [normal kernel read fault]
[29967.019415] PGD 103ffee067 P4D 103ffee067 PUD 103ffed067 PMD 0
[29967.019417] Oops: 0000 [#1] SMP PTI
[29967.019419] CPU: 10 PID: 77 Comm: khugepaged Tainted: G
   T 5.1.1 #4
[29967.019420] Hardware name: Supermicro X9SRL-F/X9SRL-F, BIOS 3.2 01/16/2015
[29967.019424] RIP: 0010:isolate_freepages_block+0xb9/0x310
[29967.019425] Code: 24 28 48 c1 e0 06 40 f6 c5 1f 48 89 44 24 20 49
8d 45 79 48 89 44 24 18 44 89 f0 4d 89 ee 45 89 fd 41 89 c7 0f 84 ef
00 00 00 <48> 8b 03 41 83 c4 01 a9 00 00 01 00 75 0c 48 8b 43 08 a8 01
0f 84
[29967.019426] RSP: 0018:ffffc900003a7860 EFLAGS: 00010246
[29967.019427] RAX: 0000000000000000 RBX: ffffea0002030000 RCX: ffffc900003a7b69
[29967.019428] RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffffffff828c4750
[29967.019429] RBP: 0000000000080c00 R08: 0000000000000001 R09: 0000000000000000
[29967.019429] R10: 0000000000000206 R11: ffffffff828c4390 R12: 0000000000000000
[29967.019430] R13: 0000000000000000 R14: ffffc900003a7af0 R15: 0000000000000000
[29967.019431] FS:  0000000000000000(0000) GS:ffff88903fa80000(0000)
knlGS:0000000000000000
[29967.019432] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[29967.019432] CR2: ffffea0002030000 CR3: 000000000280e001 CR4: 00000000001606e0
[29967.019433] Call Trace:
[29967.019436]  compaction_alloc+0x83d/0x8d0
[29967.019439]  migrate_pages+0x30d/0x750
[29967.019441]  ? isolate_migratepages_block+0xa10/0xa10
[29967.019442]  ? __reset_isolation_suitable+0x110/0x110
[29967.019443]  compact_zone+0x684/0xa70
[29967.019445]  compact_zone_order+0x109/0x150
[29967.019447]  ? schedule_timeout+0x1ba/0x290
[29967.019461]  ? record_times+0x13/0xa0
[29967.019462]  try_to_compact_pages+0x10d/0x220
[29967.019465]  __alloc_pages_direct_compact+0x93/0x180
[29967.019466]  __alloc_pages_nodemask+0x6c7/0xe20
[29967.019469]  ? __wake_up_common_lock+0xb0/0xb0
[29967.019470]  khugepaged+0x31f/0x19c0
[29967.019472]  ? __wake_up_common_lock+0xb0/0xb0
[29967.019473]  ? collapse_shmem.isra.4+0xc20/0xc20
[29967.019476]  kthread+0x100/0x120
[29967.019478]  ? __kthread_create_on_node+0x1b0/0x1b0
[29967.019480]  ret_from_fork+0x35/0x40
[29967.019481] CR2: ffffea0002030000
[29967.019482] ---[ end trace 8a9801ed5e182a98 ]---
[29967.019484] RIP: 0010:isolate_freepages_block+0xb9/0x310
[29967.019484] Code: 24 28 48 c1 e0 06 40 f6 c5 1f 48 89 44 24 20 49
8d 45 79 48 89 44 24 18 44 89 f0 4d 89 ee 45 89 fd 41 89 c7 0f 84 ef
00 00 00 <48> 8b 03 41 83 c4 01 a9 00 00 01 00 75 0c 48 8b 43 08 a8 01
0f 84
[29967.019485] RSP: 0018:ffffc900003a7860 EFLAGS: 00010246
[29967.019486] RAX: 0000000000000000 RBX: ffffea0002030000 RCX: ffffc900003a7b69
[29967.019487] RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffffffff828c4750
[29967.019487] RBP: 0000000000080c00 R08: 0000000000000001 R09: 0000000000000000
[29967.019488] R10: 0000000000000206 R11: ffffffff828c4390 R12: 0000000000000000
[29967.019489] R13: 0000000000000000 R14: ffffc900003a7af0 R15: 0000000000000000
[29967.019489] FS:  0000000000000000(0000) GS:ffff88903fa80000(0000)
knlGS:0000000000000000
[29967.019490] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[29967.019491] CR2: ffffea0002030000 CR3: 000000000280e001 CR4: 00000000001606e0

Thanks,

Justin.

^ permalink raw reply	[flat|nested] 12+ messages in thread

* RE: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000
  2019-05-12  8:27 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000 Justin Piszcz
@ 2019-05-12  8:34 ` Justin Piszcz
  2019-05-17 21:30   ` Sakari Ailus
  2019-05-20 11:56 ` Mel Gorman
  1 sibling, 1 reply; 12+ messages in thread
From: Justin Piszcz @ 2019-05-12  8:34 UTC (permalink / raw)
  To: 'LKML'



> -----Original Message-----
> From: Justin Piszcz [mailto:jpiszcz@lucidpixels.com]
> Sent: Sunday, May 12, 2019 4:28 AM
> To: LKML
> Subject: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at
> ffffea0002030000
> 
> Hello,
> 
> I've turned off zram/zswap and I am still seeing the following during
> periods of heavy I/O, I am returning to 5.0.xx in the meantime.
> 
> Kernel: 5.1.1
> Arch: x86_64
> Dist: Debian x86_64

[ .. ]

Reverting back to linux-5.0.15, will see if it recurs (I've never seen this before moving to 5.1.x)

$ diff -u linux-5.1.1/mm/compaction.c linux-5.0.15/mm/compaction.c | wc -l
1628




^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000
  2019-05-12  8:34 ` Justin Piszcz
@ 2019-05-17 21:30   ` Sakari Ailus
  0 siblings, 0 replies; 12+ messages in thread
From: Sakari Ailus @ 2019-05-17 21:30 UTC (permalink / raw)
  To: Justin Piszcz; +Cc: 'LKML', Mel Gorman

On Sun, May 12, 2019 at 04:34:55AM -0400, Justin Piszcz wrote:
> 
> 
> > -----Original Message-----
> > From: Justin Piszcz [mailto:jpiszcz@lucidpixels.com]
> > Sent: Sunday, May 12, 2019 4:28 AM
> > To: LKML
> > Subject: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at
> > ffffea0002030000
> > 
> > Hello,
> > 
> > I've turned off zram/zswap and I am still seeing the following during
> > periods of heavy I/O, I am returning to 5.0.xx in the meantime.
> > 
> > Kernel: 5.1.1
> > Arch: x86_64
> > Dist: Debian x86_64
> 
> [ .. ]
> 
> Reverting back to linux-5.0.15, will see if it recurs (I've never seen this before moving to 5.1.x)
> 
> $ diff -u linux-5.1.1/mm/compaction.c linux-5.0.15/mm/compaction.c | wc -l
> 1628

I've seen the same on v5.1.2 twice now; I wanted to see whether I'd be
lucky with applying "mm/compaction.c: correct zone boundary handling when
isolating pages from a pageblock" but it made no difference. I'm not sure
whether this is related.

Cc'ing Mel as I see he's been working on the file. :-)

The first one is plain v5.1.2 and the second one with the aforementioned
patch, and my .config is here:

<URL:https://www.retiisi.org.uk/~sakke/foo/.config.5.1>

The host isn't running anything too heavy but there are about a dozen
QEMU/KVM virtual machines.


BUG: unable to handle kernel paging request at ffffebe783ff0030
#PF error: [normal kernel read fault]
PGD 46fff9067 P4D 46fff9067 PUD 46fff8067 PMD 0 
Oops: 0000 [#1] SMP NOPTI
CPU: 1 PID: 4149 Comm: CPU 2/KVM Not tainted 5.1.2+ #133
Hardware name: empty empty/S32272NR-C538, BIOS V2.00 08/04/2017
RIP: 0010:compaction_alloc+0x4ce/0x910
Code: 48 c1 e5 06 e9 ca 00 00 00 41 80 be 3d 04 00 00 00 0f 84 dc 00 00 00 49 89 ed 4c 03 2d 9b bd ad 00 4d 85 ed 0f 84 86 00 00 00 <41> 8b 45 30 25 80 00 00 f0 3d 00 00 00 f0 0f 84 28 01 00 00 80 7b
RSP: 0018:ffffa248c0763458 EFLAGS: 00010286
RAX: 0000000000000001 RBX: ffffa248c0763640 RCX: 000000000000003c
RDX: 80000000000ffc00 RSI: 0000000000000000 RDI: ffffa0553fffa900
RBP: 0000000003ff0000 R08: 0000000000000000 R09: 0000000000000000
R10: ffffa248c07634b0 R11: ffffffff99c72438 R12: 8000000000100000
R13: ffffebe783ff0000 R14: ffffffff99c72200 R15: 80000000000ffc00
FS:  00007ff4be952700(0000) GS:ffffa0552fa80000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: ffffebe783ff0030 CR3: 000000046ab8a000 CR4: 00000000003426e0
Call Trace:
? move_freelist_tail+0xc0/0xc0
migrate_pages+0x20a/0x790
? isolate_freepages_block+0x340/0x340
compact_zone+0x4d4/0xae0
compact_zone_order+0xda/0x120
? try_to_compact_pages+0x107/0x220
try_to_compact_pages+0x107/0x220
__alloc_pages_direct_compact+0x65/0x140
__alloc_pages_slowpath+0x37f/0xc40
__alloc_pages_nodemask+0x20c/0x260
do_huge_pmd_anonymous_page+0x142/0x5a0
? cpumask_next_and+0x19/0x20
__handle_mm_fault+0xb3c/0xee0
__get_user_pages+0x178/0x6b0
get_user_pages_unlocked+0x143/0x1f0
__gfn_to_pfn_memslot+0x13b/0x3d0 [kvm]
? kvm_irq_delivery_to_apic_fast+0x124/0x3d0 [kvm]
? __switch_to_asm+0x40/0x70
? __switch_to_asm+0x34/0x70
? __switch_to_asm+0x40/0x70
try_async_pf+0x82/0x190 [kvm]
tdp_page_fault+0x12d/0x270 [kvm]
? vmx_vmexit+0xf/0x30 [kvm_intel]
kvm_mmu_page_fault+0x69/0x4a0 [kvm]
? vmx_vmexit+0xf/0x30 [kvm_intel]
? vmx_vmexit+0x1b/0x30 [kvm_intel]
? vmx_vmexit+0xf/0x30 [kvm_intel]
? vmx_vmexit+0x1b/0x30 [kvm_intel]
? vmx_vmexit+0xf/0x30 [kvm_intel]
kvm_arch_vcpu_ioctl_run+0xa77/0x1910 [kvm]
? __sched_clock_gtod_offset+0x11/0x40
? kvm_vcpu_ioctl+0x366/0x560 [kvm]
kvm_vcpu_ioctl+0x366/0x560 [kvm]
? __wake_up_common+0x91/0x170
do_vfs_ioctl+0x9c/0x620
ksys_ioctl+0x35/0x60
__x64_sys_ioctl+0x11/0x20
do_syscall_64+0x4a/0x100
entry_SYSCALL_64_after_hwframe+0x44/0xa9
RIP: 0033:0x7ff4c891c017
Code: 00 00 00 48 8b 05 81 7e 2b 00 64 c7 00 26 00 00 00 48 c7 c0 ff ff ff ff c3 66 2e 0f 1f 84 00 00 00 00 00 b8 10 00 00 00 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 8b 0d 51 7e 2b 00 f7 d8 64 89 01 48
RSP: 002b:00007ff4be9518b8 EFLAGS: 00000246 ORIG_RAX: 0000000000000010
RAX: ffffffffffffffda RBX: 000000000000ae80 RCX: 00007ff4c891c017
RDX: 0000000000000000 RSI: 000000000000ae80 RDI: 0000000000000019
RBP: 0000561754ae9460 R08: 000056175279a0b0 R09: 00005617549fac90
R10: 00007ff4b42dcc70 R11: 0000000000000246 R12: 0000000000000000
R13: 00007ff4ce157000 R14: 0000000000000000 R15: 0000561754ae9460
Modules linked in: vhost_net vhost tap tun ip6t_REJECT nf_reject_ipv6 nf_log_ipv6 ip6table_nat ip6table_mangle xt_mark xt_conntrack ipt_REJECT nf_reject_ipv4 xt_tcpudp xt_nat xt_multiport nf_log_ipv4 nf_log_common xt_LOG iptable_mangle iptable_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 ebtable_filter ebtables ip6table_filter ip6_tables iptable_filter bpfilter algif_skcipher af_alg binfmt_misc nls_utf8 nls_cp437 vfat fat bridge stp llc hid_generic iTCO_wdt evdev qat_c3xxx pnd2_edac edac_core intel_qat intel_rapl x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul crc32_pclmul snd_pcm ghash_clmulni_intel snd_timer snd efi_pstore soundcore cdc_ether usbnet pcspkr usbhid efivars r8152 hid i2c_i801 i2c_ismt sg dh_generic authenc button pcc_cpufreq acpi_cpufreq efivarfs ip_tables x_tables autofs4 ext4 crc16 mbcache jbd2 dm_crypt dm_mod raid10 raid0 multipath mii raid456 libcrc32c crc32c_generic async_raid6_recov async_memcpy async_pq
async_xor xor async_tx raid6_pq raid1 md_mod crc32c_intel aesni_intel ahci libahci aes_x86_64 crypto_simd r8169 cryptd glue_helper realtek libata libphy thermal
CR2: ffffebe783ff0030
---[ end trace f21ef296e31afc3d ]---
RIP: 0010:compaction_alloc+0x4ce/0x910
Code: 48 c1 e5 06 e9 ca 00 00 00 41 80 be 3d 04 00 00 00 0f 84 dc 00 00 00 49 89 ed 4c 03 2d 9b bd ad 00 4d 85 ed 0f 84 86 00 00 00 <41> 8b 45 30 25 80 00 00 f0 3d 00 00 00 f0 0f 84 28 01 00 00 80 7b
RSP: 0018:ffffa248c0763458 EFLAGS: 00010286
RAX: 0000000000000001 RBX: ffffa248c0763640 RCX: 000000000000003c
RDX: 80000000000ffc00 RSI: 0000000000000000 RDI: ffffa0553fffa900
RBP: 0000000003ff0000 R08: 0000000000000000 R09: 0000000000000000
R10: ffffa248c07634b0 R11: ffffffff99c72438 R12: 8000000000100000
R13: ffffebe783ff0000 R14: ffffffff99c72200 R15: 80000000000ffc00
FS:  00007ff4be952700(0000) GS:ffffa0552fa80000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: ffffebe783ff0030 CR3: 000000046ab8a000 CR4: 00000000003426e0



BUG: unable to handle kernel paging request at ffffecb4c3ff0030
#PF error: [normal kernel read fault]
PGD 46fff9067 P4D 46fff9067 PUD 46fff8067 PMD 0 
Oops: 0000 [#1] SMP NOPTI
CPU: 0 PID: 7794 Comm: CPU 2/KVM Not tainted 5.1.2+ #135
Hardware name: empty empty/S32272NR-C538, BIOS V2.00 08/04/2017
RIP: 0010:compaction_alloc+0x4ce/0x920
Code: 48 c1 e5 06 e9 ca 00 00 00 41 80 be 3d 04 00 00 00 0f 84 dc 00 00 00 49 89 ed 4c 03 2d 9b bd ad 00 4d 85 ed 0f 84 86 00 00 00 <41> 8b 45 30 25 80 00 00 f0 3d 00 00 00 f0 0f 84 28 01 00 00 80 7b
RSP: 0018:ffffa62ac16cf458 EFLAGS: 00010286
RAX: 0000000000000001 RBX: ffffa62ac16cf640 RCX: 000000000000003c
RDX: 80000000000ffc00 RSI: 0000000000000000 RDI: ffffa1283fffa900
RBP: 0000000003ff0000 R08: 0000000000000000 R09: 0000000000000000
R10: ffffa62ac16cf4b0 R11: ffffffff986723f0 R12: 8000000000100000
R13: ffffecb4c3ff0000 R14: ffffffff98672200 R15: 80000000000ffc00
FS:  00007f451e582700(0000) GS:ffffa1282fa00000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: ffffecb4c3ff0030 CR3: 00000004356cc000 CR4: 00000000003426f0
Call Trace:
? move_freelist_tail+0xc0/0xc0
migrate_pages+0x20a/0x790
? isolate_freepages_block+0x340/0x340
compact_zone+0x4d4/0xae0
compact_zone_order+0xda/0x120
? try_to_compact_pages+0x107/0x220
try_to_compact_pages+0x107/0x220
__alloc_pages_direct_compact+0x65/0x140
__alloc_pages_slowpath+0x37f/0xc40
__alloc_pages_nodemask+0x20c/0x260
do_huge_pmd_anonymous_page+0x142/0x5a0
? cpumask_next_and+0x19/0x20
__handle_mm_fault+0xb3c/0xee0
__get_user_pages+0x178/0x6b0
get_user_pages_unlocked+0x143/0x1f0
__gfn_to_pfn_memslot+0x13b/0x3d0 [kvm]
? kvm_irq_delivery_to_apic_fast+0x124/0x3d0 [kvm]
? __switch_to_asm+0x40/0x70
try_async_pf+0x82/0x190 [kvm]
tdp_page_fault+0x12d/0x270 [kvm]
? vmx_vmexit+0xf/0x30 [kvm_intel]
kvm_mmu_page_fault+0x69/0x4a0 [kvm]
? vmx_vmexit+0xf/0x30 [kvm_intel]
? vmx_vmexit+0x1b/0x30 [kvm_intel]
? vmx_vmexit+0xf/0x30 [kvm_intel]
? vmx_vmexit+0x1b/0x30 [kvm_intel]
? vmx_vmexit+0xf/0x30 [kvm_intel]
kvm_arch_vcpu_ioctl_run+0xa77/0x1910 [kvm]
? kvm_vcpu_ioctl+0x366/0x560 [kvm]
kvm_vcpu_ioctl+0x366/0x560 [kvm]
? __switch_to_asm+0x40/0x70
? __switch_to_asm+0x34/0x70
? __switch_to_asm+0x40/0x70
? __switch_to_asm+0x34/0x70
? __switch_to_asm+0x40/0x70
? __switch_to_asm+0x34/0x70
? __switch_to_asm+0x40/0x70
? __switch_to_asm+0x34/0x70
? __switch_to_asm+0x40/0x70
? __switch_to_asm+0x34/0x70
do_vfs_ioctl+0x9c/0x620
? __schedule+0x1dd/0x740
ksys_ioctl+0x35/0x60
__x64_sys_ioctl+0x11/0x20
do_syscall_64+0x4a/0x100
entry_SYSCALL_64_after_hwframe+0x44/0xa9
RIP: 0033:0x7f452854c017
Code: 00 00 00 48 8b 05 81 7e 2b 00 64 c7 00 26 00 00 00 48 c7 c0 ff ff ff ff c3 66 2e 0f 1f 84 00 00 00 00 00 b8 10 00 00 00 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 8b 0d 51 7e 2b 00 f7 d8 64 89 01 48
RSP: 002b:00007f451e5818b8 EFLAGS: 00000246 ORIG_RAX: 0000000000000010
RAX: ffffffffffffffda RBX: 000000000000ae80 RCX: 00007f452854c017
RDX: 0000000000000000 RSI: 000000000000ae80 RDI: 0000000000000018
RBP: 000055c6dc1c5460 R08: 000055c6daae50b0 R09: 000055c6dc0d6c90
R10: 00007f450c2dcce0 R11: 0000000000000246 R12: 0000000000000000
R13: 00007f452dd87000 R14: 0000000000000000 R15: 000055c6dc1c5460
Modules linked in: cpuid vhost_net vhost tap tun ip6t_REJECT nf_reject_ipv6 nf_log_ipv6 ip6table_nat ebtable_filter ebtables ip6table_filter ip6table_mangle ip6_tables xt_mark xt_conntrack ipt_REJECT nf_reject_ipv4 xt_tcpudp xt_nat xt_multiport nf_log_ipv4 nf_log_common xt_LOG iptable_mangle iptable_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 iptable_filter bpfilter algif_skcipher af_alg bridge stp llc binfmt_misc nls_utf8 nls_cp437 vfat fat hid_generic iTCO_wdt evdev qat_c3xxx intel_qat pnd2_edac edac_core intel_rapl x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul crc32_pclmul ghash_clmulni_intel snd_pcm efi_pstore snd_timer cdc_ether snd usbnet soundcore usbhid hid efivars r8152 pcspkr i2c_i801 i2c_ismt sg dh_generic authenc button pcc_cpufreq acpi_cpufreq efivarfs ip_tables x_tables autofs4 ext4 crc16 mbcache jbd2 dm_crypt dm_mod raid10 raid0 multipath mii raid456 libcrc32c crc32c_generic async_raid6_recov async_memcpy
async_pq async_xor xor async_tx raid1 raid6_pq md_mod crc32c_intel aesni_intel ahci libahci aes_x86_64 crypto_simd cryptd glue_helper r8169 libata realtek libphy thermal
CR2: ffffecb4c3ff0030
---[ end trace 8eb576fb4a5bab98 ]---
RIP: 0010:compaction_alloc+0x4ce/0x920
Code: 48 c1 e5 06 e9 ca 00 00 00 41 80 be 3d 04 00 00 00 0f 84 dc 00 00 00 49 89 ed 4c 03 2d 9b bd ad 00 4d 85 ed 0f 84 86 00 00 00 <41> 8b 45 30 25 80 00 00 f0 3d 00 00 00 f0 0f 84 28 01 00 00 80 7b
RSP: 0018:ffffa62ac16cf458 EFLAGS: 00010286
RAX: 0000000000000001 RBX: ffffa62ac16cf640 RCX: 000000000000003c
RDX: 80000000000ffc00 RSI: 0000000000000000 RDI: ffffa1283fffa900
RBP: 0000000003ff0000 R08: 0000000000000000 R09: 0000000000000000
R10: ffffa62ac16cf4b0 R11: ffffffff986723f0 R12: 8000000000100000
R13: ffffecb4c3ff0000 R14: ffffffff98672200 R15: 80000000000ffc00
FS:  00007f451e582700(0000) GS:ffffa1282fa00000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: ffffecb4c3ff0030 CR3: 00000004356cc000 CR4: 00000000003426f0

-- 
Kind regards,

Sakari Ailus

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000
  2019-05-12  8:27 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000 Justin Piszcz
  2019-05-12  8:34 ` Justin Piszcz
@ 2019-05-20 11:56 ` Mel Gorman
  2019-05-20 12:20   ` Justin Piszcz
  2019-05-21  9:01   ` Justin Piszcz
  1 sibling, 2 replies; 12+ messages in thread
From: Mel Gorman @ 2019-05-20 11:56 UTC (permalink / raw)
  To: Justin Piszcz; +Cc: LKML

On Sun, May 12, 2019 at 04:27:45AM -0400, Justin Piszcz wrote:
> Hello,
> 
> I've turned off zram/zswap and I am still seeing the following during
> periods of heavy I/O, I am returning to 5.0.xx in the meantime.
> 
> Kernel: 5.1.1
> Arch: x86_64
> Dist: Debian x86_64
> 
> [29967.019411] BUG: unable to handle kernel paging request at ffffea0002030000
> [29967.019414] #PF error: [normal kernel read fault]
> [29967.019415] PGD 103ffee067 P4D 103ffee067 PUD 103ffed067 PMD 0
> [29967.019417] Oops: 0000 [#1] SMP PTI
> [29967.019419] CPU: 10 PID: 77 Comm: khugepaged Tainted: G
>    T 5.1.1 #4
> [29967.019420] Hardware name: Supermicro X9SRL-F/X9SRL-F, BIOS 3.2 01/16/2015
> [29967.019424] RIP: 0010:isolate_freepages_block+0xb9/0x310
> [29967.019425] Code: 24 28 48 c1 e0 06 40 f6 c5 1f 48 89 44 24 20 49
> 8d 45 79 48 89 44 24 18 44 89 f0 4d 89 ee 45 89 fd 41 89 c7 0f 84 ef
> 00 00 00 <48> 8b 03 41 83 c4 01 a9 00 00 01 00 75 0c 48 8b 43 08 a8 01
> 0f 84

If you have debugging symbols installed, can you translate the faulting
address with the following?

ADDR=`nm /path/to/vmlinux-or-debuginfo-file | grep "t isolate_freepages_block\$" | awk '{print $1}'`
addr2line -i -e vmlinux `printf "0x%lX" $((0x$ADDR+0xb9))`

I haven't seen this particular error before so I want to see if the
faulting address could have anything to do with the randomising of
struct fields. Similarly a full dmesg log would be nice so I can see
what your memory layout looks like.

Thanks Justin.

-- 
Mel Gorman
SUSE Labs

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000
  2019-05-20 11:56 ` Mel Gorman
@ 2019-05-20 12:20   ` Justin Piszcz
  2019-05-21  9:01   ` Justin Piszcz
  1 sibling, 0 replies; 12+ messages in thread
From: Justin Piszcz @ 2019-05-20 12:20 UTC (permalink / raw)
  To: Mel Gorman; +Cc: LKML

On Mon, May 20, 2019 at 7:56 AM Mel Gorman <mgorman@techsingularity.net> wrote:
>
> On Sun, May 12, 2019 at 04:27:45AM -0400, Justin Piszcz wrote:
> > Hello,
> >
> > I've turned off zram/zswap and I am still seeing the following during
> > periods of heavy I/O, I am returning to 5.0.xx in the meantime.
> >
> > Kernel: 5.1.1
> > Arch: x86_64
> > Dist: Debian x86_64
> >
> > [29967.019411] BUG: unable to handle kernel paging request at ffffea0002030000
> > [29967.019414] #PF error: [normal kernel read fault]
> > [29967.019415] PGD 103ffee067 P4D 103ffee067 PUD 103ffed067 PMD 0
> > [29967.019417] Oops: 0000 [#1] SMP PTI
> > [29967.019419] CPU: 10 PID: 77 Comm: khugepaged Tainted: G
> >    T 5.1.1 #4
> > [29967.019420] Hardware name: Supermicro X9SRL-F/X9SRL-F, BIOS 3.2 01/16/2015
> > [29967.019424] RIP: 0010:isolate_freepages_block+0xb9/0x310
> > [29967.019425] Code: 24 28 48 c1 e0 06 40 f6 c5 1f 48 89 44 24 20 49
> > 8d 45 79 48 89 44 24 18 44 89 f0 4d 89 ee 45 89 fd 41 89 c7 0f 84 ef
> > 00 00 00 <48> 8b 03 41 83 c4 01 a9 00 00 01 00 75 0c 48 8b 43 08 a8 01
> > 0f 84
>
> If you have debugging symbols installed, can you translate the faulting
> address with the following?
>
> ADDR=`nm /path/to/vmlinux-or-debuginfo-file | grep "t isolate_freepages_block\$" | awk '{print $1}'`
> addr2line -i -e vmlinux `printf "0x%lX" $((0x$ADDR+0xb9))`
>
> I haven't seen this particular error before so I want to see if the
> faulting address could have anything to do with the randomising of
> struct fields. Similarly a full dmesg log would be nice so I can see
> what your memory layout looks like.
>
> Thanks Justin.

Hello,

DMESG output in the interim:
http://installkernel.tripod.com/20190520_dmesg_full.txt

Once the error happens again I will post another dmesg link and get
the addr2line output.

Justin.

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000
  2019-05-20 11:56 ` Mel Gorman
  2019-05-20 12:20   ` Justin Piszcz
@ 2019-05-21  9:01   ` Justin Piszcz
  2019-05-21 12:43     ` Mel Gorman
  1 sibling, 1 reply; 12+ messages in thread
From: Justin Piszcz @ 2019-05-21  9:01 UTC (permalink / raw)
  To: Mel Gorman; +Cc: LKML

On Mon, May 20, 2019 at 7:56 AM Mel Gorman <mgorman@techsingularity.net> wrote:
>
> On Sun, May 12, 2019 at 04:27:45AM -0400, Justin Piszcz wrote:
> > Hello,
> >
> > I've turned off zram/zswap and I am still seeing the following during
> > periods of heavy I/O, I am returning to 5.0.xx in the meantime.
> >
> > Kernel: 5.1.1
> > Arch: x86_64
> > Dist: Debian x86_64
> >
> > [29967.019411] BUG: unable to handle kernel paging request at ffffea0002030000
> > [29967.019414] #PF error: [normal kernel read fault]
> > [29967.019415] PGD 103ffee067 P4D 103ffee067 PUD 103ffed067 PMD 0
> > [29967.019417] Oops: 0000 [#1] SMP PTI
> > [29967.019419] CPU: 10 PID: 77 Comm: khugepaged Tainted: G
> >    T 5.1.1 #4
> > [29967.019420] Hardware name: Supermicro X9SRL-F/X9SRL-F, BIOS 3.2 01/16/2015
> > [29967.019424] RIP: 0010:isolate_freepages_block+0xb9/0x310
> > [29967.019425] Code: 24 28 48 c1 e0 06 40 f6 c5 1f 48 89 44 24 20 49
> > 8d 45 79 48 89 44 24 18 44 89 f0 4d 89 ee 45 89 fd 41 89 c7 0f 84 ef
> > 00 00 00 <48> 8b 03 41 83 c4 01 a9 00 00 01 00 75 0c 48 8b 43 08 a8 01
> > 0f 84
>
> If you have debugging symbols installed, can you translate the faulting
> address with the following?
>
> ADDR=`nm /path/to/vmlinux-or-debuginfo-file | grep "t isolate_freepages_block\$" | awk '{print $1}'`
> addr2line -i -e vmlinux `printf "0x%lX" $((0x$ADDR+0xb9))`

Another event this morning, this occurred when copying a single ~25GB
backup file from one block device device (3ware HW RAID) to a SW
RAID-1 (mdadm):

With this event, it was a fault and khugepaged is not stuck at 100%
but this may be related as the stack trace is similar where
compaction_alloc is utilizing most of the CPU:
https://lkml.org/lkml/2019/5/9/225

# ADDR=`nm /usr/src/linux/vmlinux | grep "t isolate_freepages_block\$"
| awk '{print $1}'`
# echo $ADDR
ffffffff812274f0
# addr2line -i -e /usr/src/linux/vmlinux `printf "0x%lX" $((0x$ADDR+0x83d))`
compaction.c:?
# addr2line -i -e /usr/src/linux/vmlinux `printf "0x%lX" $((0x$ADDR+0x8d0))`
compaction.c:?

# grep DEBUG_INFO /usr/src/linux/.config-5.1.3-2
CONFIG_DEBUG_INFO=y

I can help test again in a 2-3 weeks if needed but for now I need to
return back to disabling transparent huge pages.

[43775.068702] BUG: unable to handle kernel paging request at ffffea0002250000
[43775.068706] #PF error: [normal kernel read fault]
[43775.068707] PGD 103ffee067 P4D 103ffee067 PUD 103ffed067 PMD 0
[43775.068709] Oops: 0000 [#1] SMP PTI
[43775.068711] CPU: 1 PID: 77 Comm: khugepaged Tainted: G
  T 5.1.3 #2
[43775.068712] Hardware name: Supermicro X9SRL-F/X9SRL-F, BIOS 3.2 01/16/2015
[43775.068717] RIP: 0010:isolate_freepages_block+0xb9/0x310
[43775.068718] Code: 24 28 48 c1 e0 06 40 f6 c5 1f 48 89 44 24 20 49
8d 45 79 48 89 44 24 18 44 89 f0 4d 89 ee 45 89 fd 41 89 c7 0f 84 ef
00 00 00 <48> 8b 03 41 83 c4 01 a9 00 00 01 00 75 0c 48 8b 43 08 a8 01
0f 84
[43775.068719] RSP: 0018:ffffc900003a7860 EFLAGS: 00010246
[43775.068720] RAX: 0000000000000000 RBX: ffffea0002250000 RCX: ffffc900003a7b69
[43775.068721] RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffffffff828c4d90
[43775.068721] RBP: 0000000000089400 R08: 0000000000000001 R09: 0000000000000000
[43775.068722] R10: 0000000000000202 R11: ffffffff828c49d0 R12: 0000000000000000
[43775.068723] R13: 0000000000000000 R14: ffffc900003a7af0 R15: 0000000000000000
[43775.068724] FS:  0000000000000000(0000) GS:ffff88903f840000(0000)
knlGS:0000000000000000
[43775.068725] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[43775.068725] CR2: ffffea0002250000 CR3: 000000000280e002 CR4: 00000000001606e0
[43775.068726] Call Trace:
[43775.068730]  compaction_alloc+0x83d/0x8d0
[43775.068732]  migrate_pages+0x30d/0x750
[43775.068734]  ? isolate_migratepages_block+0xa10/0xa10
[43775.068735]  ? __reset_isolation_suitable+0x110/0x110
[43775.068736]  compact_zone+0x684/0xa70
[43775.068738]  compact_zone_order+0x109/0x150
[43775.068741]  ? schedule_timeout+0x1ba/0x290
[43775.068743]  ? record_times+0x13/0xa0
[43775.068744]  try_to_compact_pages+0x10d/0x220
[43775.068747]  __alloc_pages_direct_compact+0x93/0x180
[43775.068748]  __alloc_pages_nodemask+0x6c7/0xe20
[43775.068751]  ? __wake_up_common_lock+0xb0/0xb0
[43775.068752]  khugepaged+0x31f/0x19c0
[43775.068754]  ? __wake_up_common_lock+0xb0/0xb0
[43775.068755]  ? __wake_up_common_lock+0xb0/0xb0
[43775.068756]  ? collapse_shmem.isra.4+0xc20/0xc20
[43775.068759]  kthread+0x10a/0x120
[43775.068761]  ? __kthread_create_on_node+0x1b0/0x1b0
[43775.068762]  ret_from_fork+0x35/0x40
[43775.068763] CR2: ffffea0002250000
[43775.068764] ---[ end trace 1f63fc0e799750fe ]---
[43775.068766] RIP: 0010:isolate_freepages_block+0xb9/0x310
[43775.068767] Code: 24 28 48 c1 e0 06 40 f6 c5 1f 48 89 44 24 20 49
8d 45 79 48 89 44 24 18 44 89 f0 4d 89 ee 45 89 fd 41 89 c7 0f 84 ef
00 00 00 <48> 8b 03 41 83 c4 01 a9 00 00 01 00 75 0c 48 8b 43 08 a8 01
0f 84
[43775.068767] RSP: 0018:ffffc900003a7860 EFLAGS: 00010246
[43775.068768] RAX: 0000000000000000 RBX: ffffea0002250000 RCX: ffffc900003a7b69
[43775.068769] RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffffffff828c4d90
[43775.068770] RBP: 0000000000089400 R08: 0000000000000001 R09: 0000000000000000
[43775.068770] R10: 0000000000000202 R11: ffffffff828c49d0 R12: 0000000000000000
[43775.068771] R13: 0000000000000000 R14: ffffc900003a7af0 R15: 0000000000000000
[43775.068772] FS:  0000000000000000(0000) GS:ffff88903f840000(0000)
knlGS:0000000000000000
[43775.068773] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[43775.068773] CR2: ffffea0002250000 CR3: 000000000280e002 CR4: 00000000001606e0

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000
  2019-05-21  9:01   ` Justin Piszcz
@ 2019-05-21 12:43     ` Mel Gorman
  2019-05-24 11:43       ` Oleksandr Natalenko
  0 siblings, 1 reply; 12+ messages in thread
From: Mel Gorman @ 2019-05-21 12:43 UTC (permalink / raw)
  To: Justin Piszcz; +Cc: LKML

On Tue, May 21, 2019 at 05:01:06AM -0400, Justin Piszcz wrote:
> On Mon, May 20, 2019 at 7:56 AM Mel Gorman <mgorman@techsingularity.net> wrote:
> >
> > On Sun, May 12, 2019 at 04:27:45AM -0400, Justin Piszcz wrote:
> > > Hello,
> > >
> > > I've turned off zram/zswap and I am still seeing the following during
> > > periods of heavy I/O, I am returning to 5.0.xx in the meantime.
> > >
> > > Kernel: 5.1.1
> > > Arch: x86_64
> > > Dist: Debian x86_64
> > >
> > > [29967.019411] BUG: unable to handle kernel paging request at ffffea0002030000
> > > [29967.019414] #PF error: [normal kernel read fault]
> > > [29967.019415] PGD 103ffee067 P4D 103ffee067 PUD 103ffed067 PMD 0
> > > [29967.019417] Oops: 0000 [#1] SMP PTI
> > > [29967.019419] CPU: 10 PID: 77 Comm: khugepaged Tainted: G
> > >    T 5.1.1 #4
> > > [29967.019420] Hardware name: Supermicro X9SRL-F/X9SRL-F, BIOS 3.2 01/16/2015
> > > [29967.019424] RIP: 0010:isolate_freepages_block+0xb9/0x310
> > > [29967.019425] Code: 24 28 48 c1 e0 06 40 f6 c5 1f 48 89 44 24 20 49
> > > 8d 45 79 48 89 44 24 18 44 89 f0 4d 89 ee 45 89 fd 41 89 c7 0f 84 ef
> > > 00 00 00 <48> 8b 03 41 83 c4 01 a9 00 00 01 00 75 0c 48 8b 43 08 a8 01
> > > 0f 84
> >
> > If you have debugging symbols installed, can you translate the faulting
> > address with the following?
> >
> > ADDR=`nm /path/to/vmlinux-or-debuginfo-file | grep "t isolate_freepages_block\$" | awk '{print $1}'`
> > addr2line -i -e vmlinux `printf "0x%lX" $((0x$ADDR+0xb9))`
> 
> Another event this morning, this occurred when copying a single ~25GB
> backup file from one block device device (3ware HW RAID) to a SW
> RAID-1 (mdadm):
> 
> With this event, it was a fault and khugepaged is not stuck at 100%
> but this may be related as the stack trace is similar where
> compaction_alloc is utilizing most of the CPU:
> https://lkml.org/lkml/2019/5/9/225
> 
> # ADDR=`nm /usr/src/linux/vmlinux | grep "t isolate_freepages_block\$"
> | awk '{print $1}'`
> # echo $ADDR
> ffffffff812274f0
> # addr2line -i -e /usr/src/linux/vmlinux `printf "0x%lX" $((0x$ADDR+0x83d))`
> compaction.c:?
> # addr2line -i -e /usr/src/linux/vmlinux `printf "0x%lX" $((0x$ADDR+0x8d0))`
> compaction.c:?
> 

Please use the offset 0xb9

addr2line -i -e /usr/src/linux/vmlinux `printf "0x%lX" $((0x$ADDR+0xb9))

-- 
Mel Gorman
SUSE Labs

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000
  2019-05-21 12:43     ` Mel Gorman
@ 2019-05-24 11:43       ` Oleksandr Natalenko
  2019-05-24 12:31         ` Mel Gorman
  0 siblings, 1 reply; 12+ messages in thread
From: Oleksandr Natalenko @ 2019-05-24 11:43 UTC (permalink / raw)
  To: Mel Gorman; +Cc: Justin Piszcz, LKML

On Tue, May 21, 2019 at 01:43:10PM +0100, Mel Gorman wrote:
> On Tue, May 21, 2019 at 05:01:06AM -0400, Justin Piszcz wrote:
> > On Mon, May 20, 2019 at 7:56 AM Mel Gorman <mgorman@techsingularity.net> wrote:
> > >
> > > On Sun, May 12, 2019 at 04:27:45AM -0400, Justin Piszcz wrote:
> > > > Hello,
> > > >
> > > > I've turned off zram/zswap and I am still seeing the following during
> > > > periods of heavy I/O, I am returning to 5.0.xx in the meantime.
> > > >
> > > > Kernel: 5.1.1
> > > > Arch: x86_64
> > > > Dist: Debian x86_64
> > > >
> > > > [29967.019411] BUG: unable to handle kernel paging request at ffffea0002030000
> > > > [29967.019414] #PF error: [normal kernel read fault]
> > > > [29967.019415] PGD 103ffee067 P4D 103ffee067 PUD 103ffed067 PMD 0
> > > > [29967.019417] Oops: 0000 [#1] SMP PTI
> > > > [29967.019419] CPU: 10 PID: 77 Comm: khugepaged Tainted: G
> > > >    T 5.1.1 #4
> > > > [29967.019420] Hardware name: Supermicro X9SRL-F/X9SRL-F, BIOS 3.2 01/16/2015
> > > > [29967.019424] RIP: 0010:isolate_freepages_block+0xb9/0x310
> > > > [29967.019425] Code: 24 28 48 c1 e0 06 40 f6 c5 1f 48 89 44 24 20 49
> > > > 8d 45 79 48 89 44 24 18 44 89 f0 4d 89 ee 45 89 fd 41 89 c7 0f 84 ef
> > > > 00 00 00 <48> 8b 03 41 83 c4 01 a9 00 00 01 00 75 0c 48 8b 43 08 a8 01
> > > > 0f 84
> > >
> > > If you have debugging symbols installed, can you translate the faulting
> > > address with the following?
> > >
> > > ADDR=`nm /path/to/vmlinux-or-debuginfo-file | grep "t isolate_freepages_block\$" | awk '{print $1}'`
> > > addr2line -i -e vmlinux `printf "0x%lX" $((0x$ADDR+0xb9))`
> > 
> > Another event this morning, this occurred when copying a single ~25GB
> > backup file from one block device device (3ware HW RAID) to a SW
> > RAID-1 (mdadm):
> > 
> > With this event, it was a fault and khugepaged is not stuck at 100%
> > but this may be related as the stack trace is similar where
> > compaction_alloc is utilizing most of the CPU:
> > https://lkml.org/lkml/2019/5/9/225
> > 
> > # ADDR=`nm /usr/src/linux/vmlinux | grep "t isolate_freepages_block\$"
> > | awk '{print $1}'`
> > # echo $ADDR
> > ffffffff812274f0
> > # addr2line -i -e /usr/src/linux/vmlinux `printf "0x%lX" $((0x$ADDR+0x83d))`
> > compaction.c:?
> > # addr2line -i -e /usr/src/linux/vmlinux `printf "0x%lX" $((0x$ADDR+0x8d0))`
> > compaction.c:?
> > 
> 
> Please use the offset 0xb9
> 
> addr2line -i -e /usr/src/linux/vmlinux `printf "0x%lX" $((0x$ADDR+0xb9))
> 
> -- 
> Mel Gorman
> SUSE Labs

Cc'ing myself since i observe such a behaviour sometimes right after KVM
VM is launched. No luck with reproducing it on demand so far, though.

-- 
  Best regards,
    Oleksandr Natalenko (post-factum)
    Senior Software Maintenance Engineer

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000
  2019-05-24 11:43       ` Oleksandr Natalenko
@ 2019-05-24 12:31         ` Mel Gorman
  2019-05-24 12:38           ` Oleksandr Natalenko
  2019-05-27  6:23           ` Oleksandr Natalenko
  0 siblings, 2 replies; 12+ messages in thread
From: Mel Gorman @ 2019-05-24 12:31 UTC (permalink / raw)
  To: Oleksandr Natalenko; +Cc: Justin Piszcz, LKML

On Fri, May 24, 2019 at 01:43:29PM +0200, Oleksandr Natalenko wrote:
> > Please use the offset 0xb9
> > 
> > addr2line -i -e /usr/src/linux/vmlinux `printf "0x%lX" $((0x$ADDR+0xb9))
> > 
> > -- 
> > Mel Gorman
> > SUSE Labs
> 
> Cc'ing myself since i observe such a behaviour sometimes right after KVM
> VM is launched. No luck with reproducing it on demand so far, though.
> 

It's worth testing the patch at https://lore.kernel.org/lkml/1558689619-16891-1-git-send-email-suzuki.poulose@arm.com/T/#u

-- 
Mel Gorman
SUSE Labs

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000
  2019-05-24 12:31         ` Mel Gorman
@ 2019-05-24 12:38           ` Oleksandr Natalenko
  2019-05-27  6:23           ` Oleksandr Natalenko
  1 sibling, 0 replies; 12+ messages in thread
From: Oleksandr Natalenko @ 2019-05-24 12:38 UTC (permalink / raw)
  To: Mel Gorman; +Cc: Justin Piszcz, LKML

On Fri, May 24, 2019 at 01:31:46PM +0100, Mel Gorman wrote:
> On Fri, May 24, 2019 at 01:43:29PM +0200, Oleksandr Natalenko wrote:
> > > Please use the offset 0xb9
> > > 
> > > addr2line -i -e /usr/src/linux/vmlinux `printf "0x%lX" $((0x$ADDR+0xb9))
> > > 
> > > -- 
> > > Mel Gorman
> > > SUSE Labs
> > 
> > Cc'ing myself since i observe such a behaviour sometimes right after KVM
> > VM is launched. No luck with reproducing it on demand so far, though.
> > 
> 
> It's worth testing the patch at https://lore.kernel.org/lkml/1558689619-16891-1-git-send-email-suzuki.poulose@arm.com/T/#u

On my TODO list already, thanks.

> 
> -- 
> Mel Gorman
> SUSE Labs

-- 
  Best regards,
    Oleksandr Natalenko (post-factum)
    Senior Software Maintenance Engineer

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000
  2019-05-24 12:31         ` Mel Gorman
  2019-05-24 12:38           ` Oleksandr Natalenko
@ 2019-05-27  6:23           ` Oleksandr Natalenko
  2019-05-27  8:25             ` Mel Gorman
  1 sibling, 1 reply; 12+ messages in thread
From: Oleksandr Natalenko @ 2019-05-27  6:23 UTC (permalink / raw)
  To: Mel Gorman; +Cc: Justin Piszcz, LKML

On Fri, May 24, 2019 at 01:31:46PM +0100, Mel Gorman wrote:
> On Fri, May 24, 2019 at 01:43:29PM +0200, Oleksandr Natalenko wrote:
> > > Please use the offset 0xb9
> > > 
> > > addr2line -i -e /usr/src/linux/vmlinux `printf "0x%lX" $((0x$ADDR+0xb9))
> > > 
> > > -- 
> > > Mel Gorman
> > > SUSE Labs
> > 
> > Cc'ing myself since i observe such a behaviour sometimes right after KVM
> > VM is launched. No luck with reproducing it on demand so far, though.
> > 
> 
> It's worth testing the patch at https://lore.kernel.org/lkml/1558689619-16891-1-git-send-email-suzuki.poulose@arm.com/T/#u

It is reported [1] that this fixes the issue.

Thanks.

[1] https://github.com/NixOS/nixpkgs/issues/61909

> 
> -- 
> Mel Gorman
> SUSE Labs

-- 
  Best regards,
    Oleksandr Natalenko (post-factum)
    Senior Software Maintenance Engineer

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000
  2019-05-27  6:23           ` Oleksandr Natalenko
@ 2019-05-27  8:25             ` Mel Gorman
  0 siblings, 0 replies; 12+ messages in thread
From: Mel Gorman @ 2019-05-27  8:25 UTC (permalink / raw)
  To: Oleksandr Natalenko; +Cc: Justin Piszcz, LKML

On Mon, May 27, 2019 at 08:23:01AM +0200, Oleksandr Natalenko wrote:
> On Fri, May 24, 2019 at 01:31:46PM +0100, Mel Gorman wrote:
> > On Fri, May 24, 2019 at 01:43:29PM +0200, Oleksandr Natalenko wrote:
> > > > Please use the offset 0xb9
> > > > 
> > > > addr2line -i -e /usr/src/linux/vmlinux `printf "0x%lX" $((0x$ADDR+0xb9))
> > > > 
> > > > -- 
> > > > Mel Gorman
> > > > SUSE Labs
> > > 
> > > Cc'ing myself since i observe such a behaviour sometimes right after KVM
> > > VM is launched. No luck with reproducing it on demand so far, though.
> > > 
> > 
> > It's worth testing the patch at https://lore.kernel.org/lkml/1558689619-16891-1-git-send-email-suzuki.poulose@arm.com/T/#u
> 
> It is reported [1] that this fixes the issue.
> 
> Thanks.
> 
> [1] https://github.com/NixOS/nixpkgs/issues/61909
> 

Thanks! The patch has been picked up by Andrew and it's marked for
-stable. It should make its way into a stable release eventually.

-- 
Mel Gorman
SUSE Labs

^ permalink raw reply	[flat|nested] 12+ messages in thread

end of thread, other threads:[~2019-05-27  8:25 UTC | newest]

Thread overview: 12+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2019-05-12  8:27 5.1 and 5.1.1: BUG: unable to handle kernel paging request at ffffea0002030000 Justin Piszcz
2019-05-12  8:34 ` Justin Piszcz
2019-05-17 21:30   ` Sakari Ailus
2019-05-20 11:56 ` Mel Gorman
2019-05-20 12:20   ` Justin Piszcz
2019-05-21  9:01   ` Justin Piszcz
2019-05-21 12:43     ` Mel Gorman
2019-05-24 11:43       ` Oleksandr Natalenko
2019-05-24 12:31         ` Mel Gorman
2019-05-24 12:38           ` Oleksandr Natalenko
2019-05-27  6:23           ` Oleksandr Natalenko
2019-05-27  8:25             ` Mel Gorman

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).