[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: (Debian) Bug#1051862: server flooded with xen_mc_flush warnings with xen 4.17 + linux 6.1


  • To: Hans van Kranenburg <hans@xxxxxxxxxxx>, Radoslav Bodó <radoslav.bodo@xxxxxxxxxxx>, 1051862@xxxxxxxxxxxxxxx, Boris Ostrovsky <boris.ostrovsky@xxxxxxxxxx>
  • From: Juergen Gross <jgross@xxxxxxxx>
  • Date: Thu, 14 Sep 2023 07:43:41 +0200
  • Autocrypt: addr=jgross@xxxxxxxx; keydata= xsBNBFOMcBYBCACgGjqjoGvbEouQZw/ToiBg9W98AlM2QHV+iNHsEs7kxWhKMjrioyspZKOB ycWxw3ie3j9uvg9EOB3aN4xiTv4qbnGiTr3oJhkB1gsb6ToJQZ8uxGq2kaV2KL9650I1SJve dYm8Of8Zd621lSmoKOwlNClALZNew72NjJLEzTalU1OdT7/i1TXkH09XSSI8mEQ/ouNcMvIJ NwQpd369y9bfIhWUiVXEK7MlRgUG6MvIj6Y3Am/BBLUVbDa4+gmzDC9ezlZkTZG2t14zWPvx XP3FAp2pkW0xqG7/377qptDmrk42GlSKN4z76ELnLxussxc7I2hx18NUcbP8+uty4bMxABEB AAHNH0p1ZXJnZW4gR3Jvc3MgPGpncm9zc0BzdXNlLmNvbT7CwHkEEwECACMFAlOMcK8CGwMH CwkIBwMCAQYVCAIJCgsEFgIDAQIeAQIXgAAKCRCw3p3WKL8TL8eZB/9G0juS/kDY9LhEXseh mE9U+iA1VsLhgDqVbsOtZ/S14LRFHczNd/Lqkn7souCSoyWsBs3/wO+OjPvxf7m+Ef+sMtr0 G5lCWEWa9wa0IXx5HRPW/ScL+e4AVUbL7rurYMfwCzco+7TfjhMEOkC+va5gzi1KrErgNRHH kg3PhlnRY0Udyqx++UYkAsN4TQuEhNN32MvN0Np3WlBJOgKcuXpIElmMM5f1BBzJSKBkW0Jc Wy3h2Wy912vHKpPV/Xv7ZwVJ27v7KcuZcErtptDevAljxJtE7aJG6WiBzm+v9EswyWxwMCIO RoVBYuiocc51872tRGywc03xaQydB+9R7BHPzsBNBFOMcBYBCADLMfoA44MwGOB9YT1V4KCy vAfd7E0BTfaAurbG+Olacciz3yd09QOmejFZC6AnoykydyvTFLAWYcSCdISMr88COmmCbJzn sHAogjexXiif6ANUUlHpjxlHCCcELmZUzomNDnEOTxZFeWMTFF9Rf2k2F0Tl4E5kmsNGgtSa aMO0rNZoOEiD/7UfPP3dfh8JCQ1VtUUsQtT1sxos8Eb/HmriJhnaTZ7Hp3jtgTVkV0ybpgFg w6WMaRkrBh17mV0z2ajjmabB7SJxcouSkR0hcpNl4oM74d2/VqoW4BxxxOD1FcNCObCELfIS auZx+XT6s+CE7Qi/c44ibBMR7hyjdzWbABEBAAHCwF8EGAECAAkFAlOMcBYCGwwACgkQsN6d 1ii/Ey9D+Af/WFr3q+bg/8v5tCknCtn92d5lyYTBNt7xgWzDZX8G6/pngzKyWfedArllp0Pn fgIXtMNV+3t8Li1Tg843EXkP7+2+CQ98MB8XvvPLYAfW8nNDV85TyVgWlldNcgdv7nn1Sq8g HwB2BHdIAkYce3hEoDQXt/mKlgEGsLpzJcnLKimtPXQQy9TxUaLBe9PInPd+Ohix0XOlY+Uk QFEx50Ki3rSDl2Zt2tnkNYKUCvTJq7jvOlaPd6d/W0tZqpyy7KVay+K4aMobDsodB3dvEAs6 ScCnh03dDAFgIq5nsB11j3KPKdVoPlfucX2c7kGNH+LUMbzqV6beIENfNexkOfxHfw==
  • Cc: xen-devel <xen-devel@xxxxxxxxxxxxxxxxxxxx>
  • Delivery-date: Thu, 14 Sep 2023 05:43:57 +0000
  • List-id: Xen developer discussion <xen-devel.lists.xenproject.org>

Hi Hans,

On 13.09.23 23:38, Hans van Kranenburg wrote:
Hi Radoslav,

Thanks for your report...

Hi Juergen, Boris and xen-devel,

At Debian, we got the report below. (Also at
https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=1051862)

This hardware, with only Xen and Dom0 running is hitting the failed
multicall warning and logging in arch/x86/xen/multicalls.c. Can you help
advise what we can do to further debug this issue?

Since this looks like pretty low level Xen/hardware stuff, I'd rather
ask upstream for directions first. If needed the Debian Xen Team can
assist the end user with the debugging process.

Thanks,

More reply inline...

On 9/13/23 20:12, Radoslav Bodó wrote:
Package: xen-system-amd64
Version: 4.17.1+2-gb773c48e36-1
Severity: important

Hello,

after upgrade from Bullseye to Bookworm one of our dom0's
became unusable due to logs/system being continuously flooded
with warnings from arch/x86/xen/multicalls.c:102 xen_mc_flush, and the
system become unusable.

The issue starts at some point where system services starts to come up,
but nothing very special is on that box (dom0, nftables, fail2ban,
prometheus-node-exporter, 3x domU). We have tried to disable all domU's
and fail2ban as the name of the process would suggest, but issue is
still present. We have tried also some other elaboration but none of
them have helped so far:

* the issue arise when xen 4.17 + linux >= 6.1 is booted
* xen + bookworm-backports linux-image-6.4.0-0.deb12.2-amd64 have same isuue
* without xen hypervisor, linux 6.1 runs just fine
* systemrescue cd boot and xfs_repair rootfs did not helped
* memtest seem to be fine running for hours

Thanks for already trying out all these combinations.

As a workaround we have booted xen 4.17 + linux 5.10.0-25 (5.10.191-1)
and the system is running fine as for last few months.

Hardware:
* Dell PowerEdge R750xs
* 2x Intel Xeon Silver 4310 2.1G
* 256GB RAM
* PERC H755 Adapter, 12x 18TB HDDs

I have a few quick additional questions already:

1. For clarification.. From your text, I understand that only this one
single server is showing the problem after the Debian version upgrade.
Does this mean that this is the only server you have running with
exactly this combination of hardware (and BIOS version, CPU microcode
etc etc)? Or, is there another one with same hardware which does not
show the problem?

2. Can you reply with the output of 'xl dmesg' when the problem happens?
Or, if the system gets unusable too quick, do you have a serial console
connection to capture the output?

3. To confirm... I understand that there are many of these messages.
Since you pasted only one, does that mean that all of them look exactly
the same, with "1 of 1 multicall(s) failed: cpu 10" "call  1: op=1
arg=[ffff8888a1a9eb10] result=-22"? Or are there variations? If so, can
you reply with a few different ones?

Since this very much looks like an issue of Xen related code where the
Xen hypervisor, dom0 kernel and hardware has to work together correctly,
(and not a Debian packaging problem) I'm already asking upstream for
advice about what we should/could do next, instead of trying to make a
guess myself.

Thanks,
Hans

Any help, advice or bug confirmation would be appreciated

Best regards
bodik


(log also in attachment)

```
kernel: [   99.762402] WARNING: CPU: 10 PID: 1301 at
arch/x86/xen/multicalls.c:102 xen_mc_flush+0x196/0x220
kernel: [   99.762598] Modules linked in: nvme_fabrics nvme_core bridge
xen_acpi_processor xen_gntdev stp llc xen_evtchn xenfs xen_privcmd
binfmt_misc intel_rapl_msr ext4 intel_rapl_common crc16
intel_uncore_frequency_common mbcache ipmi_ssif jbd2 nfit libnvdimm
ghash_clmulni_intel sha512_ssse3 sha512_generic aesni_intel acpi_ipmi
nft_ct crypto_simd cryptd mei_me mgag200 ipmi_si iTCO_wdt intel_pmc_bxt
ipmi_devintf drm_shmem_helper dell_smbios nft_masq iTCO_vendor_support
isst_if_mbox_pci drm_kms_helper isst_if_mmio dcdbas mei intel_vsec
isst_if_common dell_wmi_descriptor wmi_bmof watchdog pcspkr
intel_pch_thermal ipmi_msghandler i2c_algo_bit acpi_power_meter button
nft_nat joydev evdev sg nft_chain_nat nf_nat nf_conntrack nf_defrag_ipv6
nf_defrag_ipv4 nf_tables nfnetlink drm fuse loop efi_pstore configfs
ip_tables x_tables autofs4 xfs libcrc32c crc32c_generic hid_generic
usbhid hid dm_mod sd_mod t10_pi crc64_rocksoft crc64 crc_t10dif
crct10dif_generic ahci libahci xhci_pci libata xhci_hcd
kernel: [   99.762633]  megaraid_sas tg3 crct10dif_pclmul
crct10dif_common crc32_pclmul crc32c_intel bnxt_en usbcore scsi_mod
i2c_i801 libphy i2c_smbus usb_common scsi_common wmi
kernel: [   99.764765] CPU: 10 PID: 1301 Comm: python3 Tainted: G
W          6.1.0-12-amd64 #1  Debian 6.1.52-1
kernel: [   99.764989] Hardware name: Dell Inc. PowerEdge R750xs/0441XG,
BIOS 1.8.2 09/14/2022
kernel: [   99.765214] RIP: e030:xen_mc_flush+0x196/0x220
kernel: [   99.765436] Code: e2 06 48 01 da 85 c0 0f 84 23 ff ff ff 48
8b 43 18 48 83 c3 40 48 c1 e8 3f 41 01 c5 48 39 d3 75 ec 45 85 ed 0f 84
06 ff ff ff <0f> 0b e8 e3 6e a0 00 41 8b 14 24 44 89 ee 48 c7 c7 c0 ea
33 82 89
kernel: [   99.765910] RSP: e02b:ffffc900412ffc60 EFLAGS: 00010082
kernel: [   99.766152] RAX: ffffffffffffffea RBX: ffff8888a1a9e300 RCX:
0000000000000000
kernel: [   99.766403] RDX: 0000000000000000 RSI: 0000000000000001 RDI:
ffff8888a1a9eb10
kernel: [   99.766653] RBP: 0000000080000002 R08: 0000000000000000 R09:
0000000000744f8b
kernel: [   99.766902] R10: 0000000000007ff0 R11: 0000000000000018 R12:
ffff8888a1a9e300
kernel: [   99.767153] R13: 0000000000000001 R14: ffffea0005130000 R15:
ffffea0005130000
kernel: [   99.767409] FS:  00007f59b5ba62c0(0000)
GS:ffff8888a1a80000(0000) knlGS:0000000000000000
kernel: [   99.767664] CS:  10000e030 DS: 0000 ES: 0000 CR0:
0000000080050033
kernel: [   99.767918] CR2: 00007f59b2200000 CR3: 0000000141bd0000 CR4:
0000000000050660
kernel: [   99.768181] Call Trace:
kernel: [   99.768436]  <TASK>
kernel: [   99.768691]  ? __warn+0x7d/0xc0
kernel: [   99.768947]  ? xen_mc_flush+0x196/0x220
kernel: [   99.769204]  ? report_bug+0xe6/0x170
kernel: [   99.769460]  ? handle_bug+0x41/0x70
kernel: [   99.769713]  ? exc_invalid_op+0x13/0x60
kernel: [   99.769967]  ? asm_exc_invalid_op+0x16/0x20
kernel: [   99.770223]  ? xen_mc_flush+0x196/0x220
kernel: [   99.770478]  xen_mc_issue+0x6d/0x70
kernel: [   99.770726]  xen_set_pmd_hyper+0x54/0x90
kernel: [   99.770965]  do_set_pmd+0x188/0x2a0

This looks like an attempt to map a hugepage, which isn't supported when running
as a Xen PV guest (this includes dom0).

Are transparent hugepages enabled somehow? In a Xen PV guest there should be no
/sys/kernel/mm/transparent_hugepage directory. Depending on the presence of that
directory either hugepage_init() has a bug, or a test for hugepages being
supported is missing in filemap_map_pages() or do_set_pmd().

kernel: [   99.771200]  filemap_map_pages+0x1a9/0x6e0
kernel: [   99.771434]  xfs_filemap_map_pages+0x41/0x60 [xfs]
kernel: [   99.771714]  do_fault+0x1a4/0x410
kernel: [   99.771947]  __handle_mm_fault+0x660/0xfa0
kernel: [   99.772182]  handle_mm_fault+0xdb/0x2d0
kernel: [   99.772414]  do_user_addr_fault+0x19c/0x570
kernel: [   99.772643]  exc_page_fault+0x70/0x170
kernel: [   99.772873]  asm_exc_page_fault+0x22/0x30
kernel: [   99.773102] RIP: 0033:0x7f59b502cbe2
kernel: [   99.773329] Code: 4d 8d 87 80 01 00 00 48 89 d9 45 31 c9 48
85 ff 74 5c 66 0f 1f 44 00 00 49 8d b0 80 fe ff ff 31 d2 41 0f 18 08 f3
0f 6f 04 11 <f3> 0f 6f 1c 16 f3 0f 6f 24 16 66 0f ef c3 66 0f 70 c8 31
66 0f f4
kernel: [   99.773806] RSP: 002b:00007ffc69923f70 EFLAGS: 00010246
kernel: [   99.774040] RAX: 00007ffc69923fa0 RBX: 00007f59b5038000 RCX:
00007f59b5038000
kernel: [   99.774276] RDX: 0000000000000000 RSI: 00007f59b21ffffa RDI:
0000000000000010
kernel: [   99.774504] RBP: 00000000025b8f40 R08: 00007f59b220017a R09:
0000000000000000
kernel: [   99.774727] R10: 0000000000077eca R11: 0000000000000400 R12:
00007f59b1a002fa
kernel: [   99.774947] R13: 00007ffc69923fe0 R14: 00007f59b5038080 R15:
00007f59b21ffffa
kernel: [   99.775161]  </TASK>
kernel: [   99.775365] ---[ end trace 0000000000000000 ]---
kernel: [   99.775567] 1 of 1 multicall(s) failed: cpu 10
kernel: [   99.775763]   call  1: op=1 arg=[ffff8888a1a9eb10] result=-22
```


Juergen

Attachment: OpenPGP_0xB0DE9DD628BF132F.asc
Description: OpenPGP public key

Attachment: OpenPGP_signature.asc
Description: OpenPGP digital signature


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.