[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: Serious AMD-Vi(?) issue
On Wed, Apr 17, 2024 at 02:40:09PM +0200, Jan Beulich wrote: > On 11.04.2024 04:41, Elliott Mitchell wrote: > > On Thu, Mar 28, 2024 at 07:25:02AM +0100, Jan Beulich wrote: > >> On 27.03.2024 18:27, Elliott Mitchell wrote: > >>> On Mon, Mar 25, 2024 at 02:43:44PM -0700, Elliott Mitchell wrote: > >>>> On Mon, Mar 25, 2024 at 08:55:56AM +0100, Jan Beulich wrote: > >>>>> > >>>>> In fact when running into trouble, the usual course of action would be > >>>>> to > >>>>> increase verbosity in both hypervisor and kernel, just to make sure no > >>>>> potentially relevant message is missed. > >>>> > >>>> More/better information might have been obtained if I'd been engaged > >>>> earlier. > >>> > >>> This is still true, things are in full mitigation mode and I'll be > >>> quite unhappy to go back with experiments at this point. > >> > >> Well, it very likely won't work without further experimenting by someone > >> able to observe the bad behavior. Recall we're on xen-devel here; it is > >> kind of expected that without clear (and practical) repro instructions > >> experimenting as well as info collection will remain with the reporter. > > > > After looking at the situation and considering the issues, I /may/ be > > able to setup for doing more testing. I guess I should confirm, which of > > those criteria do you think currently provided information fails at? > > > > AMD-IOMMU + Linux MD RAID1 + dual Samsung SATA (or various NVMe) + > > dbench; seems a pretty specific setup. > > Indeed. If that's the only way to observe the issue, it suggests to me > that it'll need to be mainly you to do further testing, and perhaps even > debugging. Which isn't to say we're not available to help, but from all > I have gathered so far we're pretty much in the dark even as to which > component(s) may be to blame. As can still be seen at the top in reply > context, some suggestions were given as to obtaining possible further > information (or confirming the absence thereof). There may be other ways which haven't yet been found. I've been left with the suspicion AMD was to some degree sponsoring work to ensure Xen works on their hardware. Given the severity of this problem I would kind of expect them not want to gain a reputation for having data loss issues. Assuming a suitable pair of devices weren't already on-hand, I would kind of expect this to be well within their budget. > I'd also like to come back to the vague theory you did voice, in that > you're suspecting flushes to take too long. I continue to have trouble > with this, and I would therefore like to ask that you put this down in > more technical terms, making connections to actual actions taken by > software / hardware. I'm trying to figure out a pattern. Nominally all the devices are roughly on par (only a very cheap flash device will be unable to overwhelm SATA's bandwidth). Yet why did the Crucial SATA device /seem/ not to have the issue? Why did a Crucial NVMe device demonstrate the issue. My guess is the flash controllers Samsung uses may be able to start executing commands faster than the ones Crucial uses. Meanwhile NVMe is lower overhead and latency than SATA (SATA's overhead isn't an issue for actual disks). Perhaps the IOMMU is still flushing its TLB, or hasn't loaded the new tables. I suspect when the MD-RAID1 issues block requests to a pair of devices, it likely sends the block to one device and then reuses most/all of the structures for the second device. As a result the second request would likely get a command to the device rather faster than the first request. Perhaps look into what structures the MD-RAID1 subsystem reuses are. Then see whether doing early setup of those structures triggers the issue? (okay I'm deep into speculation here, but this seems the simplest explanation for what could be occuring) -- (\___(\___(\______ --=> 8-) EHM <=-- ______/)___/)___/) \BS ( | ehem+sigmsg@xxxxxxx PGP 87145445 | ) / \_CS\ | _____ -O #include <stddisclaimer.h> O- _____ | / _/ 8A19\___\_|_/58D2 7E3D DDF4 7BA6 <-PGP-> 41D1 B375 37D0 8714\_|_/___/5445
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |