[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [xen-4.12-testing test] 169199: regressions - FAIL
- To: Jan Beulich <jbeulich@xxxxxxxx>
- From: Roger Pau Monné <roger.pau@xxxxxxxxxx>
- Date: Fri, 8 Apr 2022 10:09:50 +0200
- Arc-authentication-results: i=1; mx.microsoft.com 1; spf=pass smtp.mailfrom=citrix.com; dmarc=pass action=none header.from=citrix.com; dkim=pass header.d=citrix.com; arc=none
- Arc-message-signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-AntiSpam-MessageData-ChunkCount:X-MS-Exchange-AntiSpam-MessageData-0:X-MS-Exchange-AntiSpam-MessageData-1; bh=YhIcRCxA9o8VUGgY5jmlY4CTX4e20NJrMwY7FQoS2Gs=; b=VZJ2qHJ6hMfwJ1BQjIZ/POdOX98/4qLPX7pGf8XINPNRdDdggcoq/ndANEobSw2d7oJNXpYhM5P2ROsYXmfRhzFaBCRFXbv8Kjk79oeyI8I0B0bYdP2wNXOqOUSPELAM392H9uHmscBfKnF+wkH1tREOwq9wQBdJ89e6GyEB0LyJ6aBofXtooXTTSYXozbHbnyy8USJlwpYirR2X7lLPg2jcThE1vo9gvN/54zydZGWgIH93zdNakzJ/XWqywedf3xUng6qGFkKi3YNIytuydsoVSkilMzeUWEPixjVDUMc+bZeaL0OANber1P6wuVZagjZ81qhRz81U67sUBMEtzw==
- Arc-seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=cr417hGJu3KCSlW71NLzK5YDIufgtuihT5RAwj8SYfRQqDkyKLuUaenUgf3Mvz+kNxe4nVYRl4XoeG00wHAOURha2rvL84XEgEjxLPD0RzIQT7wkPjIvao4m9Zsde+/b80aNXtuL4lsXD/IaDsMlOsNerl5OFpQvZ7y9adz8wsYjCXCoOV/Aa65p2Jwwrbma9WDY7a1kC27q64gubiJ3waHqXgpOYriVBnFR8tL4rR28AXubbYynoAfECzlimi2Kwkva1gN7i2yxejTTMD5W780/AOGCWURVt7XSVIUpNYnd6YYArG5jPxFDMCkiBPh1c0TWWPe7ivem7/xV7qZijg==
- Authentication-results: esa5.hc3370-68.iphmx.com; dkim=pass (signature verified) header.i=@citrix.onmicrosoft.com
- Cc: <xen-devel@xxxxxxxxxxxxxxxxxxxx>, osstest service owner <osstest-admin@xxxxxxxxxxxxxx>, Andrew Cooper <andrew.cooper3@xxxxxxxxxx>, George Dunlap <george.dunlap@xxxxxxxxxx>, Julien Grall <julien@xxxxxxx>, Stefano Stabellini <sstabellini@xxxxxxxxxx>, Wei Liu <wl@xxxxxxx>, "Dario Faggioli" <dfaggioli@xxxxxxxx>
- Delivery-date: Fri, 08 Apr 2022 08:10:26 +0000
- Ironport-data: A9a23:UjaubKBCvof9JBVW/8fjw5YqxClBgxIJ4kV8jS/XYbTApG4q02FUm mJNXWqBa6yKYTP2e4t1PITg9RxTu5eGyNU2QQY4rX1jcSlH+JHPbTi7wuYcHM8wwunrFh8PA xA2M4GYRCwMZiaA4E/raNANlFEkvU2ybuOU5NXsZ2YgHWeIdA970Ug5w7Jh0tYx6TSEK1jlV e3a8pW31GCNg1aYAkpMg05UgEoy1BhakGpwUm0WPZinjneH/5UmJMt3yZWKB2n5WuFp8tuSH I4v+l0bElTxpH/BAvv9+lryn9ZjrrT6ZWBigVIOM0Sub4QrSoXfHc/XOdJFAXq7hQllkPhRi /ZojraiCj50O6LXwf8gAgllSh1XaPguFL/veRBTsOSWxkzCNXDt3+9vHAc9OohwFuRfWD8Us 6ZCcXZUM07F17neLLGTE4GAguw5K8bmJsUHs2xIxjDFF/c2B5vERs0m4PcGg2hu3JseQp4yY eI2bwNsVzb+cSRzGW8SI5MXzPyV3kXGJmgwRFW9+vNsvjm7IBZK+KP2LNPfd9iORMNUtkWVv GTL+yL+GB5yHMOb4SqI9DSrnOCntS/1UY0dFbq899ZxnUaegGcUDXU+V0a/oPS/ol6zXZRYM UN80igkoLU29UerZsLgRBD+q3mB1jYMVtwVH+Ak5QWlzqvP/x3fFmUCViRGatEtqIkxXzNC/ liShM/kHiAqubGQSHS15rqStSm1OyUeMSkFfyBscOcey4C9+sdp1EuJF4s9Vv7u5jHoJd3u6 xDJjw0FradQtJMO2L7i5m2Wkw/1mrGcG2bZ+T7rdm6i6wp4YqusaIqp9UXX4J58EWqJcrWSl CNawpbDtYjiGbnIzXXQG7tVQNlF8t7faFXhbUhT847NHthH01qqZshu7T53Py+F2e5UKGayM Cc/Ve68jaK/3UdGj4cqO+pd6OxwlMAM8OgJsNiOM7KihbArKWe6ENlGPxL44owUuBFEfVsDE Zmaa92wKn0RFL5qyjG7L89Ej+N6nHBjmDOMGsmip/hC7VZ4TCTIIVviGAHQBt3VEYve+FmFm zqhH5XiJ+pjvB3WPXCMrN97waEiJnknH5Hmw/G7hcbYSjeK7FoJUqeLqZt4ItQNt/0Myo/go yHsMmcFmQGXrSCWdm23hoVLNeqHsWBX9ilgY0TB/D+AhhAeXGpYxPtHLMtoION/rYSOD5dcF pE4RilJOdwWIhzv8DUBd5jt6otkcRWgnwWVOCS5JjM4evZdq8bhoLcIoiOHGPEyMxeK
- Ironport-hdrordr: A9a23:mS2Deai3jjhO5IodlR6WWBNHBHBQXzR13DAbv31ZSRFFG/FwyP rBoB1L73DJYWgqNE3I+erhBEGBKUmsk6KdxbNhQItKPTOWwldASbsC0WKM+UyEJ8STzJ846U 4kSdkDNDSSNykKsS+Z2njBLz9I+rDum8rE9ISurUuFDzsaEJ2Ihz0JdDpzeXcGPTWua6BJc6 Z1saF81kWdkDksH4yGL0hAe9KGi8zAlZrgbxJDLxk76DOWhTftzLLhCRCX0joXTjsKmN4ZgC L4uj28wp/mn+Cwyxfa2WOWx5NKmOH5wt8GIMCXkMAaJhjllw7tToV8XL+puiwzvYiUmR0Xue iJhy1lE9V46nvXcG3wiRzx2zP42DJr0HPmwU/wuwqVneXJABYBT+ZRj4NQdRXUr2A6ustn7a 5N12WF87JKEBLphk3Glpf1fiAvsnDxjWspkOYVgXAae5AZcqVtoYsW+14QOIscHRj99JssHI BVfY3hDc5tABKnhk3izylSKITGZAVxIv7GeDlOhiWt6UkZoJgjpHFohvD2nR87heYAotd/lq H5259T5cJzp/8tHNJA7dg6MLmK40z2MGTx2TGpUB3a/J9uAQO5l3ew2sRw2N2X
- List-id: Xen developer discussion <xen-devel.lists.xenproject.org>
On Fri, Apr 08, 2022 at 09:01:11AM +0200, Jan Beulich wrote:
> On 07.04.2022 10:45, osstest service owner wrote:
> > flight 169199 xen-4.12-testing real [real]
> > http://logs.test-lab.xenproject.org/osstest/logs/169199/
> >
> > Regressions :-(
> >
> > Tests which did not succeed and are blocking,
> > including tests which could not be run:
> > test-amd64-amd64-xl-qemut-debianhvm-i386-xsm 12 debian-hvm-install fail
> > REGR. vs. 168480
>
> While the subsequent flight passed, I thought I'd still look into
> the logs here since the earlier flight had failed too. The state of
> the machine when the debug keys were issued is somewhat odd (and
> similar to the earlier failure's): 11 of the 56 CPUs try to
> acquire (apparently) Dom0's event lock, from evtchn_move_pirqs().
> All other CPUs are idle. The test failed because the sole guest
> didn't reboot in time. Whether the failure is actually connected to
> this apparent lock contention is unclear, though.
>
> One can further see that really all about 70 ECS_PIRQ ports are
> bound to vCPU 0 (which makes me wonder about lack of balancing
> inside Dom0 itself, but that's unrelated). This means that all
> other vCPU-s have nothing at all to do in evtchn_move_pirqs().
> Since this moving of pIRQ-s is an optimization (the value of which
> has been put under question in the past, iirc), I wonder whether we
> shouldn't add a check to the function for the list being empty
> prior to actually acquiring the lock. I guess I'll make a patch and
> post it as RFC.
Seems good to me.
I think a better model would be to migrate the PIRQs when fired, or
even better when EOI is performed? So that Xen doesn't pointlessly
migrate PIRQs for vCPUs that aren't running.
> And of course in a mostly idle system the other aspect here (again)
> is: Why are vCPU-s moved across pCPU-s in the first place? I've
> observed (and reported) such seemingly over-aggressive vCPU
> migration before, most recently in the context of putting together
> 'x86: make "dom0_nodes=" work with credit2'. Is there anything that
> can be done about this in credit2?
>
> A final, osstest-related question is: Does it make sense to run Dom0
> on 56 vCPU-s, one each per pCPU? The bigger a system, the less
> useful it looks to me to actually also have a Dom0 as big, when the
> purpose of the system is to run guests, not meaningful other
> workloads in Dom0. While this is Xen's default (i.e. in the absence
> of command line options restricting Dom0), I don't think it's
> representing typical use of Xen in the field.
I could add a suitable dom0_max_vcpus parameter to osstest. XenServer
uses 16 for example.
Albeit not having such parameter has likely led you into figuring out
this issue, so it might not be so bad. I agree however it's likely
better to test scenarios closer to real world usage.
Thanks, Roger.
|