|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [RFC 3/6] sysctl: extend XEN_SYSCTL_getcpuinfo interface
On 26.07.19 15:15, Dario Faggioli wrote: Yep, I think being able to know time spent running guests could be useful.
Well, my intention was to see hypervisor run and true idle time.
With this full series I see the distinct difference in xentop depending on the
type of load in domains:
On my regular system (HW less Dom0, Linux with UI aka DomD, Android with PV
drivers aka DomA), I see following:
Idle system:
xentop - 10:10:42 Xen 4.13-unstable
3 domains: 1 running, 2 blocked, 0 paused, 0 crashed, 0 dying, 0 shutdown
%CPU(s): 7.0 gu, 2.6 hy, 390.4 id
Mem: 8257536k total, 8257536k used, 99020k free CPUs: 4 @ 8MHz
NAME STATE CPU(sec) CPU(%) MEM(k) MEM(%) MAXMEM(k) MAXMEM(%)
VCPUS NETS NETTX(k) NETRX(k) VBDS VBD_OO VBD_RD VBD_WR VBD_RSECT
VBD_WSECT SSID
DomA --b--- 76 3.3 6258456 75.8 6259712 75.8
4 0 0 0 0 0 0 0 0 0
0
Domain-0 -----r 14 1.0 262144 3.2 no limit n/a
4 0 0 0 0 0 0 0 0 0
0
DomD --b--- 111 2.8 1181972 14.3 1246208 15.1
4 0 0 0 0 0 0 0 0 0
0
System with CPU burners in all domains:
xentop - 10:12:19 Xen 4.13-unstable
3 domains: 3 running, 0 blocked, 0 paused, 0 crashed, 0 dying, 0 shutdown
%CPU(s): 389.1 gu, 10.9 hy, 0.0 id
Mem: 8257536k total, 8257536k used, 99020k free CPUs: 4 @ 8MHz
NAME STATE CPU(sec) CPU(%) MEM(k) MEM(%) MAXMEM(k) MAXMEM(%)
VCPUS NETS NETTX(k) NETRX(k) VBDS VBD_OO VBD_RD VBD_WR VBD_RSECT
VBD_WSECT SSID
DomA -----r 115 129.7 6258456 75.8 6259712 75.8
4 0 0 0 0 0 0 0 0 0
0
Domain-0 -----r 120 129.8 262144 3.2 no limit n/a
4 0 0 0 0 0 0 0 0 0
0
DomD -----r 163 129.6 1181972 14.3 1246208 15.1
4 0 0 0 0 0 0 0 0 0
0
System with GPU load run both in DomD and DomA:
xentop - 10:14:26 Xen 4.13-unstable
3 domains: 2 running, 1 blocked, 0 paused, 0 crashed, 0 dying, 0 shutdown
%CPU(s): 165.7 gu, 51.4 hy, 182.9 id
Mem: 8257536k total, 8257536k used, 99020k free CPUs: 4 @ 8MHz
NAME STATE CPU(sec) CPU(%) MEM(k) MEM(%) MAXMEM(k) MAXMEM(%)
VCPUS NETS NETTX(k) NETRX(k) VBDS VBD_OO VBD_RD VBD_WR VBD_RSECT
VBD_WSECT SSID
DomA --b--- 250 60.8 6258456 75.8 6259712 75.8
4 0 0 0 0 0 0 0 0 0
0
Domain-0 -----r 159 2.1 262144 3.2 no limit n/a
4 0 0 0 0 0 0 0 0 0
0
DomD -----r 275 102.7 1181972 14.3 1246208 15.1
4 0 0 0 0 0 0 0 0 0
0
You can see that rise of CPU used by hypervisor itself in high IRQ use-case
(GPU load).
I confirm what I said about patch 1: idle time being the time idle_vcpu spent in RUNSTATE_blocked, and hypervisor time being the time idle_vcpu spent in RUNSTATE_running sounds quite confusing to me. As I said before, think of idle_vcpu as hypervisor_vcpu ;) -- Sincerely, Andrii Anisov. _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |