[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH 2/2][Resend] Output Virtual Block Device requests information for xentop. ver.2
This patch adds VBD statistical information into xentop. Our new xentop collects I/O requests from sysfs and displays them. This is an example output of the new xentop in the case of 4 VMs. xentop - 22:26:21 Xen 3.0-unstable 2 domains: 1 running, 1 blocked, 0 paused, 0 crashed, 0 dying, 0 shutdown Mem: 2620476k total, 2128336k used, 492140k free CPUs: 2 @ 2992MHz NAME STATE CPU(sec) CPU(%) MEM(k) MEM(%) MAXMEM(k) MAXMEM(%) VCPUS NETS NETTX(k) NETRX(k) VBDS VBD_OO VBD_RD VBD_WR SSID Domain-0 -----r 279 31.4 1048628 40.0 no limit n/a 2 8 53668 3761 0 0 0 0 0 VM1 --b--- 14 0.1 1048352 40.0 1048576 40.0 2 1 1 645 1 0 1435 733 0 Delay Networks vBds VCPUs Repeat header Sort order Quit As compared with the traditional xentop, new statistical information (VBDS, VBD_OO, VBD_RD and VBD_WR) is added. In detail mode, outputs of xentop is as follows. xentop - 22:26:39 Xen 3.0-unstable 2 domains: 1 running, 1 blocked, 0 paused, 0 crashed, 0 dying, 0 shutdown Mem: 2620476k total, 2128336k used, 492140k free CPUs: 2 @ 2992MHz NAME STATE CPU(sec) CPU(%) MEM(k) MEM(%) MAXMEM(k) MAXMEM(%) VCPUS NETS NETTX(k) NETRX(k) VBDS VBD_OO VBD_RD VBD_WR SSID Domain-0 -----r 284 17.4 1048580 40.0 no limit n/a 2 8 56148 3885 0 0 0 0 0 VM1 --b--- 14 0.0 1048400 40.0 1048576 40.0 2 1 1 662 1 0 1435 736 0 VBD 2049 [ 8: 1] OO: 0 RD: 1435 WR: 736 Delay Networks vBds VCPUs Repeat header Sort order Quit Information about each VBD is displayed under a line of corresponding domain. This patch is practically no change from previous patch. Only the location of code is changed, and insert code which transfers from dev information to major and minor information. Thanks, Satoshi UCHIDA Attachment:
Add_VBD_info_to_xentop.patch _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |