[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

RE: [Xen-ia64-devel] [IPF-ia64] with Cset 10690, creating a VTImakexen0 hang



Sakai-san,
Another short term approach,
When emulating PAL_HALT_LIGHT
If(dom0)
        Ia64_pal_halt_hight();
Else  //domU
        do_block;

Thanks,
anthony
>-----Original Message-----
>From: xen-ia64-devel-bounces@xxxxxxxxxxxxxxxxxxx
>[mailto:xen-ia64-devel-bounces@xxxxxxxxxxxxxxxxxxx] On Behalf Of Atsushi
>SAKAI
>Sent: 2006?7?11? 21:46
>To: Alex Williamson; Zhang, Xiantao
>Cc: Isaku Yamahata; xen-ia64-devel@xxxxxxxxxxxxxxxxxxx
>Subject: Re: [Xen-ia64-devel] [IPF-ia64] with Cset 10690, creating a
>VTImakexen0 hang
>
>Hi, Alex
>
>Sorry for late.
>
>I found your problem(boot time difference w/ PAL_HALT_LIGHT emulation patch)
>occurred in SMP(credit).
>But, it does not occurred in UP, SMP(bvt) and SMP(credit w/ affinity).
>
>I think the emulation of pal_halt_light for domU
>does not good work for DomVTI boot up
>under credit scheduling w/o affinity.
>
>And consider the Xiantao survey,
>qemu make heavy I/O operations at the boot up.
>
>Consider the above two conditions,
>I think credit scheduler algorithm does not consider
>the block state.(caused by pal_halt_light emulation)
>So I want to switch off the vcpu migration at heavy load
>
>
>I planned as follows.
>
>1)In the short term,
>I want to avoid this problem by
>HALT the PAL_HALT_LIGHT emulation while DomVTI boot up.
>or
>Lock VCPUs migrations while DomVTI boot up.
>(when Credit scheduler runs)
>
>2)In the long term,
>I will make a patch to avoid this problem.
>(Consider the heavy io w/ vcpu migration)
>
>N.B.
>I checked under CS:10559.(original patch made)
>
>Thanks,
>Atsushi SAKAI
>
>
>
>
>
>>On Tue, 2006-07-11 at 19:42 +0800, Zhang, Xiantao wrote:
>>> Hi Alex,
>>>     Seems this issue was caused by Cset 10688. In vcpu_itr_d, the current
>>> logic purges vhpt with cpu_flush_vhpt_range but it is very heavy to
>>> xen0. When creating VTi domain @ early stage, IO operation is very
>>> excessive, so qemu was scheduled out and in very frequently and this
>>> logic was executed every time. In addition, cpu_flush_vhpt_range using
>>> identity map to purge vhpt may cause more tlb miss due to no TR map.
>>> If remove vcpu_flush_tlb_vhpt_range logic although it definitely
>>> needed, seems VTi becomes healthy. Maybe potential bugs exist there.:)
>>
>>   Thanks for investigating Xiantao.  Isaku, any thoughts on how to
>>regain VTI performance?  Thanks,
>>
>>      Alex
>>
>>--
>>Alex Williamson                             HP Open Source & Linux Org.
>>
>>
>>_______________________________________________
>>Xen-ia64-devel mailing list
>>Xen-ia64-devel@xxxxxxxxxxxxxxxxxxx
>>http://lists.xensource.com/xen-ia64-devel
>>
>
>
>
>
>
>
>
>_______________________________________________
>Xen-ia64-devel mailing list
>Xen-ia64-devel@xxxxxxxxxxxxxxxxxxx
>http://lists.xensource.com/xen-ia64-devel

_______________________________________________
Xen-ia64-devel mailing list
Xen-ia64-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-ia64-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.