[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v3 3/3] tools: introduce parameter max_wp_ram_ranges.
> -----Original Message----- > From: Jan Beulich [mailto:JBeulich@xxxxxxxx] > Sent: 03 February 2016 12:36 > To: Paul Durrant > Cc: Andrew Cooper; Ian Campbell; Ian Jackson; Stefano Stabellini; Wei Liu; > Kevin Tian; zhiyuan.lv@xxxxxxxxx; Zhang Yu; xen-devel@xxxxxxxxxxxxx; Keir > (Xen.org) > Subject: RE: [Xen-devel] [PATCH v3 3/3] tools: introduce parameter > max_wp_ram_ranges. > > >>> On 03.02.16 at 13:20, <Paul.Durrant@xxxxxxxxxx> wrote: > >> -----Original Message----- > >> From: Jan Beulich [mailto:JBeulich@xxxxxxxx] > >> Sent: 03 February 2016 08:33 > >> To: Zhang Yu > >> Cc: Andrew Cooper; Ian Campbell; Paul Durrant; Wei Liu; Ian Jackson; > Stefano > >> Stabellini; Kevin Tian; zhiyuan.lv@xxxxxxxxx; xen-devel@xxxxxxxxxxxxx; Keir > >> (Xen.org) > >> Subject: Re: [Xen-devel] [PATCH v3 3/3] tools: introduce parameter > >> max_wp_ram_ranges. > >> > >> >>> On 03.02.16 at 08:10, <yu.c.zhang@xxxxxxxxxxxxxxx> wrote: > >> > On 2/2/2016 11:21 PM, Jan Beulich wrote: > >> >>>>> On 02.02.16 at 16:00, <yu.c.zhang@xxxxxxxxxxxxxxx> wrote: > >> >>> The limit of 4G is to avoid the data missing from uint64 to uint32 > >> >>> assignment. And I can accept the 8K limit for XenGT in practice. > >> >>> After all, it is vGPU page tables we are trying to trap and emulate, > >> >>> not normal page frames. > >> >>> > >> >>> And I guess the reason that one domain exhausting Xen's memory > can > >> >>> affect another domain is because rangeset uses Xen heap, instead of > >> the > >> >>> per-domain memory. So what about we use a 8K limit by now for > XenGT, > >> >>> and in the future, if a per-domain memory allocation solution for > >> >>> rangeset is ready, we do need to limit the rangeset size. Does this > >> >>> sounds more acceptable? > >> >> > >> >> The lower the limit the better (but no matter how low the limit > >> >> it won't make this a pretty thing). Anyway I'd still like to wait > >> >> for what Ian may further say on this. > >> >> > >> > Hi Jan, I just had a discussion with my colleague. We believe 8K could > >> > be the biggest limit for the write-protected ram ranges. If in the > >> > future, number of vGPU page tables exceeds this limit, we will modify > >> > our back-end device model to find a trade-off method, instead of > >> > extending this limit. If you can accept this value as the upper bound > >> > of rangeset, maybe we do not need to add any tool stack parameters, > but > >> > define a MAX_NR_WR_RAM_RANGES for the write-protected ram > >> rangesset. As > >> > to other rangesets, we keep their limit as 256. Does this sounds OK? :) > >> > >> I'm getting the impression that we're moving in circles. A blanket > >> limit above the 256 one for all domains is _not_ going to be > >> acceptable; going to 8k will still need host admin consent. With > >> your rangeset performance improvement patch, each range is > >> going to be tracked by a 40 byte structure (up from 32), which > >> already means an overhead increase for all the other ranges. 8k > >> of wp ranges implies an overhead beyond 448k (including the > >> xmalloc() overhead), which is not _that_ much, but also not > >> negligible. > >> > > > > ... which means we are still going to need a toolstack parameter to set the > > limit. We already have a parameter for VRAM size so is having a parameter > for > > max. GTT shadow ranges such a bad thing? > > It's workable, but not nice (see also Ian's earlier response). > > > Is the fact that the memory comes > > from xenheap rather than domheap the real problem? > > Not the primary one, since except on huge memory machines > both heaps are identical. To me the primary one is the quite > more significant resource consumption in the first place (I'm not > going to repeat what I've written in already way too many > replies before). Ok. Well the only way round tracking specific ranges for emulation (and consequently suffering the overhead) is tracking by type. For XenGT I guess it would be possible to live with a situation where a single ioreq server can register all wp mem emulations for a given VM. I can't say I particularly like that way of doing things but if it's the only way forward then I guess we may have to live with it. Paul > > Jan _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |