[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [PATCH RFC 0/4] mm: place pages to the freelist tail when onling and undoing isolation
On 23.09.20 16:31, Vlastimil Babka wrote: > On 9/16/20 9:31 PM, David Hildenbrand wrote: >> >> >>> Am 16.09.2020 um 20:50 schrieb osalvador@xxxxxxx: >>> >>> On 2020-09-16 20:34, David Hildenbrand wrote: >>>> When adding separate memory blocks via add_memory*() and onlining them >>>> immediately, the metadata (especially the memmap) of the next block will be >>>> placed onto one of the just added+onlined block. This creates a chain >>>> of unmovable allocations: If the last memory block cannot get >>>> offlined+removed() so will all dependant ones. We directly have unmovable >>>> allocations all over the place. >>>> This can be observed quite easily using virtio-mem, however, it can also >>>> be observed when using DIMMs. The freshly onlined pages will usually be >>>> placed to the head of the freelists, meaning they will be allocated next, >>>> turning the just-added memory usually immediately un-removable. The >>>> fresh pages are cold, prefering to allocate others (that might be hot) >>>> also feels to be the natural thing to do. >>>> It also applies to the hyper-v balloon xen-balloon, and ppc64 dlpar: when >>>> adding separate, successive memory blocks, each memory block will have >>>> unmovable allocations on them - for example gigantic pages will fail to >>>> allocate. >>>> While the ZONE_NORMAL doesn't provide any guarantees that memory can get >>>> offlined+removed again (any kind of fragmentation with unmovable >>>> allocations is possible), there are many scenarios (hotplugging a lot of >>>> memory, running workload, hotunplug some memory/as much as possible) where >>>> we can offline+remove quite a lot with this patchset. >>> >>> Hi David, >>> >> >> Hi Oscar. >> >>> I did not read through the patchset yet, so sorry if the question is >>> nonsense, but is this not trying to fix the same issue the vmemmap patches >>> did? [1] >> >> Not nonesense at all. It only helps to some degree, though. It solves the >> dependencies due to the memmap. However, it‘s not completely ideal, >> especially for single memory blocks. >> >> With single memory blocks (virtio-mem, xen-balloon, hv balloon, ppc dlpar) >> you still have unmovable (vmemmap chunks) all over the physical address >> space. Consider the gigantic page example after hotplug. You directly >> fragmented all hotplugged memory. >> >> Of course, there might be (less extreme) dependencies due page tables for >> the identity mapping, extended struct pages and similar. >> >> Having that said, there are other benefits when preferring other memory over >> just hotplugged memory. Think about adding+onlining memory during boot >> (dimms under QEMU, virtio-mem), once the system is up you will have most >> (all) of that memory completely untouched. >> >> So while vmemmap on hotplugged memory would tackle some part of the issue, >> there are cases where this approach is better, and there are even benefits >> when combining both. > Hi Vlastimil, > I see the point, but I don't think the head/tail mechanism is great for this. > It > might sort of work, but with other interfering activity there are no > guarantees > and it relies on a subtle implementation detail. There are better mechanisms For the specified use case of adding+onlining a whole bunch of memory this works just fine. We don't care too much about "other interfering activity" as you mention here, or about guarantees - this is a pure optimization that seems to work just fine in practice. I'm not sure about the "subtle implementation detail" - buddy merging, and head/tail of buddy lists are a basic concept of our page allocator. If that would ever change, the optimization here would be lost and we would have to think of something else. Nothing would actually break - and it's all kept directly in page_alloc.c I'd like to stress that what I propose here is both simple and powerful. > possible I think, such as preparing a larger MIGRATE_UNMOVABLE area in the > existing memory before we allocate those long-term management structures. Or > onlining a bunch of blocks as zone_movable first and only later convert to > zone_normal in a controlled way when existing normal zone becomes depeted? I see the following (more or less complicated) alternatives 1) Having a larger MIGRATE_UNMOVABLE area a) Sizing it is difficult. I mean you would have to plan ahead for all memory you might eventually hotplug later - and that could even be impossible if you hotplug quite a lot of memory to a smaller machine. (I've seen people in the vm/container world trying to hotplug 128GB DIMMs to 2GB VMs ... and failing for obvious reasons) b) not really desired. You usually want to have most memory movable, not the opposite (just because you might hotplug memory in small chunks later). 2) smarter onlining I have prototype patches for better auto-onlining (which I'll share at some point), where I balance between ZONE_NORMAL and ZONE_MOVABLE in a defined ratio. Assuming something very simple, adding separate memory blocks and onlining them based on the current zone ratio (assuming a 1:4 normal:movable target ratio) would (without some other policies I have in place) result in something like this for hotplugged memory (via virtio-mem): [N][M][M][M][M][N][M][M][M][M][N][M][M][M][M]... (note: layout is suboptimal, just a simple example) But even here, all [N] memory blocks would immediately be use for allocations for the memmap of successive blocks. It doesn't solve the dependency issues. Now assume we would want to group [N] in a way to allow for gigantic pages, like [N][N][N][N][N][N][N][N][M][M][M][M] .... we would, once again, never be able to allocate a gigantic page because all [N] would contain a memmap. 3) conversion from MOVABLE -> NORMAL While a conversion from MOVABLE to NORMAL would be interesting to see, it's going to be a challenging task to actually implement (people expect that page_zone() remains stable). Without any hacks, we'd have to 1. offline the selected (MOVABLE) memory block/chunk 2. online the selected memory block/chunk to the NORMAL zone This is not something we can do out of random context (for example, we need both, the device hotplug lock and the memory hotplug lock, as we might race with user space) - so there might still be a chance of corner-case OOMs. (I assume there could also be quite a negative performance impact when always relying on the conversion, and not properly planning ahead as in 2.) > > I guess it's an issue that the e.g. 128M block onlines are so disconnected > from > each other it's hard to employ a strategy that works best for e.g. a whole > bunch > of GB onlined at once. But I noticed some effort towards new API, so maybe > that > will be solved there too? While new interfaces might make it easier to identify boundaries of separate DIMMs (e.g., to online a single DIMM either movable or unmovable - which can partially be done right now when going via memory resource boundaries), it doesn't help for the use case of adding separate memory blocks. So while having an automatic conversion from MOVABLE -> NORMAL would be interesting, I doubt we'll see it in the foreseeable future. Are there any similarly simple alternatives to optimize this? Thanks! -- Thanks, David / dhildenb
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |