[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] Tmem [PATCH 1/5] (Take 3): Core API between kernel and tmem
Tmem [PATCH 1/5] (Take 3): Core API between kernel and tmem Declares tmem_ops accessors and initializes them to be no-ops (returning -ENOSYS). By itself, this API is useless; it requires a layer such as cleancache and/or frontswap (or similar) on top to make use of the API and a layer below to declare non-no-op tmem accessors that interface to a tmem implementation (e.g. Xen hypercalls). (Many thanks to Jeremy Fitzhardinge for suggesting this approach.) Signed-off-by: Dan Magenheimer <dan.magenheimer@xxxxxxxxxx> Documentation/transcendent-memory.txt | 176 +++++++++++++++++++++ include/linux/tmem.h | 88 ++++++++++ mm/tmem.c | 93 +++++++++++ 3 files changed, 357 insertions(+) --- linux-2.6.32/include/linux/tmem.h 1969-12-31 17:00:00.000000000 -0700 +++ linux-2.6.32-tmem/include/linux/tmem.h 2009-12-10 11:10:17.000000000 -0700 @@ -0,0 +1,88 @@ +/* + * linux/tmem.h + * + * Interface to transcendent memory + * + * Copyright (C) 2008,2009 Dan Magenheimer, Oracle Corp. + */ + +#include <linux/errno.h> + +struct tmem_pool_uuid { + u64 uuid_lo; + u64 uuid_hi; +}; + +#define TMEM_POOL_PRIVATE_UUID { 0, 0 } + +struct tmem_ops { + int (*new_pool)(struct tmem_pool_uuid uuid, u32 flags); + int (*put_page)(u32 pool_id, u64 object, u32 index, unsigned long gmfn); + int (*get_page)(u32 pool_id, u64 object, u32 index, unsigned long gmfn); + int (*flush_page)(u32 pool_id, u64 object, u32 index); + int (*flush_object)(u32 pool_id, u64 object); + int (*destroy_pool)(u32 pool_id); +}; + +extern struct tmem_ops *tmem_ops; +extern void tmem_set_ops(struct tmem_ops *ops); + +/* flags for tmem_ops.new_pool */ +#define TMEM_POOL_PERSIST 1 +#define TMEM_POOL_SHARED 2 + +static inline int tmem_new_pool(struct tmem_pool_uuid uuid, u32 flags) +{ + int ret = -ENOSYS; +#ifdef CONFIG_TMEM + ret = (*tmem_ops->new_pool)(uuid, flags); +#endif + return ret; +} + +static inline int tmem_put_page(u32 pool_id, u64 object, u32 index, + unsigned long gmfn) +{ + int ret = -ENOSYS; +#ifdef CONFIG_TMEM + ret = (*tmem_ops->put_page)(pool_id, object, index, gmfn); +#endif + return ret; +} + +static inline int tmem_get_page(u32 pool_id, u64 object, u32 index, + unsigned long gmfn) +{ + int ret = -ENOSYS; +#ifdef CONFIG_TMEM + ret = (*tmem_ops->get_page)(pool_id, object, index, gmfn); +#endif + return ret; +} + +static inline int tmem_flush_page(u32 pool_id, u64 object, u32 index) +{ + int ret = -ENOSYS; +#ifdef CONFIG_TMEM + ret = (*tmem_ops->flush_page)(pool_id, object, index); +#endif + return ret; +} + +static inline int tmem_flush_object(u32 pool_id, u64 object) +{ + int ret = -ENOSYS; +#ifdef CONFIG_TMEM + ret = (*tmem_ops->flush_object)(pool_id, object); +#endif + return ret; +} + +static inline int tmem_destroy_pool(u32 pool_id) +{ + int ret = -ENOSYS; +#ifdef CONFIG_TMEM + ret = (*tmem_ops->destroy_pool)(pool_id); +#endif + return ret; +} --- linux-2.6.32/mm/tmem.c 1969-12-31 17:00:00.000000000 -0700 +++ linux-2.6.32-tmem/mm/tmem.c 2009-12-17 13:58:59.000000000 -0700 @@ -0,0 +1,93 @@ +/* + * Default implementation for transcendent memory (tmem) + * + * Copyright (C) 2008, 2009 Dan Magenheimer, Oracle Corp. + */ + +#include <linux/types.h> +#include <linux/init.h> +#include <linux/errno.h> +#include <linux/tmem.h> +#include <linux/bug.h> + +static int default_tmem_new_pool(struct tmem_pool_uuid uuid, u32 flags) +{ + return -ENOSYS; +} + +static int default_tmem_put_page(u32 pool_id, u64 object, u32 index, + unsigned long gmfn) +{ + return -ENOSYS; +} + +static int default_tmem_get_page(u32 pool_id, u64 object, u32 index, + unsigned long gmfn) +{ + return -ENOSYS; +} + +static int default_tmem_flush_page(u32 pool_id, u64 object, u32 index) +{ + return -ENOSYS; +} + +static int default_tmem_flush_object(u32 pool_id, u64 object) +{ + return -ENOSYS; +} + +static int default_tmem_destroy_pool(u32 pool_id) +{ + return -ENOSYS; +} + +static struct tmem_ops default_tmem_ops = { + .new_pool = default_tmem_new_pool, + .put_page = default_tmem_put_page, + .get_page = default_tmem_get_page, + .flush_page = default_tmem_flush_page, + .flush_object = default_tmem_flush_object, + .destroy_pool = default_tmem_destroy_pool +}; + +struct tmem_ops *tmem_ops = &default_tmem_ops; + +void __init tmem_set_ops(struct tmem_ops *ops) +{ + /* should only ever be set once */ + WARN_ON(tmem_ops != &default_tmem_ops); + + tmem_ops = ops; +} + +#ifdef CONFIG_SYSCTL +#include <linux/sysctl.h> + +#ifdef CONFIG_CLEANCACHE +extern struct ctl_table cleancache_table[]; +#endif +#ifdef CONFIG_FRONTSWAP +extern struct ctl_table frontswap_table[]; +#endif + +ctl_table tmem_table[] = { +#ifdef CONFIG_CLEANCACHE + { + .ctl_name = CTL_UNNUMBERED, + .procname = "cleancache", + .mode = 0555, + .child = cleancache_table, + }, +#endif +#ifdef CONFIG_FRONTSWAP + { + .ctl_name = CTL_UNNUMBERED, + .procname = "frontswap", + .mode = 0555, + .child = frontswap_table, + }, +#endif + { .ctl_name = 0 } +}; +#endif /* CONFIG_SYSCTL */ --- linux-2.6.32/Documentation/transcendent-memory.txt 1969-12-31 17:00:00.000000000 -0700 +++ linux-2.6.32-tmem/Documentation/transcendent-memory.txt 2009-12-17 15:29:10.000000000 -0700 @@ -0,0 +1,176 @@ +Normal memory is directly addressable by the kernel, of a known +normally-fixed size, synchronously accessible, and persistent (though +not across a reboot). + +What if there was a class of memory that is of unknown and dynamically +variable size, is addressable only indirectly by the kernel, can be +configured either as persistent or as "ephemeral" (meaning it will be +around for awhile, but might disappear without warning), and is still +fast enough to be synchronously accessible? + +We call this latter class "transcendent memory" and it provides an +interesting opportunity to more efficiently utilize RAM in a virtualized +environment. However this "memory but not really memory" may also have +applications in NON-virtualized environments, such as hotplug-memory +deletion, SSDs, and page cache compression. Others have suggested ideas +such as allowing use of highmem memory without a highmem kernel, or use +of spare video memory. + +Transcendent memory, or "tmem" for short, provides a well-defined API to +access this unusual class of memory. (A summary of the API is provided +below.) The basic operations are page-copy-based and use a flexible +object-oriented addressing mechanism. Tmem assumes that some "privileged +entity" is capable of executing tmem requests and storing pages of data; +this entity is currently a hypervisor and operations are performed via +hypercalls, but the entity could be a kernel policy, or perhaps a +"memory node" in a cluster of blades connected by a high-speed +interconnect such as hypertransport or QPI. + +Since tmem is not directly accessible and because page copying is done +to/from physical pageframes, it more suitable for in-kernel memory needs +than for userland applications. However, there may be yet undiscovered +userland possibilities. + +With the tmem concept outlined vaguely and its broader potential hinted, +we will overview two existing examples of how tmem can be used by the +kernel. + +"Cleancache" can be thought of as a page-granularity victim cache for clean +pages that the kernel's pageframe replacement algorithm (PFRA) would like +to keep around, but can't since there isn't enough memory. So when the +PFRA "evicts" a page, it first puts it into the cleancache via a call to +tmem. And any time a filesystem reads a page from disk, it first attempts +to get the page from cleancache. If it's there, a disk access is eliminated. +If not, the filesystem just goes to the disk like normal. Cleancache is +"ephemeral" so whether a page is kept in cleancache (between the "put" and +the "get") is dependent on a number of factors that are invisible to +the kernel. + +"Frontswap" is so named because it can be thought of as the opposite of +a "backing store". Frontswap IS persistent, but for various reasons may not +always be available for use, again due to factors that may not be visible to +the kernel. (But, briefly, if the kernel is being "good" and has shared its +resources nicely, then it will be able to use frontswap, else it will not.) +Once a page is put, a get on the page will always succeed. So when the +kernel finds itself in a situation where it needs to swap out a page, it +first attempts to use frontswap. If the put works, a disk write and +(usually) a disk read are avoided. If it doesn't, the page is written +to swap as usual. Unlike cleancache, whether a page is stored in frontswap +vs swap is recorded in kernel data structures, so when a page needs to +be fetched, the kernel does a get if it is in frontswap and reads from +swap if it is not in frontswap. + +Both cleancache and frontswap may be optionally compressed, trading off 2x +space reduction vs 10x performance for access. Cleancache also has a +sharing feature, which allows different nodes in a "virtual cluster" +to share a local page cache. + +Tmem has some similarity to IBM's Collaborative Memory Management, but +creates more of a partnership between the kernel and the "privileged +entity" and is not very invasive. Tmem may be applicable for KVM and +containers; there is some disagreement on the extent of its value. +Tmem is highly complementary to ballooning (aka page granularity hot +plug) and memory deduplication (aka transparent content-based page +sharing) but still has value when neither are present. + +Performance is difficult to quantify because some benchmarks respond +very favorably to increases in memory and tmem may do quite well on +those, depending on how much tmem is available which may vary widely +and dynamically, depending on conditions completely outside of the +system being measured. Ideas on how best to provide useful metrics +would be appreciated. + +Tmem is supported starting in Xen 4.0 and is in Xen's Linux 2.6.18-xen +source tree. It is also released as a technology preview in Oracle's +Xen-based virtualization product, Oracle VM 2.2. Again, Xen is not +necessarily a requirement, but currently provides the only existing +implementation of tmem. + +Lots more information about tmem can be found at: + http://oss.oracle.com/projects/tmem +and there was a talk about it on the first day of Linux Symposium in +July 2009; an updated talk is planned at linux.conf.au in January 2010. +Tmem is the result of a group effort, including Dan Magenheimer, +Chris Mason, Dave McCracken, Kurt Hackel and Zhigang Wang, with helpful +input from Jeremy Fitzhardinge, Keir Fraser, Ian Pratt, Sunil Mushran, +Joel Becker, and Jan Beulich. + +THE TRANSCENDENT MEMORY API + +Transcendent memory is made up of a set of pools. Each pool is made +up of a set of objects. And each object contains a set of pages. +The combination of a 32-bit pool id, a 64-bit object id, and a 32-bit +page id, uniquely identify a page of tmem data, and this tuple is called +a "handle." Commonly, the three parts of a handle are used to address +a filesystem, a file within that filesystem, and a page within that file; +however an OS can use any values as long as they uniquely identify +a page of data. + +When a tmem pool is created, it is given certain attributes: It can +be private or shared, and it can be persistent or ephemeral. Each +combination of these attributes provides a different set of useful +functionality and also defines a slightly different set of semantics +for the various operations on the pool. Other pool attributes include +the size of the page and a version number. + +Once a pool is created, operations are performed on the pool. Pages +are copied between the OS and tmem and are addressed using a handle. +Pages and/or objects may also be flushed from the pool. When all +operations are completed, a pool can be destroyed. + +The specific tmem functions are called in Linux through a set of +accessor functions: + +int (*new_pool)(struct tmem_pool_uuid uuid, u32 flags); +int (*destroy_pool)(u32 pool_id); +int (*put_page)(u32 pool_id, u64 object, u32 index, unsigned long pfn); +int (*get_page)(u32 pool_id, u64 object, u32 index, unsigned long pfn); +int (*flush_page)(u32 pool_id, u64 object, u32 index); +int (*flush_object)(u32 pool_id, u64 object); + +The new_pool accessor creates a new pool and returns a pool id +which is a non-negative 32-bit integer. If the flags parameter +specifies that the pool is to be shared, the uuid is a 128-bit "shared +secret" else it is ignored. The destroy_pool accessor destroys the pool. +(Note: shared pools are not supported until security implications +are better understood.) + +The put_page accessor copies a page of data from the specified pageframe +and associates it with the specified handle. + +The get_page accessor looks up a page of data in tmem associated with +the specified handle and, if found, copies it to the specified pageframe. + +The flush_page accessor ensures that subsequent gets of a page with +the specified handle will fail. The flush_object accessor ensures +that subsequent gets of any page matching the pool id and object +will fail. + +There are many subtle but critical behaviors for get_page and put_page: +- Any put_page (with one notable exception) may be rejected and the client + must be prepared to deal with that failure. A put_page copies, NOT moves, + data; that is the data exists in both places. Linux is responsible for + destroying or overwriting its own copy, or alternately managing any + coherency between the copies. +- Every page successfully put to a persistent pool must be found by a + subsequent get_page that specifies the same handle. A page successfully + put to an ephemeral pool has an indeterminate lifetime and even an + immediately subsequent get_page may fail. +- A get_page to a private pool is destructive, that is it behaves as if + the get_page were atomically followed by a flush_page. A get_page + to a shared pool is non-destructive. A flush_page behaves just like + a get_page to a private pool except the data is thrown away. +- Put-put-get coherency is guaranteed. For example, after the sequence: + put_page(ABC,D1); + put_page(ABC,D2); + get_page(ABC,E) + E may never contain the data from D1. However, even for a persistent + pool, the get_page may fail if the second put_page indicates failure. +- Get-get coherency is guaranteed. For example, in the sequence: + put_page(ABC,D); + get_page(ABC,E1); + get_page(ABC,E2) + if the first get_page fails, the second must also fail. +- A tmem implementation provides no serialization guarantees (e.g. to + an SMP Linux). So if different Linux threads are putting and flushing + the same page, the results are indeterminate. _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |