[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Minios-devel] [UNIKRAFT PATCH v4 1/7] plat/xen: Add support for communication with Xenstore daemon



Reviewed-by: Yuri Volchkov <yuri.volchkov@xxxxxxxxx>

Costin Lupu <costin.lupu@xxxxxxxxx> writes:

> Add support for communication with Xenstore daemon via the shared
> page. In Unikraft, the incoming messages are processed by the
> Xenstore thread.
>
> Signed-off-by: Costin Lupu <costin.lupu@xxxxxxxxx>
> ---
>  plat/xen/Makefile.uk       |   1 +
>  plat/xen/xenbus/xenbus.c   |   8 +
>  plat/xen/xenbus/xs_comms.c | 610 
> +++++++++++++++++++++++++++++++++++++++++++++
>  plat/xen/xenbus/xs_comms.h |  64 +++++
>  4 files changed, 683 insertions(+)
>  create mode 100644 plat/xen/xenbus/xs_comms.c
>  create mode 100644 plat/xen/xenbus/xs_comms.h
>
> diff --git a/plat/xen/Makefile.uk b/plat/xen/Makefile.uk
> index b8c70e1..2703a54 100644
> --- a/plat/xen/Makefile.uk
> +++ b/plat/xen/Makefile.uk
> @@ -79,4 +79,5 @@ LIBXENBUS_ASINCLUDES-y         += $(LIBXENPLAT_ASINCLUDES-y)
>  LIBXENBUS_CFLAGS-y             += $(LIBXENPLAT_CFLAGS-y)
>  LIBXENBUS_CINCLUDES-y          += $(LIBXENPLAT_CINCLUDES-y)
>  LIBXENBUS_SRCS-y               += $(LIBXENPLAT_BASE)/xenbus/xenbus.c
> +LIBXENBUS_SRCS-y               += $(LIBXENPLAT_BASE)/xenbus/xs_comms.c
>  endif
> diff --git a/plat/xen/xenbus/xenbus.c b/plat/xen/xenbus/xenbus.c
> index 1bc57c3..a20546b 100644
> --- a/plat/xen/xenbus/xenbus.c
> +++ b/plat/xen/xenbus/xenbus.c
> @@ -43,6 +43,7 @@
>  #include <uk/errptr.h>
>  #include <uk/assert.h>
>  #include <xenbus/xenbus.h>
> +#include "xs_comms.h"
>  
>  static struct xenbus_handler xbh;
>  
> @@ -86,6 +87,13 @@ static int xenbus_init(struct uk_alloc *a)
>  
>       xbh.a = a;
>  
> +     ret = xs_comms_init();
> +     if (ret) {
> +             uk_printd(DLVL_ERR,
> +                     "Error initializing Xenstore communication.");
> +             return ret;
> +     }
> +
>       UK_TAILQ_FOREACH_SAFE(drv, &xbh.drv_list, next, drv_next) {
>               if (drv->init) {
>                       ret = drv->init(a);
> diff --git a/plat/xen/xenbus/xs_comms.c b/plat/xen/xenbus/xs_comms.c
> new file mode 100644
> index 0000000..40909c3
> --- /dev/null
> +++ b/plat/xen/xenbus/xs_comms.c
> @@ -0,0 +1,610 @@
> +/* SPDX-License-Identifier: BSD-3-Clause */
> +/*
> + * Authors: Steven Smith (sos22@xxxxxxxxx)
> + *          Grzegorz Milos (gm281@xxxxxxxxx)
> + *          John D. Ramsdell
> + *          Costin Lupu <costin.lupu@xxxxxxxxx>
> + *
> + * Copyright (c) 2006, Cambridge University
> + *               2018, NEC Europe Ltd., NEC Corporation. All rights reserved.
> + *
> + * Redistribution and use in source and binary forms, with or without
> + * modification, are permitted provided that the following conditions
> + * are met:
> + *
> + * 1. Redistributions of source code must retain the above copyright
> + *    notice, this list of conditions and the following disclaimer.
> + * 2. Redistributions in binary form must reproduce the above copyright
> + *    notice, this list of conditions and the following disclaimer in the
> + *    documentation and/or other materials provided with the distribution.
> + * 3. Neither the name of the copyright holder nor the names of its
> + *    contributors may be used to endorse or promote products derived from
> + *    this software without specific prior written permission.
> + *
> + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS 
> IS"
> + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
> + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
> + * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
> + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
> + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
> + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
> + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
> + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
> + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
> + * POSSIBILITY OF SUCH DAMAGE.
> + *
> + * THIS HEADER MAY NOT BE EXTRACTED OR MODIFIED IN ANY WAY.
> + */
> +/*
> + * Communication with Xenstore
> + * Ported from Mini-OS xenbus.c
> + */
> +
> +#include <string.h>
> +#include <uk/errptr.h>
> +#include <uk/bitmap.h>
> +#include <uk/wait.h>
> +#include <uk/arch/spinlock.h>
> +#include <common/events.h>
> +#include <xen-x86/mm.h>
> +#include <xen-x86/setup.h>
> +#include "xs_comms.h"
> +
> +
> +/*
> + * Xenstore handler structure
> + */
> +struct xs_handler {
> +     /**< Communication: event channel */
> +     evtchn_port_t evtchn;
> +     /**< Communication: shared memory */
> +     struct xenstore_domain_interface *buf;
> +     /**< Thread processing incoming xs replies */
> +     struct uk_thread *thread;
> +     /**< Waiting queue for notifying incoming xs replies */
> +     struct uk_waitq waitq;
> +};
> +
> +static struct xs_handler xsh = {
> +     .waitq = __WAIT_QUEUE_INITIALIZER(xsh.waitq),
> +};
> +
> +/*
> + * In-flight request structure.
> + */
> +struct xs_request {
> +     /**< used when queueing requests */
> +     UK_TAILQ_ENTRY(struct xs_request) next;
> +     /**< Waiting queue for incoming reply notification */
> +     struct uk_waitq waitq;
> +     /**< Request header */
> +     struct xsd_sockmsg hdr;
> +     /**< Request payload iovecs */
> +     const struct xs_iovec *payload_iovecs;
> +     /**< Received reply */
> +     struct {
> +             /**< Reply string + size */
> +             struct xs_iovec iovec;
> +             /**< Error number */
> +             int errornum;
> +             /**< Non-zero for incoming replies */
> +             int recvd;
> +     } reply;
> +};
> +UK_TAILQ_HEAD(xs_request_list, struct xs_request);
> +
> +/*
> + * Pool of in-flight requests.
> + * Request IDs are reused, hence the limited set of entries.
> + */
> +struct xs_request_pool {
> +     /**< Number of live requests */
> +     __u32 num_live;
> +     /**< Last probed request index */
> +     __u32 last_probed;
> +     /**< Lock */
> +     spinlock_t lock;
> +     /**< Waiting queue for 'not-full' notifications */
> +     struct uk_waitq waitq;
> +     /**< Queue for requests to be sent */
> +     struct xs_request_list queued;
> +
> +     /* Map size is power of 2 */
> +#define XS_REQ_POOL_SHIFT  5
> +#define XS_REQ_POOL_SIZE   (1 << XS_REQ_POOL_SHIFT)
> +#define XS_REQ_POOL_MASK   (XS_REQ_POOL_SIZE - 1)
> +     unsigned long entries_bm[UK_BITS_TO_LONGS(XS_REQ_POOL_SIZE)
> +                     * sizeof(long)];
> +     /**< Entries */
> +     struct xs_request entries[XS_REQ_POOL_SIZE];
> +};
> +
> +static struct xs_request_pool xs_req_pool;
> +
> +static void xs_request_pool_init(struct xs_request_pool *pool)
> +{
> +     struct xs_request *xs_req;
> +
> +     pool->num_live = 0;
> +     pool->last_probed = -1;
> +     ukarch_spin_lock_init(&pool->lock);
> +     uk_waitq_init(&pool->waitq);
> +     UK_TAILQ_INIT(&pool->queued);
> +     uk_bitmap_zero(pool->entries_bm, XS_REQ_POOL_SIZE);
> +     for (int i = 0; i < XS_REQ_POOL_SIZE; i++) {
> +             xs_req = &pool->entries[i];
> +             xs_req->hdr.req_id = i;
> +             uk_waitq_init(&xs_req->waitq);
> +     }
> +}
> +
> +/*
> + * Allocate an identifier for a Xenstore request.
> + * Blocks if none are available.
> + */
> +static struct xs_request *xs_request_get(void)
> +{
> +     unsigned long entry_idx;
> +
> +     /* wait for an available entry */
> +     while (1) {
> +             ukarch_spin_lock(&xs_req_pool.lock);
> +
> +             if (xs_req_pool.num_live < XS_REQ_POOL_SIZE)
> +                     break;
> +
> +             ukarch_spin_unlock(&xs_req_pool.lock);
> +
> +             uk_waitq_wait_event(&xs_req_pool.waitq,
> +                     (xs_req_pool.num_live < XS_REQ_POOL_SIZE));
> +     }
> +
> +     /* find an available entry */
> +     entry_idx =
> +             uk_find_next_zero_bit(xs_req_pool.entries_bm, XS_REQ_POOL_SIZE,
> +                     (xs_req_pool.last_probed + 1) & XS_REQ_POOL_MASK);
> +
> +     if (entry_idx == XS_REQ_POOL_SIZE)
> +             entry_idx = uk_find_next_zero_bit(xs_req_pool.entries_bm,
> +                     XS_REQ_POOL_SIZE, 0);
> +
> +     uk_set_bit(entry_idx, xs_req_pool.entries_bm);
> +     xs_req_pool.last_probed = entry_idx;
> +     xs_req_pool.num_live++;
> +
> +     ukarch_spin_unlock(&xs_req_pool.lock);
> +
> +     return &xs_req_pool.entries[entry_idx];
> +}
> +
> +/* Release a request identifier */
> +static void xs_request_put(struct xs_request *xs_req)
> +{
> +     __u32 reqid = xs_req->hdr.req_id;
> +
> +     ukarch_spin_lock(&xs_req_pool.lock);
> +
> +     UK_ASSERT(ukarch_test_bit(reqid, xs_req_pool.entries_bm) == 1);
> +
> +     uk_clear_bit(reqid, xs_req_pool.entries_bm);
> +     xs_req_pool.num_live--;
> +
> +     if (xs_req_pool.num_live == XS_REQ_POOL_SIZE - 1)
> +             uk_waitq_wake_up(&xs_req_pool.waitq);
> +
> +     ukarch_spin_unlock(&xs_req_pool.lock);
> +}
> +
> +static struct xs_request *xs_request_peek(void)
> +{
> +     struct xs_request *xs_req;
> +
> +     ukarch_spin_lock(&xs_req_pool.lock);
> +     xs_req = UK_TAILQ_FIRST(&xs_req_pool.queued);
> +     ukarch_spin_unlock(&xs_req_pool.lock);
> +
> +     return xs_req;
> +}
> +
> +static void xs_request_enqueue(struct xs_request *xs_req)
> +{
> +     ukarch_spin_lock(&xs_req_pool.lock);
> +     UK_TAILQ_INSERT_TAIL(&xs_req_pool.queued, xs_req, next);
> +     ukarch_spin_unlock(&xs_req_pool.lock);
> +}
> +
> +static struct xs_request *xs_request_dequeue(void)
> +{
> +     struct xs_request *xs_req;
> +
> +     ukarch_spin_lock(&xs_req_pool.lock);
> +     xs_req = UK_TAILQ_FIRST(&xs_req_pool.queued);
> +     if (xs_req)
> +             UK_TAILQ_REMOVE(&xs_req_pool.queued, xs_req, next);
> +     ukarch_spin_unlock(&xs_req_pool.lock);
> +
> +     return xs_req;
> +}
> +
> +static int xs_avail_to_read(void)
> +{
> +     return (xsh.buf->rsp_prod != xsh.buf->rsp_cons);
> +}
> +
> +static int xs_avail_space_for_read(unsigned int size)
> +{
> +     return (xsh.buf->rsp_prod - xsh.buf->rsp_cons >= size);
> +}
> +
> +static int xs_avail_to_write(void)
> +{
> +     return (xsh.buf->req_prod - xsh.buf->req_cons != XENSTORE_RING_SIZE &&
> +             !UK_TAILQ_EMPTY(&xs_req_pool.queued));
> +}
> +
> +static int xs_avail_space_for_write(unsigned int size)
> +{
> +     return (xsh.buf->req_prod - xsh.buf->req_cons +
> +             size <= XENSTORE_RING_SIZE);
> +}
> +
> +static int xs_avail_work(void)
> +{
> +     return (xs_avail_to_read() || xs_avail_to_write());
> +}
> +
> +/*
> + * Send request to Xenstore. A request is made of multiple iovecs which are
> + * preceded by a single iovec referencing the request header. The iovecs are
> + * seen by Xenstore as if sent atomically. This can block.
> + */
> +static int xs_msg_write(struct xsd_sockmsg *xsd_req,
> +     const struct xs_iovec *iovec)
> +{
> +     XENSTORE_RING_IDX prod;
> +     const struct xs_iovec *crnt_iovec;
> +     struct xs_iovec hdr_iovec;
> +     unsigned int req_size, req_off;
> +     unsigned int buf_off;
> +     unsigned int this_chunk_len;
> +     int rc;
> +
> +     req_size = sizeof(*xsd_req) + xsd_req->len;
> +     if (req_size > XENSTORE_RING_SIZE)
> +             return -ENOSPC;
> +
> +     if (!xs_avail_space_for_write(req_size))
> +             return -ENOSPC;
> +
> +     /* We must write requests after reading the consumer index. */
> +     mb();
> +
> +     /*
> +      * We're now guaranteed to be able to send the message
> +      * without overflowing the ring. Do so.
> +      */
> +
> +     hdr_iovec.data = xsd_req;
> +     hdr_iovec.len  = sizeof(*xsd_req);
> +
> +     /* The batched iovecs are preceded by a single header. */
> +     crnt_iovec = &hdr_iovec;
> +
> +     prod = xsh.buf->req_prod;
> +     req_off = 0;
> +     buf_off = 0;
> +     while (req_off < req_size) {
> +             this_chunk_len = MIN(crnt_iovec->len - buf_off,
> +                     XENSTORE_RING_SIZE - MASK_XENSTORE_IDX(prod));
> +
> +             memcpy(
> +                     (char *) xsh.buf->req + MASK_XENSTORE_IDX(prod),
> +                     (char *) crnt_iovec->data + buf_off,
> +                     this_chunk_len
> +             );
> +
> +             prod += this_chunk_len;
> +             req_off += this_chunk_len;
> +             buf_off += this_chunk_len;
> +
> +             if (buf_off == crnt_iovec->len) {
> +                     buf_off = 0;
> +                     if (crnt_iovec == &hdr_iovec)
> +                             crnt_iovec = iovec;
> +                     else
> +                             crnt_iovec++;
> +             }
> +     }
> +
> +     uk_printd(DLVL_EXTRA, "Complete main loop of %s.\n", __func__);
> +     UK_ASSERT(buf_off == 0);
> +     UK_ASSERT(req_off == req_size);
> +     UK_ASSERT(prod <= xsh.buf->req_cons + XENSTORE_RING_SIZE);
> +
> +     /* Remote must see entire message before updating indexes */
> +     wmb();
> +
> +     xsh.buf->req_prod += req_size;
> +
> +     /* Send evtchn to notify remote */
> +     rc = notify_remote_via_evtchn(xsh.evtchn);
> +     UK_ASSERT(rc == 0);
> +
> +     return 0;
> +}
> +
> +int xs_msg_reply(enum xsd_sockmsg_type msg_type, xenbus_transaction_t xbt,
> +     const struct xs_iovec *req_iovecs, int req_iovecs_num,
> +     struct xs_iovec *rep_iovec)
> +{
> +     struct xs_request *xs_req;
> +     int err;
> +
> +     if (req_iovecs == NULL)
> +             return -EINVAL;
> +
> +     xs_req = xs_request_get();
> +     xs_req->hdr.type = msg_type;
> +     /* req_id was set on pool init  */
> +     xs_req->hdr.tx_id = xbt;
> +     xs_req->hdr.len = 0;
> +     for (int i = 0; i < req_iovecs_num; i++)
> +             xs_req->hdr.len += req_iovecs[i].len;
> +
> +     xs_req->payload_iovecs = req_iovecs;
> +     xs_req->reply.recvd = 0;
> +
> +     /* enqueue the request */
> +     xs_request_enqueue(xs_req);
> +     /* wake xenstore thread to send it */
> +     uk_waitq_wake_up(&xsh.waitq);
> +
> +     /* wait reply */
> +     uk_waitq_wait_event(&xs_req->waitq,
> +             xs_req->reply.recvd != 0);
> +
> +     err = -xs_req->reply.errornum;
> +     if (err == 0) {
> +             if (rep_iovec)
> +                     *rep_iovec = xs_req->reply.iovec;
> +             else
> +                     free(xs_req->reply.iovec.data);
> +     }
> +
> +     xs_request_put(xs_req);
> +
> +     return err;
> +}
> +
> +void xs_send(void)
> +{
> +     struct xs_request *xs_req;
> +     int err;
> +
> +     xs_req = xs_request_peek();
> +     while (xs_req != NULL) {
> +             err = xs_msg_write(&xs_req->hdr, xs_req->payload_iovecs);
> +             if (err) {
> +                     if (err != -ENOSPC)
> +                             uk_printd(DLVL_WARN,
> +                                     "Error sending message err=%d\n", err);
> +                     break;
> +             }
> +
> +             /* remove it from queue */
> +             xs_request_dequeue();
> +
> +             xs_req = xs_request_peek();
> +     }
> +}
> +
> +/*
> + * Converts a Xenstore reply error to a positive error number.
> + * Returns 0 if the reply is successful.
> + */
> +static int reply_to_errno(const char *reply)
> +{
> +     int err = 0;
> +
> +     for (int i = 0; i < (int) ARRAY_SIZE(xsd_errors); i++) {
> +             if (!strcmp(reply, xsd_errors[i].errstring)) {
> +                     err = xsd_errors[i].errnum;
> +                     goto out;
> +             }
> +     }
> +
> +     uk_printd(DLVL_WARN, "Unknown Xenstore error: %s\n", reply);
> +     err = EINVAL;
> +
> +out:
> +     return err;
> +}
> +
> +/* Process an incoming xs reply */
> +static void process_reply(struct xsd_sockmsg *hdr, char *payload)
> +{
> +     struct xs_request *xs_req;
> +
> +     if (!ukarch_test_bit(hdr->req_id, xs_req_pool.entries_bm)) {
> +             uk_printd(DLVL_WARN, "Invalid reply id=%d\n", hdr->req_id);
> +             free(payload);
> +             return;
> +     }
> +
> +     xs_req = &xs_req_pool.entries[hdr->req_id];
> +
> +     if (hdr->type == XS_ERROR) {
> +             xs_req->reply.errornum = reply_to_errno(payload);
> +             free(payload);
> +
> +     } else if (hdr->type != xs_req->hdr.type) {
> +             uk_printd(DLVL_WARN,
> +                     "Mismatching message type: %d\n", hdr->type);
> +             free(payload);
> +             return;
> +
> +     } else {
> +             /* set reply */
> +             xs_req->reply.iovec.data = payload;
> +             xs_req->reply.iovec.len = hdr->len;
> +             xs_req->reply.errornum = 0;
> +     }
> +
> +     xs_req->reply.recvd = 1;
> +
> +     /* notify waiting requester */
> +     uk_waitq_wake_up(&xs_req->waitq);
> +}
> +
> +/* Process an incoming xs watch event */
> +static void process_watch_event(char *watch_msg)
> +{
> +     /* TODO */
> +}
> +
> +static void memcpy_from_ring(const char *ring, char *dest, int off, int len)
> +{
> +     int c1, c2;
> +
> +     c1 = MIN(len, XENSTORE_RING_SIZE - off);
> +     c2 = len - c1;
> +
> +     memcpy(dest, ring + off, c1);
> +     if (c2)
> +             memcpy(dest + c1, ring, c2);
> +}
> +
> +static void xs_msg_read(struct xsd_sockmsg *hdr)
> +{
> +     XENSTORE_RING_IDX cons;
> +     char *payload;
> +
> +     payload = malloc(hdr->len + 1);
> +     if (payload == NULL) {
> +             uk_printd(DLVL_WARN,
> +                     "No memory available for saving Xenstore message!\n");
> +             return;
> +     }
> +
> +     cons = xsh.buf->rsp_cons;
> +
> +     /* copy payload */
> +     memcpy_from_ring(
> +             xsh.buf->rsp,
> +             payload,
> +             MASK_XENSTORE_IDX(cons + sizeof(*hdr)),
> +             hdr->len
> +     );
> +     payload[hdr->len] = '\0';
> +
> +     /* Remote must not see available space until we've copied the reply */
> +     mb();
> +     xsh.buf->rsp_cons += sizeof(*hdr) + hdr->len;
> +
> +     if (xsh.buf->rsp_prod - cons >= XENSTORE_RING_SIZE)
> +             notify_remote_via_evtchn(xsh.evtchn);
> +
> +     if (hdr->type == XS_WATCH_EVENT)
> +             process_watch_event(payload);
> +     else
> +             process_reply(hdr, payload);
> +}
> +
> +static void xs_recv(void)
> +{
> +     struct xsd_sockmsg msg;
> +
> +     while (1) {
> +             uk_printd(DLVL_EXTRA, "Rsp_cons %d, rsp_prod %d.\n",
> +                     xsh.buf->rsp_cons, xsh.buf->rsp_prod);
> +
> +             if (!xs_avail_space_for_read(sizeof(msg)))
> +                     break;
> +
> +             /* Make sure data is read after reading the indexes */
> +             rmb();
> +
> +             /* copy the message header */
> +             memcpy_from_ring(
> +                     xsh.buf->rsp,
> +                     (char *) &msg,
> +                     MASK_XENSTORE_IDX(xsh.buf->rsp_cons),
> +                     sizeof(msg)
> +             );
> +
> +             uk_printd(DLVL_EXTRA, "Msg len %lu, %u avail, id %u.\n",
> +                     msg.len + sizeof(msg),
> +                     xsh.buf->rsp_prod - xsh.buf->rsp_cons,
> +                     msg.req_id);
> +
> +             if (!xs_avail_space_for_read(sizeof(msg) + msg.len))
> +                     break;
> +
> +             /* Make sure data is read after reading the indexes */
> +             rmb();
> +
> +             uk_printd(DLVL_EXTRA, "Message is good.\n");
> +             xs_msg_read(&msg);
> +     }
> +}
> +
> +static void xs_thread_func(void *ign __unused)
> +{
> +     for (;;) {
> +             uk_waitq_wait_event(&xsh.waitq, xs_avail_work());
> +
> +             if (xs_avail_to_write())
> +                     xs_send();
> +
> +             if (xs_avail_to_read())
> +                     xs_recv();
> +     }
> +}
> +
> +static void xs_evtchn_handler(evtchn_port_t port,
> +             struct __regs *regs __unused, void *ign __unused)
> +{
> +     UK_ASSERT(xsh.evtchn == port);
> +     uk_waitq_wake_up(&xsh.waitq);
> +}
> +
> +int xs_comms_init(void)
> +{
> +     struct uk_thread *thread;
> +     evtchn_port_t port;
> +
> +     xs_request_pool_init(&xs_req_pool);
> +
> +     uk_waitq_init(&xsh.waitq);
> +
> +     thread = uk_thread_create("xenstore", xs_thread_func, NULL);
> +     if (PTRISERR(thread))
> +             return PTR2ERR(thread);
> +
> +     xsh.thread = thread;
> +
> +     xsh.evtchn = HYPERVISOR_start_info->store_evtchn;
> +     xsh.buf = mfn_to_virt(HYPERVISOR_start_info->store_mfn);
> +
> +     port = bind_evtchn(xsh.evtchn, xs_evtchn_handler, NULL);
> +     UK_ASSERT(port == xsh.evtchn);
> +     unmask_evtchn(xsh.evtchn);
> +
> +     uk_printd(DLVL_INFO,
> +             "Xenstore connection initialised on port %d, buf %p (mfn 
> %#lx)\n",
> +             port, xsh.buf, HYPERVISOR_start_info->store_mfn);
> +
> +     return 0;
> +}
> +
> +void xs_comms_fini(void)
> +{
> +     mask_evtchn(xsh.evtchn);
> +     unbind_evtchn(xsh.evtchn);
> +
> +     xsh.buf = NULL;
> +
> +     /* TODO stop thread, instead of killing it */
> +     uk_thread_destroy(xsh.thread);
> +     xsh.thread = NULL;
> +}
> diff --git a/plat/xen/xenbus/xs_comms.h b/plat/xen/xenbus/xs_comms.h
> new file mode 100644
> index 0000000..cae6093
> --- /dev/null
> +++ b/plat/xen/xenbus/xs_comms.h
> @@ -0,0 +1,64 @@
> +/* SPDX-License-Identifier: BSD-3-Clause */
> +/*
> + * Authors: Costin Lupu <costin.lupu@xxxxxxxxx>
> + *
> + * Copyright (c) 2018, NEC Europe Ltd., NEC Corporation. All rights reserved.
> + *
> + * Redistribution and use in source and binary forms, with or without
> + * modification, are permitted provided that the following conditions
> + * are met:
> + *
> + * 1. Redistributions of source code must retain the above copyright
> + *    notice, this list of conditions and the following disclaimer.
> + * 2. Redistributions in binary form must reproduce the above copyright
> + *    notice, this list of conditions and the following disclaimer in the
> + *    documentation and/or other materials provided with the distribution.
> + * 3. Neither the name of the copyright holder nor the names of its
> + *    contributors may be used to endorse or promote products derived from
> + *    this software without specific prior written permission.
> + *
> + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS 
> IS"
> + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
> + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
> + * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
> + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
> + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
> + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
> + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
> + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
> + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
> + * POSSIBILITY OF SUCH DAMAGE.
> + *
> + * THIS HEADER MAY NOT BE EXTRACTED OR MODIFIED IN ANY WAY.
> + */
> +
> +#ifndef __XS_COMMS_H__
> +#define __XS_COMMS_H__
> +
> +#include <xen/io/xs_wire.h>
> +#include <xenbus/xenbus.h>
> +
> +int  xs_comms_init(void);
> +void xs_comms_fini(void);
> +
> +struct xs_iovec {
> +     void *data;
> +     unsigned int len;
> +};
> +
> +/*
> + * Sends a message to Xenstore and blocks waiting for a reply.
> + * The reply is malloc'ed and should be freed by the caller.
> + *
> + * @param msg_type Xenstore message type
> + * @param xbt Xenbus transaction id
> + * @param req_iovecs Array of request strings buffers
> + * @param req_iovecs_num Request strings buffers number
> + * @param rep_iovec Incoming reply string buffer (optional)
> + * @return 0 on success, a negative errno value on error.
> + */
> +int xs_msg_reply(enum xsd_sockmsg_type msg_type, xenbus_transaction_t xbt,
> +     const struct xs_iovec *req_iovecs, int req_iovecs_num,
> +     struct xs_iovec *rep_iovec);
> +
> +#endif /* __XS_COMMS_H__ */
> -- 
> 2.11.0
>

-- 
Yuri Volchkov
Software Specialist

NEC Europe Ltd
Kurfürsten-Anlage 36
D-69115 Heidelberg

_______________________________________________
Minios-devel mailing list
Minios-devel@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/mailman/listinfo/minios-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.