[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH 13/14] tmem: reorg the shared pool allocate path



On Tue, Jan 28, 2014 at 12:28:32PM +0800, Bob Liu wrote:
> Reorg the code to make it more readable.
> Check the return value of shared_pool_join() and drop a unneeded call to
> it. Disable creating a shared&persistant pool in an advance place.
> 
> Signed-off-by: Bob Liu <bob.liu@xxxxxxxxxx>
> ---
>  xen/common/tmem.c |  104 
> +++++++++++++++++++++++++++++++++++------------------
>  1 file changed, 70 insertions(+), 34 deletions(-)
> 
> diff --git a/xen/common/tmem.c b/xen/common/tmem.c
> index 37f4e8f..ab5bba7 100644
> --- a/xen/common/tmem.c
> +++ b/xen/common/tmem.c
> @@ -1021,21 +1021,24 @@ static void pool_free(struct tmem_pool *pool)
>      xfree(pool);
>  }
>  
> -/* register new_client as a user of this shared pool and return new
> -   total number of registered users */
> +/*
> + * Register new_client as a user of this shared pool and return 0 on succ.
> + */
>  static int shared_pool_join(struct tmem_pool *pool, struct client 
> *new_client)
>  {
>      struct share_list *sl;
> -
>      ASSERT(is_shared(pool));
> +

Stray?

>      if ( (sl = tmem_malloc(sizeof(struct share_list), NULL)) == NULL )
>          return -1;
>      sl->client = new_client;
>      list_add_tail(&sl->share_list, &pool->share_list);
>      if ( new_client->cli_id != pool->client->cli_id )
>          tmem_client_info("adding new %s %d to shared pool owned by %s %d\n",
> -            tmem_client_str, new_client->cli_id, tmem_client_str, 
> pool->client->cli_id);
> -    return ++pool->shared_count;
> +                    tmem_client_str, new_client->cli_id, tmem_client_str,
> +                    pool->client->cli_id);
> +    ++pool->shared_count;
> +    return 0;
>  }
>  
>  /* reassign "ownership" of the pool to another client that shares this pool 
> */
> @@ -1841,8 +1844,7 @@ static int do_tmem_new_pool(domid_t this_cli_id,
>      int specversion = (flags >> TMEM_POOL_VERSION_SHIFT)
>           & TMEM_POOL_VERSION_MASK;
>      struct tmem_pool *pool, *shpool;
> -    int s_poolid, first_unused_s_poolid;
> -    int i;
> +    int i, first_unused_s_poolid;
>  
>      if ( this_cli_id == TMEM_CLI_ID_NULL )
>          cli_id = current->domain->domain_id;
> @@ -1856,6 +1858,11 @@ static int do_tmem_new_pool(domid_t this_cli_id,
>          tmem_client_err("failed... unsupported spec version\n");
>          return -EPERM;
>      }
> +    if ( shared && persistent )
> +    {
> +        tmem_client_err("failed... unable to create a shared-persistant 
> pool\n");
> +        return -EPERM;
> +    }
>      if ( pagebits != (PAGE_SHIFT - 12) )
>      {
>          tmem_client_err("failed... unsupported pagesize %d\n",
> @@ -1872,17 +1879,12 @@ static int do_tmem_new_pool(domid_t this_cli_id,
>          tmem_client_err("failed... reserved bits must be zero\n");
>          return -EPERM;
>      }
> -    if ( (pool = pool_alloc()) == NULL )
> -    {
> -        tmem_client_err("failed... out of memory\n");
> -        return -ENOMEM;
> -    }
>      if ( this_cli_id != TMEM_CLI_ID_NULL )
>      {
>          if ( (client = tmem_client_from_cli_id(this_cli_id)) == NULL
>               || d_poolid >= MAX_POOLS_PER_DOMAIN
>               || client->pools[d_poolid] != NULL )
> -            goto fail;
> +            return -EPERM;
>      }
>      else
>      {
> @@ -1895,13 +1897,35 @@ static int do_tmem_new_pool(domid_t this_cli_id,
>          {
>              tmem_client_err("failed... no more pool slots available for this 
> %s\n",
>                     tmem_client_str);
> -            goto fail;
> +            return -EPERM;
>          }
>      }
> +
> +    if ( (pool = pool_alloc()) == NULL )
> +    {
> +        tmem_client_err("failed... out of memory\n");
> +        return -ENOMEM;
> +    }
> +    client->pools[d_poolid] = pool;
> +    pool->client = client;
> +    pool->pool_id = d_poolid;
> +    pool->shared = shared;
> +    pool->persistent = persistent;
> +    pool->uuid[0] = uuid_lo;
> +    pool->uuid[1] = uuid_hi;
> +
> +    /*
> +     * Already created a pool when arrived here, but need some special 
> process
> +     * for shared pool.
> +     */
>      if ( shared )
>      {
>          if ( uuid_lo == -1L && uuid_hi == -1L )
> -            shared = 0;
> +        {
> +         tmem_client_info("Invalid uuid, create non shared pool instead!\n");
> +            pool->shared = 0;

spaces.
> +         goto out;

We don't set any errors? Should we do that and return?

(And potentially do the pool_alloc even further down)

> +        }
>          if ( client->shared_auth_required && !global_shared_auth )
>          {
>              for ( i = 0; i < MAX_GLOBAL_SHARED_POOLS; i++)
> @@ -1909,48 +1933,60 @@ static int do_tmem_new_pool(domid_t this_cli_id,
>                       (client->shared_auth_uuid[i][1] == uuid_hi) )
>                      break;
>              if ( i == MAX_GLOBAL_SHARED_POOLS )
> -                shared = 0;
> +         {
> +                tmem_client_info("Shared auth failed, create non shared pool 
> instead!\n");
> +                pool->shared = 0;
> +                goto out;

Ditto. Should we set errors and return?

(And naturally do the pool_allocation even further down?)

> +            }
>          }
> -    }
> -    pool->shared = shared;
> -    pool->client = client;
> -    if ( shared )
> -    {
> +
> +     /*
> +      * Authorize okay, match a global shared pool or use the newly allocated
> +      * one
> +      */
>          first_unused_s_poolid = MAX_GLOBAL_SHARED_POOLS;
> -        for ( s_poolid = 0; s_poolid < MAX_GLOBAL_SHARED_POOLS; s_poolid++ )
> +        for ( i = 0; i < MAX_GLOBAL_SHARED_POOLS; i++ )
>          {
> -            if ( (shpool = global_shared_pools[s_poolid]) != NULL )
> +            if ( (shpool = global_shared_pools[i]) != NULL )
>              {
>                  if ( shpool->uuid[0] == uuid_lo && shpool->uuid[1] == 
> uuid_hi )
>                  {
> +                 /* Succ to match a global shared pool */
>                      tmem_client_info("(matches shared pool 
> uuid=%"PRIx64".%"PRIx64") pool_id=%d\n",
>                          uuid_hi, uuid_lo, d_poolid);
> -                    client->pools[d_poolid] = global_shared_pools[s_poolid];
> -                    shared_pool_join(global_shared_pools[s_poolid], client);
> -                    pool_free(pool);
> -                    return d_poolid;
> +                    client->pools[d_poolid] = shpool;
> +                    if ( !shared_pool_join(shpool, client) )

Yeey! Error checking.
> +                    {
> +                        pool_free(pool);
> +                        goto out;
> +                 }
> +                    else
> +                        goto fail;

At this point we would return d_poolid. Is that OK? Should we instead
return an error?
>                  }
>              }
> -            else if ( first_unused_s_poolid == MAX_GLOBAL_SHARED_POOLS )
> -                first_unused_s_poolid = s_poolid;
> +            else
> +            {
> +                if ( first_unused_s_poolid == MAX_GLOBAL_SHARED_POOLS )
> +                    first_unused_s_poolid = i;
> +            }
>          }
> +
> +     /* Failed to find a global shard pool slot */

Spaces.
>          if ( first_unused_s_poolid == MAX_GLOBAL_SHARED_POOLS )
>          {
>              tmem_client_warn("tmem: failed... no global shared pool slots 
> available\n");
>              goto fail;
>          }
> +     /* Add pool to global shard pool */

Spaces.
>          else
>          {
>              INIT_LIST_HEAD(&pool->share_list);
>              pool->shared_count = 0;
>              global_shared_pools[first_unused_s_poolid] = pool;
> -            (void)shared_pool_join(pool,client);
>          }
>      }
> -    client->pools[d_poolid] = pool;
> -    pool->pool_id = d_poolid;
> -    pool->persistent = persistent;
> -    pool->uuid[0] = uuid_lo; pool->uuid[1] = uuid_hi;
> +
> +out:
>      tmem_client_info("pool_id=%d\n", d_poolid);
>      return d_poolid;
>  
> -- 
> 1.7.10.4
> 

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.