[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-devel] [PATCH v7 3/3] xen/blkback: Remove unnecessary static variable name prefixes



A few of static variables in blkback have 'xen_blkif_' prefix, though it
is unnecessary for static variables.  This commit removes such prefixes.

Reviewed-by: Roger Pau Monné <roger.pau@xxxxxxxxxx>
Signed-off-by: SeongJae Park <sjpark@xxxxxxxxx>
---
 drivers/block/xen-blkback/blkback.c | 37 +++++++++++++----------------
 1 file changed, 17 insertions(+), 20 deletions(-)

diff --git a/drivers/block/xen-blkback/blkback.c 
b/drivers/block/xen-blkback/blkback.c
index 98823d150905..f41c698dd854 100644
--- a/drivers/block/xen-blkback/blkback.c
+++ b/drivers/block/xen-blkback/blkback.c
@@ -62,8 +62,8 @@
  * IO workloads.
  */
 
-static int xen_blkif_max_buffer_pages = 1024;
-module_param_named(max_buffer_pages, xen_blkif_max_buffer_pages, int, 0644);
+static int max_buffer_pages = 1024;
+module_param_named(max_buffer_pages, max_buffer_pages, int, 0644);
 MODULE_PARM_DESC(max_buffer_pages,
 "Maximum number of free pages to keep in each block backend buffer");
 
@@ -78,8 +78,8 @@ MODULE_PARM_DESC(max_buffer_pages,
  * algorithm.
  */
 
-static int xen_blkif_max_pgrants = 1056;
-module_param_named(max_persistent_grants, xen_blkif_max_pgrants, int, 0644);
+static int max_pgrants = 1056;
+module_param_named(max_persistent_grants, max_pgrants, int, 0644);
 MODULE_PARM_DESC(max_persistent_grants,
                  "Maximum number of grants to map persistently");
 
@@ -88,8 +88,8 @@ MODULE_PARM_DESC(max_persistent_grants,
  * use. The time is in seconds, 0 means indefinitely long.
  */
 
-static unsigned int xen_blkif_pgrant_timeout = 60;
-module_param_named(persistent_grant_unused_seconds, xen_blkif_pgrant_timeout,
+static unsigned int pgrant_timeout = 60;
+module_param_named(persistent_grant_unused_seconds, pgrant_timeout,
                   uint, 0644);
 MODULE_PARM_DESC(persistent_grant_unused_seconds,
                 "Time in seconds an unused persistent grant is allowed to "
@@ -137,9 +137,8 @@ module_param(log_stats, int, 0644);
 
 static inline bool persistent_gnt_timeout(struct persistent_gnt 
*persistent_gnt)
 {
-       return xen_blkif_pgrant_timeout &&
-              (jiffies - persistent_gnt->last_used >=
-               HZ * xen_blkif_pgrant_timeout);
+       return pgrant_timeout && (jiffies - persistent_gnt->last_used >=
+                       HZ * pgrant_timeout);
 }
 
 /* Once a memory pressure is detected, squeeze free page pools for a while. */
@@ -249,7 +248,7 @@ static int add_persistent_gnt(struct xen_blkif_ring *ring,
        struct persistent_gnt *this;
        struct xen_blkif *blkif = ring->blkif;
 
-       if (ring->persistent_gnt_c >= xen_blkif_max_pgrants) {
+       if (ring->persistent_gnt_c >= max_pgrants) {
                if (!blkif->vbd.overflow_max_grants)
                        blkif->vbd.overflow_max_grants = 1;
                return -EBUSY;
@@ -412,14 +411,13 @@ static void purge_persistent_gnt(struct xen_blkif_ring 
*ring)
                goto out;
        }
 
-       if (ring->persistent_gnt_c < xen_blkif_max_pgrants ||
-           (ring->persistent_gnt_c == xen_blkif_max_pgrants &&
+       if (ring->persistent_gnt_c < max_pgrants ||
+           (ring->persistent_gnt_c == max_pgrants &&
            !ring->blkif->vbd.overflow_max_grants)) {
                num_clean = 0;
        } else {
-               num_clean = (xen_blkif_max_pgrants / 100) * LRU_PERCENT_CLEAN;
-               num_clean = ring->persistent_gnt_c - xen_blkif_max_pgrants +
-                           num_clean;
+               num_clean = (max_pgrants / 100) * LRU_PERCENT_CLEAN;
+               num_clean = ring->persistent_gnt_c - max_pgrants + num_clean;
                num_clean = min(ring->persistent_gnt_c, num_clean);
                pr_debug("Going to purge at least %u persistent grants\n",
                         num_clean);
@@ -614,8 +612,7 @@ static void print_stats(struct xen_blkif_ring *ring)
                 current->comm, ring->st_oo_req,
                 ring->st_rd_req, ring->st_wr_req,
                 ring->st_f_req, ring->st_ds_req,
-                ring->persistent_gnt_c,
-                xen_blkif_max_pgrants);
+                ring->persistent_gnt_c, max_pgrants);
        ring->st_print = jiffies + msecs_to_jiffies(10 * 1000);
        ring->st_rd_req = 0;
        ring->st_wr_req = 0;
@@ -675,7 +672,7 @@ int xen_blkif_schedule(void *arg)
                if (time_before(jiffies, buffer_squeeze_end))
                        shrink_free_pagepool(ring, 0);
                else
-                       shrink_free_pagepool(ring, xen_blkif_max_buffer_pages);
+                       shrink_free_pagepool(ring, max_buffer_pages);
 
                if (log_stats && time_after(jiffies, ring->st_print))
                        print_stats(ring);
@@ -902,7 +899,7 @@ static int xen_blkbk_map(struct xen_blkif_ring *ring,
                        continue;
                }
                if (use_persistent_gnts &&
-                   ring->persistent_gnt_c < xen_blkif_max_pgrants) {
+                   ring->persistent_gnt_c < max_pgrants) {
                        /*
                         * We are using persistent grants, the grant is
                         * not mapped but we might have room for it.
@@ -929,7 +926,7 @@ static int xen_blkbk_map(struct xen_blkif_ring *ring,
                        pages[seg_idx]->persistent_gnt = persistent_gnt;
                        pr_debug("grant %u added to the tree of persistent 
grants, using %u/%u\n",
                                 persistent_gnt->gnt, ring->persistent_gnt_c,
-                                xen_blkif_max_pgrants);
+                                max_pgrants);
                        goto next;
                }
                if (use_persistent_gnts && !blkif->vbd.overflow_max_grants) {
-- 
2.17.1


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/mailman/listinfo/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.