|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH 1/3] arinc: whitespace and formatting fixes
On 18/11/2013 20:16, Nathan Studer wrote:
> From: Nathan Studer <nate.studer@xxxxxxxxxxxxxxx>
>
> Remove the excessive amount of trailing whitespace in the
> arinc653 scheduler file and add a local variables block.
>
> Suggested-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
> Signed-off-by: Nathan Studer <nate.studer@xxxxxxxxxxxxxxx>
That's much better, thanks.
Reviewed-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
> ---
> xen/common/sched_arinc653.c | 96
> ++++++++++++++++++++++++-------------------
> 1 file changed, 53 insertions(+), 43 deletions(-)
>
> diff --git a/xen/common/sched_arinc653.c b/xen/common/sched_arinc653.c
> index 7b7b387..f4eb943 100644
> --- a/xen/common/sched_arinc653.c
> +++ b/xen/common/sched_arinc653.c
> @@ -40,9 +40,9 @@
> * Private Macros *
> **************************************************************************/
>
> -/**
> - * Retrieve the idle VCPU for a given physical CPU
> - */
> +/**
> + * Retrieve the idle VCPU for a given physical CPU
> + */
> #define IDLETASK(cpu) (idle_vcpu[cpu])
>
> /**
> @@ -76,7 +76,7 @@ typedef struct arinc653_vcpu_s
> struct list_head list;
> } arinc653_vcpu_t;
>
> -/**
> +/**
> * The sched_entry_t structure holds a single entry of the
> * ARINC 653 schedule.
> */
> @@ -101,8 +101,8 @@ typedef struct sched_entry_s
> typedef struct a653sched_priv_s
> {
> /**
> - * This array holds the active ARINC 653 schedule.
> - *
> + * This array holds the active ARINC 653 schedule.
> + *
> * When the system tries to start a new VCPU, this schedule is scanned
> * to look for a matching (handle, VCPU #) pair. If both the handle
> (UUID)
> * and VCPU number match, then the VCPU is allowed to run. Its run time
> @@ -112,12 +112,12 @@ typedef struct a653sched_priv_s
>
> /**
> * This variable holds the number of entries that are valid in
> - * the arinc653_schedule table.
> - *
> + * the arinc653_schedule table.
> + *
> * This is not necessarily the same as the number of domains in the
> * schedule. A domain could be listed multiple times within the schedule,
> * or a domain with multiple VCPUs could have a different
> - * schedule entry for each VCPU.
> + * schedule entry for each VCPU.
> */
> int num_schedule_entries;
>
> @@ -131,9 +131,9 @@ typedef struct a653sched_priv_s
> */
> s_time_t next_major_frame;
>
> - /**
> - * pointers to all Xen VCPU structures for iterating through
> - */
> + /**
> + * pointers to all Xen VCPU structures for iterating through
> + */
> struct list_head vcpu_list;
> } a653sched_priv_t;
>
> @@ -143,14 +143,14 @@ typedef struct a653sched_priv_s
>
> /**
> * This function compares two domain handles.
> - *
> + *
> * @param h1 Pointer to handle 1
> * @param h2 Pointer to handle 2
> - *
> + *
> * @return <ul>
> - * <li> <0: handle 1 is less than handle 2
> - * <li> 0: handle 1 is equal to handle 2
> - * <li> >0: handle 1 is greater than handle 2
> + * <li> <0: handle 1 is less than handle 2
> + * <li> 0: handle 1 is equal to handle 2
> + * <li> >0: handle 1 is greater than handle 2
> * </ul>
> */
> static int dom_handle_cmp(const xen_domain_handle_t h1,
> @@ -162,11 +162,11 @@ static int dom_handle_cmp(const xen_domain_handle_t h1,
> /**
> * This function searches the vcpu list to find a VCPU that matches
> * the domain handle and VCPU ID specified.
> - *
> + *
> * @param ops Pointer to this instance of the scheduler structure
> * @param handle Pointer to handler
> * @param vcpu_id VCPU ID
> - *
> + *
> * @return <ul>
> * <li> Pointer to the matching VCPU if one is found
> * <li> NULL otherwise
> @@ -191,7 +191,7 @@ static struct vcpu *find_vcpu(
> /**
> * This function updates the pointer to the Xen VCPU structure for each entry
> * in the ARINC 653 schedule.
> - *
> + *
> * @param ops Pointer to this instance of the scheduler structure
> * @return <None>
> */
> @@ -211,7 +211,7 @@ static void update_schedule_vcpus(const struct scheduler
> *ops)
> * in place a new ARINC653 schedule.
> *
> * @param ops Pointer to this instance of the scheduler structure
> - *
> + *
> * @return <ul>
> * <li> 0 = success
> * <li> !0 = error
> @@ -253,10 +253,10 @@ arinc653_sched_set(
> if ( !found_dom0 )
> goto fail;
>
> - /*
> + /*
> * Error if the major frame is not large enough to run all entries as
> * indicated by comparing the total run time to the major frame length.
> - */
> + */
> if ( total_runtime > schedule->major_frame )
> goto fail;
>
> @@ -276,10 +276,10 @@ arinc653_sched_set(
> update_schedule_vcpus(ops);
>
> /*
> - * The newly-installed schedule takes effect immediately. We do not even
> + * The newly-installed schedule takes effect immediately. We do not even
> * wait for the current major frame to expire.
> *
> - * Signal a new major frame to begin. The next major frame is set up by
> + * Signal a new major frame to begin. The next major frame is set up by
> * the do_schedule callback function when it is next invoked.
> */
> sched_priv->next_major_frame = NOW();
> @@ -392,8 +392,8 @@ a653sched_alloc_vdata(const struct scheduler *ops, struct
> vcpu *vc, void *dd)
>
> /*
> * Initialize our ARINC 653 scheduler-specific information for the VCPU.
> - * The VCPU starts "asleep." When Xen is ready for the VCPU to run, it
> - * will call the vcpu_wake scheduler callback function and our scheduler
> + * The VCPU starts "asleep." When Xen is ready for the VCPU to run, it
> + * will call the vcpu_wake scheduler callback function and our scheduler
> * will mark the VCPU awake.
> */
> svc->vc = vc;
> @@ -483,7 +483,7 @@ a653sched_free_domdata(const struct scheduler *ops, void
> *data)
>
> /**
> * Xen scheduler callback function to sleep a VCPU
> - *
> + *
> * @param ops Pointer to this instance of the scheduler structure
> * @param vc Pointer to the VCPU structure for the current domain
> */
> @@ -503,7 +503,7 @@ a653sched_vcpu_sleep(const struct scheduler *ops, struct
> vcpu *vc)
>
> /**
> * Xen scheduler callback function to wake up a VCPU
> - *
> + *
> * @param ops Pointer to this instance of the scheduler structure
> * @param vc Pointer to the VCPU structure for the current domain
> */
> @@ -519,10 +519,10 @@ a653sched_vcpu_wake(const struct scheduler *ops, struct
> vcpu *vc)
> /**
> * Xen scheduler callback function to select a VCPU to run.
> * This is the main scheduler routine.
> - *
> + *
> * @param ops Pointer to this instance of the scheduler structure
> * @param now Current time
> - *
> + *
> * @return Address of the VCPU structure scheduled to be run next
> * Amount of time to execute the returned VCPU
> * Flag for whether the VCPU was migrated
> @@ -559,7 +559,7 @@ a653sched_do_schedule(
> }
> }
>
> - /*
> + /*
> * If we exhausted the domains in the schedule and still have time left
> * in the major frame then switch next at the next major frame.
> */
> @@ -567,10 +567,10 @@ a653sched_do_schedule(
> next_switch_time = sched_priv->next_major_frame;
>
> /*
> - * If there are more domains to run in the current major frame, set
> - * new_task equal to the address of next domain's VCPU structure.
> - * Otherwise, set new_task equal to the address of the idle task's VCPU
> - * structure.
> + * If there are more domains to run in the current major frame, set
> + * new_task equal to the address of next domain's VCPU structure.
> + * Otherwise, set new_task equal to the address of the idle task's VCPU
> + * structure.
> */
> new_task = (sched_index < sched_priv->num_schedule_entries)
> ? sched_priv->schedule[sched_index].vc
> @@ -584,10 +584,10 @@ a653sched_do_schedule(
> new_task = IDLETASK(0);
> BUG_ON(new_task == NULL);
>
> - /*
> + /*
> * Check to make sure we did not miss a major frame.
> - * This is a good test for robust partitioning.
> - */
> + * This is a good test for robust partitioning.
> + */
> BUG_ON(now >= sched_priv->next_major_frame);
>
> /* Tasklet work (which runs in idle VCPU context) overrides all else. */
> @@ -595,8 +595,8 @@ a653sched_do_schedule(
> new_task = IDLETASK(0);
>
> /*
> - * Return the amount of time the next domain has to run and the address
> - * of the selected task's VCPU structure.
> + * Return the amount of time the next domain has to run and the address
> + * of the selected task's VCPU structure.
> */
> ret.time = next_switch_time - now;
> ret.task = new_task;
> @@ -609,10 +609,10 @@ a653sched_do_schedule(
>
> /**
> * Xen scheduler callback function to select a CPU for the VCPU to run on
> - *
> + *
> * @param ops Pointer to this instance of the scheduler structure
> * @param v Pointer to the VCPU structure for the current domain
> - *
> + *
> * @return Number of selected physical CPU
> */
> static int
> @@ -709,3 +709,13 @@ const struct scheduler sched_arinc653_def = {
> .tick_suspend = NULL,
> .tick_resume = NULL,
> };
> +
> +/*
> + * Local variables:
> + * mode: C
> + * c-file-style: "BSD"
> + * c-basic-offset: 4
> + * tab-width: 4
> + * indent-tabs-mode: nil
> + * End:
> + */
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |