[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH 11/14] xenalyze: handle scheduling events
On Fri, Feb 05, 2016 at 07:35:22PM +0100, Dario Faggioli wrote: > so the trace will show properly decoded info, > rather than just a bunch of hex codes. > > Signed-off-by: Dario Faggioli <dario.faggioli@xxxxxxxxxx> > --- > Cc: George Dunlap <george.dunlap@xxxxxxxxxxxxx> > Cc: Ian Jackson <ian.jackson@xxxxxxxxxxxxx> > Cc: Ian Campbell <ian.campbell@xxxxxxxxxx> > Cc: Wei Liu <wei.liu2@xxxxxxxxxx> > Cc: Olaf Hering <olaf@xxxxxxxxx> > --- > tools/xentrace/xenalyze.c | 156 > ++++++++++++++++++++++++++++++++++++--------- > 1 file changed, 126 insertions(+), 30 deletions(-) > > diff --git a/tools/xentrace/xenalyze.c b/tools/xentrace/xenalyze.c > index 6520790..be698aa 100644 > --- a/tools/xentrace/xenalyze.c > +++ b/tools/xentrace/xenalyze.c > @@ -1519,27 +1519,6 @@ struct pv_data { > }; > > /* Sched data */ > - > -enum { > - SCHED_DOM_ADD=1, > - SCHED_DOM_REM, > - SCHED_SLEEP, > - SCHED_WAKE, > - SCHED_YIELD, > - SCHED_BLOCK, > - SCHED_SHUTDOWN, > - SCHED_CTL, > - SCHED_ADJDOM, > - SCHED_SWITCH, > - SCHED_S_TIMER_FN, > - SCHED_T_TIMER_FN, > - SCHED_DOM_TIMER_FN, > - SCHED_SWITCH_INFPREV, > - SCHED_SWITCH_INFNEXT, > - SCHED_SHUTDOWN_CODE, > - SCHED_MAX > -}; > - > enum { > RUNSTATE_RUNNING=0, > RUNSTATE_RUNNABLE, > @@ -7431,6 +7410,17 @@ no_update: > return; > } > > +void dump_sched_switch(struct record_info *ri) > +{ > + struct { > + unsigned int prev_dom, prev_vcpu, next_dom, next_vcpu; > + } * r = (typeof(r))ri->d; > + > + printf(" %s sched_switch prev d%uv%u next d%uv%u\n", > + ri->dump_header, r->prev_dom, r->prev_vcpu, > + r->next_dom, r->next_vcpu); > +} > + > void sched_switch_process(struct pcpu_info *p) > { > struct vcpu_data *prev, *next; > @@ -7440,10 +7430,7 @@ void sched_switch_process(struct pcpu_info *p) > } * r = (typeof(r))ri->d; > > if(opt.dump_all) > - printf("%s sched_switch prev d%uv%u next d%uv%u\n", > - ri->dump_header, > - r->prev_dom, r->prev_vcpu, > - r->next_dom, r->next_vcpu); > + dump_sched_switch(ri); > > if(r->prev_vcpu > MAX_CPUS) > { > @@ -7559,6 +7546,14 @@ void sched_summary_domain(struct domain_data *d) > } > } > > +void dump_sched_vcpu_action(struct record_info *ri, const char *action) > +{ > + struct { > + unsigned int domid, vcpuid; > + } * r = (typeof(r))ri->d; > + > + printf(" %s %s d%uv%u\n", ri->dump_header, action, r->domid, r->vcpuid); > +} I know that 'sched_switch_process' used to have '} * r =' type but there are different pieces of code that do: "} *r =" (mem_pod_zero_reclaim_process) Could all of the patches use the '*r = ' style? There are more of them: [konrad@x230 xentrace]$ cat xenalyze.c | grep "} \*r =" | wc 32 128 942 [konrad@x230 xentrace]$ cat xenalyze.c | grep "} \* r =" | wc 3 17 92 Besides that: Reviewed-by: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx> on this and the other patches that touch xenalyze.c (that is with the modification to * r = .. > > void sched_process(struct pcpu_info *p) > { > @@ -7573,13 +7568,114 @@ void sched_process(struct pcpu_info *p) > default: > process_generic(&p->ri); > } > - } else { > - if(ri->evt.sub == 1) > - sched_runstate_process(p); > - else { > - UPDATE_VOLUME(p, sched_verbose, ri->size); > + return; > + } > + > + if(ri->evt.sub == 1) { > + /* TRC_SCHED_MIN */ > + sched_runstate_process(p); > + } else if (ri->evt.sub == 8) { > + /* TRC_SCHED_VERBOSE */ > + switch(ri->event) > + { > + case TRC_SCHED_DOM_ADD: > + if(opt.dump_all) { > + struct { > + unsigned int domid; > + } * r = (typeof(r))ri->d; > + > + printf(" %s domain create d%u\n", ri->dump_header, r->domid); > + } > + break; > + case TRC_SCHED_DOM_REM: > + if(opt.dump_all) { > + struct { > + unsigned int domid, vcpuid; > + } * r = (typeof(r))ri->d; > + > + printf(" %s domain destroy d%u\n", ri->dump_header, > r->domid); > + } > + break; > + case TRC_SCHED_SLEEP: > + if(opt.dump_all) > + dump_sched_vcpu_action(ri, "vcpu_sleep"); > + break; > + case TRC_SCHED_WAKE: > + if(opt.dump_all) > + dump_sched_vcpu_action(ri, "vcpu_wake"); > + break; > + case TRC_SCHED_YIELD: > + if(opt.dump_all) > + dump_sched_vcpu_action(ri, "vcpu_yield"); > + break; > + case TRC_SCHED_BLOCK: > + if(opt.dump_all) > + dump_sched_vcpu_action(ri, "vcpu_block"); > + break; > + case TRC_SCHED_SHUTDOWN: > + case TRC_SCHED_SHUTDOWN_CODE: > + if(opt.dump_all) { > + struct { > + unsigned int domid, vcpuid, reason; > + } * r = (typeof(r))ri->d; > + > + printf(" %s %s d%uv%u, reason = %u\n", ri->dump_header, > + ri->event == TRC_SCHED_SHUTDOWN ? "sched_shutdown" : > + "sched_shutdown_code", r->domid, r->vcpuid, > r->reason); > + } > + break; > + case TRC_SCHED_ADJDOM: > + if(opt.dump_all) { > + struct { > + unsigned int domid; > + } * r = (typeof(r))ri->d; > + > + printf(" %s sched_adjust d%u\n", ri->dump_header, r->domid); > + } > + break; > + case TRC_SCHED_SWITCH: > + dump_sched_switch(ri); > + break; > + case TRC_SCHED_SWITCH_INFPREV: > + if(opt.dump_all) { > + struct { > + unsigned int domid, runtime; > + } * r = (typeof(r))ri->d; > + > + printf(" %s sched_switch prev d%u, run for %u.%uus\n", > + ri->dump_header, r->domid, r->runtime / 1000, > + r->runtime % 1000); > + } > + break; > + case TRC_SCHED_SWITCH_INFNEXT: > + if(opt.dump_all) > + { > + struct { > + unsigned int domid, rsince; > + int slice; > + } * r = (typeof(r))ri->d; > + > + printf(" %s sched_switch next d%u", ri->dump_header, > r->domid); > + if ( r->rsince != 0 ) > + printf(", was runnable for %u.%uus, ", r->rsince / 1000, > + r->rsince % 1000); > + if ( r->slice > 0 ) > + printf("next slice %u.%uus\n", r->slice / 1000, > + r->slice % 1000); > + printf("\n"); > + } > + break; > + case TRC_SCHED_CTL: > + case TRC_SCHED_S_TIMER_FN: > + case TRC_SCHED_T_TIMER_FN: > + case TRC_SCHED_DOM_TIMER_FN: > + break; > + default: > process_generic(&p->ri); > } > + } else { > + UPDATE_VOLUME(p, sched_verbose, ri->size); > + process_generic(&p->ri); > } > } > > > > _______________________________________________ > Xen-devel mailing list > Xen-devel@xxxxxxxxxxxxx > http://lists.xen.org/xen-devel _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |