[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [PATCH v2 2/2] types: replace remaining uses of s64
On Thu, Sep 12, 2024 at 1:20 PM Jan Beulich <jbeulich@xxxxxxxx> wrote: > Minor for the subject, there are also some removal of u64 and u8, not only s64 > ... and move the type itself to linux-compat.h. > > While doing so > - correct the type of union uu's uq field in lib/divmod.c, > - switch a few adjacent types as well, for (a little bit of) > consistency. > > Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx> > --- > v2: Split off ubsan.[ch] adjustments. Re-base. > > --- a/xen/arch/arm/arm64/cpufeature.c > +++ b/xen/arch/arm/arm64/cpufeature.c > @@ -484,8 +484,8 @@ static const struct arm64_ftr_bits ftr_r > ARM64_FTR_END, > }; > > -static u64 arm64_ftr_set_value(const struct arm64_ftr_bits *ftrp, s64 reg, > - s64 ftr_val) > +static uint64_t arm64_ftr_set_value(const struct arm64_ftr_bits *ftrp, > + int64_t reg, int64_t ftr_val) > { > u64 mask = arm64_ftr_mask(ftrp); > > @@ -494,10 +494,10 @@ static u64 arm64_ftr_set_value(const str > return reg; > } > > -static s64 arm64_ftr_safe_value(const struct arm64_ftr_bits *ftrp, s64 new, > - s64 cur) > +static int64_t arm64_ftr_safe_value(const struct arm64_ftr_bits *ftrp, > + int64_t new, int64_t cur) > { > - s64 ret = 0; > + int64_t ret = 0; > > switch (ftrp->type) { > case FTR_EXACT: > @@ -532,8 +532,8 @@ static void sanitize_reg(u64 *cur_reg, u > > for (;ftrp->width != 0;ftrp++) > { > - s64 cur_field = arm64_ftr_value(ftrp, *cur_reg); > - s64 new_field = arm64_ftr_value(ftrp, new_reg); > + int64_t cur_field = arm64_ftr_value(ftrp, *cur_reg); > + int64_t new_field = arm64_ftr_value(ftrp, new_reg); > > if (cur_field == new_field) > continue; > --- a/xen/arch/arm/arm64/livepatch.c > +++ b/xen/arch/arm/arm64/livepatch.c > @@ -119,7 +119,7 @@ static u64 do_reloc(enum aarch64_reloc_o > > static int reloc_data(enum aarch64_reloc_op op, void *place, u64 val, int > len) > { > - s64 sval = do_reloc(op, place, val); > + int64_t sval = do_reloc(op, place, val); > > switch ( len ) > { > @@ -136,7 +136,7 @@ static int reloc_data(enum aarch64_reloc > break; > > case 64: > - *(s64 *)place = sval; > + *(int64_t *)place = sval; > break; > > default: > @@ -155,9 +155,9 @@ enum aarch64_insn_movw_imm_type { > static int reloc_insn_movw(enum aarch64_reloc_op op, void *dest, u64 val, > int lsb, enum aarch64_insn_movw_imm_type imm_type) > { > - u64 imm; > - s64 sval; > - u32 insn = *(u32 *)dest; > + uint64_t imm; > + int64_t sval; > + uint32_t insn = *(uint32_t *)dest; > > sval = do_reloc(op, dest, val); > imm = sval >> lsb; > @@ -200,9 +200,9 @@ static int reloc_insn_movw(enum aarch64_ > static int reloc_insn_imm(enum aarch64_reloc_op op, void *dest, u64 val, > int lsb, int len, enum aarch64_insn_imm_type > imm_type) > { > - u64 imm, imm_mask; > - s64 sval; > - u32 insn = *(u32 *)dest; > + uint64_t imm, imm_mask; > + int64_t sval; > + uint32_t insn = *(uint32_t *)dest; > > /* Calculate the relocation value. */ > sval = do_reloc(op, dest, val); > @@ -220,7 +220,7 @@ static int reloc_insn_imm(enum aarch64_r > * Extract the upper value bits (including the sign bit) and > * shift them to bit 0. > */ > - sval = (s64)(sval & ~(imm_mask >> 1)) >> (len - 1); > + sval = (int64_t)(sval & ~(imm_mask >> 1)) >> (len - 1); > > /* > * Overflow has occurred if the upper bits are not all equal to > --- a/xen/arch/arm/include/asm/arm64/cpufeature.h > +++ b/xen/arch/arm/include/asm/arm64/cpufeature.h > @@ -39,15 +39,15 @@ struct arm64_ftr_bits { > bool visible; > bool strict; /* CPU Sanity check: strict matching required > ? */ > enum ftr_type type; > - u8 shift; > - u8 width; > - s64 safe_val; /* safe value for FTR_EXACT features */ > + uint8_t shift; > + uint8_t width; > + int64_t safe_val; /* safe value for FTR_EXACT features */ > }; > > static inline int attr_const > cpuid_feature_extract_signed_field_width(u64 features, int field, int width) > { > - return (s64)(features << (64 - width - field)) >> (64 - width); > + return (int64_t)(features << (64 - width - field)) >> (64 - width); > } > > static inline int attr_const > @@ -87,9 +87,12 @@ cpuid_feature_extract_field(u64 features > return cpuid_feature_extract_field_width(features, field, 4, sign); > } > > -static inline s64 arm64_ftr_value(const struct arm64_ftr_bits *ftrp, u64 val) > +static inline int64_t arm64_ftr_value(const struct arm64_ftr_bits *ftrp, > + uint64_t val) > { > - return (s64)cpuid_feature_extract_field_width(val, ftrp->shift, > ftrp->width, ftrp->sign); > + return (int64_t)cpuid_feature_extract_field_width(val, ftrp->shift, > + ftrp->width, > + ftrp->sign); > } > > #endif /* _ASM_ARM_ARM64_CPUFEATURES_H */ > --- a/xen/arch/x86/hvm/vlapic.c > +++ b/xen/arch/x86/hvm/vlapic.c > @@ -1208,7 +1208,7 @@ void vlapic_tdt_msr_set(struct vlapic *v > if ( value > guest_tsc ) > { > uint64_t delta = gtsc_to_gtime(v->domain, value - guest_tsc); > - delta = max_t(s64, delta, 0); > + delta = max_t(int64_t, delta, 0); > > HVM_DBG_LOG(DBG_LEVEL_VLAPIC_TIMER, "delta[0x%016"PRIx64"]", delta); > > --- a/xen/arch/x86/include/asm/hvm/vcpu.h > +++ b/xen/arch/x86/include/asm/hvm/vcpu.h > @@ -134,8 +134,8 @@ struct hvm_vcpu { > unsigned long hw_cr[5]; > > struct vlapic vlapic; > - s64 cache_tsc_offset; > - u64 guest_time; > + int64_t cache_tsc_offset; > + uint64_t guest_time; > > /* Lock and list for virtual platform timers. */ > spinlock_t tm_lock; > --- a/xen/arch/x86/time.c > +++ b/xen/arch/x86/time.c > @@ -66,10 +66,10 @@ struct cpu_time { > struct platform_timesource { > const char *id; > const char *name; > - u64 frequency; > + uint64_t frequency; > /* Post-init this hook may only be invoked via the read_counter() > wrapper! */ > - u64 (*read_counter)(void); > - s64 (*init)(struct platform_timesource *); > + uint64_t (*read_counter)(void); > + int64_t (*init)(struct platform_timesource *); > void (*resume)(struct platform_timesource *); > int counter_bits; > }; > @@ -368,7 +368,7 @@ static u64 cf_check read_pit_count(void) > return count32; > } > > -static s64 __init cf_check init_pit(struct platform_timesource *pts) > +static int64_t __init cf_check init_pit(struct platform_timesource *pts) > { > u8 portb = inb(0x61); > u64 start, end; > @@ -610,7 +610,7 @@ static u64 cf_check read_pmtimer_count(v > return inl(pmtmr_ioport); > } > > -static s64 __init cf_check init_pmtimer(struct platform_timesource *pts) > +static int64_t __init cf_check init_pmtimer(struct platform_timesource *pts) > { > if ( !pmtmr_ioport || (pmtmr_width != 24 && pmtmr_width != 32) ) > return 0; > @@ -655,7 +655,7 @@ static unsigned int __initdata tsc_flags > * Called in verify_tsc_reliability() under reliable TSC conditions > * thus reusing all the checks already performed there. > */ > -static s64 __init cf_check init_tsc(struct platform_timesource *pts) > +static int64_t __init cf_check init_tsc(struct platform_timesource *pts) > { > u64 ret = pts->frequency; > > @@ -1034,9 +1034,9 @@ static void __init reset_platform_timer( > spin_unlock_irq(&platform_timer_lock); > } > > -static s64 __init try_platform_timer(struct platform_timesource *pts) > +static int64_t __init try_platform_timer(struct platform_timesource *pts) > { > - s64 rc = pts->init(pts); > + int64_t rc = pts->init(pts); > > if ( rc <= 0 ) > return rc; > @@ -1070,7 +1070,7 @@ static u64 __init init_platform_timer(vo > > struct platform_timesource *pts = NULL; > unsigned int i; > - s64 rc = -1; > + int64_t rc = -1; > > /* clocksource=tsc is initialized via __initcalls (when CPUs are up). */ > if ( (opt_clocksource[0] != '\0') && strcmp(opt_clocksource, "tsc") ) > @@ -1923,7 +1923,7 @@ static void cf_check local_time_calibrat > * Weirdness can happen if we lose sync with the platform timer. > * We could be smarter here: resync platform timer with local timer? > */ > - if ( ((s64)stime_elapsed64 < (EPOCH / 2)) ) > + if ( ((int64_t)stime_elapsed64 < (EPOCH / 2)) ) > goto out; > > /* > @@ -2398,7 +2398,7 @@ static void __init tsc_check_writability > > write_tsc(tsc | (1ULL << 32)); > tmp = rdtsc(); > - if ( ABS((s64)tmp - (s64)tmp2) < (1LL << 31) ) > + if ( ABS((int64_t)tmp - (int64_t)tmp2) < (1LL << 31) ) > what = "only partially"; > } > else > --- a/xen/drivers/acpi/apei/erst.c > +++ b/xen/drivers/acpi/apei/erst.c > @@ -105,7 +105,7 @@ static inline int erst_errno(int command > > static int erst_timedout(u64 *t, u64 spin_unit) > { > - if ((s64)*t < spin_unit) { > + if ((int64_t)*t < spin_unit) { > printk(XENLOG_WARNING "Firmware does not respond in time\n"); > return 1; > } > --- a/xen/include/acpi/actypes.h > +++ b/xen/include/acpi/actypes.h > @@ -147,8 +147,8 @@ typedef int INT32; > > /*! [End] no source code translation !*/ > > -typedef u64 acpi_native_uint; > -typedef s64 acpi_native_int; > +typedef uint64_t acpi_native_uint; > +typedef int64_t acpi_native_int; > > typedef u64 acpi_io_address; > typedef u64 acpi_physical_address; > --- a/xen/include/xen/linux-compat.h > +++ b/xen/include/xen/linux-compat.h > @@ -15,7 +15,7 @@ typedef int8_t s8, __s8; > typedef uint8_t __u8; > typedef int16_t s16, __s16; > typedef int32_t s32, __s32; > -typedef int64_t __s64; > +typedef int64_t s64, __s64; > > typedef paddr_t phys_addr_t; > > --- a/xen/include/xen/time.h > +++ b/xen/include/xen/time.h > @@ -29,7 +29,7 @@ struct vcpu; > * of real time into system time > */ > > -typedef s64 s_time_t; > +typedef int64_t s_time_t; > #define PRI_stime PRId64 > > s_time_t get_s_time_fixed(u64 at_tick); > --- a/xen/include/xen/types.h > +++ b/xen/include/xen/types.h > @@ -8,7 +8,6 @@ > typedef uint8_t u8; > typedef uint16_t u16, __u16; > typedef uint32_t u32, __u32; > -typedef int64_t s64; > typedef uint64_t u64, __u64; > > #include <asm/types.h> > --- a/xen/lib/divmod.c > +++ b/xen/lib/divmod.c > @@ -46,8 +46,8 @@ > * one or more of the following formats. > */ > union uu { > - s64 q; /* as a (signed) quad */ > - s64 uq; /* as an unsigned quad */ > + int64_t q; /* as a (signed) quad */ > + uint64_t uq; /* as an unsigned quad */ > long sl[2]; /* as two signed longs */ > unsigned long ul[2]; /* as two unsigned longs */ > }; > @@ -72,7 +72,7 @@ union uu { > * and assembly. > */ > #define CHAR_BIT 8 /* number of bits in a char */ > -#define QUAD_BITS (sizeof(s64) * CHAR_BIT) > +#define QUAD_BITS (sizeof(int64_t) * CHAR_BIT) > #define LONG_BITS (sizeof(long) * CHAR_BIT) > #define HALF_BITS (sizeof(long) * CHAR_BIT / 2) > > @@ -324,7 +324,7 @@ u64 __qdivrem(u64 uq, u64 vq, u64 *arq) > * Divide two signed quads. > * Truncates towards zero, as required by C99. > */ > -s64 __divdi3(s64 a, s64 b) > +int64_t __divdi3(int64_t a, int64_t b) > { > u64 ua, ub, uq; > int neg = (a < 0) ^ (b < 0); > @@ -361,7 +361,7 @@ u64 __umoddi3(u64 a, u64 b) > * 11 % -5 = 1 > * -11 % -5 = -1 > */ > -s64 __moddi3(s64 a, s64 b) > +int64_t __moddi3(int64_t a, int64_t b) > { > u64 ua, ub, urem; > int neg = (a < 0); > @@ -374,7 +374,7 @@ s64 __moddi3(s64 a, s64 b) > /* > * Quotient and remainder of unsigned long long division > */ > -s64 __ldivmod_helper(s64 a, s64 b, s64 *r) > +int64_t __ldivmod_helper(int64_t a, int64_t b, int64_t *r) > { > u64 ua, ub, rem, quot; > Frediano
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |