|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [xen master] x86/entry: Avoid register spilling in cr4_pv32_restore()
commit b24b540de613dad0f432bb606ea1eb935a8a89cf
Author: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
AuthorDate: Fri Feb 10 21:20:42 2023 +0000
Commit: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
CommitDate: Thu Jan 25 17:46:57 2024 +0000
x86/entry: Avoid register spilling in cr4_pv32_restore()
cr4_pv32_restore() needs two registers. Right now, it spills %rdx and
clobbers %rax.
However, %rcx is free to use at all callsites. Annotate CR4_PV32_RESTORE
with
our usual clobber comments, and swap %rdx for %rcx in the non-fatal paths
Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
Reviewed-by: Jan Beulich <jbeulich@xxxxxxxx>
---
xen/arch/x86/x86_64/compat/entry.S | 17 +++++++----------
xen/arch/x86/x86_64/entry.S | 8 ++++----
2 files changed, 11 insertions(+), 14 deletions(-)
diff --git a/xen/arch/x86/x86_64/compat/entry.S
b/xen/arch/x86/x86_64/compat/entry.S
index 49811a56e9..d4f0e48040 100644
--- a/xen/arch/x86/x86_64/compat/entry.S
+++ b/xen/arch/x86/x86_64/compat/entry.S
@@ -23,7 +23,7 @@ FUNC(entry_int82)
sti
- CR4_PV32_RESTORE
+ CR4_PV32_RESTORE /* Clob: ac */
GET_CURRENT(bx)
@@ -163,17 +163,15 @@ FUNC(compat_restore_all_guest)
_ASM_PRE_EXTABLE(.Lft0, handle_exception)
END(compat_restore_all_guest)
-/* This mustn't modify registers other than %rax. */
+/* Callers can cope with both %rax and %rcx being clobbered. */
FUNC(cr4_pv32_restore)
- push %rdx
- GET_CPUINFO_FIELD(cr4, dx)
- mov (%rdx), %rax
+ GET_CPUINFO_FIELD(cr4, cx)
+ mov (%rcx), %rax
test $XEN_CR4_PV32_BITS, %eax
jnz 0f
or cr4_pv32_mask(%rip), %rax
mov %rax, %cr4
- mov %rax, (%rdx)
- pop %rdx
+ mov %rax, (%rcx)
ret
0:
#ifndef NDEBUG
@@ -191,7 +189,6 @@ FUNC(cr4_pv32_restore)
BUG
1:
#endif
- pop %rdx
xor %eax, %eax
ret
END(cr4_pv32_restore)
@@ -227,7 +224,7 @@ UNLIKELY_END(compat_syscall_gpf)
END(compat_syscall)
FUNC(compat_sysenter)
- CR4_PV32_RESTORE
+ CR4_PV32_RESTORE /* Clob: ac */
movq VCPU_trap_ctxt(%rbx),%rcx
cmpb $X86_EXC_GP, UREGS_entry_vector(%rsp)
movzwl VCPU_sysenter_sel(%rbx),%eax
@@ -242,7 +239,7 @@ FUNC(compat_sysenter)
END(compat_sysenter)
FUNC(compat_int80_direct_trap)
- CR4_PV32_RESTORE
+ CR4_PV32_RESTORE /* Clob: ac */
call compat_create_bounce_frame
jmp compat_test_all_events
END(compat_int80_direct_trap)
diff --git a/xen/arch/x86/x86_64/entry.S b/xen/arch/x86/x86_64/entry.S
index 5dab3e4832..44f25def86 100644
--- a/xen/arch/x86/x86_64/entry.S
+++ b/xen/arch/x86/x86_64/entry.S
@@ -309,7 +309,7 @@ FUNC(cstar_enter)
.Lcstar_cr3_okay:
sti
- CR4_PV32_RESTORE
+ CR4_PV32_RESTORE /* Clob: ac */
movq STACK_CPUINFO_FIELD(current_vcpu)(%rbx), %rbx
@@ -712,7 +712,7 @@ FUNC(common_interrupt)
cmovnz %r12d, %ebx
.Lintr_cr3_okay:
- CR4_PV32_RESTORE
+ CR4_PV32_RESTORE /* Clob: ac */
movq %rsp,%rdi
callq do_IRQ
mov %r15, STACK_CPUINFO_FIELD(xen_cr3)(%r14)
@@ -762,7 +762,7 @@ handle_exception_saved:
jz .Lcr4_pv32_done
cmpb $0,DOMAIN_is_32bit_pv(%rax)
je .Lcr4_pv32_done
- call cr4_pv32_restore
+ call cr4_pv32_restore /* Clob: ac */
/*
* An NMI or #MC may occur between clearing CR4.SMEP / CR4.SMAP in
* compat_restore_all_guest and it actually returning to guest
@@ -1046,7 +1046,7 @@ FUNC(handle_ist_exception)
.List_cr3_okay:
#ifdef CONFIG_PV
- CR4_PV32_RESTORE
+ CR4_PV32_RESTORE /* Clob: ac */
testb $3,UREGS_cs(%rsp)
jz 1f
/*
--
generated by git-patchbot for /home/xen/git/xen.git#master
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |