[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [xen stable-4.18] x86/entry: Introduce EFRAME_* constants
commit fe1869a569bab56e44c35d1522ee064bab6286da Author: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> AuthorDate: Sat Jan 27 17:52:09 2024 +0000 Commit: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> CommitDate: Tue Mar 12 11:36:41 2024 +0000 x86/entry: Introduce EFRAME_* constants restore_all_guest() does a lot of manipulation of the stack after popping the GPRs, and uses raw %rsp displacements to do so. Also, almost all entrypaths use raw %rsp displacements prior to pushing GPRs. Provide better mnemonics, to aid readability and reduce the chance of errors when editing. No functional change. The resulting binary is identical. Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> Reviewed-by: Jan Beulich <jbeulich@xxxxxxxx> (cherry picked from commit 37541208f119a9c552c6c6c3246ea61be0d44035) --- xen/arch/x86/x86_64/asm-offsets.c | 17 +++++++++ xen/arch/x86/x86_64/compat/entry.S | 2 +- xen/arch/x86/x86_64/entry.S | 70 +++++++++++++++++++------------------- 3 files changed, 53 insertions(+), 36 deletions(-) diff --git a/xen/arch/x86/x86_64/asm-offsets.c b/xen/arch/x86/x86_64/asm-offsets.c index 57b73a4e62..2fc4d9130a 100644 --- a/xen/arch/x86/x86_64/asm-offsets.c +++ b/xen/arch/x86/x86_64/asm-offsets.c @@ -51,6 +51,23 @@ void __dummy__(void) OFFSET(UREGS_kernel_sizeof, struct cpu_user_regs, es); BLANK(); + /* + * EFRAME_* is for the entry/exit logic where %rsp is pointing at + * UREGS_error_code and GPRs are still/already guest values. + */ +#define OFFSET_EF(sym, mem) \ + DEFINE(sym, offsetof(struct cpu_user_regs, mem) - \ + offsetof(struct cpu_user_regs, error_code)) + + OFFSET_EF(EFRAME_entry_vector, entry_vector); + OFFSET_EF(EFRAME_rip, rip); + OFFSET_EF(EFRAME_cs, cs); + OFFSET_EF(EFRAME_eflags, eflags); + OFFSET_EF(EFRAME_rsp, rsp); + BLANK(); + +#undef OFFSET_EF + OFFSET(VCPU_processor, struct vcpu, processor); OFFSET(VCPU_domain, struct vcpu, domain); OFFSET(VCPU_vcpu_info, struct vcpu, vcpu_info_area.map); diff --git a/xen/arch/x86/x86_64/compat/entry.S b/xen/arch/x86/x86_64/compat/entry.S index fcc3a721f1..cb473f08ee 100644 --- a/xen/arch/x86/x86_64/compat/entry.S +++ b/xen/arch/x86/x86_64/compat/entry.S @@ -15,7 +15,7 @@ ENTRY(entry_int82) ENDBR64 ALTERNATIVE "", clac, X86_FEATURE_XEN_SMAP pushq $0 - movl $HYPERCALL_VECTOR, 4(%rsp) + movl $HYPERCALL_VECTOR, EFRAME_entry_vector(%rsp) SAVE_ALL compat=1 /* DPL1 gate, restricted to 32bit PV guests only. */ SPEC_CTRL_ENTRY_FROM_PV /* Req: %rsp=regs/cpuinfo, %rdx=0, Clob: acd */ diff --git a/xen/arch/x86/x86_64/entry.S b/xen/arch/x86/x86_64/entry.S index 9a7b129aa7..968da9d727 100644 --- a/xen/arch/x86/x86_64/entry.S +++ b/xen/arch/x86/x86_64/entry.S @@ -190,15 +190,15 @@ restore_all_guest: SPEC_CTRL_EXIT_TO_PV /* Req: a=spec_ctrl %rsp=regs/cpuinfo, Clob: cd */ RESTORE_ALL - testw $TRAP_syscall,4(%rsp) + testw $TRAP_syscall, EFRAME_entry_vector(%rsp) jz iret_exit_to_guest - movq 24(%rsp),%r11 # RFLAGS + mov EFRAME_eflags(%rsp), %r11 andq $~(X86_EFLAGS_IOPL | X86_EFLAGS_VM), %r11 orq $X86_EFLAGS_IF,%r11 /* Don't use SYSRET path if the return address is not canonical. */ - movq 8(%rsp),%rcx + mov EFRAME_rip(%rsp), %rcx sarq $47,%rcx incl %ecx cmpl $1,%ecx @@ -213,20 +213,20 @@ restore_all_guest: ALTERNATIVE "", rag_clrssbsy, X86_FEATURE_XEN_SHSTK #endif - movq 8(%rsp), %rcx # RIP - cmpw $FLAT_USER_CS32,16(%rsp)# CS - movq 32(%rsp),%rsp # RSP + mov EFRAME_rip(%rsp), %rcx + cmpw $FLAT_USER_CS32, EFRAME_cs(%rsp) + mov EFRAME_rsp(%rsp), %rsp je 1f sysretq 1: sysretl ALIGN .Lrestore_rcx_iret_exit_to_guest: - movq 8(%rsp), %rcx # RIP + mov EFRAME_rip(%rsp), %rcx /* No special register assumptions. */ iret_exit_to_guest: - andl $~(X86_EFLAGS_IOPL | X86_EFLAGS_VM), 24(%rsp) - orl $X86_EFLAGS_IF,24(%rsp) + andl $~(X86_EFLAGS_IOPL | X86_EFLAGS_VM), EFRAME_eflags(%rsp) + orl $X86_EFLAGS_IF, EFRAME_eflags(%rsp) addq $8,%rsp .Lft0: iretq _ASM_PRE_EXTABLE(.Lft0, handle_exception) @@ -257,7 +257,7 @@ ENTRY(lstar_enter) pushq $FLAT_KERNEL_CS64 pushq %rcx pushq $0 - movl $TRAP_syscall, 4(%rsp) + movl $TRAP_syscall, EFRAME_entry_vector(%rsp) SAVE_ALL SPEC_CTRL_ENTRY_FROM_PV /* Req: %rsp=regs/cpuinfo, %rdx=0, Clob: acd */ @@ -294,7 +294,7 @@ ENTRY(cstar_enter) pushq $FLAT_USER_CS32 pushq %rcx pushq $0 - movl $TRAP_syscall, 4(%rsp) + movl $TRAP_syscall, EFRAME_entry_vector(%rsp) SAVE_ALL SPEC_CTRL_ENTRY_FROM_PV /* Req: %rsp=regs/cpuinfo, %rdx=0, Clob: acd */ @@ -335,7 +335,7 @@ GLOBAL(sysenter_eflags_saved) pushq $3 /* ring 3 null cs */ pushq $0 /* null rip */ pushq $0 - movl $TRAP_syscall, 4(%rsp) + movl $TRAP_syscall, EFRAME_entry_vector(%rsp) SAVE_ALL SPEC_CTRL_ENTRY_FROM_PV /* Req: %rsp=regs/cpuinfo, %rdx=0, Clob: acd */ @@ -389,7 +389,7 @@ ENTRY(int80_direct_trap) ENDBR64 ALTERNATIVE "", clac, X86_FEATURE_XEN_SMAP pushq $0 - movl $0x80, 4(%rsp) + movl $0x80, EFRAME_entry_vector(%rsp) SAVE_ALL SPEC_CTRL_ENTRY_FROM_PV /* Req: %rsp=regs/cpuinfo, %rdx=0, Clob: acd */ @@ -649,7 +649,7 @@ ret_from_intr: .section .init.text, "ax", @progbits ENTRY(early_page_fault) ENDBR64 - movl $X86_EXC_PF, 4(%rsp) + movl $X86_EXC_PF, EFRAME_entry_vector(%rsp) SAVE_ALL movq %rsp, %rdi call do_early_page_fault @@ -716,7 +716,7 @@ ENTRY(common_interrupt) ENTRY(entry_PF) ENDBR64 - movl $X86_EXC_PF, 4(%rsp) + movl $X86_EXC_PF, EFRAME_entry_vector(%rsp) /* No special register assumptions. */ GLOBAL(handle_exception) ALTERNATIVE "", clac, X86_FEATURE_XEN_SMAP @@ -890,90 +890,90 @@ FATAL_exception_with_ints_disabled: ENTRY(entry_DE) ENDBR64 pushq $0 - movl $X86_EXC_DE, 4(%rsp) + movl $X86_EXC_DE, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_MF) ENDBR64 pushq $0 - movl $X86_EXC_MF, 4(%rsp) + movl $X86_EXC_MF, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_XM) ENDBR64 pushq $0 - movl $X86_EXC_XM, 4(%rsp) + movl $X86_EXC_XM, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_NM) ENDBR64 pushq $0 - movl $X86_EXC_NM, 4(%rsp) + movl $X86_EXC_NM, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_DB) ENDBR64 pushq $0 - movl $X86_EXC_DB, 4(%rsp) + movl $X86_EXC_DB, EFRAME_entry_vector(%rsp) jmp handle_ist_exception ENTRY(entry_BP) ENDBR64 pushq $0 - movl $X86_EXC_BP, 4(%rsp) + movl $X86_EXC_BP, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_OF) ENDBR64 pushq $0 - movl $X86_EXC_OF, 4(%rsp) + movl $X86_EXC_OF, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_BR) ENDBR64 pushq $0 - movl $X86_EXC_BR, 4(%rsp) + movl $X86_EXC_BR, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_UD) ENDBR64 pushq $0 - movl $X86_EXC_UD, 4(%rsp) + movl $X86_EXC_UD, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_TS) ENDBR64 - movl $X86_EXC_TS, 4(%rsp) + movl $X86_EXC_TS, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_NP) ENDBR64 - movl $X86_EXC_NP, 4(%rsp) + movl $X86_EXC_NP, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_SS) ENDBR64 - movl $X86_EXC_SS, 4(%rsp) + movl $X86_EXC_SS, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_GP) ENDBR64 - movl $X86_EXC_GP, 4(%rsp) + movl $X86_EXC_GP, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_AC) ENDBR64 - movl $X86_EXC_AC, 4(%rsp) + movl $X86_EXC_AC, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_CP) ENDBR64 - movl $X86_EXC_CP, 4(%rsp) + movl $X86_EXC_CP, EFRAME_entry_vector(%rsp) jmp handle_exception ENTRY(entry_DF) ENDBR64 - movl $X86_EXC_DF, 4(%rsp) + movl $X86_EXC_DF, EFRAME_entry_vector(%rsp) /* Set AC to reduce chance of further SMAP faults */ ALTERNATIVE "", stac, X86_FEATURE_XEN_SMAP SAVE_ALL @@ -998,7 +998,7 @@ ENTRY(entry_DF) ENTRY(entry_NMI) ENDBR64 pushq $0 - movl $X86_EXC_NMI, 4(%rsp) + movl $X86_EXC_NMI, EFRAME_entry_vector(%rsp) handle_ist_exception: ALTERNATIVE "", clac, X86_FEATURE_XEN_SMAP SAVE_ALL @@ -1130,7 +1130,7 @@ handle_ist_exception: ENTRY(entry_MC) ENDBR64 pushq $0 - movl $X86_EXC_MC, 4(%rsp) + movl $X86_EXC_MC, EFRAME_entry_vector(%rsp) jmp handle_ist_exception /* No op trap handler. Required for kexec crash path. */ @@ -1167,7 +1167,7 @@ autogen_stubs: /* Automatically generated stubs. */ 1: ENDBR64 pushq $0 - movb $vec,4(%rsp) + movb $vec, EFRAME_entry_vector(%rsp) jmp common_interrupt entrypoint 1b @@ -1181,7 +1181,7 @@ autogen_stubs: /* Automatically generated stubs. */ test $8,%spl /* 64bit exception frames are 16 byte aligned, but the word */ jz 2f /* size is 8 bytes. Check whether the processor gave us an */ pushq $0 /* error code, and insert an empty one if not. */ -2: movb $vec,4(%rsp) +2: movb $vec, EFRAME_entry_vector(%rsp) jmp handle_exception entrypoint 1b -- generated by git-patchbot for /home/xen/git/xen.git#stable-4.18
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |