[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-devel] [PATCH] x86: hide assembly labels



Prevent the fault recovery labels to clutter the symbol table and the
disassembly.

Signed-off-by: Jan Beulich <jbeulich@xxxxxxxxxx>

Index: 2007-02-07/xen/arch/x86/x86_32/entry.S
===================================================================
--- 2007-02-07.orig/xen/arch/x86/x86_32/entry.S 2006-12-11 15:09:57.000000000 
+0100
+++ 2007-02-07/xen/arch/x86/x86_32/entry.S      2007-02-07 16:30:09.000000000 
+0100
@@ -84,10 +84,10 @@ restore_all_guest:
         jmp   restore_all_vm86
 1:
 #endif
-FLT1:   mov  UREGS_ds(%esp),%ds
-FLT2:   mov  UREGS_es(%esp),%es
-FLT3:   mov  UREGS_fs(%esp),%fs
-FLT4:   mov  UREGS_gs(%esp),%gs
+.LFT1:  mov  UREGS_ds(%esp),%ds
+.LFT2:  mov  UREGS_es(%esp),%es
+.LFT3:  mov  UREGS_fs(%esp),%fs
+.LFT4:  mov  UREGS_gs(%esp),%gs
 restore_all_vm86:
         popl %ebx
         popl %ecx
@@ -97,9 +97,9 @@ restore_all_vm86:
         popl %ebp
         popl %eax
         addl $4,%esp
-FLT5:   iret
+.LFT5:  iret
 .section .fixup,"ax"
-FIX5:   subl  $28,%esp
+.LFX5:  subl  $28,%esp
         pushl 28(%esp)                 # error_code/entry_vector
         movl  %eax,UREGS_eax+4(%esp)
         movl  %ebp,UREGS_ebp+4(%esp)
@@ -108,7 +108,7 @@ FIX5:   subl  $28,%esp
         movl  %edx,UREGS_edx+4(%esp)
         movl  %ecx,UREGS_ecx+4(%esp)
         movl  %ebx,UREGS_ebx+4(%esp)
-FIX1:   SET_XEN_SEGMENTS(a)
+.LFX1:  SET_XEN_SEGMENTS(a)
         movl  %eax,%fs
         movl  %eax,%gs
         sti
@@ -116,11 +116,11 @@ FIX1:   SET_XEN_SEGMENTS(a)
         pushfl                         # EFLAGS
         movl  $__HYPERVISOR_CS,%eax
         pushl %eax                     # CS
-        movl  $DBLFLT1,%eax
+        movl  $.LDF1,%eax
         pushl %eax                     # EIP
         pushl %esi                     # error_code/entry_vector
         jmp   handle_exception
-DBLFLT1:GET_CURRENT(%ebx)
+.LDF1:  GET_CURRENT(%ebx)
         jmp   test_all_events
 failsafe_callback:
         GET_CURRENT(%ebx)
@@ -142,14 +142,14 @@ failsafe_callback:
         jmp   test_all_events
 .previous
 .section __pre_ex_table,"a"
-        .long FLT1,FIX1
-        .long FLT2,FIX1
-        .long FLT3,FIX1
-        .long FLT4,FIX1
-        .long FLT5,FIX5
+        .long .LFT1,.LFX1
+        .long .LFT2,.LFX1
+        .long .LFT3,.LFX1
+        .long .LFT4,.LFX1
+        .long .LFT5,.LFX5
 .previous
 .section __ex_table,"a"
-        .long DBLFLT1,failsafe_callback
+        .long .LDF1,failsafe_callback
 .previous
 
         ALIGN
@@ -288,32 +288,33 @@ create_bounce_frame:
         testl $(2|X86_EFLAGS_VM),%ecx
         jz   ring1 /* jump if returning to an existing ring-1 activation */
         movl VCPU_kernel_sp(%ebx),%esi
-FLT6:   mov  VCPU_kernel_ss(%ebx),%gs
+.LFT6:  mov  VCPU_kernel_ss(%ebx),%gs
         testl $X86_EFLAGS_VM,UREGS_eflags+4(%esp)
-        jz   nvm86_1
+        jz   .Lnvm86_1
         subl $16,%esi       /* push ES/DS/FS/GS (VM86 stack frame) */
         movl UREGS_es+4(%esp),%eax
-FLT7:   movl %eax,%gs:(%esi)
+.LFT7:  movl %eax,%gs:(%esi)
         movl UREGS_ds+4(%esp),%eax
-FLT8:   movl %eax,%gs:4(%esi)
+.LFT8:  movl %eax,%gs:4(%esi)
         movl UREGS_fs+4(%esp),%eax
-FLT9:   movl %eax,%gs:8(%esi)
+.LFT9:  movl %eax,%gs:8(%esi)
         movl UREGS_gs+4(%esp),%eax
-FLT10:  movl %eax,%gs:12(%esi)
-nvm86_1:subl $8,%esi        /* push SS/ESP (inter-priv iret) */
+.LFT10: movl %eax,%gs:12(%esi)
+.Lnvm86_1:
+        subl $8,%esi        /* push SS/ESP (inter-priv iret) */
         movl UREGS_esp+4(%esp),%eax
-FLT11:  movl %eax,%gs:(%esi) 
+.LFT11: movl %eax,%gs:(%esi)
         movl UREGS_ss+4(%esp),%eax
-FLT12:  movl %eax,%gs:4(%esi) 
+.LFT12: movl %eax,%gs:4(%esi)
         jmp 1f
 ring1:  /* obtain ss/esp from oldss/oldesp -- a ring-1 activation exists */
         movl UREGS_esp+4(%esp),%esi
-FLT13:  mov  UREGS_ss+4(%esp),%gs 
+.LFT13: mov  UREGS_ss+4(%esp),%gs
 1:      /* Construct a stack frame: EFLAGS, CS/EIP */
         movb TRAPBOUNCE_flags(%edx),%cl
         subl $12,%esi
         movl UREGS_eip+4(%esp),%eax
-FLT14:  movl %eax,%gs:(%esi) 
+.LFT14: movl %eax,%gs:(%esi)
         movl VCPU_vcpu_info(%ebx),%eax
         pushl VCPUINFO_upcall_mask(%eax)
         testb $TBF_INTERRUPT,%cl
@@ -324,49 +325,51 @@ FLT14:  movl %eax,%gs:(%esi) 
         movw UREGS_cs+4(%esp),%ax        # Bits  0-15: CS
 #ifdef CONFIG_X86_SUPERVISOR_MODE_KERNEL
         testw $2,%ax
-        jnz  FLT15
+        jnz  .LFT15
         and  $~3,%ax                     # RPL 1 -> RPL 0
 #endif
-FLT15:  movl %eax,%gs:4(%esi) 
+.LFT15: movl %eax,%gs:4(%esi)
         test $0x00FF0000,%eax            # Bits 16-23: saved_upcall_mask
         setz %ch                         # %ch == !saved_upcall_mask
         movl UREGS_eflags+4(%esp),%eax
         andl $~X86_EFLAGS_IF,%eax
         shlb $1,%ch                      # Bit 9 (EFLAGS.IF)
         orb  %ch,%ah                     # Fold EFLAGS.IF into %eax
-FLT16:  movl %eax,%gs:8(%esi)
+.LFT16: movl %eax,%gs:8(%esi)
         test $TBF_EXCEPTION_ERRCODE,%cl
         jz   1f
         subl $4,%esi                    # push error_code onto guest frame
         movl TRAPBOUNCE_error_code(%edx),%eax
-FLT17:  movl %eax,%gs:(%esi)
+.LFT17: movl %eax,%gs:(%esi)
 1:      testb $TBF_FAILSAFE,%cl
         jz   2f
         subl $16,%esi                # add DS/ES/FS/GS to failsafe stack frame
         testl $X86_EFLAGS_VM,UREGS_eflags+4(%esp)
-        jz   nvm86_2
+        jz   .Lnvm86_2
         xorl %eax,%eax               # VM86: we write zero selector values
-FLT18:  movl %eax,%gs:(%esi) 
-FLT19:  movl %eax,%gs:4(%esi)
-FLT20:  movl %eax,%gs:8(%esi) 
-FLT21:  movl %eax,%gs:12(%esi)
+.LFT18: movl %eax,%gs:(%esi)
+.LFT19: movl %eax,%gs:4(%esi)
+.LFT20: movl %eax,%gs:8(%esi)
+.LFT21: movl %eax,%gs:12(%esi)
         jmp  2f
-nvm86_2:movl UREGS_ds+4(%esp),%eax   # non-VM86: write real selector values
-FLT22:  movl %eax,%gs:(%esi) 
+.Lnvm86_2:
+        movl UREGS_ds+4(%esp),%eax   # non-VM86: write real selector values
+.LFT22: movl %eax,%gs:(%esi)
         movl UREGS_es+4(%esp),%eax
-FLT23:  movl %eax,%gs:4(%esi)
+.LFT23: movl %eax,%gs:4(%esi)
         movl UREGS_fs+4(%esp),%eax
-FLT24:  movl %eax,%gs:8(%esi) 
+.LFT24: movl %eax,%gs:8(%esi)
         movl UREGS_gs+4(%esp),%eax
-FLT25:  movl %eax,%gs:12(%esi)
+.LFT25: movl %eax,%gs:12(%esi)
 2:      testl $X86_EFLAGS_VM,UREGS_eflags+4(%esp)
-        jz   nvm86_3
+        jz   .Lnvm86_3
         xorl %eax,%eax      /* zero DS-GS, just as a real CPU would */
         movl %eax,UREGS_ds+4(%esp)
         movl %eax,UREGS_es+4(%esp)
         movl %eax,UREGS_fs+4(%esp)
         movl %eax,UREGS_gs+4(%esp)
-nvm86_3:/* Rewrite our stack frame and return to ring 1. */
+.Lnvm86_3:
+        /* Rewrite our stack frame and return to ring 1. */
         /* IA32 Ref. Vol. 3: TF, VM, RF and NT flags are cleared on trap. */
         andl $~(X86_EFLAGS_VM|X86_EFLAGS_RF|\
                 X86_EFLAGS_NT|X86_EFLAGS_TF),UREGS_eflags+4(%esp)
@@ -382,16 +385,16 @@ nvm86_3:/* Rewrite our stack frame and r
         movb $0,TRAPBOUNCE_flags(%edx)
         ret
 .section __ex_table,"a"
-        .long  FLT6,domain_crash_synchronous ,  FLT7,domain_crash_synchronous
-        .long  FLT8,domain_crash_synchronous ,  FLT9,domain_crash_synchronous
-        .long FLT10,domain_crash_synchronous , FLT11,domain_crash_synchronous
-        .long FLT12,domain_crash_synchronous , FLT13,domain_crash_synchronous
-        .long FLT14,domain_crash_synchronous , FLT15,domain_crash_synchronous
-        .long FLT16,domain_crash_synchronous , FLT17,domain_crash_synchronous
-        .long FLT18,domain_crash_synchronous , FLT19,domain_crash_synchronous
-        .long FLT20,domain_crash_synchronous , FLT21,domain_crash_synchronous
-        .long FLT22,domain_crash_synchronous , FLT23,domain_crash_synchronous
-        .long FLT24,domain_crash_synchronous , FLT25,domain_crash_synchronous
+        .long  .LFT6,domain_crash_synchronous ,  .LFT7,domain_crash_synchronous
+        .long  .LFT8,domain_crash_synchronous ,  .LFT9,domain_crash_synchronous
+        .long .LFT10,domain_crash_synchronous , .LFT11,domain_crash_synchronous
+        .long .LFT12,domain_crash_synchronous , .LFT13,domain_crash_synchronous
+        .long .LFT14,domain_crash_synchronous , .LFT15,domain_crash_synchronous
+        .long .LFT16,domain_crash_synchronous , .LFT17,domain_crash_synchronous
+        .long .LFT18,domain_crash_synchronous , .LFT19,domain_crash_synchronous
+        .long .LFT20,domain_crash_synchronous , .LFT21,domain_crash_synchronous
+        .long .LFT22,domain_crash_synchronous , .LFT23,domain_crash_synchronous
+        .long .LFT24,domain_crash_synchronous , .LFT25,domain_crash_synchronous
 .previous
 
 domain_crash_synchronous_string:
Index: 2007-02-07/xen/arch/x86/x86_64/compat/entry.S
===================================================================
--- 2007-02-07.orig/xen/arch/x86/x86_64/compat/entry.S  2007-02-01 
17:41:08.000000000 +0100
+++ 2007-02-07/xen/arch/x86/x86_64/compat/entry.S       2007-02-07 
16:28:28.000000000 +0100
@@ -123,10 +123,10 @@ compat_bad_hypercall:
 compat_restore_all_guest:
         RESTORE_ALL
         addq  $8,%rsp
-CFLT0:  iretq
+.Lft0:  iretq
 
 .section .fixup,"ax"
-CFIX0:  popq  -15*8-8(%rsp)            # error_code/entry_vector
+.Lfx0:  popq  -15*8-8(%rsp)            # error_code/entry_vector
         SAVE_ALL                       # 15*8 bytes pushed
         movq  -8(%rsp),%rsi            # error_code/entry_vector
         sti                            # after stack abuse (-1024(%rsp))
@@ -135,11 +135,11 @@ CFIX0:  popq  -15*8-8(%rsp)            #
         pushq %rax                     # RSP
         pushfq                         # RFLAGS
         pushq $__HYPERVISOR_CS         # CS
-        leaq  CDBLFLT0(%rip),%rax
+        leaq  .Ldf0(%rip),%rax
         pushq %rax                     # RIP
         pushq %rsi                     # error_code/entry_vector
         jmp   handle_exception
-CDBLFLT0:GET_CURRENT(%rbx)
+.Ldf0:  GET_CURRENT(%rbx)
         jmp   compat_test_all_events
 compat_failsafe_callback:
         GET_CURRENT(%rbx)
@@ -157,10 +157,10 @@ compat_failsafe_callback:
         jmp   compat_test_all_events
 .previous
 .section __pre_ex_table,"a"
-       .quad CFLT0,CFIX0
+       .quad .Lft0,.Lfx0
 .previous
 .section __ex_table,"a"
-        .quad CDBLFLT0,compat_failsafe_callback
+        .quad .Ldf0,compat_failsafe_callback
 .previous
 
 /* %rdx: trap_bounce, %rbx: struct vcpu */
@@ -180,16 +180,16 @@ compat_create_bounce_frame:
         jz    1f
         /* Push new frame at registered guest-OS stack base. */
         movl  VCPU_kernel_sp(%rbx),%esi
-CFLT1:  mov   VCPU_kernel_ss(%rbx),%fs
+.Lft1:  mov   VCPU_kernel_ss(%rbx),%fs
         subl  $2*4,%esi
         movl  UREGS_rsp+8(%rsp),%eax
-CFLT2:  movl  %eax,%fs:(%rsi)
+.Lft2:  movl  %eax,%fs:(%rsi)
         movl  UREGS_ss+8(%rsp),%eax
-CFLT3:  movl  %eax,%fs:4(%rsi)
+.Lft3:  movl  %eax,%fs:4(%rsi)
         jmp   2f
 1:      /* In kernel context already: push new frame at existing %rsp. */
         movl  UREGS_rsp+8(%rsp),%esi
-CFLT4:  mov   UREGS_ss+8(%rsp),%fs
+.Lft4:  mov   UREGS_ss+8(%rsp),%fs
 2:
         movb  TRAPBOUNCE_flags(%rdx),%cl
         subl  $3*4,%esi
@@ -201,7 +201,7 @@ CFLT4:  mov   UREGS_ss+8(%rsp),%fs
         popq  %rax
         shll  $16,%eax                  # Bits 16-23: saved_upcall_mask
         movw  UREGS_cs+8(%rsp),%ax      # Bits  0-15: CS
-CFLT5:  movl  %eax,%fs:4(%rsi)          # CS / saved_upcall_mask
+.Lft5:  movl  %eax,%fs:4(%rsi)          # CS / saved_upcall_mask
         shrl  $16,%eax
         testb %al,%al                   # Bits 0-7: saved_upcall_mask
         setz  %ch                       # %ch == !saved_upcall_mask
@@ -209,25 +209,25 @@ CFLT5:  movl  %eax,%fs:4(%rsi)          
         andl  $~X86_EFLAGS_IF,%eax
         shlb  $1,%ch                    # Bit 9 (EFLAGS.IF)
         orb   %ch,%ah                   # Fold EFLAGS.IF into %eax
-CFLT6:  movl  %eax,%fs:2*4(%rsi)        # EFLAGS
+.Lft6:  movl  %eax,%fs:2*4(%rsi)        # EFLAGS
         movl  UREGS_rip+8(%rsp),%eax
-CFLT7:  movl  %eax,%fs:(%rsi)           # EIP
+.Lft7:  movl  %eax,%fs:(%rsi)           # EIP
         testb $TBF_EXCEPTION_ERRCODE,%cl
         jz    1f
         subl  $4,%esi
         movl  TRAPBOUNCE_error_code(%rdx),%eax
-CFLT8:  movl  %eax,%fs:(%rsi)           # ERROR CODE
+.Lft8:  movl  %eax,%fs:(%rsi)           # ERROR CODE
 1:
         testb $TBF_FAILSAFE,%cl
         jz    2f
         subl  $4*4,%esi
         movl  %gs,%eax
-CFLT9:  movl  %eax,%fs:3*4(%rsi)        # GS
-CFLT10: movl  %edi,%fs:2*4(%rsi)        # FS
+.Lft9:  movl  %eax,%fs:3*4(%rsi)        # GS
+.Lft10: movl  %edi,%fs:2*4(%rsi)        # FS
         movl  %es,%eax
-CFLT11: movl  %eax,%fs:1*4(%rsi)        # ES
+.Lft11: movl  %eax,%fs:1*4(%rsi)        # ES
         movl  %ds,%eax
-CFLT12: movl  %eax,%fs:0*4(%rsi)        # DS
+.Lft12: movl  %eax,%fs:0*4(%rsi)        # DS
 2:
         /* Rewrite our stack frame and return to guest-OS mode. */
         /* IA32 Ref. Vol. 3: TF, VM, RF and NT flags are cleared on trap. */
@@ -236,7 +236,7 @@ CFLT12: movl  %eax,%fs:0*4(%rsi)        
                  X86_EFLAGS_NT|X86_EFLAGS_TF),UREGS_eflags+8(%rsp)
         mov   %fs,UREGS_ss+8(%rsp)
         movl  %esi,UREGS_rsp+8(%rsp)
-CFLT13: mov   %edi,%fs
+.Lft13: mov   %edi,%fs
         movzwl TRAPBOUNCE_cs(%rdx),%eax
         /* Null selectors (0-3) are not allowed. */
         testl $~3,%eax
@@ -247,18 +247,18 @@ CFLT13: mov   %edi,%fs
         movb  $0,TRAPBOUNCE_flags(%rdx)
         ret
 .section .fixup,"ax"
-CFIX13:
+.Lfx13:
         xorl  %edi,%edi
-        jmp   CFLT13
+        jmp   .Lft13
 .previous
 .section __ex_table,"a"
-        .quad  CFLT1,domain_crash_synchronous  ,  CFLT2,compat_crash_page_fault
-        .quad  CFLT3,compat_crash_page_fault_4 ,  
CFLT4,domain_crash_synchronous
-        .quad  CFLT5,compat_crash_page_fault_4 ,  
CFLT6,compat_crash_page_fault_8
-        .quad  CFLT7,compat_crash_page_fault   ,  CFLT8,compat_crash_page_fault
-        .quad  CFLT9,compat_crash_page_fault_12, 
CFLT10,compat_crash_page_fault_8
-        .quad CFLT11,compat_crash_page_fault_4 , CFLT12,compat_crash_page_fault
-        .quad CFLT13,CFIX13
+        .quad  .Lft1,domain_crash_synchronous  ,  .Lft2,compat_crash_page_fault
+        .quad  .Lft3,compat_crash_page_fault_4 ,  
.Lft4,domain_crash_synchronous
+        .quad  .Lft5,compat_crash_page_fault_4 ,  
.Lft6,compat_crash_page_fault_8
+        .quad  .Lft7,compat_crash_page_fault   ,  .Lft8,compat_crash_page_fault
+        .quad  .Lft9,compat_crash_page_fault_12, 
.Lft10,compat_crash_page_fault_8
+        .quad .Lft11,compat_crash_page_fault_4 , .Lft12,compat_crash_page_fault
+        .quad .Lft13,.Lfx13
 .previous
 
 compat_crash_page_fault_12:
@@ -268,17 +268,17 @@ compat_crash_page_fault_8:
 compat_crash_page_fault_4:
         addl  $4,%esi
 compat_crash_page_fault:
-CFLT14: mov   %edi,%fs
+.Lft14: mov   %edi,%fs
         movl  %esi,%edi
         call  show_page_walk
         jmp   domain_crash_synchronous
 .section .fixup,"ax"
-CFIX14:
+.Lfx14:
         xorl  %edi,%edi
-        jmp   CFLT14
+        jmp   .Lft14
 .previous
 .section __ex_table,"a"
-        .quad CFLT14,CFIX14
+        .quad .Lft14,.Lfx14
 .previous
 
 .section .rodata, "a", @progbits
Index: 2007-02-07/xen/arch/x86/x86_64/entry.S
===================================================================
--- 2007-02-07.orig/xen/arch/x86/x86_64/entry.S 2007-01-08 14:15:31.000000000 
+0100
+++ 2007-02-07/xen/arch/x86/x86_64/entry.S      2007-02-07 16:31:20.000000000 
+0100
@@ -56,10 +56,10 @@ restore_all_guest:
 /* No special register assumptions. */
 iret_exit_to_guest:
         addq  $8,%rsp
-FLT1:   iretq
+.LFT1:  iretq
 
 .section .fixup,"ax"
-FIX1:   popq  -15*8-8(%rsp)            # error_code/entry_vector
+.LFX1:  popq  -15*8-8(%rsp)            # error_code/entry_vector
         SAVE_ALL                       # 15*8 bytes pushed
         movq  -8(%rsp),%rsi            # error_code/entry_vector
         sti                            # after stack abuse (-1024(%rsp))
@@ -68,11 +68,11 @@ FIX1:   popq  -15*8-8(%rsp)            #
         pushq %rax                     # RSP
         pushf                          # RFLAGS
         pushq $__HYPERVISOR_CS         # CS
-        leaq  DBLFLT1(%rip),%rax
+        leaq  .LDF1(%rip),%rax
         pushq %rax                     # RIP
         pushq %rsi                     # error_code/entry_vector
         jmp   handle_exception
-DBLFLT1:GET_CURRENT(%rbx)
+.LDF1:  GET_CURRENT(%rbx)
         jmp   test_all_events
 failsafe_callback:
         GET_CURRENT(%rbx)
@@ -87,10 +87,10 @@ failsafe_callback:
         jmp   test_all_events
 .previous
 .section __pre_ex_table,"a"
-        .quad FLT1,FIX1
+        .quad .LFT1,.LFX1
 .previous
 .section __ex_table,"a"
-        .quad DBLFLT1,failsafe_callback
+        .quad .LDF1,failsafe_callback
 .previous
 
         ALIGN
@@ -249,9 +249,9 @@ create_bounce_frame:
 1:      movb  TRAPBOUNCE_flags(%rdx),%cl
         subq  $40,%rsi
         movq  UREGS_ss+8(%rsp),%rax
-FLT2:   movq  %rax,32(%rsi)             # SS
+.LFT2:  movq  %rax,32(%rsi)             # SS
         movq  UREGS_rsp+8(%rsp),%rax
-FLT3:   movq  %rax,24(%rsi)             # RSP
+.LFT3:  movq  %rax,24(%rsi)             # RSP
         movq  VCPU_vcpu_info(%rbx),%rax
         pushq VCPUINFO_upcall_mask(%rax)
         testb $TBF_INTERRUPT,%cl
@@ -260,7 +260,7 @@ FLT3:   movq  %rax,24(%rsi)             
         popq  %rax
         shlq  $32,%rax                  # Bits 32-39: saved_upcall_mask
         movw  UREGS_cs+8(%rsp),%ax      # Bits  0-15: CS
-FLT4:   movq  %rax,8(%rsi)              # CS / saved_upcall_mask
+.LFT4:  movq  %rax,8(%rsi)              # CS / saved_upcall_mask
         shrq  $32,%rax
         testb $0xFF,%al                 # Bits 0-7: saved_upcall_mask
         setz  %ch                       # %ch == !saved_upcall_mask
@@ -268,30 +268,30 @@ FLT4:   movq  %rax,8(%rsi)              
         andq  $~X86_EFLAGS_IF,%rax
         shlb  $1,%ch                    # Bit 9 (EFLAGS.IF)
         orb   %ch,%ah                   # Fold EFLAGS.IF into %eax
-FLT5:   movq  %rax,16(%rsi)             # RFLAGS
+.LFT5:  movq  %rax,16(%rsi)             # RFLAGS
         movq  UREGS_rip+8(%rsp),%rax
-FLT6:   movq  %rax,(%rsi)               # RIP
+.LFT6:  movq  %rax,(%rsi)               # RIP
         testb $TBF_EXCEPTION_ERRCODE,%cl
         jz    1f
         subq  $8,%rsi
         movl  TRAPBOUNCE_error_code(%rdx),%eax
-FLT7:   movq  %rax,(%rsi)               # ERROR CODE
+.LFT7:  movq  %rax,(%rsi)               # ERROR CODE
 1:      testb $TBF_FAILSAFE,%cl
         jz    2f
         subq  $32,%rsi
         movl  %gs,%eax
-FLT8:   movq  %rax,24(%rsi)             # GS
+.LFT8:  movq  %rax,24(%rsi)             # GS
         movl  %fs,%eax
-FLT9:   movq  %rax,16(%rsi)             # FS
+.LFT9:  movq  %rax,16(%rsi)             # FS
         movl  %es,%eax
-FLT10:  movq  %rax,8(%rsi)              # ES
+.LFT10: movq  %rax,8(%rsi)              # ES
         movl  %ds,%eax
-FLT11:  movq  %rax,(%rsi)               # DS
+.LFT11: movq  %rax,(%rsi)               # DS
 2:      subq  $16,%rsi
         movq  UREGS_r11+8(%rsp),%rax
-FLT12:  movq  %rax,8(%rsi)              # R11
+.LFT12: movq  %rax,8(%rsi)              # R11
         movq  UREGS_rcx+8(%rsp),%rax
-FLT13:  movq  %rax,(%rsi)               # RCX
+.LFT13: movq  %rax,(%rsi)               # RCX
         /* Rewrite our stack frame and return to guest-OS mode. */
         /* IA32 Ref. Vol. 3: TF, VM, RF and NT flags are cleared on trap. */
         /* Also clear AC: alignment checks shouldn't trigger in kernel mode. */
@@ -308,12 +308,12 @@ FLT13:  movq  %rax,(%rsi)               
         movb  $0,TRAPBOUNCE_flags(%rdx)
         ret
 .section __ex_table,"a"
-        .quad  FLT2,domain_crash_synchronous ,  FLT3,domain_crash_synchronous
-        .quad  FLT4,domain_crash_synchronous ,  FLT5,domain_crash_synchronous
-        .quad  FLT6,domain_crash_synchronous ,  FLT7,domain_crash_synchronous
-        .quad  FLT8,domain_crash_synchronous ,  FLT9,domain_crash_synchronous
-        .quad FLT10,domain_crash_synchronous , FLT11,domain_crash_synchronous
-        .quad FLT12,domain_crash_synchronous , FLT13,domain_crash_synchronous
+        .quad  .LFT2,domain_crash_synchronous ,  .LFT3,domain_crash_synchronous
+        .quad  .LFT4,domain_crash_synchronous ,  .LFT5,domain_crash_synchronous
+        .quad  .LFT6,domain_crash_synchronous ,  .LFT7,domain_crash_synchronous
+        .quad  .LFT8,domain_crash_synchronous ,  .LFT9,domain_crash_synchronous
+        .quad .LFT10,domain_crash_synchronous , .LFT11,domain_crash_synchronous
+        .quad .LFT12,domain_crash_synchronous , .LFT13,domain_crash_synchronous
 .previous
 
 domain_crash_synchronous_string:



_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.