x86/asm/32: Change all ENTRY+END to SYM_CODE_*
Change all assembly code which is marked using END (and not ENDPROC) to appropriate new markings SYM_CODE_START and SYM_CODE_END. And since the last user of END on X86 is gone now, make sure that END is not defined there. Signed-off-by: Jiri Slaby <jslaby@suse.cz> Signed-off-by: Borislav Petkov <bp@suse.de> Cc: Andrey Ryabinin <aryabinin@virtuozzo.com> Cc: Andy Lutomirski <luto@kernel.org> Cc: Boris Ostrovsky <boris.ostrovsky@oracle.com> Cc: Herbert Xu <herbert@gondor.apana.org.au> Cc: "H. Peter Anvin" <hpa@zytor.com> Cc: Ingo Molnar <mingo@redhat.com> Cc: Josh Poimboeuf <jpoimboe@redhat.com> Cc: linux-arch@vger.kernel.org Cc: Mark Rutland <mark.rutland@arm.com> Cc: Peter Zijlstra <peterz@infradead.org> Cc: "Rafael J. Wysocki" <rafael.j.wysocki@intel.com> Cc: "Steven Rostedt (VMware)" <rostedt@goodmis.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: x86-ml <x86@kernel.org> Link: https://lkml.kernel.org/r/20191011115108.12392-27-jslaby@suse.cz
This commit is contained in:
parent
78762b0e79
commit
5e63306f16
|
@ -709,7 +709,7 @@
|
|||
* %eax: prev task
|
||||
* %edx: next task
|
||||
*/
|
||||
ENTRY(__switch_to_asm)
|
||||
SYM_CODE_START(__switch_to_asm)
|
||||
/*
|
||||
* Save callee-saved registers
|
||||
* This must match the order in struct inactive_task_frame
|
||||
|
@ -748,7 +748,7 @@ ENTRY(__switch_to_asm)
|
|||
popl %ebp
|
||||
|
||||
jmp __switch_to
|
||||
END(__switch_to_asm)
|
||||
SYM_CODE_END(__switch_to_asm)
|
||||
|
||||
/*
|
||||
* The unwinder expects the last frame on the stack to always be at the same
|
||||
|
@ -774,7 +774,7 @@ ENDPROC(schedule_tail_wrapper)
|
|||
* ebx: kernel thread func (NULL for user thread)
|
||||
* edi: kernel thread arg
|
||||
*/
|
||||
ENTRY(ret_from_fork)
|
||||
SYM_CODE_START(ret_from_fork)
|
||||
call schedule_tail_wrapper
|
||||
|
||||
testl %ebx, %ebx
|
||||
|
@ -797,7 +797,7 @@ ENTRY(ret_from_fork)
|
|||
*/
|
||||
movl $0, PT_EAX(%esp)
|
||||
jmp 2b
|
||||
END(ret_from_fork)
|
||||
SYM_CODE_END(ret_from_fork)
|
||||
|
||||
/*
|
||||
* Return to user mode is not as complex as all this looks,
|
||||
|
@ -1161,7 +1161,7 @@ ENDPROC(entry_INT80_32)
|
|||
* We pack 1 stub into every 8-byte block.
|
||||
*/
|
||||
.align 8
|
||||
ENTRY(irq_entries_start)
|
||||
SYM_CODE_START(irq_entries_start)
|
||||
vector=FIRST_EXTERNAL_VECTOR
|
||||
.rept (FIRST_SYSTEM_VECTOR - FIRST_EXTERNAL_VECTOR)
|
||||
pushl $(~vector+0x80) /* Note: always in signed byte range */
|
||||
|
@ -1169,11 +1169,11 @@ ENTRY(irq_entries_start)
|
|||
jmp common_interrupt
|
||||
.align 8
|
||||
.endr
|
||||
END(irq_entries_start)
|
||||
SYM_CODE_END(irq_entries_start)
|
||||
|
||||
#ifdef CONFIG_X86_LOCAL_APIC
|
||||
.align 8
|
||||
ENTRY(spurious_entries_start)
|
||||
SYM_CODE_START(spurious_entries_start)
|
||||
vector=FIRST_SYSTEM_VECTOR
|
||||
.rept (NR_VECTORS - FIRST_SYSTEM_VECTOR)
|
||||
pushl $(~vector+0x80) /* Note: always in signed byte range */
|
||||
|
@ -1181,7 +1181,7 @@ ENTRY(spurious_entries_start)
|
|||
jmp common_spurious
|
||||
.align 8
|
||||
.endr
|
||||
END(spurious_entries_start)
|
||||
SYM_CODE_END(spurious_entries_start)
|
||||
|
||||
SYM_CODE_START_LOCAL(common_spurious)
|
||||
ASM_CLAC
|
||||
|
@ -1230,14 +1230,14 @@ ENDPROC(name)
|
|||
/* The include is where all of the SMP etc. interrupts come from */
|
||||
#include <asm/entry_arch.h>
|
||||
|
||||
ENTRY(coprocessor_error)
|
||||
SYM_CODE_START(coprocessor_error)
|
||||
ASM_CLAC
|
||||
pushl $0
|
||||
pushl $do_coprocessor_error
|
||||
jmp common_exception
|
||||
END(coprocessor_error)
|
||||
SYM_CODE_END(coprocessor_error)
|
||||
|
||||
ENTRY(simd_coprocessor_error)
|
||||
SYM_CODE_START(simd_coprocessor_error)
|
||||
ASM_CLAC
|
||||
pushl $0
|
||||
#ifdef CONFIG_X86_INVD_BUG
|
||||
|
@ -1249,96 +1249,96 @@ ENTRY(simd_coprocessor_error)
|
|||
pushl $do_simd_coprocessor_error
|
||||
#endif
|
||||
jmp common_exception
|
||||
END(simd_coprocessor_error)
|
||||
SYM_CODE_END(simd_coprocessor_error)
|
||||
|
||||
ENTRY(device_not_available)
|
||||
SYM_CODE_START(device_not_available)
|
||||
ASM_CLAC
|
||||
pushl $-1 # mark this as an int
|
||||
pushl $do_device_not_available
|
||||
jmp common_exception
|
||||
END(device_not_available)
|
||||
SYM_CODE_END(device_not_available)
|
||||
|
||||
#ifdef CONFIG_PARAVIRT
|
||||
ENTRY(native_iret)
|
||||
SYM_CODE_START(native_iret)
|
||||
iret
|
||||
_ASM_EXTABLE(native_iret, iret_exc)
|
||||
END(native_iret)
|
||||
SYM_CODE_END(native_iret)
|
||||
#endif
|
||||
|
||||
ENTRY(overflow)
|
||||
SYM_CODE_START(overflow)
|
||||
ASM_CLAC
|
||||
pushl $0
|
||||
pushl $do_overflow
|
||||
jmp common_exception
|
||||
END(overflow)
|
||||
SYM_CODE_END(overflow)
|
||||
|
||||
ENTRY(bounds)
|
||||
SYM_CODE_START(bounds)
|
||||
ASM_CLAC
|
||||
pushl $0
|
||||
pushl $do_bounds
|
||||
jmp common_exception
|
||||
END(bounds)
|
||||
SYM_CODE_END(bounds)
|
||||
|
||||
ENTRY(invalid_op)
|
||||
SYM_CODE_START(invalid_op)
|
||||
ASM_CLAC
|
||||
pushl $0
|
||||
pushl $do_invalid_op
|
||||
jmp common_exception
|
||||
END(invalid_op)
|
||||
SYM_CODE_END(invalid_op)
|
||||
|
||||
ENTRY(coprocessor_segment_overrun)
|
||||
SYM_CODE_START(coprocessor_segment_overrun)
|
||||
ASM_CLAC
|
||||
pushl $0
|
||||
pushl $do_coprocessor_segment_overrun
|
||||
jmp common_exception
|
||||
END(coprocessor_segment_overrun)
|
||||
SYM_CODE_END(coprocessor_segment_overrun)
|
||||
|
||||
ENTRY(invalid_TSS)
|
||||
SYM_CODE_START(invalid_TSS)
|
||||
ASM_CLAC
|
||||
pushl $do_invalid_TSS
|
||||
jmp common_exception
|
||||
END(invalid_TSS)
|
||||
SYM_CODE_END(invalid_TSS)
|
||||
|
||||
ENTRY(segment_not_present)
|
||||
SYM_CODE_START(segment_not_present)
|
||||
ASM_CLAC
|
||||
pushl $do_segment_not_present
|
||||
jmp common_exception
|
||||
END(segment_not_present)
|
||||
SYM_CODE_END(segment_not_present)
|
||||
|
||||
ENTRY(stack_segment)
|
||||
SYM_CODE_START(stack_segment)
|
||||
ASM_CLAC
|
||||
pushl $do_stack_segment
|
||||
jmp common_exception
|
||||
END(stack_segment)
|
||||
SYM_CODE_END(stack_segment)
|
||||
|
||||
ENTRY(alignment_check)
|
||||
SYM_CODE_START(alignment_check)
|
||||
ASM_CLAC
|
||||
pushl $do_alignment_check
|
||||
jmp common_exception
|
||||
END(alignment_check)
|
||||
SYM_CODE_END(alignment_check)
|
||||
|
||||
ENTRY(divide_error)
|
||||
SYM_CODE_START(divide_error)
|
||||
ASM_CLAC
|
||||
pushl $0 # no error code
|
||||
pushl $do_divide_error
|
||||
jmp common_exception
|
||||
END(divide_error)
|
||||
SYM_CODE_END(divide_error)
|
||||
|
||||
#ifdef CONFIG_X86_MCE
|
||||
ENTRY(machine_check)
|
||||
SYM_CODE_START(machine_check)
|
||||
ASM_CLAC
|
||||
pushl $0
|
||||
pushl machine_check_vector
|
||||
jmp common_exception
|
||||
END(machine_check)
|
||||
SYM_CODE_END(machine_check)
|
||||
#endif
|
||||
|
||||
ENTRY(spurious_interrupt_bug)
|
||||
SYM_CODE_START(spurious_interrupt_bug)
|
||||
ASM_CLAC
|
||||
pushl $0
|
||||
pushl $do_spurious_interrupt_bug
|
||||
jmp common_exception
|
||||
END(spurious_interrupt_bug)
|
||||
SYM_CODE_END(spurious_interrupt_bug)
|
||||
|
||||
#ifdef CONFIG_XEN_PV
|
||||
ENTRY(xen_hypervisor_callback)
|
||||
|
@ -1442,11 +1442,11 @@ BUILD_INTERRUPT3(hv_stimer0_callback_vector, HYPERV_STIMER0_VECTOR,
|
|||
|
||||
#endif /* CONFIG_HYPERV */
|
||||
|
||||
ENTRY(page_fault)
|
||||
SYM_CODE_START(page_fault)
|
||||
ASM_CLAC
|
||||
pushl $do_page_fault
|
||||
jmp common_exception_read_cr2
|
||||
END(page_fault)
|
||||
SYM_CODE_END(page_fault)
|
||||
|
||||
SYM_CODE_START_LOCAL_NOALIGN(common_exception_read_cr2)
|
||||
/* the function address is in %gs's slot on the stack */
|
||||
|
@ -1495,7 +1495,7 @@ SYM_CODE_START_LOCAL_NOALIGN(common_exception)
|
|||
jmp ret_from_exception
|
||||
SYM_CODE_END(common_exception)
|
||||
|
||||
ENTRY(debug)
|
||||
SYM_CODE_START(debug)
|
||||
/*
|
||||
* Entry from sysenter is now handled in common_exception
|
||||
*/
|
||||
|
@ -1503,7 +1503,7 @@ ENTRY(debug)
|
|||
pushl $-1 # mark this as an int
|
||||
pushl $do_debug
|
||||
jmp common_exception
|
||||
END(debug)
|
||||
SYM_CODE_END(debug)
|
||||
|
||||
/*
|
||||
* NMI is doubly nasty. It can happen on the first instruction of
|
||||
|
@ -1512,7 +1512,7 @@ END(debug)
|
|||
* switched stacks. We handle both conditions by simply checking whether we
|
||||
* interrupted kernel code running on the SYSENTER stack.
|
||||
*/
|
||||
ENTRY(nmi)
|
||||
SYM_CODE_START(nmi)
|
||||
ASM_CLAC
|
||||
|
||||
#ifdef CONFIG_X86_ESPFIX32
|
||||
|
@ -1577,9 +1577,9 @@ ENTRY(nmi)
|
|||
lss 12+4(%esp), %esp # back to espfix stack
|
||||
jmp .Lirq_return
|
||||
#endif
|
||||
END(nmi)
|
||||
SYM_CODE_END(nmi)
|
||||
|
||||
ENTRY(int3)
|
||||
SYM_CODE_START(int3)
|
||||
ASM_CLAC
|
||||
pushl $-1 # mark this as an int
|
||||
|
||||
|
@ -1590,22 +1590,22 @@ ENTRY(int3)
|
|||
movl %esp, %eax # pt_regs pointer
|
||||
call do_int3
|
||||
jmp ret_from_exception
|
||||
END(int3)
|
||||
SYM_CODE_END(int3)
|
||||
|
||||
ENTRY(general_protection)
|
||||
SYM_CODE_START(general_protection)
|
||||
pushl $do_general_protection
|
||||
jmp common_exception
|
||||
END(general_protection)
|
||||
SYM_CODE_END(general_protection)
|
||||
|
||||
#ifdef CONFIG_KVM_GUEST
|
||||
ENTRY(async_page_fault)
|
||||
SYM_CODE_START(async_page_fault)
|
||||
ASM_CLAC
|
||||
pushl $do_async_page_fault
|
||||
jmp common_exception_read_cr2
|
||||
END(async_page_fault)
|
||||
SYM_CODE_END(async_page_fault)
|
||||
#endif
|
||||
|
||||
ENTRY(rewind_stack_do_exit)
|
||||
SYM_CODE_START(rewind_stack_do_exit)
|
||||
/* Prevent any naive code from trying to unwind to our caller. */
|
||||
xorl %ebp, %ebp
|
||||
|
||||
|
@ -1614,4 +1614,4 @@ ENTRY(rewind_stack_do_exit)
|
|||
|
||||
call do_exit
|
||||
1: jmp 1b
|
||||
END(rewind_stack_do_exit)
|
||||
SYM_CODE_END(rewind_stack_do_exit)
|
||||
|
|
|
@ -25,7 +25,7 @@ SYM_FUNC_START(function_hook)
|
|||
ret
|
||||
SYM_FUNC_END(function_hook)
|
||||
|
||||
ENTRY(ftrace_caller)
|
||||
SYM_CODE_START(ftrace_caller)
|
||||
|
||||
#ifdef CONFIG_FRAME_POINTER
|
||||
/*
|
||||
|
@ -87,7 +87,7 @@ ftrace_graph_call:
|
|||
/* This is weak to keep gas from relaxing the jumps */
|
||||
WEAK(ftrace_stub)
|
||||
ret
|
||||
END(ftrace_caller)
|
||||
SYM_CODE_END(ftrace_caller)
|
||||
|
||||
SYM_CODE_START(ftrace_regs_caller)
|
||||
/*
|
||||
|
@ -166,7 +166,7 @@ SYM_INNER_LABEL(ftrace_regs_call, SYM_L_GLOBAL)
|
|||
SYM_CODE_END(ftrace_regs_caller)
|
||||
|
||||
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
|
||||
ENTRY(ftrace_graph_caller)
|
||||
SYM_CODE_START(ftrace_graph_caller)
|
||||
pushl %eax
|
||||
pushl %ecx
|
||||
pushl %edx
|
||||
|
@ -180,7 +180,7 @@ ENTRY(ftrace_graph_caller)
|
|||
popl %ecx
|
||||
popl %eax
|
||||
ret
|
||||
END(ftrace_graph_caller)
|
||||
SYM_CODE_END(ftrace_graph_caller)
|
||||
|
||||
.globl return_to_handler
|
||||
return_to_handler:
|
||||
|
|
|
@ -129,11 +129,13 @@
|
|||
SYM_FUNC_START_WEAK(name)
|
||||
#endif
|
||||
|
||||
#ifndef CONFIG_X86
|
||||
#ifndef END
|
||||
/* deprecated, use SYM_FUNC_END, SYM_DATA_END, or SYM_END */
|
||||
#define END(name) \
|
||||
.size name, .-name
|
||||
#endif
|
||||
#endif /* CONFIG_X86 */
|
||||
|
||||
#ifndef CONFIG_X86_64
|
||||
/* If symbol 'name' is treated as a subroutine (gets called, and returns)
|
||||
|
|
Loading…
Reference in New Issue