136 lines
3.3 KiB
C
136 lines
3.3 KiB
C
|
#ifndef _ASM_IA64_ASMMACRO_H
|
||
|
#define _ASM_IA64_ASMMACRO_H
|
||
|
|
||
|
/*
|
||
|
* Copyright (C) 2000-2001, 2003-2004 Hewlett-Packard Co
|
||
|
* David Mosberger-Tang <davidm@hpl.hp.com>
|
||
|
*/
|
||
|
|
||
|
|
||
|
#define ENTRY(name) \
|
||
|
.align 32; \
|
||
|
.proc name; \
|
||
|
name:
|
||
|
|
||
|
#define ENTRY_MIN_ALIGN(name) \
|
||
|
.align 16; \
|
||
|
.proc name; \
|
||
|
name:
|
||
|
|
||
|
#define GLOBAL_ENTRY(name) \
|
||
|
.global name; \
|
||
|
ENTRY(name)
|
||
|
|
||
|
#define END(name) \
|
||
|
.endp name
|
||
|
|
||
|
/*
|
||
|
* Helper macros to make unwind directives more readable:
|
||
|
*/
|
||
|
|
||
|
/* prologue_gr: */
|
||
|
#define ASM_UNW_PRLG_RP 0x8
|
||
|
#define ASM_UNW_PRLG_PFS 0x4
|
||
|
#define ASM_UNW_PRLG_PSP 0x2
|
||
|
#define ASM_UNW_PRLG_PR 0x1
|
||
|
#define ASM_UNW_PRLG_GRSAVE(ninputs) (32+(ninputs))
|
||
|
|
||
|
/*
|
||
|
* Helper macros for accessing user memory.
|
||
|
*
|
||
|
* When adding any new .section/.previous entries here, make sure to
|
||
|
* also add it to the DISCARD section in arch/ia64/kernel/gate.lds.S or
|
||
|
* unpleasant things will happen.
|
||
|
*/
|
||
|
|
||
|
.section "__ex_table", "a" // declare section & section attributes
|
||
|
.previous
|
||
|
|
||
|
# define EX(y,x...) \
|
||
|
.xdata4 "__ex_table", 99f-., y-.; \
|
||
|
[99:] x
|
||
|
# define EXCLR(y,x...) \
|
||
|
.xdata4 "__ex_table", 99f-., y-.+4; \
|
||
|
[99:] x
|
||
|
|
||
|
/*
|
||
|
* Tag MCA recoverable instruction ranges.
|
||
|
*/
|
||
|
|
||
|
.section "__mca_table", "a" // declare section & section attributes
|
||
|
.previous
|
||
|
|
||
|
# define MCA_RECOVER_RANGE(y) \
|
||
|
.xdata4 "__mca_table", y-., 99f-.; \
|
||
|
[99:]
|
||
|
|
||
|
/*
|
||
|
* Mark instructions that need a load of a virtual address patched to be
|
||
|
* a load of a physical address. We use this either in critical performance
|
||
|
* path (ivt.S - TLB miss processing) or in places where it might not be
|
||
|
* safe to use a "tpa" instruction (mca_asm.S - error recovery).
|
||
|
*/
|
||
|
.section ".data..patch.vtop", "a" // declare section & section attributes
|
||
|
.previous
|
||
|
|
||
|
#define LOAD_PHYSICAL(pr, reg, obj) \
|
||
|
[1:](pr)movl reg = obj; \
|
||
|
.xdata4 ".data..patch.vtop", 1b-.
|
||
|
|
||
|
/*
|
||
|
* For now, we always put in the McKinley E9 workaround. On CPUs that don't need it,
|
||
|
* we'll patch out the work-around bundles with NOPs, so their impact is minimal.
|
||
|
*/
|
||
|
#define DO_MCKINLEY_E9_WORKAROUND
|
||
|
|
||
|
#ifdef DO_MCKINLEY_E9_WORKAROUND
|
||
|
.section ".data..patch.mckinley_e9", "a"
|
||
|
.previous
|
||
|
/* workaround for Itanium 2 Errata 9: */
|
||
|
# define FSYS_RETURN \
|
||
|
.xdata4 ".data..patch.mckinley_e9", 1f-.; \
|
||
|
1:{ .mib; \
|
||
|
nop.m 0; \
|
||
|
mov r16=ar.pfs; \
|
||
|
br.call.sptk.many b7=2f;; \
|
||
|
}; \
|
||
|
2:{ .mib; \
|
||
|
nop.m 0; \
|
||
|
mov ar.pfs=r16; \
|
||
|
br.ret.sptk.many b6;; \
|
||
|
}
|
||
|
#else
|
||
|
# define FSYS_RETURN br.ret.sptk.many b6
|
||
|
#endif
|
||
|
|
||
|
/*
|
||
|
* If physical stack register size is different from DEF_NUM_STACK_REG,
|
||
|
* dynamically patch the kernel for correct size.
|
||
|
*/
|
||
|
.section ".data..patch.phys_stack_reg", "a"
|
||
|
.previous
|
||
|
#define LOAD_PHYS_STACK_REG_SIZE(reg) \
|
||
|
[1:] adds reg=IA64_NUM_PHYS_STACK_REG*8+8,r0; \
|
||
|
.xdata4 ".data..patch.phys_stack_reg", 1b-.
|
||
|
|
||
|
/*
|
||
|
* Up until early 2004, use of .align within a function caused bad unwind info.
|
||
|
* TEXT_ALIGN(n) expands into ".align n" if a fixed GAS is available or into nothing
|
||
|
* otherwise.
|
||
|
*/
|
||
|
#ifdef HAVE_WORKING_TEXT_ALIGN
|
||
|
# define TEXT_ALIGN(n) .align n
|
||
|
#else
|
||
|
# define TEXT_ALIGN(n)
|
||
|
#endif
|
||
|
|
||
|
#ifdef HAVE_SERIALIZE_DIRECTIVE
|
||
|
# define dv_serialize_data .serialize.data
|
||
|
# define dv_serialize_instruction .serialize.instruction
|
||
|
#else
|
||
|
# define dv_serialize_data
|
||
|
# define dv_serialize_instruction
|
||
|
#endif
|
||
|
|
||
|
#endif /* _ASM_IA64_ASMMACRO_H */
|