480 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			480 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /******************************************************************************
 | |
|  * Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp>
 | |
|  *                    VA Linux Systems Japan K.K.
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or modify
 | |
|  * it under the terms of the GNU General Public License as published by
 | |
|  * the Free Software Foundation; either version 2 of the License, or
 | |
|  * (at your option) any later version.
 | |
|  *
 | |
|  * This program is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
|  * GNU General Public License for more details.
 | |
|  *
 | |
|  * You should have received a copy of the GNU General Public License
 | |
|  * along with this program; if not, write to the Free Software
 | |
|  * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
 | |
|  *
 | |
|  */
 | |
| 
 | |
| #ifndef _ASM_IA64_PARAVIRT_PRIVOP_H
 | |
| #define _ASM_IA64_PARAVIRT_PRIVOP_H
 | |
| 
 | |
| #ifdef CONFIG_PARAVIRT
 | |
| 
 | |
| #ifndef __ASSEMBLY__
 | |
| 
 | |
| #include <linux/types.h>
 | |
| #include <asm/kregs.h> /* for IA64_PSR_I */
 | |
| 
 | |
| /******************************************************************************
 | |
|  * replacement of intrinsics operations.
 | |
|  */
 | |
| 
 | |
| struct pv_cpu_ops {
 | |
| 	void (*fc)(void *addr);
 | |
| 	unsigned long (*thash)(unsigned long addr);
 | |
| 	unsigned long (*get_cpuid)(int index);
 | |
| 	unsigned long (*get_pmd)(int index);
 | |
| 	unsigned long (*getreg)(int reg);
 | |
| 	void (*setreg)(int reg, unsigned long val);
 | |
| 	void (*ptcga)(unsigned long addr, unsigned long size);
 | |
| 	unsigned long (*get_rr)(unsigned long index);
 | |
| 	void (*set_rr)(unsigned long index, unsigned long val);
 | |
| 	void (*set_rr0_to_rr4)(unsigned long val0, unsigned long val1,
 | |
| 			       unsigned long val2, unsigned long val3,
 | |
| 			       unsigned long val4);
 | |
| 	void (*ssm_i)(void);
 | |
| 	void (*rsm_i)(void);
 | |
| 	unsigned long (*get_psr_i)(void);
 | |
| 	void (*intrin_local_irq_restore)(unsigned long flags);
 | |
| };
 | |
| 
 | |
| extern struct pv_cpu_ops pv_cpu_ops;
 | |
| 
 | |
| extern void ia64_native_setreg_func(int regnum, unsigned long val);
 | |
| extern unsigned long ia64_native_getreg_func(int regnum);
 | |
| 
 | |
| /************************************************/
 | |
| /* Instructions paravirtualized for performance */
 | |
| /************************************************/
 | |
| 
 | |
| #ifndef ASM_SUPPORTED
 | |
| #define paravirt_ssm_i()	pv_cpu_ops.ssm_i()
 | |
| #define paravirt_rsm_i()	pv_cpu_ops.rsm_i()
 | |
| #define __paravirt_getreg()	pv_cpu_ops.getreg()
 | |
| #endif
 | |
| 
 | |
| /* mask for ia64_native_ssm/rsm() must be constant.("i" constraing).
 | |
|  * static inline function doesn't satisfy it. */
 | |
| #define paravirt_ssm(mask)			\
 | |
| 	do {					\
 | |
| 		if ((mask) == IA64_PSR_I)	\
 | |
| 			paravirt_ssm_i();	\
 | |
| 		else				\
 | |
| 			ia64_native_ssm(mask);	\
 | |
| 	} while (0)
 | |
| 
 | |
| #define paravirt_rsm(mask)			\
 | |
| 	do {					\
 | |
| 		if ((mask) == IA64_PSR_I)	\
 | |
| 			paravirt_rsm_i();	\
 | |
| 		else				\
 | |
| 			ia64_native_rsm(mask);	\
 | |
| 	} while (0)
 | |
| 
 | |
| /* returned ip value should be the one in the caller,
 | |
|  * not in __paravirt_getreg() */
 | |
| #define paravirt_getreg(reg)					\
 | |
| 	({							\
 | |
| 		unsigned long res;				\
 | |
| 		if ((reg) == _IA64_REG_IP)			\
 | |
| 			res = ia64_native_getreg(_IA64_REG_IP); \
 | |
| 		else						\
 | |
| 			res = __paravirt_getreg(reg);		\
 | |
| 		res;						\
 | |
| 	})
 | |
| 
 | |
| /******************************************************************************
 | |
|  * replacement of hand written assembly codes.
 | |
|  */
 | |
| struct pv_cpu_asm_switch {
 | |
| 	unsigned long switch_to;
 | |
| 	unsigned long leave_syscall;
 | |
| 	unsigned long work_processed_syscall;
 | |
| 	unsigned long leave_kernel;
 | |
| };
 | |
| void paravirt_cpu_asm_init(const struct pv_cpu_asm_switch *cpu_asm_switch);
 | |
| 
 | |
| #endif /* __ASSEMBLY__ */
 | |
| 
 | |
| #define IA64_PARAVIRT_ASM_FUNC(name)	paravirt_ ## name
 | |
| 
 | |
| #else
 | |
| 
 | |
| /* fallback for native case */
 | |
| #define IA64_PARAVIRT_ASM_FUNC(name)	ia64_native_ ## name
 | |
| 
 | |
| #endif /* CONFIG_PARAVIRT */
 | |
| 
 | |
| #if defined(CONFIG_PARAVIRT) && defined(ASM_SUPPORTED)
 | |
| #define paravirt_dv_serialize_data()	ia64_dv_serialize_data()
 | |
| #else
 | |
| #define paravirt_dv_serialize_data()	/* nothing */
 | |
| #endif
 | |
| 
 | |
| /* these routines utilize privilege-sensitive or performance-sensitive
 | |
|  * privileged instructions so the code must be replaced with
 | |
|  * paravirtualized versions */
 | |
| #define ia64_switch_to			IA64_PARAVIRT_ASM_FUNC(switch_to)
 | |
| #define ia64_leave_syscall		IA64_PARAVIRT_ASM_FUNC(leave_syscall)
 | |
| #define ia64_work_processed_syscall	\
 | |
| 	IA64_PARAVIRT_ASM_FUNC(work_processed_syscall)
 | |
| #define ia64_leave_kernel		IA64_PARAVIRT_ASM_FUNC(leave_kernel)
 | |
| 
 | |
| 
 | |
| #if defined(CONFIG_PARAVIRT)
 | |
| /******************************************************************************
 | |
|  * binary patching infrastructure
 | |
|  */
 | |
| #define PARAVIRT_PATCH_TYPE_FC				1
 | |
| #define PARAVIRT_PATCH_TYPE_THASH			2
 | |
| #define PARAVIRT_PATCH_TYPE_GET_CPUID			3
 | |
| #define PARAVIRT_PATCH_TYPE_GET_PMD			4
 | |
| #define PARAVIRT_PATCH_TYPE_PTCGA			5
 | |
| #define PARAVIRT_PATCH_TYPE_GET_RR			6
 | |
| #define PARAVIRT_PATCH_TYPE_SET_RR			7
 | |
| #define PARAVIRT_PATCH_TYPE_SET_RR0_TO_RR4		8
 | |
| #define PARAVIRT_PATCH_TYPE_SSM_I			9
 | |
| #define PARAVIRT_PATCH_TYPE_RSM_I			10
 | |
| #define PARAVIRT_PATCH_TYPE_GET_PSR_I			11
 | |
| #define PARAVIRT_PATCH_TYPE_INTRIN_LOCAL_IRQ_RESTORE	12
 | |
| 
 | |
| /* PARAVIRT_PATY_TYPE_[GS]ETREG + _IA64_REG_xxx */
 | |
| #define PARAVIRT_PATCH_TYPE_GETREG			0x10000000
 | |
| #define PARAVIRT_PATCH_TYPE_SETREG			0x20000000
 | |
| 
 | |
| /*
 | |
|  * struct task_struct* (*ia64_switch_to)(void* next_task);
 | |
|  * void *ia64_leave_syscall;
 | |
|  * void *ia64_work_processed_syscall
 | |
|  * void *ia64_leave_kernel;
 | |
|  */
 | |
| 
 | |
| #define PARAVIRT_PATCH_TYPE_BR_START			0x30000000
 | |
| #define PARAVIRT_PATCH_TYPE_BR_SWITCH_TO		\
 | |
| 	(PARAVIRT_PATCH_TYPE_BR_START + 0)
 | |
| #define PARAVIRT_PATCH_TYPE_BR_LEAVE_SYSCALL		\
 | |
| 	(PARAVIRT_PATCH_TYPE_BR_START + 1)
 | |
| #define PARAVIRT_PATCH_TYPE_BR_WORK_PROCESSED_SYSCALL	\
 | |
| 	(PARAVIRT_PATCH_TYPE_BR_START + 2)
 | |
| #define PARAVIRT_PATCH_TYPE_BR_LEAVE_KERNEL		\
 | |
| 	(PARAVIRT_PATCH_TYPE_BR_START + 3)
 | |
| 
 | |
| #ifdef ASM_SUPPORTED
 | |
| #include <asm/paravirt_patch.h>
 | |
| 
 | |
| /*
 | |
|  * pv_cpu_ops calling stub.
 | |
|  * normal function call convension can't be written by gcc
 | |
|  * inline assembly.
 | |
|  *
 | |
|  * from the caller's point of view,
 | |
|  * the following registers will be clobbered.
 | |
|  * r2, r3
 | |
|  * r8-r15
 | |
|  * r16, r17
 | |
|  * b6, b7
 | |
|  * p6-p15
 | |
|  * ar.ccv
 | |
|  *
 | |
|  * from the callee's point of view ,
 | |
|  * the following registers can be used.
 | |
|  * r2, r3: scratch
 | |
|  * r8: scratch, input argument0 and return value
 | |
|  * r0-r15: scratch, input argument1-5
 | |
|  * b6: return pointer
 | |
|  * b7: scratch
 | |
|  * p6-p15: scratch
 | |
|  * ar.ccv: scratch
 | |
|  *
 | |
|  * other registers must not be changed. especially
 | |
|  * b0: rp: preserved. gcc ignores b0 in clobbered register.
 | |
|  * r16: saved gp
 | |
|  */
 | |
| /* 5 bundles */
 | |
| #define __PARAVIRT_BR							\
 | |
| 	";;\n"								\
 | |
| 	"{ .mlx\n"							\
 | |
| 	"nop 0\n"							\
 | |
| 	"movl r2 = %[op_addr]\n"/* get function pointer address */	\
 | |
| 	";;\n"								\
 | |
| 	"}\n"								\
 | |
| 	"1:\n"								\
 | |
| 	"{ .mii\n"							\
 | |
| 	"ld8 r2 = [r2]\n"	/* load function descriptor address */	\
 | |
| 	"mov r17 = ip\n"	/* get ip to calc return address */	\
 | |
| 	"mov r16 = gp\n"	/* save gp */				\
 | |
| 	";;\n"								\
 | |
| 	"}\n"								\
 | |
| 	"{ .mii\n"							\
 | |
| 	"ld8 r3 = [r2], 8\n"	/* load entry address */		\
 | |
| 	"adds r17 =  1f - 1b, r17\n"	/* calculate return address */	\
 | |
| 	";;\n"								\
 | |
| 	"mov b7 = r3\n"		/* set entry address */			\
 | |
| 	"}\n"								\
 | |
| 	"{ .mib\n"							\
 | |
| 	"ld8 gp = [r2]\n"	/* load gp value */			\
 | |
| 	"mov b6 = r17\n"	/* set return address */		\
 | |
| 	"br.cond.sptk.few b7\n"	/* intrinsics are very short isns */	\
 | |
| 	"}\n"								\
 | |
| 	"1:\n"								\
 | |
| 	"{ .mii\n"							\
 | |
| 	"mov gp = r16\n"	/* restore gp value */			\
 | |
| 	"nop 0\n"							\
 | |
| 	"nop 0\n"							\
 | |
| 	";;\n"								\
 | |
| 	"}\n"
 | |
| 
 | |
| #define PARAVIRT_OP(op)				\
 | |
| 	[op_addr] "i"(&pv_cpu_ops.op)
 | |
| 
 | |
| #define PARAVIRT_TYPE(type)			\
 | |
| 	PARAVIRT_PATCH_TYPE_ ## type
 | |
| 
 | |
| #define PARAVIRT_REG_CLOBBERS0					\
 | |
| 	"r2", "r3", /*"r8",*/ "r9", "r10", "r11", "r14",	\
 | |
| 		"r15", "r16", "r17"
 | |
| 
 | |
| #define PARAVIRT_REG_CLOBBERS1					\
 | |
| 	"r2","r3", /*"r8",*/ "r9", "r10", "r11", "r14",		\
 | |
| 		"r15", "r16", "r17"
 | |
| 
 | |
| #define PARAVIRT_REG_CLOBBERS2					\
 | |
| 	"r2", "r3", /*"r8", "r9",*/ "r10", "r11", "r14",	\
 | |
| 		"r15", "r16", "r17"
 | |
| 
 | |
| #define PARAVIRT_REG_CLOBBERS5					\
 | |
| 	"r2", "r3", /*"r8", "r9", "r10", "r11", "r14",*/	\
 | |
| 		"r15", "r16", "r17"
 | |
| 
 | |
| #define PARAVIRT_BR_CLOBBERS			\
 | |
| 	"b6", "b7"
 | |
| 
 | |
| #define PARAVIRT_PR_CLOBBERS						\
 | |
| 	"p6", "p7", "p8", "p9", "p10", "p11", "p12", "p13", "p14", "p15"
 | |
| 
 | |
| #define PARAVIRT_AR_CLOBBERS			\
 | |
| 	"ar.ccv"
 | |
| 
 | |
| #define PARAVIRT_CLOBBERS0			\
 | |
| 		PARAVIRT_REG_CLOBBERS0,		\
 | |
| 		PARAVIRT_BR_CLOBBERS,		\
 | |
| 		PARAVIRT_PR_CLOBBERS,		\
 | |
| 		PARAVIRT_AR_CLOBBERS,		\
 | |
| 		"memory"
 | |
| 
 | |
| #define PARAVIRT_CLOBBERS1			\
 | |
| 		PARAVIRT_REG_CLOBBERS1,		\
 | |
| 		PARAVIRT_BR_CLOBBERS,		\
 | |
| 		PARAVIRT_PR_CLOBBERS,		\
 | |
| 		PARAVIRT_AR_CLOBBERS,		\
 | |
| 		"memory"
 | |
| 
 | |
| #define PARAVIRT_CLOBBERS2			\
 | |
| 		PARAVIRT_REG_CLOBBERS2,		\
 | |
| 		PARAVIRT_BR_CLOBBERS,		\
 | |
| 		PARAVIRT_PR_CLOBBERS,		\
 | |
| 		PARAVIRT_AR_CLOBBERS,		\
 | |
| 		"memory"
 | |
| 
 | |
| #define PARAVIRT_CLOBBERS5			\
 | |
| 		PARAVIRT_REG_CLOBBERS5,		\
 | |
| 		PARAVIRT_BR_CLOBBERS,		\
 | |
| 		PARAVIRT_PR_CLOBBERS,		\
 | |
| 		PARAVIRT_AR_CLOBBERS,		\
 | |
| 		"memory"
 | |
| 
 | |
| #define PARAVIRT_BR0(op, type)					\
 | |
| 	register unsigned long ia64_clobber asm ("r8");		\
 | |
| 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
 | |
| 					  PARAVIRT_TYPE(type))	\
 | |
| 		      :	"=r"(ia64_clobber)			\
 | |
| 		      : PARAVIRT_OP(op)				\
 | |
| 		      : PARAVIRT_CLOBBERS0)
 | |
| 
 | |
| #define PARAVIRT_BR0_RET(op, type)				\
 | |
| 	register unsigned long ia64_intri_res asm ("r8");	\
 | |
| 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
 | |
| 					  PARAVIRT_TYPE(type))	\
 | |
| 		      : "=r"(ia64_intri_res)			\
 | |
| 		      : PARAVIRT_OP(op)				\
 | |
| 		      : PARAVIRT_CLOBBERS0)
 | |
| 
 | |
| #define PARAVIRT_BR1(op, type, arg1)				\
 | |
| 	register unsigned long __##arg1 asm ("r8") = arg1;	\
 | |
| 	register unsigned long ia64_clobber asm ("r8");		\
 | |
| 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
 | |
| 					  PARAVIRT_TYPE(type))	\
 | |
| 		      :	"=r"(ia64_clobber)			\
 | |
| 		      : PARAVIRT_OP(op), "0"(__##arg1)		\
 | |
| 		      : PARAVIRT_CLOBBERS1)
 | |
| 
 | |
| #define PARAVIRT_BR1_RET(op, type, arg1)			\
 | |
| 	register unsigned long ia64_intri_res asm ("r8");	\
 | |
| 	register unsigned long __##arg1 asm ("r8") = arg1;	\
 | |
| 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
 | |
| 					  PARAVIRT_TYPE(type))	\
 | |
| 		      : "=r"(ia64_intri_res)			\
 | |
| 		      : PARAVIRT_OP(op), "0"(__##arg1)		\
 | |
| 		      : PARAVIRT_CLOBBERS1)
 | |
| 
 | |
| #define PARAVIRT_BR1_VOID(op, type, arg1)			\
 | |
| 	register void *__##arg1 asm ("r8") = arg1;		\
 | |
| 	register unsigned long ia64_clobber asm ("r8");		\
 | |
| 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
 | |
| 					  PARAVIRT_TYPE(type))	\
 | |
| 		      :	"=r"(ia64_clobber)			\
 | |
| 		      : PARAVIRT_OP(op), "0"(__##arg1)		\
 | |
| 		      : PARAVIRT_CLOBBERS1)
 | |
| 
 | |
| #define PARAVIRT_BR2(op, type, arg1, arg2)				\
 | |
| 	register unsigned long __##arg1 asm ("r8") = arg1;		\
 | |
| 	register unsigned long __##arg2 asm ("r9") = arg2;		\
 | |
| 	register unsigned long ia64_clobber1 asm ("r8");		\
 | |
| 	register unsigned long ia64_clobber2 asm ("r9");		\
 | |
| 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,		\
 | |
| 					  PARAVIRT_TYPE(type))		\
 | |
| 		      : "=r"(ia64_clobber1), "=r"(ia64_clobber2)	\
 | |
| 		      : PARAVIRT_OP(op), "0"(__##arg1), "1"(__##arg2)	\
 | |
| 		      : PARAVIRT_CLOBBERS2)
 | |
| 
 | |
| 
 | |
| #define PARAVIRT_DEFINE_CPU_OP0(op, type)		\
 | |
| 	static inline void				\
 | |
| 	paravirt_ ## op (void)				\
 | |
| 	{						\
 | |
| 		PARAVIRT_BR0(op, type);			\
 | |
| 	}
 | |
| 
 | |
| #define PARAVIRT_DEFINE_CPU_OP0_RET(op, type)		\
 | |
| 	static inline unsigned long			\
 | |
| 	paravirt_ ## op (void)				\
 | |
| 	{						\
 | |
| 		PARAVIRT_BR0_RET(op, type);		\
 | |
| 		return ia64_intri_res;			\
 | |
| 	}
 | |
| 
 | |
| #define PARAVIRT_DEFINE_CPU_OP1_VOID(op, type)		\
 | |
| 	static inline void				\
 | |
| 	paravirt_ ## op (void *arg1)			\
 | |
| 	{						\
 | |
| 		PARAVIRT_BR1_VOID(op, type, arg1);	\
 | |
| 	}
 | |
| 
 | |
| #define PARAVIRT_DEFINE_CPU_OP1(op, type)		\
 | |
| 	static inline void				\
 | |
| 	paravirt_ ## op (unsigned long arg1)		\
 | |
| 	{						\
 | |
| 		PARAVIRT_BR1(op, type, arg1);		\
 | |
| 	}
 | |
| 
 | |
| #define PARAVIRT_DEFINE_CPU_OP1_RET(op, type)		\
 | |
| 	static inline unsigned long			\
 | |
| 	paravirt_ ## op (unsigned long arg1)		\
 | |
| 	{						\
 | |
| 		PARAVIRT_BR1_RET(op, type, arg1);	\
 | |
| 		return ia64_intri_res;			\
 | |
| 	}
 | |
| 
 | |
| #define PARAVIRT_DEFINE_CPU_OP2(op, type)		\
 | |
| 	static inline void				\
 | |
| 	paravirt_ ## op (unsigned long arg1,		\
 | |
| 			 unsigned long arg2)		\
 | |
| 	{						\
 | |
| 		PARAVIRT_BR2(op, type, arg1, arg2);	\
 | |
| 	}
 | |
| 
 | |
| 
 | |
| PARAVIRT_DEFINE_CPU_OP1_VOID(fc, FC);
 | |
| PARAVIRT_DEFINE_CPU_OP1_RET(thash, THASH)
 | |
| PARAVIRT_DEFINE_CPU_OP1_RET(get_cpuid, GET_CPUID)
 | |
| PARAVIRT_DEFINE_CPU_OP1_RET(get_pmd, GET_PMD)
 | |
| PARAVIRT_DEFINE_CPU_OP2(ptcga, PTCGA)
 | |
| PARAVIRT_DEFINE_CPU_OP1_RET(get_rr, GET_RR)
 | |
| PARAVIRT_DEFINE_CPU_OP2(set_rr, SET_RR)
 | |
| PARAVIRT_DEFINE_CPU_OP0(ssm_i, SSM_I)
 | |
| PARAVIRT_DEFINE_CPU_OP0(rsm_i, RSM_I)
 | |
| PARAVIRT_DEFINE_CPU_OP0_RET(get_psr_i, GET_PSR_I)
 | |
| PARAVIRT_DEFINE_CPU_OP1(intrin_local_irq_restore, INTRIN_LOCAL_IRQ_RESTORE)
 | |
| 
 | |
| static inline void
 | |
| paravirt_set_rr0_to_rr4(unsigned long val0, unsigned long val1,
 | |
| 			unsigned long val2, unsigned long val3,
 | |
| 			unsigned long val4)
 | |
| {
 | |
| 	register unsigned long __val0 asm ("r8") = val0;
 | |
| 	register unsigned long __val1 asm ("r9") = val1;
 | |
| 	register unsigned long __val2 asm ("r10") = val2;
 | |
| 	register unsigned long __val3 asm ("r11") = val3;
 | |
| 	register unsigned long __val4 asm ("r14") = val4;
 | |
| 
 | |
| 	register unsigned long ia64_clobber0 asm ("r8");
 | |
| 	register unsigned long ia64_clobber1 asm ("r9");
 | |
| 	register unsigned long ia64_clobber2 asm ("r10");
 | |
| 	register unsigned long ia64_clobber3 asm ("r11");
 | |
| 	register unsigned long ia64_clobber4 asm ("r14");
 | |
| 
 | |
| 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,
 | |
| 					  PARAVIRT_TYPE(SET_RR0_TO_RR4))
 | |
| 		      : "=r"(ia64_clobber0),
 | |
| 			"=r"(ia64_clobber1),
 | |
| 			"=r"(ia64_clobber2),
 | |
| 			"=r"(ia64_clobber3),
 | |
| 			"=r"(ia64_clobber4)
 | |
| 		      : PARAVIRT_OP(set_rr0_to_rr4),
 | |
| 			"0"(__val0), "1"(__val1), "2"(__val2),
 | |
| 			"3"(__val3), "4"(__val4)
 | |
| 		      : PARAVIRT_CLOBBERS5);
 | |
| }
 | |
| 
 | |
| /* unsigned long paravirt_getreg(int reg) */
 | |
| #define __paravirt_getreg(reg)						\
 | |
| 	({								\
 | |
| 		register unsigned long ia64_intri_res asm ("r8");	\
 | |
| 		register unsigned long __reg asm ("r8") = (reg);	\
 | |
| 									\
 | |
| 		asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
 | |
| 						  PARAVIRT_TYPE(GETREG) \
 | |
| 						  + (reg))		\
 | |
| 			      : "=r"(ia64_intri_res)			\
 | |
| 			      : PARAVIRT_OP(getreg), "0"(__reg)		\
 | |
| 			      : PARAVIRT_CLOBBERS1);			\
 | |
| 									\
 | |
| 		ia64_intri_res;						\
 | |
| 	})
 | |
| 
 | |
| /* void paravirt_setreg(int reg, unsigned long val) */
 | |
| #define paravirt_setreg(reg, val)					\
 | |
| 	do {								\
 | |
| 		register unsigned long __val asm ("r8") = val;		\
 | |
| 		register unsigned long __reg asm ("r9") = reg;		\
 | |
| 		register unsigned long ia64_clobber1 asm ("r8");	\
 | |
| 		register unsigned long ia64_clobber2 asm ("r9");	\
 | |
| 									\
 | |
| 		asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
 | |
| 						  PARAVIRT_TYPE(SETREG) \
 | |
| 						  + (reg))		\
 | |
| 			      : "=r"(ia64_clobber1),			\
 | |
| 				"=r"(ia64_clobber2)			\
 | |
| 			      : PARAVIRT_OP(setreg),			\
 | |
| 				"1"(__reg), "0"(__val)			\
 | |
| 			      : PARAVIRT_CLOBBERS2);			\
 | |
| 	} while (0)
 | |
| 
 | |
| #endif /* ASM_SUPPORTED */
 | |
| #endif /* CONFIG_PARAVIRT && ASM_SUPPOTED */
 | |
| 
 | |
| #endif /* _ASM_IA64_PARAVIRT_PRIVOP_H */
 |