Current Path : /sys/amd64/compile/hs32/modules/usr/src/sys/modules/cxgbe/if_cxgbe/@/ia64/include/ |
FreeBSD hs32.drive.ne.jp 9.1-RELEASE FreeBSD 9.1-RELEASE #1: Wed Jan 14 12:18:08 JST 2015 root@hs32.drive.ne.jp:/sys/amd64/compile/hs32 amd64 |
Current File : //sys/amd64/compile/hs32/modules/usr/src/sys/modules/cxgbe/if_cxgbe/@/ia64/include/ia64_cpu.h |
/*- * Copyright (c) 2007 Marcel Moolenaar * Copyright (c) 2000 Doug Rabson * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * * $FreeBSD: release/9.1.0/sys/ia64/include/ia64_cpu.h 223170 2011-06-17 04:26:03Z marcel $ */ #ifndef _MACHINE_IA64_CPU_H_ #define _MACHINE_IA64_CPU_H_ /* * Local Interrupt ID. */ #define IA64_LID_GET_SAPIC_ID(x) ((u_int)((x) >> 16) & 0xffff) #define IA64_LID_SET_SAPIC_ID(x) ((u_int)((x) & 0xffff) << 16) /* * Definition of DCR bits. */ #define IA64_DCR_PP 0x0000000000000001 #define IA64_DCR_BE 0x0000000000000002 #define IA64_DCR_LC 0x0000000000000004 #define IA64_DCR_DM 0x0000000000000100 #define IA64_DCR_DP 0x0000000000000200 #define IA64_DCR_DK 0x0000000000000400 #define IA64_DCR_DX 0x0000000000000800 #define IA64_DCR_DR 0x0000000000001000 #define IA64_DCR_DA 0x0000000000002000 #define IA64_DCR_DD 0x0000000000004000 #define IA64_DCR_DEFAULT \ (IA64_DCR_DM | IA64_DCR_DP | IA64_DCR_DK | IA64_DCR_DX | \ IA64_DCR_DR | IA64_DCR_DA | IA64_DCR_DD) /* * Definition of PSR and IPSR bits. */ #define IA64_PSR_BE 0x0000000000000002 #define IA64_PSR_UP 0x0000000000000004 #define IA64_PSR_AC 0x0000000000000008 #define IA64_PSR_MFL 0x0000000000000010 #define IA64_PSR_MFH 0x0000000000000020 #define IA64_PSR_IC 0x0000000000002000 #define IA64_PSR_I 0x0000000000004000 #define IA64_PSR_PK 0x0000000000008000 #define IA64_PSR_DT 0x0000000000020000 #define IA64_PSR_DFL 0x0000000000040000 #define IA64_PSR_DFH 0x0000000000080000 #define IA64_PSR_SP 0x0000000000100000 #define IA64_PSR_PP 0x0000000000200000 #define IA64_PSR_DI 0x0000000000400000 #define IA64_PSR_SI 0x0000000000800000 #define IA64_PSR_DB 0x0000000001000000 #define IA64_PSR_LP 0x0000000002000000 #define IA64_PSR_TB 0x0000000004000000 #define IA64_PSR_RT 0x0000000008000000 #define IA64_PSR_CPL 0x0000000300000000 #define IA64_PSR_CPL_KERN 0x0000000000000000 #define IA64_PSR_CPL_1 0x0000000100000000 #define IA64_PSR_CPL_2 0x0000000200000000 #define IA64_PSR_CPL_USER 0x0000000300000000 #define IA64_PSR_IS 0x0000000400000000 #define IA64_PSR_MC 0x0000000800000000 #define IA64_PSR_IT 0x0000001000000000 #define IA64_PSR_ID 0x0000002000000000 #define IA64_PSR_DA 0x0000004000000000 #define IA64_PSR_DD 0x0000008000000000 #define IA64_PSR_SS 0x0000010000000000 #define IA64_PSR_RI 0x0000060000000000 #define IA64_PSR_RI_0 0x0000000000000000 #define IA64_PSR_RI_1 0x0000020000000000 #define IA64_PSR_RI_2 0x0000040000000000 #define IA64_PSR_ED 0x0000080000000000 #define IA64_PSR_BN 0x0000100000000000 #define IA64_PSR_IA 0x0000200000000000 /* * Definition of ISR bits. */ #define IA64_ISR_CODE 0x000000000000ffff #define IA64_ISR_VECTOR 0x0000000000ff0000 #define IA64_ISR_X 0x0000000100000000 #define IA64_ISR_W 0x0000000200000000 #define IA64_ISR_R 0x0000000400000000 #define IA64_ISR_NA 0x0000000800000000 #define IA64_ISR_SP 0x0000001000000000 #define IA64_ISR_RS 0x0000002000000000 #define IA64_ISR_IR 0x0000004000000000 #define IA64_ISR_NI 0x0000008000000000 #define IA64_ISR_SO 0x0000010000000000 #define IA64_ISR_EI 0x0000060000000000 #define IA64_ISR_EI_0 0x0000000000000000 #define IA64_ISR_EI_1 0x0000020000000000 #define IA64_ISR_EI_2 0x0000040000000000 #define IA64_ISR_ED 0x0000080000000000 /* * Vector numbers for various ia64 interrupts. */ #define IA64_VEC_VHPT 0 #define IA64_VEC_ITLB 1 #define IA64_VEC_DTLB 2 #define IA64_VEC_ALT_ITLB 3 #define IA64_VEC_ALT_DTLB 4 #define IA64_VEC_NESTED_DTLB 5 #define IA64_VEC_IKEY_MISS 6 #define IA64_VEC_DKEY_MISS 7 #define IA64_VEC_DIRTY_BIT 8 #define IA64_VEC_INST_ACCESS 9 #define IA64_VEC_DATA_ACCESS 10 #define IA64_VEC_BREAK 11 #define IA64_VEC_EXT_INTR 12 #define IA64_VEC_PAGE_NOT_PRESENT 20 #define IA64_VEC_KEY_PERMISSION 21 #define IA64_VEC_INST_ACCESS_RIGHTS 22 #define IA64_VEC_DATA_ACCESS_RIGHTS 23 #define IA64_VEC_GENERAL_EXCEPTION 24 #define IA64_VEC_DISABLED_FP 25 #define IA64_VEC_NAT_CONSUMPTION 26 #define IA64_VEC_SPECULATION 27 #define IA64_VEC_DEBUG 29 #define IA64_VEC_UNALIGNED_REFERENCE 30 #define IA64_VEC_UNSUPP_DATA_REFERENCE 31 #define IA64_VEC_FLOATING_POINT_FAULT 32 #define IA64_VEC_FLOATING_POINT_TRAP 33 #define IA64_VEC_LOWER_PRIVILEGE_TRANSFER 34 #define IA64_VEC_TAKEN_BRANCH_TRAP 35 #define IA64_VEC_SINGLE_STEP_TRAP 36 #define IA64_VEC_IA32_EXCEPTION 45 #define IA64_VEC_IA32_INTERCEPT 46 #define IA64_VEC_IA32_INTERRUPT 47 /* * IA-32 exceptions. */ #define IA32_EXCEPTION_DIVIDE 0 #define IA32_EXCEPTION_DEBUG 1 #define IA32_EXCEPTION_BREAK 3 #define IA32_EXCEPTION_OVERFLOW 4 #define IA32_EXCEPTION_BOUND 5 #define IA32_EXCEPTION_DNA 7 #define IA32_EXCEPTION_NOT_PRESENT 11 #define IA32_EXCEPTION_STACK_FAULT 12 #define IA32_EXCEPTION_GPFAULT 13 #define IA32_EXCEPTION_FPERROR 16 #define IA32_EXCEPTION_ALIGNMENT_CHECK 17 #define IA32_EXCEPTION_STREAMING_SIMD 19 #define IA32_INTERCEPT_INSTRUCTION 0 #define IA32_INTERCEPT_GATE 1 #define IA32_INTERCEPT_SYSTEM_FLAG 2 #define IA32_INTERCEPT_LOCK 4 #ifndef LOCORE /* * Various special ia64 instructions. */ /* * Memory Fence. */ static __inline void ia64_mf(void) { __asm __volatile("mf"); } static __inline void ia64_mf_a(void) { __asm __volatile("mf.a"); } /* * Flush Cache. */ static __inline void ia64_fc(uint64_t va) { __asm __volatile("fc %0" :: "r"(va)); } static __inline void ia64_fc_i(uint64_t va) { __asm __volatile("fc.i %0" :: "r"(va)); } /* * Sync instruction stream. */ static __inline void ia64_sync_i(void) { __asm __volatile("sync.i"); } /* * Calculate address in VHPT for va. */ static __inline uint64_t ia64_thash(uint64_t va) { uint64_t result; __asm __volatile("thash %0=%1" : "=r" (result) : "r" (va)); return result; } /* * Calculate VHPT tag for va. */ static __inline uint64_t ia64_ttag(uint64_t va) { uint64_t result; __asm __volatile("ttag %0=%1" : "=r" (result) : "r" (va)); return result; } /* * Convert virtual address to physical. */ static __inline uint64_t ia64_tpa(uint64_t va) { uint64_t result; __asm __volatile("tpa %0=%1" : "=r" (result) : "r" (va)); return result; } /* * Generate a ptc.e instruction. */ static __inline void ia64_ptc_e(uint64_t v) { __asm __volatile("ptc.e %0;; srlz.i;;" :: "r"(v)); } /* * Generate a ptc.g instruction. */ static __inline void ia64_ptc_g(uint64_t va, uint64_t log2size) { __asm __volatile("ptc.g %0,%1;;" :: "r"(va), "r"(log2size)); } /* * Generate a ptc.ga instruction. */ static __inline void ia64_ptc_ga(uint64_t va, uint64_t log2size) { __asm __volatile("ptc.ga %0,%1;;" :: "r"(va), "r"(log2size)); } /* * Generate a ptc.l instruction. */ static __inline void ia64_ptc_l(uint64_t va, uint64_t log2size) { __asm __volatile("ptc.l %0,%1;; srlz.i;;" :: "r"(va), "r"(log2size)); } /* * Invalidate the ALAT on the local processor. */ static __inline void ia64_invala(void) { __asm __volatile("invala;;"); } /* * Unordered memory load. */ static __inline uint8_t ia64_ld1(uint8_t *p) { uint8_t v; __asm __volatile("ld1 %0=[%1];;" : "=r"(v) : "r"(p)); return (v); } static __inline uint16_t ia64_ld2(uint16_t *p) { uint16_t v; __asm __volatile("ld2 %0=[%1];;" : "=r"(v) : "r"(p)); return (v); } static __inline uint32_t ia64_ld4(uint32_t *p) { uint32_t v; __asm __volatile("ld4 %0=[%1];;" : "=r"(v) : "r"(p)); return (v); } static __inline uint64_t ia64_ld8(uint64_t *p) { uint64_t v; __asm __volatile("ld8 %0=[%1];;" : "=r"(v) : "r"(p)); return (v); } /* * Unordered memory store. */ static __inline void ia64_st1(uint8_t *p, uint8_t v) { __asm __volatile("st1 [%0]=%1;;" :: "r"(p), "r"(v)); } static __inline void ia64_st2(uint16_t *p, uint16_t v) { __asm __volatile("st2 [%0]=%1;;" :: "r"(p), "r"(v)); } static __inline void ia64_st4(uint32_t *p, uint32_t v) { __asm __volatile("st4 [%0]=%1;;" :: "r"(p), "r"(v)); } static __inline void ia64_st8(uint64_t *p, uint64_t v) { __asm __volatile("st8 [%0]=%1;;" :: "r"(p), "r"(v)); } /* * Read the value of psr. */ static __inline uint64_t ia64_get_psr(void) { uint64_t result; __asm __volatile("mov %0=psr;;" : "=r" (result)); return result; } /* * Define accessors for application registers. */ #define IA64_AR(name) \ \ static __inline uint64_t \ ia64_get_##name(void) \ { \ uint64_t result; \ __asm __volatile("mov %0=ar." #name : "=r" (result)); \ return result; \ } \ \ static __inline void \ ia64_set_##name(uint64_t v) \ { \ __asm __volatile("mov ar." #name "=%0;;" :: "r" (v)); \ } IA64_AR(k0) IA64_AR(k1) IA64_AR(k2) IA64_AR(k3) IA64_AR(k4) IA64_AR(k5) IA64_AR(k6) IA64_AR(k7) IA64_AR(rsc) IA64_AR(bsp) IA64_AR(bspstore) IA64_AR(rnat) IA64_AR(fcr) IA64_AR(eflag) IA64_AR(csd) IA64_AR(ssd) IA64_AR(cflg) IA64_AR(fsr) IA64_AR(fir) IA64_AR(fdr) IA64_AR(ccv) IA64_AR(unat) IA64_AR(fpsr) IA64_AR(itc) IA64_AR(pfs) IA64_AR(lc) IA64_AR(ec) /* * Define accessors for control registers. */ #define IA64_CR(name) \ \ static __inline uint64_t \ ia64_get_##name(void) \ { \ uint64_t result; \ __asm __volatile("mov %0=cr." #name : "=r" (result)); \ return result; \ } \ \ static __inline void \ ia64_set_##name(uint64_t v) \ { \ __asm __volatile("mov cr." #name "=%0;;" :: "r" (v)); \ } IA64_CR(dcr) IA64_CR(itm) IA64_CR(iva) IA64_CR(pta) IA64_CR(ipsr) IA64_CR(isr) IA64_CR(iip) IA64_CR(ifa) IA64_CR(itir) IA64_CR(iipa) IA64_CR(ifs) IA64_CR(iim) IA64_CR(iha) IA64_CR(lid) IA64_CR(ivr) IA64_CR(tpr) IA64_CR(eoi) IA64_CR(irr0) IA64_CR(irr1) IA64_CR(irr2) IA64_CR(irr3) IA64_CR(itv) IA64_CR(pmv) IA64_CR(cmcv) IA64_CR(lrr0) IA64_CR(lrr1) /* * Write a region register. */ static __inline void ia64_set_rr(uint64_t rrbase, uint64_t v) { __asm __volatile("mov rr[%0]=%1" :: "r"(rrbase), "r"(v) : "memory"); } /* * Read a CPUID register. */ static __inline uint64_t ia64_get_cpuid(int i) { uint64_t result; __asm __volatile("mov %0=cpuid[%1]" : "=r" (result) : "r"(i)); return result; } static __inline void ia64_disable_highfp(void) { __asm __volatile("ssm psr.dfh;; srlz.d"); } static __inline void ia64_enable_highfp(void) { __asm __volatile("rsm psr.dfh;; srlz.d"); } /* * Avoid inline functions for the following so that they still work * correctly when inlining is not enabled (e.g. -O0). Function calls * need data serialization after setting psr, which results in a * hazard. */ #define ia64_srlz_d() __asm __volatile("srlz.d") #define ia64_srlz_i() __asm __volatile("srlz.i;;") #endif /* !LOCORE */ #endif /* _MACHINE_IA64_CPU_H_ */