xref: /f-stack/freebsd/arm64/include/cpu.h (revision 22ce4aff)
1a9643ea8Slogwang /*-
2a9643ea8Slogwang  * Copyright (c) 1990 The Regents of the University of California.
3a9643ea8Slogwang  * Copyright (c) 2014-2016 The FreeBSD Foundation
4a9643ea8Slogwang  * All rights reserved.
5a9643ea8Slogwang  *
6a9643ea8Slogwang  * This code is derived from software contributed to Berkeley by
7a9643ea8Slogwang  * William Jolitz.
8a9643ea8Slogwang  *
9a9643ea8Slogwang  * Portions of this software were developed by Andrew Turner
10a9643ea8Slogwang  * under sponsorship from the FreeBSD Foundation
11a9643ea8Slogwang  *
12a9643ea8Slogwang  * Redistribution and use in source and binary forms, with or without
13a9643ea8Slogwang  * modification, are permitted provided that the following conditions
14a9643ea8Slogwang  * are met:
15a9643ea8Slogwang  * 1. Redistributions of source code must retain the above copyright
16a9643ea8Slogwang  *    notice, this list of conditions and the following disclaimer.
17a9643ea8Slogwang  * 2. Redistributions in binary form must reproduce the above copyright
18a9643ea8Slogwang  *    notice, this list of conditions and the following disclaimer in the
19a9643ea8Slogwang  *    documentation and/or other materials provided with the distribution.
20a9643ea8Slogwang  * 3. Neither the name of the University nor the names of its contributors
21a9643ea8Slogwang  *    may be used to endorse or promote products derived from this software
22a9643ea8Slogwang  *    without specific prior written permission.
23a9643ea8Slogwang  *
24a9643ea8Slogwang  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
25a9643ea8Slogwang  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
26a9643ea8Slogwang  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
27a9643ea8Slogwang  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
28a9643ea8Slogwang  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
29a9643ea8Slogwang  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
30a9643ea8Slogwang  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
31a9643ea8Slogwang  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
32a9643ea8Slogwang  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
33a9643ea8Slogwang  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
34a9643ea8Slogwang  * SUCH DAMAGE.
35a9643ea8Slogwang  *
36a9643ea8Slogwang  *	from: @(#)cpu.h 5.4 (Berkeley) 5/9/91
37a9643ea8Slogwang  *	from: FreeBSD: src/sys/i386/include/cpu.h,v 1.62 2001/06/29
38a9643ea8Slogwang  * $FreeBSD$
39a9643ea8Slogwang  */
40a9643ea8Slogwang 
41a9643ea8Slogwang #ifndef _MACHINE_CPU_H_
42a9643ea8Slogwang #define	_MACHINE_CPU_H_
43a9643ea8Slogwang 
44a9643ea8Slogwang #include <machine/atomic.h>
45a9643ea8Slogwang #include <machine/frame.h>
46a9643ea8Slogwang #include <machine/armreg.h>
47a9643ea8Slogwang 
48a9643ea8Slogwang #define	TRAPF_PC(tfp)		((tfp)->tf_lr)
49a9643ea8Slogwang #define	TRAPF_USERMODE(tfp)	(((tfp)->tf_spsr & PSR_M_MASK) == PSR_M_EL0t)
50a9643ea8Slogwang 
51a9643ea8Slogwang #define	cpu_getstack(td)	((td)->td_frame->tf_sp)
52a9643ea8Slogwang #define	cpu_setstack(td, sp)	((td)->td_frame->tf_sp = (sp))
53a9643ea8Slogwang #define	cpu_spinwait()		__asm __volatile("yield" ::: "memory")
54*22ce4affSfengbojiang #define	cpu_lock_delay()	DELAY(1)
55a9643ea8Slogwang 
56a9643ea8Slogwang /* Extract CPU affinity levels 0-3 */
57a9643ea8Slogwang #define	CPU_AFF0(mpidr)	(u_int)(((mpidr) >> 0) & 0xff)
58a9643ea8Slogwang #define	CPU_AFF1(mpidr)	(u_int)(((mpidr) >> 8) & 0xff)
59a9643ea8Slogwang #define	CPU_AFF2(mpidr)	(u_int)(((mpidr) >> 16) & 0xff)
60a9643ea8Slogwang #define	CPU_AFF3(mpidr)	(u_int)(((mpidr) >> 32) & 0xff)
61a9643ea8Slogwang #define	CPU_AFF0_MASK	0xffUL
62a9643ea8Slogwang #define	CPU_AFF1_MASK	0xff00UL
63a9643ea8Slogwang #define	CPU_AFF2_MASK	0xff0000UL
64a9643ea8Slogwang #define	CPU_AFF3_MASK	0xff00000000UL
65a9643ea8Slogwang #define	CPU_AFF_MASK	(CPU_AFF0_MASK | CPU_AFF1_MASK | \
66a9643ea8Slogwang     CPU_AFF2_MASK| CPU_AFF3_MASK)	/* Mask affinity fields in MPIDR_EL1 */
67a9643ea8Slogwang 
68a9643ea8Slogwang #ifdef _KERNEL
69a9643ea8Slogwang 
70a9643ea8Slogwang #define	CPU_IMPL_ARM		0x41
71a9643ea8Slogwang #define	CPU_IMPL_BROADCOM	0x42
72a9643ea8Slogwang #define	CPU_IMPL_CAVIUM		0x43
73a9643ea8Slogwang #define	CPU_IMPL_DEC		0x44
74a9643ea8Slogwang #define	CPU_IMPL_INFINEON	0x49
75a9643ea8Slogwang #define	CPU_IMPL_FREESCALE	0x4D
76a9643ea8Slogwang #define	CPU_IMPL_NVIDIA		0x4E
77a9643ea8Slogwang #define	CPU_IMPL_APM		0x50
78a9643ea8Slogwang #define	CPU_IMPL_QUALCOMM	0x51
79a9643ea8Slogwang #define	CPU_IMPL_MARVELL	0x56
80a9643ea8Slogwang #define	CPU_IMPL_INTEL		0x69
81a9643ea8Slogwang 
82*22ce4affSfengbojiang /* ARM Part numbers */
83a9643ea8Slogwang #define	CPU_PART_FOUNDATION	0xD00
84a9643ea8Slogwang #define	CPU_PART_CORTEX_A53	0xD03
85*22ce4affSfengbojiang #define	CPU_PART_CORTEX_A35	0xD04
86*22ce4affSfengbojiang #define	CPU_PART_CORTEX_A55	0xD05
87*22ce4affSfengbojiang #define	CPU_PART_CORTEX_A65	0xD06
88a9643ea8Slogwang #define	CPU_PART_CORTEX_A57	0xD07
89*22ce4affSfengbojiang #define	CPU_PART_CORTEX_A72	0xD08
90*22ce4affSfengbojiang #define	CPU_PART_CORTEX_A73	0xD09
91*22ce4affSfengbojiang #define	CPU_PART_CORTEX_A75	0xD0A
92*22ce4affSfengbojiang #define	CPU_PART_CORTEX_A76	0xD0B
93*22ce4affSfengbojiang #define	CPU_PART_NEOVERSE_N1	0xD0C
94*22ce4affSfengbojiang #define	CPU_PART_CORTEX_A77	0xD0D
95*22ce4affSfengbojiang #define	CPU_PART_CORTEX_A76AE	0xD0E
96a9643ea8Slogwang 
97*22ce4affSfengbojiang /* Cavium Part numbers */
98*22ce4affSfengbojiang #define	CPU_PART_THUNDERX	0x0A1
99*22ce4affSfengbojiang #define	CPU_PART_THUNDERX_81XX	0x0A2
100*22ce4affSfengbojiang #define	CPU_PART_THUNDERX_83XX	0x0A3
101*22ce4affSfengbojiang #define	CPU_PART_THUNDERX2	0x0AF
102*22ce4affSfengbojiang 
103*22ce4affSfengbojiang #define	CPU_REV_THUNDERX_1_0	0x00
104*22ce4affSfengbojiang #define	CPU_REV_THUNDERX_1_1	0x01
105*22ce4affSfengbojiang 
106*22ce4affSfengbojiang #define	CPU_REV_THUNDERX2_0	0x00
107*22ce4affSfengbojiang 
108*22ce4affSfengbojiang /* APM / Ampere Part Number */
109*22ce4affSfengbojiang #define CPU_PART_EMAG8180	0x000
110a9643ea8Slogwang 
111a9643ea8Slogwang #define	CPU_IMPL(midr)	(((midr) >> 24) & 0xff)
112a9643ea8Slogwang #define	CPU_PART(midr)	(((midr) >> 4) & 0xfff)
113a9643ea8Slogwang #define	CPU_VAR(midr)	(((midr) >> 20) & 0xf)
114a9643ea8Slogwang #define	CPU_REV(midr)	(((midr) >> 0) & 0xf)
115a9643ea8Slogwang 
116a9643ea8Slogwang #define	CPU_IMPL_TO_MIDR(val)	(((val) & 0xff) << 24)
117a9643ea8Slogwang #define	CPU_PART_TO_MIDR(val)	(((val) & 0xfff) << 4)
118a9643ea8Slogwang #define	CPU_VAR_TO_MIDR(val)	(((val) & 0xf) << 20)
119a9643ea8Slogwang #define	CPU_REV_TO_MIDR(val)	(((val) & 0xf) << 0)
120a9643ea8Slogwang 
121a9643ea8Slogwang #define	CPU_IMPL_MASK	(0xff << 24)
122a9643ea8Slogwang #define	CPU_PART_MASK	(0xfff << 4)
123a9643ea8Slogwang #define	CPU_VAR_MASK	(0xf << 20)
124a9643ea8Slogwang #define	CPU_REV_MASK	(0xf << 0)
125a9643ea8Slogwang 
126a9643ea8Slogwang #define	CPU_ID_RAW(impl, part, var, rev)		\
127a9643ea8Slogwang     (CPU_IMPL_TO_MIDR((impl)) |				\
128a9643ea8Slogwang     CPU_PART_TO_MIDR((part)) | CPU_VAR_TO_MIDR((var)) |	\
129a9643ea8Slogwang     CPU_REV_TO_MIDR((rev)))
130a9643ea8Slogwang 
131a9643ea8Slogwang #define	CPU_MATCH(mask, impl, part, var, rev)		\
132a9643ea8Slogwang     (((mask) & PCPU_GET(midr)) ==			\
133a9643ea8Slogwang     ((mask) & CPU_ID_RAW((impl), (part), (var), (rev))))
134a9643ea8Slogwang 
135a9643ea8Slogwang #define	CPU_MATCH_RAW(mask, devid)			\
136a9643ea8Slogwang     (((mask) & PCPU_GET(midr)) == ((mask) & (devid)))
137a9643ea8Slogwang 
138a9643ea8Slogwang /*
139a9643ea8Slogwang  * Chip-specific errata. This defines are intended to be
140a9643ea8Slogwang  * booleans used within if statements. When an appropriate
141a9643ea8Slogwang  * kernel option is disabled, these defines must be defined
142a9643ea8Slogwang  * as 0 to allow the compiler to remove a dead code thus
143a9643ea8Slogwang  * produce better optimized kernel image.
144a9643ea8Slogwang  */
145a9643ea8Slogwang /*
146a9643ea8Slogwang  * Vendor:	Cavium
147a9643ea8Slogwang  * Chip:	ThunderX
148a9643ea8Slogwang  * Revision(s):	Pass 1.0, Pass 1.1
149a9643ea8Slogwang  */
150a9643ea8Slogwang #ifdef THUNDERX_PASS_1_1_ERRATA
151*22ce4affSfengbojiang #define	CPU_MATCH_ERRATA_CAVIUM_THUNDERX_1_1				\
152a9643ea8Slogwang     (CPU_MATCH(CPU_IMPL_MASK | CPU_PART_MASK | CPU_REV_MASK,		\
153*22ce4affSfengbojiang     CPU_IMPL_CAVIUM, CPU_PART_THUNDERX, 0, CPU_REV_THUNDERX_1_0) ||	\
154a9643ea8Slogwang     CPU_MATCH(CPU_IMPL_MASK | CPU_PART_MASK | CPU_REV_MASK,		\
155*22ce4affSfengbojiang     CPU_IMPL_CAVIUM, CPU_PART_THUNDERX, 0, CPU_REV_THUNDERX_1_1))
156a9643ea8Slogwang #else
157*22ce4affSfengbojiang #define	CPU_MATCH_ERRATA_CAVIUM_THUNDERX_1_1	0
158a9643ea8Slogwang #endif
159a9643ea8Slogwang 
160a9643ea8Slogwang extern char btext[];
161a9643ea8Slogwang extern char etext[];
162a9643ea8Slogwang 
163a9643ea8Slogwang extern uint64_t __cpu_affinity[];
164a9643ea8Slogwang 
165a9643ea8Slogwang void	cpu_halt(void) __dead2;
166a9643ea8Slogwang void	cpu_reset(void) __dead2;
167a9643ea8Slogwang void	fork_trampoline(void);
168*22ce4affSfengbojiang void	identify_cache(uint64_t);
169*22ce4affSfengbojiang void	identify_cpu(u_int);
170*22ce4affSfengbojiang void	install_cpu_errata(void);
171a9643ea8Slogwang void	swi_vm(void *v);
172a9643ea8Slogwang 
173*22ce4affSfengbojiang /* Functions to read the sanitised view of the special registers */
174*22ce4affSfengbojiang void	update_special_regs(u_int);
175*22ce4affSfengbojiang bool	extract_user_id_field(u_int, u_int, uint8_t *);
176*22ce4affSfengbojiang bool	get_kernel_reg(u_int, uint64_t *);
177*22ce4affSfengbojiang 
178a9643ea8Slogwang #define	CPU_AFFINITY(cpu)	__cpu_affinity[(cpu)]
179a9643ea8Slogwang #define	CPU_CURRENT_SOCKET				\
180a9643ea8Slogwang     (CPU_AFF2(CPU_AFFINITY(PCPU_GET(cpuid))))
181a9643ea8Slogwang 
182a9643ea8Slogwang static __inline uint64_t
get_cyclecount(void)183a9643ea8Slogwang get_cyclecount(void)
184a9643ea8Slogwang {
185a9643ea8Slogwang 	uint64_t ret;
186a9643ea8Slogwang 
187a9643ea8Slogwang 	ret = READ_SPECIALREG(cntvct_el0);
188a9643ea8Slogwang 
189a9643ea8Slogwang 	return (ret);
190a9643ea8Slogwang }
191a9643ea8Slogwang 
192a9643ea8Slogwang #define	ADDRESS_TRANSLATE_FUNC(stage)				\
193a9643ea8Slogwang static inline uint64_t						\
194a9643ea8Slogwang arm64_address_translate_ ##stage (uint64_t addr)		\
195a9643ea8Slogwang {								\
196a9643ea8Slogwang 	uint64_t ret;						\
197a9643ea8Slogwang 								\
198a9643ea8Slogwang 	__asm __volatile(					\
199a9643ea8Slogwang 	    "at " __STRING(stage) ", %1 \n"					\
200a9643ea8Slogwang 	    "mrs %0, par_el1" : "=r"(ret) : "r"(addr));		\
201a9643ea8Slogwang 								\
202a9643ea8Slogwang 	return (ret);						\
203a9643ea8Slogwang }
204a9643ea8Slogwang 
205a9643ea8Slogwang ADDRESS_TRANSLATE_FUNC(s1e0r)
206a9643ea8Slogwang ADDRESS_TRANSLATE_FUNC(s1e0w)
207a9643ea8Slogwang ADDRESS_TRANSLATE_FUNC(s1e1r)
208a9643ea8Slogwang ADDRESS_TRANSLATE_FUNC(s1e1w)
209a9643ea8Slogwang 
210a9643ea8Slogwang #endif
211a9643ea8Slogwang 
212a9643ea8Slogwang #endif /* !_MACHINE_CPU_H_ */
213