xref: /f-stack/freebsd/sys/systm.h (revision 22ce4aff)
1 /*-
2  * SPDX-License-Identifier: BSD-3-Clause
3  *
4  * Copyright (c) 1982, 1988, 1991, 1993
5  *	The Regents of the University of California.  All rights reserved.
6  * (c) UNIX System Laboratories, Inc.
7  * All or some portions of this file are derived from material licensed
8  * to the University of California by American Telephone and Telegraph
9  * Co. or Unix System Laboratories, Inc. and are reproduced herein with
10  * the permission of UNIX System Laboratories, Inc.
11  *
12  * Redistribution and use in source and binary forms, with or without
13  * modification, are permitted provided that the following conditions
14  * are met:
15  * 1. Redistributions of source code must retain the above copyright
16  *    notice, this list of conditions and the following disclaimer.
17  * 2. Redistributions in binary form must reproduce the above copyright
18  *    notice, this list of conditions and the following disclaimer in the
19  *    documentation and/or other materials provided with the distribution.
20  * 3. Neither the name of the University nor the names of its contributors
21  *    may be used to endorse or promote products derived from this software
22  *    without specific prior written permission.
23  *
24  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
25  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
26  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
27  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
28  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
29  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
30  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
31  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
32  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
33  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
34  * SUCH DAMAGE.
35  *
36  *	@(#)systm.h	8.7 (Berkeley) 3/29/95
37  * $FreeBSD$
38  */
39 
40 #ifndef _SYS_SYSTM_H_
41 #define	_SYS_SYSTM_H_
42 
43 #include <sys/cdefs.h>
44 #include <machine/atomic.h>
45 #include <machine/cpufunc.h>
46 #include <sys/callout.h>
47 #include <sys/queue.h>
48 #include <sys/stdint.h>		/* for people using printf mainly */
49 
50 __NULLABILITY_PRAGMA_PUSH
51 
52 #ifdef _KERNEL
53 extern int cold;		/* nonzero if we are doing a cold boot */
54 extern int suspend_blocked;	/* block suspend due to pending shutdown */
55 extern int rebooting;		/* kern_reboot() has been called. */
56 extern const char *panicstr;	/* panic message */
57 extern bool panicked;
58 #define	KERNEL_PANICKED()	__predict_false(panicked)
59 extern char version[];		/* system version */
60 extern char compiler_version[];	/* compiler version */
61 extern char copyright[];	/* system copyright */
62 extern int kstack_pages;	/* number of kernel stack pages */
63 
64 extern u_long pagesizes[];	/* supported page sizes */
65 extern long physmem;		/* physical memory */
66 extern long realmem;		/* 'real' memory */
67 
68 extern char *rootdevnames[2];	/* names of possible root devices */
69 
70 extern int boothowto;		/* reboot flags, from console subsystem */
71 extern int bootverbose;		/* nonzero to print verbose messages */
72 
73 extern int maxusers;		/* system tune hint */
74 extern int ngroups_max;		/* max # of supplemental groups */
75 extern int vm_guest;		/* Running as virtual machine guest? */
76 
77 extern u_long maxphys;		/* max raw I/O transfer size */
78 
79 /*
80  * Detected virtual machine guest types. The intention is to expand
81  * and/or add to the VM_GUEST_VM type if specific VM functionality is
82  * ever implemented (e.g. vendor-specific paravirtualization features).
83  * Keep in sync with vm_guest_sysctl_names[].
84  */
85 enum VM_GUEST { VM_GUEST_NO = 0, VM_GUEST_VM, VM_GUEST_XEN, VM_GUEST_HV,
86 		VM_GUEST_VMWARE, VM_GUEST_KVM, VM_GUEST_BHYVE, VM_GUEST_VBOX,
87 		VM_GUEST_PARALLELS, VM_LAST };
88 
89 #ifdef	INVARIANTS		/* The option is always available */
90 #define	VNASSERT(exp, vp, msg) do {					\
91 	if (__predict_false(!(exp))) {					\
92 		vn_printf(vp, "VNASSERT failed: %s not true at %s:%d (%s)\n",\
93 		   #exp, __FILE__, __LINE__, __func__);	 		\
94 		kassert_panic msg;					\
95 	}								\
96 } while (0)
97 #define	VNPASS(exp, vp)	do {						\
98 	const char *_exp = #exp;					\
99 	VNASSERT(exp, vp, ("condition %s not met at %s:%d (%s)",	\
100 	    _exp, __FILE__, __LINE__, __func__));			\
101 } while (0)
102 #define	__assert_unreachable() \
103 	panic("executing segment marked as unreachable at %s:%d (%s)\n", \
104 	    __FILE__, __LINE__, __func__)
105 #else
106 #define	VNASSERT(exp, vp, msg) do { \
107 } while (0)
108 #define	VNPASS(exp, vp) do { \
109 } while (0)
110 #define	__assert_unreachable()	__unreachable()
111 #endif
112 
113 #ifndef CTASSERT	/* Allow lint to override */
114 #define	CTASSERT(x)	_Static_assert(x, "compile-time assertion failed")
115 #endif
116 #endif /* KERNEL */
117 
118 /*
119  * These functions need to be declared before the KASSERT macro is invoked in
120  * !KASSERT_PANIC_OPTIONAL builds, so their declarations are sort of out of
121  * place compared to other function definitions in this header.  On the other
122  * hand, this header is a bit disorganized anyway.
123  */
124 void	panic(const char *, ...) __dead2 __printflike(1, 2);
125 void	vpanic(const char *, __va_list) __dead2 __printflike(1, 0);
126 
127 
128 #if defined(_STANDALONE)
129 struct ucred;
130 /*
131  * Until we have more experience with KASSERTS that are called
132  * from the boot loader, they are off. The bootloader does this
133  * a little differently than the kernel (we just call printf atm).
134  * we avoid most of the common functions in the boot loader, so
135  * declare printf() here too.
136  */
137 int	printf(const char *, ...) __printflike(1, 2);
138 #  define kassert_panic printf
139 #else /* !_STANDALONE */
140 #  if defined(WITNESS) || defined(INVARIANT_SUPPORT)
141 #    ifdef KASSERT_PANIC_OPTIONAL
142 void	kassert_panic(const char *fmt, ...)  __printflike(1, 2);
143 #    else
144 #      define kassert_panic	panic
145 #    endif /* KASSERT_PANIC_OPTIONAL */
146 #  endif /* defined(WITNESS) || defined(INVARIANT_SUPPORT) */
147 #endif /* _STANDALONE */
148 
149 #if defined(INVARIANTS) || defined(_STANDALONE)
150 #define	KASSERT(exp,msg) do {						\
151 	if (__predict_false(!(exp)))					\
152 		kassert_panic msg;					\
153 } while (0)
154 #else /* !INVARIANTS && !_STANDALONE */
155 #define	KASSERT(exp,msg) do { \
156 } while (0)
157 #endif /* INVARIANTS || _STANDALONE */
158 
159 /*
160  * Helpful macros for quickly coming up with assertions with informative
161  * panic messages.
162  */
163 #define MPASS(ex)		MPASS4(ex, #ex, __FILE__, __LINE__)
164 #define MPASS2(ex, what)	MPASS4(ex, what, __FILE__, __LINE__)
165 #define MPASS3(ex, file, line)	MPASS4(ex, #ex, file, line)
166 #define MPASS4(ex, what, file, line)					\
167 	KASSERT((ex), ("Assertion %s failed at %s:%d", what, file, line))
168 
169 /*
170  * Align variables.
171  */
172 #define	__read_mostly		__section(".data.read_mostly")
173 #define	__read_frequently	__section(".data.read_frequently")
174 #define	__exclusive_cache_line	__aligned(CACHE_LINE_SIZE) \
175 				    __section(".data.exclusive_cache_line")
176 #ifdef _KERNEL
177 #include <sys/param.h>		/* MAXCPU */
178 #include <sys/pcpu.h>		/* curthread */
179 #ifndef FSTACK
180 #include <sys/kpilite.h>
181 #endif
182 
183 /*
184  * Assert that a pointer can be loaded from memory atomically.
185  *
186  * This assertion enforces stronger alignment than necessary.  For example,
187  * on some architectures, atomicity for unaligned loads will depend on
188  * whether or not the load spans multiple cache lines.
189  */
190 #define	ASSERT_ATOMIC_LOAD_PTR(var, msg)				\
191 	KASSERT(sizeof(var) == sizeof(void *) &&			\
192 	    ((uintptr_t)&(var) & (sizeof(void *) - 1)) == 0, msg)
193 
194 /*
195  * Assert that a thread is in critical(9) section.
196  */
197 #define	CRITICAL_ASSERT(td)						\
198 	KASSERT((td)->td_critnest >= 1, ("Not in critical section"));
199 
200 /*
201  * If we have already panic'd and this is the thread that called
202  * panic(), then don't block on any mutexes but silently succeed.
203  * Otherwise, the kernel will deadlock since the scheduler isn't
204  * going to run the thread that holds any lock we need.
205  */
206 #define	SCHEDULER_STOPPED_TD(td)  ({					\
207 	MPASS((td) == curthread);					\
208 	__predict_false((td)->td_stopsched);				\
209 })
210 #define	SCHEDULER_STOPPED() SCHEDULER_STOPPED_TD(curthread)
211 
212 extern int osreldate;
213 
214 extern const void *zero_region;	/* address space maps to a zeroed page	*/
215 
216 extern int unmapped_buf_allowed;
217 
218 #ifdef __LP64__
219 #define	IOSIZE_MAX		iosize_max()
220 #define	DEVFS_IOSIZE_MAX	devfs_iosize_max()
221 #else
222 #define	IOSIZE_MAX		SSIZE_MAX
223 #define	DEVFS_IOSIZE_MAX	SSIZE_MAX
224 #endif
225 
226 /*
227  * General function declarations.
228  */
229 
230 struct inpcb;
231 struct lock_object;
232 struct malloc_type;
233 struct mtx;
234 struct proc;
235 struct socket;
236 struct thread;
237 struct tty;
238 struct ucred;
239 struct uio;
240 struct _jmp_buf;
241 struct trapframe;
242 struct eventtimer;
243 
244 int	setjmp(struct _jmp_buf *) __returns_twice;
245 void	longjmp(struct _jmp_buf *, int) __dead2;
246 int	dumpstatus(vm_offset_t addr, off_t count);
247 int	nullop(void);
248 int	eopnotsupp(void);
249 int	ureadc(int, struct uio *);
250 void	hashdestroy(void *, struct malloc_type *, u_long);
251 void	*hashinit(int count, struct malloc_type *type, u_long *hashmask);
252 void	*hashinit_flags(int count, struct malloc_type *type,
253     u_long *hashmask, int flags);
254 #define	HASH_NOWAIT	0x00000001
255 #define	HASH_WAITOK	0x00000002
256 
257 void	*phashinit(int count, struct malloc_type *type, u_long *nentries);
258 void	*phashinit_flags(int count, struct malloc_type *type, u_long *nentries,
259     int flags);
260 void	g_waitidle(void);
261 
262 void	cpu_flush_dcache(void *, size_t);
263 void	cpu_rootconf(void);
264 void	critical_enter_KBI(void);
265 void	critical_exit_KBI(void);
266 void	critical_exit_preempt(void);
267 void	init_param1(void);
268 void	init_param2(long physpages);
269 void	init_static_kenv(char *, size_t);
270 void	tablefull(const char *);
271 
272 /*
273  * Allocate per-thread "current" state in the linuxkpi
274  */
275 extern int (*lkpi_alloc_current)(struct thread *, int);
276 int linux_alloc_current_noop(struct thread *, int);
277 
278 #if defined(KLD_MODULE) || defined(KTR_CRITICAL) || !defined(_KERNEL) || defined(GENOFFSET)
279 #define critical_enter() critical_enter_KBI()
280 #define critical_exit() critical_exit_KBI()
281 #else
282 static __inline void
critical_enter(void)283 critical_enter(void)
284 {
285 #ifndef FSTACK
286 	struct thread_lite *td;
287 
288 	td = (struct thread_lite *)curthread;
289 	td->td_critnest++;
290 	atomic_interrupt_fence();
291 #endif
292 }
293 
294 static __inline void
critical_exit(void)295 critical_exit(void)
296 {
297 #ifndef FSTACK
298 	struct thread_lite *td;
299 
300 	td = (struct thread_lite *)curthread;
301 	KASSERT(td->td_critnest != 0,
302 	    ("critical_exit: td_critnest == 0"));
303 	atomic_interrupt_fence();
304 	td->td_critnest--;
305 	atomic_interrupt_fence();
306 	if (__predict_false(td->td_owepreempt))
307 		critical_exit_preempt();
308 #endif
309 }
310 #endif
311 
312 #ifdef  EARLY_PRINTF
313 typedef void early_putc_t(int ch);
314 extern early_putc_t *early_putc;
315 #endif
316 int	kvprintf(char const *, void (*)(int, void*), void *, int,
317 	    __va_list) __printflike(1, 0);
318 void	log(int, const char *, ...) __printflike(2, 3);
319 void	log_console(struct uio *);
320 void	vlog(int, const char *, __va_list) __printflike(2, 0);
321 int	asprintf(char **ret, struct malloc_type *mtp, const char *format,
322 	    ...) __printflike(3, 4);
323 int	printf(const char *, ...) __printflike(1, 2);
324 int	snprintf(char *, size_t, const char *, ...) __printflike(3, 4);
325 int	sprintf(char *buf, const char *, ...) __printflike(2, 3);
326 int	uprintf(const char *, ...) __printflike(1, 2);
327 int	vprintf(const char *, __va_list) __printflike(1, 0);
328 int	vasprintf(char **ret, struct malloc_type *mtp, const char *format,
329 	    __va_list ap) __printflike(3, 0);
330 int	vsnprintf(char *, size_t, const char *, __va_list) __printflike(3, 0);
331 int	vsnrprintf(char *, size_t, int, const char *, __va_list) __printflike(4, 0);
332 int	vsprintf(char *buf, const char *, __va_list) __printflike(2, 0);
333 int	sscanf(const char *, char const * _Nonnull, ...) __scanflike(2, 3);
334 int	vsscanf(const char * _Nonnull, char const * _Nonnull, __va_list)  __scanflike(2, 0);
335 long	strtol(const char *, char **, int);
336 u_long	strtoul(const char *, char **, int);
337 quad_t	strtoq(const char *, char **, int);
338 u_quad_t strtouq(const char *, char **, int);
339 void	tprintf(struct proc *p, int pri, const char *, ...) __printflike(3, 4);
340 void	vtprintf(struct proc *, int, const char *, __va_list) __printflike(3, 0);
341 void	hexdump(const void *ptr, int length, const char *hdr, int flags);
342 #define	HD_COLUMN_MASK	0xff
343 #define	HD_DELIM_MASK	0xff00
344 #define	HD_OMIT_COUNT	(1 << 16)
345 #define	HD_OMIT_HEX	(1 << 17)
346 #define	HD_OMIT_CHARS	(1 << 18)
347 
348 #define ovbcopy(f, t, l) bcopy((f), (t), (l))
349 void	bcopy(const void * _Nonnull from, void * _Nonnull to, size_t len);
350 void	bzero(void * _Nonnull buf, size_t len);
351 void	explicit_bzero(void * _Nonnull, size_t);
352 int	bcmp(const void *b1, const void *b2, size_t len);
353 
354 void	*memset(void * _Nonnull buf, int c, size_t len);
355 void	*memcpy(void * _Nonnull to, const void * _Nonnull from, size_t len);
356 void	*memmove(void * _Nonnull dest, const void * _Nonnull src, size_t n);
357 int	memcmp(const void *b1, const void *b2, size_t len);
358 
359 #ifdef KCSAN
360 void	*kcsan_memset(void *, int, size_t);
361 void	*kcsan_memcpy(void *, const void *, size_t);
362 void	*kcsan_memmove(void *, const void *, size_t);
363 int	kcsan_memcmp(const void *, const void *, size_t);
364 #define bcopy(from, to, len) kcsan_memmove((to), (from), (len))
365 #define bzero(buf, len) kcsan_memset((buf), 0, (len))
366 #define bcmp(b1, b2, len) kcsan_memcmp((b1), (b2), (len))
367 #define memset(buf, c, len) kcsan_memset((buf), (c), (len))
368 #define memcpy(to, from, len) kcsan_memcpy((to), (from), (len))
369 #define memmove(dest, src, n) kcsan_memmove((dest), (src), (n))
370 #define memcmp(b1, b2, len) kcsan_memcmp((b1), (b2), (len))
371 #else
372 #define bcopy(from, to, len) __builtin_memmove((to), (from), (len))
373 #define bzero(buf, len) __builtin_memset((buf), 0, (len))
374 #define bcmp(b1, b2, len) __builtin_memcmp((b1), (b2), (len))
375 #define memset(buf, c, len) __builtin_memset((buf), (c), (len))
376 #define memcpy(to, from, len) __builtin_memcpy((to), (from), (len))
377 #define memmove(dest, src, n) __builtin_memmove((dest), (src), (n))
378 #define memcmp(b1, b2, len) __builtin_memcmp((b1), (b2), (len))
379 #endif
380 
381 void	*memset_early(void * _Nonnull buf, int c, size_t len);
382 #define bzero_early(buf, len) memset_early((buf), 0, (len))
383 void	*memcpy_early(void * _Nonnull to, const void * _Nonnull from, size_t len);
384 void	*memmove_early(void * _Nonnull dest, const void * _Nonnull src, size_t n);
385 #define bcopy_early(from, to, len) memmove_early((to), (from), (len))
386 
387 #define	copystr(src, dst, len, outlen)	({			\
388 	size_t __r, __len, *__outlen;				\
389 								\
390 	__len = (len);						\
391 	__outlen = (outlen);					\
392 	__r = strlcpy((dst), (src), __len);			\
393 	if (__outlen != NULL)					\
394 		*__outlen = ((__r >= __len) ? __len : __r + 1);	\
395 	((__r >= __len) ? ENAMETOOLONG : 0);			\
396 })
397 
398 int	copyinstr(const void * __restrict udaddr,
399 	    void * _Nonnull __restrict kaddr, size_t len,
400 	    size_t * __restrict lencopied);
401 int	copyin(const void * __restrict udaddr,
402 	    void * _Nonnull __restrict kaddr, size_t len);
403 int	copyin_nofault(const void * __restrict udaddr,
404 	    void * _Nonnull __restrict kaddr, size_t len);
405 int	copyout(const void * _Nonnull __restrict kaddr,
406 	    void * __restrict udaddr, size_t len);
407 int	copyout_nofault(const void * _Nonnull __restrict kaddr,
408 	    void * __restrict udaddr, size_t len);
409 
410 #ifdef KCSAN
411 int	kcsan_copyin(const void *, void *, size_t);
412 int	kcsan_copyinstr(const void *, void *, size_t, size_t *);
413 int	kcsan_copyout(const void *, void *, size_t);
414 #define	copyin(u, k, l) kcsan_copyin((u), (k), (l))
415 #define	copyinstr(u, k, l, lc) kcsan_copyinstr((u), (k), (l), (lc))
416 #define	copyout(k, u, l) kcsan_copyout((k), (u), (l))
417 #endif
418 
419 int	fubyte(volatile const void *base);
420 long	fuword(volatile const void *base);
421 int	fuword16(volatile const void *base);
422 int32_t	fuword32(volatile const void *base);
423 int64_t	fuword64(volatile const void *base);
424 int	fueword(volatile const void *base, long *val);
425 int	fueword32(volatile const void *base, int32_t *val);
426 int	fueword64(volatile const void *base, int64_t *val);
427 int	subyte(volatile void *base, int byte);
428 int	suword(volatile void *base, long word);
429 int	suword16(volatile void *base, int word);
430 int	suword32(volatile void *base, int32_t word);
431 int	suword64(volatile void *base, int64_t word);
432 uint32_t casuword32(volatile uint32_t *base, uint32_t oldval, uint32_t newval);
433 u_long	casuword(volatile u_long *p, u_long oldval, u_long newval);
434 int	casueword32(volatile uint32_t *base, uint32_t oldval, uint32_t *oldvalp,
435 	    uint32_t newval);
436 int	casueword(volatile u_long *p, u_long oldval, u_long *oldvalp,
437 	    u_long newval);
438 
439 void	realitexpire(void *);
440 
441 int	sysbeep(int hertz, int period);
442 
443 void	hardclock(int cnt, int usermode);
444 void	hardclock_sync(int cpu);
445 void	softclock(void *);
446 void	statclock(int cnt, int usermode);
447 void	profclock(int cnt, int usermode, uintfptr_t pc);
448 
449 int	hardclockintr(void);
450 
451 void	startprofclock(struct proc *);
452 void	stopprofclock(struct proc *);
453 void	cpu_startprofclock(void);
454 void	cpu_stopprofclock(void);
455 void	suspendclock(void);
456 void	resumeclock(void);
457 sbintime_t 	cpu_idleclock(void);
458 void	cpu_activeclock(void);
459 void	cpu_new_callout(int cpu, sbintime_t bt, sbintime_t bt_opt);
460 void	cpu_et_frequency(struct eventtimer *et, uint64_t newfreq);
461 extern int	cpu_disable_c2_sleep;
462 extern int	cpu_disable_c3_sleep;
463 
464 char	*kern_getenv(const char *name);
465 void	freeenv(char *env);
466 int	getenv_int(const char *name, int *data);
467 int	getenv_uint(const char *name, unsigned int *data);
468 int	getenv_long(const char *name, long *data);
469 int	getenv_ulong(const char *name, unsigned long *data);
470 int	getenv_string(const char *name, char *data, int size);
471 int	getenv_int64(const char *name, int64_t *data);
472 int	getenv_uint64(const char *name, uint64_t *data);
473 int	getenv_quad(const char *name, quad_t *data);
474 int	getenv_bool(const char *name, bool *data);
475 bool	getenv_is_true(const char *name);
476 bool	getenv_is_false(const char *name);
477 int	kern_setenv(const char *name, const char *value);
478 int	kern_unsetenv(const char *name);
479 int	testenv(const char *name);
480 
481 int	getenv_array(const char *name, void *data, int size, int *psize,
482     int type_size, bool allow_signed);
483 #define	GETENV_UNSIGNED	false	/* negative numbers not allowed */
484 #define	GETENV_SIGNED	true	/* negative numbers allowed */
485 
486 typedef uint64_t (cpu_tick_f)(void);
487 void set_cputicker(cpu_tick_f *func, uint64_t freq, unsigned var);
488 extern cpu_tick_f *cpu_ticks;
489 uint64_t cpu_tickrate(void);
490 uint64_t cputick2usec(uint64_t tick);
491 
492 #include <sys/libkern.h>
493 
494 /* Initialize the world */
495 void	consinit(void);
496 void	cpu_initclocks(void);
497 void	cpu_initclocks_bsp(void);
498 void	cpu_initclocks_ap(void);
499 void	usrinfoinit(void);
500 
501 /* Finalize the world */
502 void	kern_reboot(int) __dead2;
503 void	shutdown_nice(int);
504 
505 /* Stubs for obsolete functions that used to be for interrupt management */
splhigh(void)506 static __inline intrmask_t	splhigh(void)		{ return 0; }
splimp(void)507 static __inline intrmask_t	splimp(void)		{ return 0; }
splnet(void)508 static __inline intrmask_t	splnet(void)		{ return 0; }
spltty(void)509 static __inline intrmask_t	spltty(void)		{ return 0; }
splx(intrmask_t ipl __unused)510 static __inline void		splx(intrmask_t ipl __unused)	{ return; }
511 
512 /*
513  * Common `proc' functions are declared here so that proc.h can be included
514  * less often.
515  */
516 int	_sleep(const void * _Nonnull chan, struct lock_object *lock, int pri,
517 	   const char *wmesg, sbintime_t sbt, sbintime_t pr, int flags);
518 #define	msleep(chan, mtx, pri, wmesg, timo)				\
519 	_sleep((chan), &(mtx)->lock_object, (pri), (wmesg),		\
520 	    tick_sbt * (timo), 0, C_HARDCLOCK)
521 #define	msleep_sbt(chan, mtx, pri, wmesg, bt, pr, flags)		\
522 	_sleep((chan), &(mtx)->lock_object, (pri), (wmesg), (bt), (pr),	\
523 	    (flags))
524 int	msleep_spin_sbt(const void * _Nonnull chan, struct mtx *mtx,
525 	    const char *wmesg, sbintime_t sbt, sbintime_t pr, int flags);
526 #define	msleep_spin(chan, mtx, wmesg, timo)				\
527 	msleep_spin_sbt((chan), (mtx), (wmesg), tick_sbt * (timo),	\
528 	    0, C_HARDCLOCK)
529 int	pause_sbt(const char *wmesg, sbintime_t sbt, sbintime_t pr,
530 	    int flags);
531 #define	pause(wmesg, timo)						\
532 	pause_sbt((wmesg), tick_sbt * (timo), 0, C_HARDCLOCK)
533 #define	pause_sig(wmesg, timo)						\
534 	pause_sbt((wmesg), tick_sbt * (timo), 0, C_HARDCLOCK | C_CATCH)
535 #define	tsleep(chan, pri, wmesg, timo)					\
536 	_sleep((chan), NULL, (pri), (wmesg), tick_sbt * (timo),		\
537 	    0, C_HARDCLOCK)
538 #define	tsleep_sbt(chan, pri, wmesg, bt, pr, flags)			\
539 	_sleep((chan), NULL, (pri), (wmesg), (bt), (pr), (flags))
540 void	wakeup(const void *chan);
541 void	wakeup_one(const void *chan);
542 void	wakeup_any(const void *chan);
543 
544 /*
545  * Common `struct cdev *' stuff are declared here to avoid #include poisoning
546  */
547 
548 struct cdev;
549 dev_t dev2udev(struct cdev *x);
550 const char *devtoname(struct cdev *cdev);
551 
552 #ifdef __LP64__
553 size_t	devfs_iosize_max(void);
554 size_t	iosize_max(void);
555 #endif
556 
557 int poll_no_poll(int events);
558 
559 /* XXX: Should be void nanodelay(u_int nsec); */
560 void	DELAY(int usec);
561 
562 /* Root mount holdback API */
563 struct root_hold_token {
564 	int				flags;
565 	const char			*who;
566 	TAILQ_ENTRY(root_hold_token)	list;
567 };
568 
569 struct root_hold_token *root_mount_hold(const char *identifier);
570 void root_mount_hold_token(const char *identifier, struct root_hold_token *h);
571 void root_mount_rel(struct root_hold_token *h);
572 int root_mounted(void);
573 
574 /*
575  * Unit number allocation API. (kern/subr_unit.c)
576  */
577 struct unrhdr;
578 struct unrhdr *new_unrhdr(int low, int high, struct mtx *mutex);
579 void init_unrhdr(struct unrhdr *uh, int low, int high, struct mtx *mutex);
580 void delete_unrhdr(struct unrhdr *uh);
581 void clear_unrhdr(struct unrhdr *uh);
582 void clean_unrhdr(struct unrhdr *uh);
583 void clean_unrhdrl(struct unrhdr *uh);
584 int alloc_unr(struct unrhdr *uh);
585 int alloc_unr_specific(struct unrhdr *uh, u_int item);
586 int alloc_unrl(struct unrhdr *uh);
587 void free_unr(struct unrhdr *uh, u_int item);
588 
589 #ifndef __LP64__
590 #define UNR64_LOCKED
591 #endif
592 
593 struct unrhdr64 {
594         uint64_t	counter;
595 };
596 
597 static __inline void
new_unrhdr64(struct unrhdr64 * unr64,uint64_t low)598 new_unrhdr64(struct unrhdr64 *unr64, uint64_t low)
599 {
600 
601 	unr64->counter = low;
602 }
603 
604 #ifdef UNR64_LOCKED
605 uint64_t alloc_unr64(struct unrhdr64 *);
606 #else
607 static __inline uint64_t
alloc_unr64(struct unrhdr64 * unr64)608 alloc_unr64(struct unrhdr64 *unr64)
609 {
610 
611 	return (atomic_fetchadd_64(&unr64->counter, 1));
612 }
613 #endif
614 
615 void	intr_prof_stack_use(struct thread *td, struct trapframe *frame);
616 
617 void counted_warning(unsigned *counter, const char *msg);
618 
619 /*
620  * APIs to manage deprecation and obsolescence.
621  */
622 struct device;
623 void _gone_in(int major, const char *msg);
624 void _gone_in_dev(struct device *dev, int major, const char *msg);
625 #ifdef NO_OBSOLETE_CODE
626 #define __gone_ok(m, msg)					 \
627 	_Static_assert(m < P_OSREL_MAJOR(__FreeBSD_version)),	 \
628 	    "Obsolete code: " msg);
629 #else
630 #define	__gone_ok(m, msg)
631 #endif
632 #define gone_in(major, msg)		__gone_ok(major, msg) _gone_in(major, msg)
633 #define gone_in_dev(dev, major, msg)	__gone_ok(major, msg) _gone_in_dev(dev, major, msg)
634 #endif /* _KERNEL */
635 
636 __NULLABILITY_PRAGMA_POP
637 
638 #endif /* !_SYS_SYSTM_H_ */
639