2 * SPDX-License-Identifier: BSD-3-Clause
4 * Copyright (c) 1982, 1988, 1991, 1993
5 * The Regents of the University of California. All rights reserved.
6 * (c) UNIX System Laboratories, Inc.
7 * All or some portions of this file are derived from material licensed
8 * to the University of California by American Telephone and Telegraph
9 * Co. or Unix System Laboratories, Inc. and are reproduced herein with
10 * the permission of UNIX System Laboratories, Inc.
12 * Redistribution and use in source and binary forms, with or without
13 * modification, are permitted provided that the following conditions
15 * 1. Redistributions of source code must retain the above copyright
16 * notice, this list of conditions and the following disclaimer.
17 * 2. Redistributions in binary form must reproduce the above copyright
18 * notice, this list of conditions and the following disclaimer in the
19 * documentation and/or other materials provided with the distribution.
20 * 3. Neither the name of the University nor the names of its contributors
21 * may be used to endorse or promote products derived from this software
22 * without specific prior written permission.
24 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
25 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
26 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
27 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
28 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
29 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
30 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
31 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
32 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
33 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
36 * @(#)systm.h 8.7 (Berkeley) 3/29/95
43 #include <sys/cdefs.h>
44 #include <machine/atomic.h>
45 #include <machine/cpufunc.h>
46 #include <sys/callout.h>
47 #include <sys/queue.h>
48 #include <sys/stdint.h> /* for people using printf mainly */
50 __NULLABILITY_PRAGMA_PUSH
53 extern int cold; /* nonzero if we are doing a cold boot */
54 extern int suspend_blocked; /* block suspend due to pending shutdown */
55 extern int rebooting; /* kern_reboot() has been called. */
56 extern const char *panicstr; /* panic message */
58 #define KERNEL_PANICKED() __predict_false(panicked)
59 extern char version[]; /* system version */
60 extern char compiler_version[]; /* compiler version */
61 extern char copyright[]; /* system copyright */
62 extern int kstack_pages; /* number of kernel stack pages */
64 extern u_long pagesizes[]; /* supported page sizes */
65 extern long physmem; /* physical memory */
66 extern long realmem; /* 'real' memory */
68 extern char *rootdevnames[2]; /* names of possible root devices */
70 extern int boothowto; /* reboot flags, from console subsystem */
71 extern int bootverbose; /* nonzero to print verbose messages */
73 extern int maxusers; /* system tune hint */
74 extern int ngroups_max; /* max # of supplemental groups */
75 extern int vm_guest; /* Running as virtual machine guest? */
77 extern u_long maxphys; /* max raw I/O transfer size */
80 * Detected virtual machine guest types. The intention is to expand
81 * and/or add to the VM_GUEST_VM type if specific VM functionality is
82 * ever implemented (e.g. vendor-specific paravirtualization features).
83 * Keep in sync with vm_guest_sysctl_names[].
85 enum VM_GUEST { VM_GUEST_NO = 0, VM_GUEST_VM, VM_GUEST_XEN, VM_GUEST_HV,
86 VM_GUEST_VMWARE, VM_GUEST_KVM, VM_GUEST_BHYVE, VM_GUEST_VBOX,
87 VM_GUEST_PARALLELS, VM_LAST };
89 #ifdef INVARIANTS /* The option is always available */
90 #define VNASSERT(exp, vp, msg) do { \
91 if (__predict_false(!(exp))) { \
92 vn_printf(vp, "VNASSERT failed: %s not true at %s:%d (%s)\n",\
93 #exp, __FILE__, __LINE__, __func__); \
97 #define VNPASS(exp, vp) do { \
98 const char *_exp = #exp; \
99 VNASSERT(exp, vp, ("condition %s not met at %s:%d (%s)", \
100 _exp, __FILE__, __LINE__, __func__)); \
102 #define __assert_unreachable() \
103 panic("executing segment marked as unreachable at %s:%d (%s)\n", \
104 __FILE__, __LINE__, __func__)
106 #define VNASSERT(exp, vp, msg) do { \
108 #define VNPASS(exp, vp) do { \
110 #define __assert_unreachable() __unreachable()
113 #ifndef CTASSERT /* Allow lint to override */
114 #define CTASSERT(x) _Static_assert(x, "compile-time assertion failed")
119 * These functions need to be declared before the KASSERT macro is invoked in
120 * !KASSERT_PANIC_OPTIONAL builds, so their declarations are sort of out of
121 * place compared to other function definitions in this header. On the other
122 * hand, this header is a bit disorganized anyway.
124 void panic(const char *, ...) __dead2 __printflike(1, 2);
125 void vpanic(const char *, __va_list) __dead2 __printflike(1, 0);
128 #if defined(_STANDALONE)
131 * Until we have more experience with KASSERTS that are called
132 * from the boot loader, they are off. The bootloader does this
133 * a little differently than the kernel (we just call printf atm).
134 * we avoid most of the common functions in the boot loader, so
135 * declare printf() here too.
137 int printf(const char *, ...) __printflike(1, 2);
138 # define kassert_panic printf
139 #else /* !_STANDALONE */
140 # if defined(WITNESS) || defined(INVARIANT_SUPPORT)
141 # ifdef KASSERT_PANIC_OPTIONAL
142 void kassert_panic(const char *fmt, ...) __printflike(1, 2);
144 # define kassert_panic panic
145 # endif /* KASSERT_PANIC_OPTIONAL */
146 # endif /* defined(WITNESS) || defined(INVARIANT_SUPPORT) */
147 #endif /* _STANDALONE */
149 #if defined(INVARIANTS) || defined(_STANDALONE)
150 #define KASSERT(exp,msg) do { \
151 if (__predict_false(!(exp))) \
154 #else /* !INVARIANTS && !_STANDALONE */
155 #define KASSERT(exp,msg) do { \
157 #endif /* INVARIANTS || _STANDALONE */
160 * Helpful macros for quickly coming up with assertions with informative
163 #define MPASS(ex) MPASS4(ex, #ex, __FILE__, __LINE__)
164 #define MPASS2(ex, what) MPASS4(ex, what, __FILE__, __LINE__)
165 #define MPASS3(ex, file, line) MPASS4(ex, #ex, file, line)
166 #define MPASS4(ex, what, file, line) \
167 KASSERT((ex), ("Assertion %s failed at %s:%d", what, file, line))
172 #define __read_mostly __section(".data.read_mostly")
173 #define __read_frequently __section(".data.read_frequently")
174 #define __exclusive_cache_line __aligned(CACHE_LINE_SIZE) \
175 __section(".data.exclusive_cache_line")
177 #include <sys/param.h> /* MAXCPU */
178 #include <sys/pcpu.h> /* curthread */
179 #include <sys/kpilite.h>
182 * Assert that a pointer can be loaded from memory atomically.
184 * This assertion enforces stronger alignment than necessary. For example,
185 * on some architectures, atomicity for unaligned loads will depend on
186 * whether or not the load spans multiple cache lines.
188 #define ASSERT_ATOMIC_LOAD_PTR(var, msg) \
189 KASSERT(sizeof(var) == sizeof(void *) && \
190 ((uintptr_t)&(var) & (sizeof(void *) - 1)) == 0, msg)
193 * Assert that a thread is in critical(9) section.
195 #define CRITICAL_ASSERT(td) \
196 KASSERT((td)->td_critnest >= 1, ("Not in critical section"))
199 * If we have already panic'd and this is the thread that called
200 * panic(), then don't block on any mutexes but silently succeed.
201 * Otherwise, the kernel will deadlock since the scheduler isn't
202 * going to run the thread that holds any lock we need.
204 #define SCHEDULER_STOPPED_TD(td) ({ \
205 MPASS((td) == curthread); \
206 __predict_false((td)->td_stopsched); \
208 #define SCHEDULER_STOPPED() SCHEDULER_STOPPED_TD(curthread)
210 extern int osreldate;
212 extern const void *zero_region; /* address space maps to a zeroed page */
214 extern int unmapped_buf_allowed;
217 #define IOSIZE_MAX iosize_max()
218 #define DEVFS_IOSIZE_MAX devfs_iosize_max()
220 #define IOSIZE_MAX SSIZE_MAX
221 #define DEVFS_IOSIZE_MAX SSIZE_MAX
225 * General function declarations.
242 int setjmp(struct _jmp_buf *) __returns_twice;
243 void longjmp(struct _jmp_buf *, int) __dead2;
244 int dumpstatus(vm_offset_t addr, off_t count);
246 int eopnotsupp(void);
247 int ureadc(int, struct uio *);
248 void hashdestroy(void *, struct malloc_type *, u_long);
249 void *hashinit(int count, struct malloc_type *type, u_long *hashmask);
250 void *hashinit_flags(int count, struct malloc_type *type,
251 u_long *hashmask, int flags);
252 #define HASH_NOWAIT 0x00000001
253 #define HASH_WAITOK 0x00000002
255 void *phashinit(int count, struct malloc_type *type, u_long *nentries);
256 void *phashinit_flags(int count, struct malloc_type *type, u_long *nentries,
258 void g_waitidle(void);
260 void cpu_flush_dcache(void *, size_t);
261 void cpu_rootconf(void);
262 void critical_enter_KBI(void);
263 void critical_exit_KBI(void);
264 void critical_exit_preempt(void);
265 void init_param1(void);
266 void init_param2(long physpages);
267 void init_static_kenv(char *, size_t);
268 void tablefull(const char *);
271 * Allocate per-thread "current" state in the linuxkpi
273 extern int (*lkpi_alloc_current)(struct thread *, int);
274 int linux_alloc_current_noop(struct thread *, int);
276 #if defined(KLD_MODULE) || defined(KTR_CRITICAL) || !defined(_KERNEL) || defined(GENOFFSET)
277 #define critical_enter() critical_enter_KBI()
278 #define critical_exit() critical_exit_KBI()
283 struct thread_lite *td;
285 td = (struct thread_lite *)curthread;
287 atomic_interrupt_fence();
293 struct thread_lite *td;
295 td = (struct thread_lite *)curthread;
296 KASSERT(td->td_critnest != 0,
297 ("critical_exit: td_critnest == 0"));
298 atomic_interrupt_fence();
300 atomic_interrupt_fence();
301 if (__predict_false(td->td_owepreempt))
302 critical_exit_preempt();
308 typedef void early_putc_t(int ch);
309 extern early_putc_t *early_putc;
311 int kvprintf(char const *, void (*)(int, void*), void *, int,
312 __va_list) __printflike(1, 0);
313 void log(int, const char *, ...) __printflike(2, 3);
314 void log_console(struct uio *);
315 void vlog(int, const char *, __va_list) __printflike(2, 0);
316 int asprintf(char **ret, struct malloc_type *mtp, const char *format,
317 ...) __printflike(3, 4);
318 int printf(const char *, ...) __printflike(1, 2);
319 int snprintf(char *, size_t, const char *, ...) __printflike(3, 4);
320 int sprintf(char *buf, const char *, ...) __printflike(2, 3);
321 int uprintf(const char *, ...) __printflike(1, 2);
322 int vprintf(const char *, __va_list) __printflike(1, 0);
323 int vasprintf(char **ret, struct malloc_type *mtp, const char *format,
324 __va_list ap) __printflike(3, 0);
325 int vsnprintf(char *, size_t, const char *, __va_list) __printflike(3, 0);
326 int vsnrprintf(char *, size_t, int, const char *, __va_list) __printflike(4, 0);
327 int vsprintf(char *buf, const char *, __va_list) __printflike(2, 0);
328 int sscanf(const char *, char const * _Nonnull, ...) __scanflike(2, 3);
329 int vsscanf(const char * _Nonnull, char const * _Nonnull, __va_list) __scanflike(2, 0);
330 long strtol(const char *, char **, int);
331 u_long strtoul(const char *, char **, int);
332 quad_t strtoq(const char *, char **, int);
333 u_quad_t strtouq(const char *, char **, int);
334 void tprintf(struct proc *p, int pri, const char *, ...) __printflike(3, 4);
335 void vtprintf(struct proc *, int, const char *, __va_list) __printflike(3, 0);
336 void hexdump(const void *ptr, int length, const char *hdr, int flags);
337 #define HD_COLUMN_MASK 0xff
338 #define HD_DELIM_MASK 0xff00
339 #define HD_OMIT_COUNT (1 << 16)
340 #define HD_OMIT_HEX (1 << 17)
341 #define HD_OMIT_CHARS (1 << 18)
343 #define ovbcopy(f, t, l) bcopy((f), (t), (l))
344 void explicit_bzero(void * _Nonnull, size_t);
346 void *memset(void * _Nonnull buf, int c, size_t len);
347 void *memcpy(void * _Nonnull to, const void * _Nonnull from, size_t len);
348 void *memmove(void * _Nonnull dest, const void * _Nonnull src, size_t n);
349 int memcmp(const void *b1, const void *b2, size_t len);
351 #ifdef SAN_NEEDS_INTERCEPTORS
352 #define SAN_INTERCEPTOR(func) \
353 __CONCAT(SAN_INTERCEPTOR_PREFIX, __CONCAT(_, func))
354 void *SAN_INTERCEPTOR(memset)(void *, int, size_t);
355 void *SAN_INTERCEPTOR(memcpy)(void *, const void *, size_t);
356 void *SAN_INTERCEPTOR(memmove)(void *, const void *, size_t);
357 int SAN_INTERCEPTOR(memcmp)(const void *, const void *, size_t);
359 #define bcopy(from, to, len) SAN_INTERCEPTOR(memmove)((to), (from), (len))
360 #define bzero(buf, len) SAN_INTERCEPTOR(memset)((buf), 0, (len))
361 #define bcmp(b1, b2, len) SAN_INTERCEPTOR(memcmp)((b1), (b2), (len))
362 #define memset(buf, c, len) SAN_INTERCEPTOR(memset)((buf), (c), (len))
363 #define memcpy(to, from, len) SAN_INTERCEPTOR(memcpy)((to), (from), (len))
364 #define memmove(dest, src, n) SAN_INTERCEPTOR(memmove)((dest), (src), (n))
365 #define memcmp(b1, b2, len) SAN_INTERCEPTOR(memcmp)((b1), (b2), (len))
366 #endif /* !SAN_RUNTIME */
367 #else /* !SAN_NEEDS_INTERCEPTORS */
368 #define bcopy(from, to, len) __builtin_memmove((to), (from), (len))
369 #define bzero(buf, len) __builtin_memset((buf), 0, (len))
370 #define bcmp(b1, b2, len) __builtin_memcmp((b1), (b2), (len))
371 #define memset(buf, c, len) __builtin_memset((buf), (c), (len))
372 #define memcpy(to, from, len) __builtin_memcpy((to), (from), (len))
373 #define memmove(dest, src, n) __builtin_memmove((dest), (src), (n))
374 #define memcmp(b1, b2, len) __builtin_memcmp((b1), (b2), (len))
375 #endif /* SAN_NEEDS_INTERCEPTORS */
377 void *memset_early(void * _Nonnull buf, int c, size_t len);
378 #define bzero_early(buf, len) memset_early((buf), 0, (len))
379 void *memcpy_early(void * _Nonnull to, const void * _Nonnull from, size_t len);
380 void *memmove_early(void * _Nonnull dest, const void * _Nonnull src, size_t n);
381 #define bcopy_early(from, to, len) memmove_early((to), (from), (len))
383 #define copystr(src, dst, len, outlen) ({ \
384 size_t __r, __len, *__outlen; \
387 __outlen = (outlen); \
388 __r = strlcpy((dst), (src), __len); \
389 if (__outlen != NULL) \
390 *__outlen = ((__r >= __len) ? __len : __r + 1); \
391 ((__r >= __len) ? ENAMETOOLONG : 0); \
394 int copyinstr(const void * __restrict udaddr,
395 void * _Nonnull __restrict kaddr, size_t len,
396 size_t * __restrict lencopied);
397 int copyin(const void * __restrict udaddr,
398 void * _Nonnull __restrict kaddr, size_t len);
399 int copyin_nofault(const void * __restrict udaddr,
400 void * _Nonnull __restrict kaddr, size_t len);
401 int copyout(const void * _Nonnull __restrict kaddr,
402 void * __restrict udaddr, size_t len);
403 int copyout_nofault(const void * _Nonnull __restrict kaddr,
404 void * __restrict udaddr, size_t len);
406 #ifdef SAN_NEEDS_INTERCEPTORS
407 int SAN_INTERCEPTOR(copyin)(const void *, void *, size_t);
408 int SAN_INTERCEPTOR(copyinstr)(const void *, void *, size_t, size_t *);
409 int SAN_INTERCEPTOR(copyout)(const void *, void *, size_t);
411 #define copyin(u, k, l) SAN_INTERCEPTOR(copyin)((u), (k), (l))
412 #define copyinstr(u, k, l, lc) SAN_INTERCEPTOR(copyinstr)((u), (k), (l), (lc))
413 #define copyout(k, u, l) SAN_INTERCEPTOR(copyout)((k), (u), (l))
414 #endif /* !SAN_RUNTIME */
415 #endif /* SAN_NEEDS_INTERCEPTORS */
417 int fubyte(volatile const void *base);
418 long fuword(volatile const void *base);
419 int fuword16(volatile const void *base);
420 int32_t fuword32(volatile const void *base);
421 int64_t fuword64(volatile const void *base);
422 int fueword(volatile const void *base, long *val);
423 int fueword32(volatile const void *base, int32_t *val);
424 int fueword64(volatile const void *base, int64_t *val);
425 int subyte(volatile void *base, int byte);
426 int suword(volatile void *base, long word);
427 int suword16(volatile void *base, int word);
428 int suword32(volatile void *base, int32_t word);
429 int suword64(volatile void *base, int64_t word);
430 uint32_t casuword32(volatile uint32_t *base, uint32_t oldval, uint32_t newval);
431 u_long casuword(volatile u_long *p, u_long oldval, u_long newval);
432 int casueword32(volatile uint32_t *base, uint32_t oldval, uint32_t *oldvalp,
434 int casueword(volatile u_long *p, u_long oldval, u_long *oldvalp,
437 #if defined(SAN_NEEDS_INTERCEPTORS) && !defined(KCSAN)
438 int SAN_INTERCEPTOR(fubyte)(volatile const void *base);
439 int SAN_INTERCEPTOR(fuword16)(volatile const void *base);
440 int SAN_INTERCEPTOR(fueword)(volatile const void *base, long *val);
441 int SAN_INTERCEPTOR(fueword32)(volatile const void *base, int32_t *val);
442 int SAN_INTERCEPTOR(fueword64)(volatile const void *base, int64_t *val);
443 int SAN_INTERCEPTOR(subyte)(volatile void *base, int byte);
444 int SAN_INTERCEPTOR(suword)(volatile void *base, long word);
445 int SAN_INTERCEPTOR(suword16)(volatile void *base, int word);
446 int SAN_INTERCEPTOR(suword32)(volatile void *base, int32_t word);
447 int SAN_INTERCEPTOR(suword64)(volatile void *base, int64_t word);
448 int SAN_INTERCEPTOR(casueword32)(volatile uint32_t *base, uint32_t oldval,
449 uint32_t *oldvalp, uint32_t newval);
450 int SAN_INTERCEPTOR(casueword)(volatile u_long *p, u_long oldval,
451 u_long *oldvalp, u_long newval);
453 #define fubyte(b) SAN_INTERCEPTOR(fubyte)((b))
454 #define fuword16(b) SAN_INTERCEPTOR(fuword16)((b))
455 #define fueword(b, v) SAN_INTERCEPTOR(fueword)((b), (v))
456 #define fueword32(b, v) SAN_INTERCEPTOR(fueword32)((b), (v))
457 #define fueword64(b, v) SAN_INTERCEPTOR(fueword64)((b), (v))
458 #define subyte(b, w) SAN_INTERCEPTOR(subyte)((b), (w))
459 #define suword(b, w) SAN_INTERCEPTOR(suword)((b), (w))
460 #define suword16(b, w) SAN_INTERCEPTOR(suword16)((b), (w))
461 #define suword32(b, w) SAN_INTERCEPTOR(suword32)((b), (w))
462 #define suword64(b, w) SAN_INTERCEPTOR(suword64)((b), (w))
463 #define casueword32(b, o, p, n) SAN_INTERCEPTOR(casueword32)((b), (o), (p), (n))
464 #define casueword(b, o, p, n) SAN_INTERCEPTOR(casueword)((b), (o), (p), (n))
465 #endif /* !SAN_RUNTIME */
466 #endif /* SAN_NEEDS_INTERCEPTORS && !KCSAN */
468 void realitexpire(void *);
470 int sysbeep(int hertz, int period);
472 void hardclock(int cnt, int usermode);
473 void hardclock_sync(int cpu);
474 void softclock(void *);
475 void statclock(int cnt, int usermode);
476 void profclock(int cnt, int usermode, uintfptr_t pc);
478 int hardclockintr(void);
480 void startprofclock(struct proc *);
481 void stopprofclock(struct proc *);
482 void cpu_startprofclock(void);
483 void cpu_stopprofclock(void);
484 void suspendclock(void);
485 void resumeclock(void);
486 sbintime_t cpu_idleclock(void);
487 void cpu_activeclock(void);
488 void cpu_new_callout(int cpu, sbintime_t bt, sbintime_t bt_opt);
489 void cpu_et_frequency(struct eventtimer *et, uint64_t newfreq);
490 extern int cpu_disable_c2_sleep;
491 extern int cpu_disable_c3_sleep;
493 char *kern_getenv(const char *name);
494 void freeenv(char *env);
495 int getenv_int(const char *name, int *data);
496 int getenv_uint(const char *name, unsigned int *data);
497 int getenv_long(const char *name, long *data);
498 int getenv_ulong(const char *name, unsigned long *data);
499 int getenv_string(const char *name, char *data, int size);
500 int getenv_int64(const char *name, int64_t *data);
501 int getenv_uint64(const char *name, uint64_t *data);
502 int getenv_quad(const char *name, quad_t *data);
503 int getenv_bool(const char *name, bool *data);
504 bool getenv_is_true(const char *name);
505 bool getenv_is_false(const char *name);
506 int kern_setenv(const char *name, const char *value);
507 int kern_unsetenv(const char *name);
508 int testenv(const char *name);
510 int getenv_array(const char *name, void *data, int size, int *psize,
511 int type_size, bool allow_signed);
512 #define GETENV_UNSIGNED false /* negative numbers not allowed */
513 #define GETENV_SIGNED true /* negative numbers allowed */
515 typedef uint64_t (cpu_tick_f)(void);
516 void set_cputicker(cpu_tick_f *func, uint64_t freq, unsigned var);
517 extern cpu_tick_f *cpu_ticks;
518 uint64_t cpu_tickrate(void);
519 uint64_t cputick2usec(uint64_t tick);
521 #include <sys/libkern.h>
523 /* Initialize the world */
525 void cpu_initclocks(void);
526 void cpu_initclocks_bsp(void);
527 void cpu_initclocks_ap(void);
528 void usrinfoinit(void);
530 /* Finalize the world */
531 void kern_reboot(int) __dead2;
532 void shutdown_nice(int);
534 /* Stubs for obsolete functions that used to be for interrupt management */
535 static __inline intrmask_t splhigh(void) { return 0; }
536 static __inline intrmask_t splimp(void) { return 0; }
537 static __inline intrmask_t splnet(void) { return 0; }
538 static __inline intrmask_t spltty(void) { return 0; }
539 static __inline void splx(intrmask_t ipl __unused) { return; }
542 * Common `proc' functions are declared here so that proc.h can be included
545 int _sleep(const void * _Nonnull chan, struct lock_object *lock, int pri,
546 const char *wmesg, sbintime_t sbt, sbintime_t pr, int flags);
547 #define msleep(chan, mtx, pri, wmesg, timo) \
548 _sleep((chan), &(mtx)->lock_object, (pri), (wmesg), \
549 tick_sbt * (timo), 0, C_HARDCLOCK)
550 #define msleep_sbt(chan, mtx, pri, wmesg, bt, pr, flags) \
551 _sleep((chan), &(mtx)->lock_object, (pri), (wmesg), (bt), (pr), \
553 int msleep_spin_sbt(const void * _Nonnull chan, struct mtx *mtx,
554 const char *wmesg, sbintime_t sbt, sbintime_t pr, int flags);
555 #define msleep_spin(chan, mtx, wmesg, timo) \
556 msleep_spin_sbt((chan), (mtx), (wmesg), tick_sbt * (timo), \
558 int pause_sbt(const char *wmesg, sbintime_t sbt, sbintime_t pr,
560 #define pause(wmesg, timo) \
561 pause_sbt((wmesg), tick_sbt * (timo), 0, C_HARDCLOCK)
562 #define pause_sig(wmesg, timo) \
563 pause_sbt((wmesg), tick_sbt * (timo), 0, C_HARDCLOCK | C_CATCH)
564 #define tsleep(chan, pri, wmesg, timo) \
565 _sleep((chan), NULL, (pri), (wmesg), tick_sbt * (timo), \
567 #define tsleep_sbt(chan, pri, wmesg, bt, pr, flags) \
568 _sleep((chan), NULL, (pri), (wmesg), (bt), (pr), (flags))
569 void wakeup(const void *chan);
570 void wakeup_one(const void *chan);
571 void wakeup_any(const void *chan);
574 * Common `struct cdev *' stuff are declared here to avoid #include poisoning
578 dev_t dev2udev(struct cdev *x);
579 const char *devtoname(struct cdev *cdev);
582 size_t devfs_iosize_max(void);
583 size_t iosize_max(void);
586 int poll_no_poll(int events);
588 /* XXX: Should be void nanodelay(u_int nsec); */
589 void DELAY(int usec);
591 /* Root mount holdback API */
592 struct root_hold_token {
595 TAILQ_ENTRY(root_hold_token) list;
598 struct root_hold_token *root_mount_hold(const char *identifier);
599 void root_mount_hold_token(const char *identifier, struct root_hold_token *h);
600 void root_mount_rel(struct root_hold_token *h);
601 int root_mounted(void);
604 * Unit number allocation API. (kern/subr_unit.c)
607 struct unrhdr *new_unrhdr(int low, int high, struct mtx *mutex);
608 void init_unrhdr(struct unrhdr *uh, int low, int high, struct mtx *mutex);
609 void delete_unrhdr(struct unrhdr *uh);
610 void clear_unrhdr(struct unrhdr *uh);
611 void clean_unrhdr(struct unrhdr *uh);
612 void clean_unrhdrl(struct unrhdr *uh);
613 int alloc_unr(struct unrhdr *uh);
614 int alloc_unr_specific(struct unrhdr *uh, u_int item);
615 int alloc_unrl(struct unrhdr *uh);
616 void free_unr(struct unrhdr *uh, u_int item);
627 new_unrhdr64(struct unrhdr64 *unr64, uint64_t low)
630 unr64->counter = low;
634 uint64_t alloc_unr64(struct unrhdr64 *);
636 static __inline uint64_t
637 alloc_unr64(struct unrhdr64 *unr64)
640 return (atomic_fetchadd_64(&unr64->counter, 1));
644 void intr_prof_stack_use(struct thread *td, struct trapframe *frame);
646 void counted_warning(unsigned *counter, const char *msg);
649 * APIs to manage deprecation and obsolescence.
651 void _gone_in(int major, const char *msg);
652 void _gone_in_dev(device_t dev, int major, const char *msg);
653 #ifdef NO_OBSOLETE_CODE
654 #define __gone_ok(m, msg) \
655 _Static_assert(m < P_OSREL_MAJOR(__FreeBSD_version)), \
656 "Obsolete code: " msg);
658 #define __gone_ok(m, msg)
660 #define gone_in(major, msg) __gone_ok(major, msg) _gone_in(major, msg)
661 #define gone_in_dev(dev, major, msg) __gone_ok(major, msg) _gone_in_dev(dev, major, msg)
663 #if defined(INVARIANTS) || defined(WITNESS)
666 #define __diagused __unused
671 __NULLABILITY_PRAGMA_POP
672 #endif /* !_SYS_SYSTM_H_ */