mirror of
https://github.com/freebsd/freebsd-src.git
synced 2024-12-04 03:49:02 +00:00
105 lines
3.1 KiB
C
105 lines
3.1 KiB
C
|
#ifndef _MACHINE_IPL_H_
|
||
|
#define _MACHINE_IPL_H_
|
||
|
|
||
|
#include "machine/../isa/ipl.h" /* XXX "machine" means cpu for i386 */
|
||
|
|
||
|
/*
|
||
|
* Software interrupt bit numbers in priority order. The priority only
|
||
|
* determines which swi will be dispatched next; a higher priority swi
|
||
|
* may be dispatched when a nested h/w interrupt handler returns.
|
||
|
*/
|
||
|
#define SWI_TTY (NHWI + 0)
|
||
|
#define SWI_NET (NHWI + 1)
|
||
|
#define SWI_CLOCK 30
|
||
|
#define SWI_AST 31
|
||
|
|
||
|
/*
|
||
|
* Corresponding interrupt-pending bits for ipending.
|
||
|
*/
|
||
|
#define SWI_TTY_PENDING (1 << SWI_TTY)
|
||
|
#define SWI_NET_PENDING (1 << SWI_NET)
|
||
|
#define SWI_CLOCK_PENDING (1 << SWI_CLOCK)
|
||
|
#define SWI_AST_PENDING (1 << SWI_AST)
|
||
|
|
||
|
/*
|
||
|
* Corresponding interrupt-disable masks for cpl. The ordering is now by
|
||
|
* inclusion (where each mask is considered as a set of bits). Everything
|
||
|
* except SWI_AST_MASK includes SWI_CLOCK_MASK so that softclock() doesn't
|
||
|
* run while other swi handlers are running and timeout routines can call
|
||
|
* swi handlers. Everything includes SWI_AST_MASK so that AST's are masked
|
||
|
* until just before return to user mode.
|
||
|
*/
|
||
|
#define SWI_TTY_MASK (SWI_TTY_PENDING | SWI_CLOCK_MASK)
|
||
|
#define SWI_NET_MASK (SWI_NET_PENDING | SWI_CLOCK_MASK)
|
||
|
#define SWI_CLOCK_MASK (SWI_CLOCK_PENDING | SWI_AST_MASK)
|
||
|
#define SWI_AST_MASK SWI_AST_PENDING
|
||
|
#define SWI_MASK (~HWI_MASK)
|
||
|
|
||
|
#ifndef LOCORE
|
||
|
|
||
|
extern unsigned bio_imask; /* group of interrupts masked with splbio() */
|
||
|
extern unsigned cpl; /* current priority level mask */
|
||
|
extern unsigned high_imask; /* group of interrupts masked with splhigh() */
|
||
|
extern unsigned net_imask; /* group of interrupts masked with splimp() */
|
||
|
extern volatile unsigned ipending; /* active interrupts masked by cpl */
|
||
|
extern volatile unsigned netisr;
|
||
|
extern unsigned tty_imask; /* group of interrupts masked with spltty() */
|
||
|
|
||
|
/*
|
||
|
* ipending has to be volatile so that it is read every time it is accessed
|
||
|
* in splx() and spl0(), but we don't want it to be read nonatomically when
|
||
|
* it is changed. Pretending that ipending is a plain int happens to give
|
||
|
* suitable atomic code for "ipending |= constant;".
|
||
|
*/
|
||
|
#define setsoftast() (*(unsigned *)&ipending |= SWI_AST_PENDING)
|
||
|
#define setsoftclock() (*(unsigned *)&ipending |= SWI_CLOCK_PENDING)
|
||
|
#define setsoftnet() (*(unsigned *)&ipending |= SWI_NET_PENDING)
|
||
|
#define setsofttty() (*(unsigned *)&ipending |= SWI_TTY_PENDING)
|
||
|
|
||
|
void unpend_V __P((void));
|
||
|
|
||
|
#ifdef __GNUC__
|
||
|
|
||
|
void splz __P((void));
|
||
|
|
||
|
#define GENSPL(name, set_cpl) \
|
||
|
static __inline int name(void) \
|
||
|
{ \
|
||
|
unsigned x; \
|
||
|
\
|
||
|
x = cpl; \
|
||
|
set_cpl; \
|
||
|
return (x); \
|
||
|
}
|
||
|
|
||
|
GENSPL(splbio, cpl |= bio_imask)
|
||
|
GENSPL(splclock, cpl = HWI_MASK | SWI_MASK)
|
||
|
GENSPL(splhigh, cpl = HWI_MASK | SWI_MASK)
|
||
|
GENSPL(splimp, cpl |= net_imask)
|
||
|
GENSPL(splnet, cpl |= SWI_NET_MASK)
|
||
|
GENSPL(splsoftclock, cpl = SWI_CLOCK_MASK)
|
||
|
GENSPL(splsofttty, cpl |= SWI_TTY_MASK)
|
||
|
GENSPL(spltty, cpl |= tty_imask)
|
||
|
|
||
|
static __inline void
|
||
|
spl0(void)
|
||
|
{
|
||
|
cpl = SWI_AST_MASK;
|
||
|
if (ipending & ~SWI_AST_MASK)
|
||
|
splz();
|
||
|
}
|
||
|
|
||
|
static __inline void
|
||
|
splx(int ipl)
|
||
|
{
|
||
|
cpl = ipl;
|
||
|
if (ipending & ~ipl)
|
||
|
splz();
|
||
|
}
|
||
|
|
||
|
#endif /* __GNUC__ */
|
||
|
|
||
|
#endif /* LOCORE */
|
||
|
|
||
|
#endif /* _MACHINE_IPL_H_ */
|