|  | /* Copyright (c) 2010-13 The Regents of the University of California | 
|  | * Barret Rhoden <brho@cs.berkeley.edu> | 
|  | * See LICENSE for details. | 
|  | * | 
|  | * Kernel threading.  These are for blocking within the kernel for whatever | 
|  | * reason, usually during blocking IO operations.  Check out | 
|  | * Documentation/kthreads.txt for more info than you care about. */ | 
|  |  | 
|  | #pragma once | 
|  |  | 
|  | #include <ros/common.h> | 
|  | #include <ros/syscall.h> | 
|  | #include <trap.h> | 
|  | #include <sys/queue.h> | 
|  | #include <atomic.h> | 
|  | #include <setjmp.h> | 
|  |  | 
|  | struct errbuf { | 
|  | struct jmpbuf jmpbuf; | 
|  | }; | 
|  |  | 
|  | struct proc; | 
|  | struct kthread; | 
|  | struct kth_db_info; | 
|  | TAILQ_HEAD(kthread_tailq, kthread); | 
|  | TAILQ_HEAD(kth_db_tailq, kth_db_info); | 
|  |  | 
|  | #define GENBUF_SZ 128	/* plan9 uses this as a scratch space, per syscall */ | 
|  |  | 
|  | #define KTH_IS_KTASK			(1 << 0) | 
|  | #define KTH_SAVE_ADDR_SPACE		(1 << 1) | 
|  | #define KTH_IS_RCU_KTASK		(1 << 2) | 
|  |  | 
|  | /* These flag sets are for toggling between ktasks and default/process ktasks */ | 
|  | /* These are the flags for *any* ktask */ | 
|  | #define KTH_KTASK_FLAGS			(KTH_IS_KTASK) | 
|  | /* These are the flags used for normal process context */ | 
|  | #define KTH_DEFAULT_FLAGS		(KTH_SAVE_ADDR_SPACE) | 
|  |  | 
|  | /* This captures the essence of a kernel context that we want to suspend.  When | 
|  | * a kthread is running, we make sure its stacktop is the default kernel stack, | 
|  | * meaning it will receive the interrupts from userspace. */ | 
|  | struct kthread { | 
|  | struct jmpbuf			context; | 
|  | uintptr_t			stacktop; | 
|  | struct proc			*proc; | 
|  | struct syscall			*sysc; | 
|  | struct errbuf			*errbuf; | 
|  | TAILQ_ENTRY(kthread)		link; | 
|  | /* ID, other shit, etc */ | 
|  | int				flags; | 
|  | char				*name; | 
|  | char				generic_buf[GENBUF_SZ]; | 
|  | int				errno; | 
|  | char				errstr[MAX_ERRSTR_LEN]; | 
|  | struct systrace_record		*strace; | 
|  | }; | 
|  |  | 
|  | #define KTH_DB_SEM			1 | 
|  | #define KTH_DB_CV			2 | 
|  |  | 
|  | #ifdef CONFIG_SEMAPHORE_DEBUG | 
|  |  | 
|  | struct kth_db_info { | 
|  | TAILQ_ENTRY(kth_db_info)	link; | 
|  | unsigned int			type; | 
|  | bool				on_list; | 
|  | }; | 
|  |  | 
|  | #define KTH_DB_INIT .db         = { .type = KTH_DB_SEM }, | 
|  |  | 
|  | #else | 
|  |  | 
|  | struct kth_db_info { | 
|  | }; | 
|  |  | 
|  | #define KTH_DB_INIT | 
|  |  | 
|  | #endif | 
|  |  | 
|  |  | 
|  | /* Semaphore for kthreads to sleep on.  0 or less means you need to sleep */ | 
|  | struct semaphore { | 
|  | struct kth_db_info		db; | 
|  | struct kthread_tailq		waiters; | 
|  | int 				nr_signals; | 
|  | spinlock_t 			lock; | 
|  | }; | 
|  |  | 
|  | #define SEMAPHORE_INITIALIZER(name, n)                                         \ | 
|  | {                                                                              \ | 
|  | .waiters    = TAILQ_HEAD_INITIALIZER((name).waiters),                      \ | 
|  | .nr_signals = (n),                                                         \ | 
|  | .lock       = SPINLOCK_INITIALIZER,                                        \ | 
|  | KTH_DB_INIT                                                                \ | 
|  | } | 
|  |  | 
|  | #define SEMAPHORE_INITIALIZER_IRQSAVE(name, n)                                 \ | 
|  | {                                                                              \ | 
|  | .waiters    = TAILQ_HEAD_INITIALIZER((name).waiters),                      \ | 
|  | .nr_signals = (n),                                                         \ | 
|  | .lock       = SPINLOCK_INITIALIZER_IRQSAVE,                                \ | 
|  | KTH_DB_INIT                                                                \ | 
|  | } | 
|  |  | 
|  | struct cond_var { | 
|  | struct kth_db_info		db; | 
|  | struct kthread_tailq		waiters; | 
|  | spinlock_t 			*lock;	/* usually points to internal */ | 
|  | spinlock_t 			internal_lock; | 
|  | unsigned long			nr_waiters; | 
|  | }; | 
|  |  | 
|  | struct cv_lookup_elm { | 
|  | TAILQ_ENTRY(cv_lookup_elm)	link; | 
|  | TAILQ_ENTRY(cv_lookup_elm)	abortall_link; | 
|  | struct cond_var			*cv; | 
|  | struct kthread			*kthread; | 
|  | struct syscall			*sysc; | 
|  | struct proc			*proc; | 
|  | atomic_t			abort_in_progress;	/* 0 = no */ | 
|  | }; | 
|  | TAILQ_HEAD(cv_lookup_tailq, cv_lookup_elm); | 
|  |  | 
|  | uintptr_t get_kstack(void); | 
|  | void put_kstack(uintptr_t stacktop); | 
|  | uintptr_t *kstack_bottom_addr(uintptr_t stacktop); | 
|  | void kthread_init(void); | 
|  | struct kthread *__kthread_zalloc(void); | 
|  | void __use_real_kstack(void (*f)(void *arg)); | 
|  | void restart_kthread(struct kthread *kthread); | 
|  | void kthread_runnable(struct kthread *kthread); | 
|  | void kthread_yield(void); | 
|  | void kthread_usleep(uint64_t usec); | 
|  | void ktask(char *name, void (*fn)(void*), void *arg); | 
|  |  | 
|  | static inline bool is_ktask(struct kthread *kthread) | 
|  | { | 
|  | return kthread->flags & KTH_IS_KTASK; | 
|  | } | 
|  |  | 
|  | static inline bool is_rcu_ktask(struct kthread *kthread) | 
|  | { | 
|  | return kthread->flags & KTH_IS_RCU_KTASK; | 
|  | } | 
|  |  | 
|  | void sem_init(struct semaphore *sem, int signals); | 
|  | void sem_init_irqsave(struct semaphore *sem, int signals); | 
|  | bool sem_trydown_bulk(struct semaphore *sem, int nr_signals); | 
|  | bool sem_trydown(struct semaphore *sem); | 
|  | void sem_down_bulk(struct semaphore *sem, int nr_signals); | 
|  | void sem_down(struct semaphore *sem); | 
|  | bool sem_up(struct semaphore *sem); | 
|  | bool sem_trydown_bulk_irqsave(struct semaphore *sem, int nr_signals); | 
|  | bool sem_trydown_irqsave(struct semaphore *sem); | 
|  | void sem_down_bulk_irqsave(struct semaphore *sem, int nr_signals); | 
|  | void sem_down_irqsave(struct semaphore *sem); | 
|  | bool sem_up_irqsave(struct semaphore *sem); | 
|  | void print_db_blk_info(pid_t pid); | 
|  |  | 
|  | void cv_init(struct cond_var *cv); | 
|  | void cv_init_irqsave(struct cond_var *cv); | 
|  | void cv_init_with_lock(struct cond_var *cv, spinlock_t *lock); | 
|  | void cv_init_irqsave_with_lock(struct cond_var *cv, spinlock_t *lock); | 
|  | void cv_lock(struct cond_var *cv); | 
|  | void cv_unlock(struct cond_var *cv); | 
|  | void cv_lock_irqsave(struct cond_var *cv, int8_t *irq_state); | 
|  | void cv_unlock_irqsave(struct cond_var *cv, int8_t *irq_state); | 
|  | void cv_wait_and_unlock(struct cond_var *cv);	/* does not mess with irqs */ | 
|  | void cv_wait(struct cond_var *cv); | 
|  | void __cv_signal(struct cond_var *cv); | 
|  | void __cv_broadcast(struct cond_var *cv); | 
|  | void cv_signal(struct cond_var *cv); | 
|  | void cv_broadcast(struct cond_var *cv); | 
|  | void cv_signal_irqsave(struct cond_var *cv, int8_t *irq_state); | 
|  | void cv_broadcast_irqsave(struct cond_var *cv, int8_t *irq_state); | 
|  |  | 
|  | bool abort_sysc(struct proc *p, uintptr_t sysc); | 
|  | void abort_all_sysc(struct proc *p); | 
|  | int abort_all_sysc_fd(struct proc *p, int fd); | 
|  | void __reg_abortable_cv(struct cv_lookup_elm *cle, struct cond_var *cv); | 
|  | void dereg_abortable_cv(struct cv_lookup_elm *cle); | 
|  | bool should_abort(struct cv_lookup_elm *cle); | 
|  |  | 
|  | uintptr_t switch_to_ktask(void); | 
|  | void switch_back_from_ktask(uintptr_t old_ret); | 
|  |  | 
|  | /* qlocks are plan9's binary sempahore, which are wrappers around our sems. | 
|  | * Not sure if they'll need irqsave or normal sems. */ | 
|  | typedef struct semaphore qlock_t; | 
|  | #define qlock_init(x) sem_init((x), 1) | 
|  | #define qlock(x) sem_down(x) | 
|  | #define qunlock(x) sem_up(x) | 
|  | #define canqlock(x) sem_trydown(x) | 
|  | #define QLOCK_INITIALIZER(name) SEMAPHORE_INITIALIZER(name, 1) |