diff --git a/sys/kern/kern_sig.c b/sys/kern/kern_sig.c --- a/sys/kern/kern_sig.c +++ b/sys/kern/kern_sig.c @@ -108,7 +108,7 @@ static void reschedule_signals(struct proc *p, sigset_t block, int flags); static int sigprop(int sig); static void tdsigwakeup(struct thread *, int, sig_t, int); -static int sig_suspend_threads(struct thread *, struct proc *); +static void sig_suspend_threads(struct thread *, struct proc *); static int filt_sigattach(struct knote *kn); static void filt_sigdetach(struct knote *kn); static int filt_signal(struct knote *kn, long hint); @@ -2260,16 +2260,15 @@ (void) tdsendsignal(td->td_proc, td, sig, ksi); } -static int +static void sig_sleepq_abort(struct thread *td, int intrval) { THREAD_LOCK_ASSERT(td, MA_OWNED); - if (intrval == 0 && (td->td_flags & TDF_SIGWAIT) == 0) { + if (intrval == 0 && (td->td_flags & TDF_SIGWAIT) == 0) thread_unlock(td); - return (0); - } - return (sleepq_abort(td, intrval)); + else + sleepq_abort(td, intrval); } int @@ -2281,7 +2280,6 @@ struct sigacts *ps; int intrval; int ret = 0; - int wakeup_swapper; MPASS(td == NULL || p == td->td_proc); PROC_LOCK_ASSERT(p, MA_OWNED); @@ -2383,8 +2381,6 @@ !((prop & SIGPROP_CONT) && (p->p_flag & P_STOPPED_SIG))) return (ret); - wakeup_swapper = 0; - /* * Some signals have a process-wide effect and a per-thread * component. Most processing occurs when the process next @@ -2490,7 +2486,7 @@ PROC_SLOCK(p); thread_lock(td); if (TD_CAN_ABORT(td)) - wakeup_swapper = sig_sleepq_abort(td, intrval); + sig_sleepq_abort(td, intrval); else thread_unlock(td); PROC_SUNLOCK(p); @@ -2513,7 +2509,7 @@ p->p_flag |= P_STOPPED_SIG; p->p_xsig = sig; PROC_SLOCK(p); - wakeup_swapper = sig_suspend_threads(td, p); + sig_suspend_threads(td, p); if (p->p_numthreads == p->p_suspcount) { /* * only thread sending signal to another @@ -2550,8 +2546,6 @@ out: /* If we jump here, proc slock should not be owned. */ PROC_SLOCK_ASSERT(p, MA_NOTOWNED); - if (wakeup_swapper) - kick_proc0(); return (ret); } @@ -2565,7 +2559,7 @@ tdsigwakeup(struct thread *td, int sig, sig_t action, int intrval) { struct proc *p = td->td_proc; - int prop, wakeup_swapper; + int prop; PROC_LOCK_ASSERT(p, MA_OWNED); prop = sigprop(sig); @@ -2621,10 +2615,8 @@ if (td->td_priority > PUSER && !TD_IS_IDLETHREAD(td)) sched_prio(td, PUSER); - wakeup_swapper = sig_sleepq_abort(td, intrval); + sig_sleepq_abort(td, intrval); PROC_SUNLOCK(p); - if (wakeup_swapper) - kick_proc0(); return; } @@ -2778,16 +2770,14 @@ wakeup(p); } -static int +static void sig_suspend_threads(struct thread *td, struct proc *p) { struct thread *td2; - int wakeup_swapper; PROC_LOCK_ASSERT(p, MA_OWNED); PROC_SLOCK_ASSERT(p, MA_OWNED); - wakeup_swapper = 0; FOREACH_THREAD_IN_PROC(p, td2) { thread_lock(td2); ast_sched_locked(td2, TDA_SUSPEND); @@ -2802,8 +2792,7 @@ KASSERT(!TD_IS_SUSPENDED(td2), ("thread with deferred stops suspended")); if (TD_SBDRY_INTR(td2)) { - wakeup_swapper |= sleepq_abort(td2, - TD_SBDRY_ERRNO(td2)); + sleepq_abort(td2, TD_SBDRY_ERRNO(td2)); continue; } } else if (!TD_IS_SUSPENDED(td2)) @@ -2816,7 +2805,6 @@ } thread_unlock(td2); } - return (wakeup_swapper); } /* diff --git a/sys/kern/kern_synch.c b/sys/kern/kern_synch.c --- a/sys/kern/kern_synch.c +++ b/sys/kern/kern_synch.c @@ -562,7 +562,7 @@ * * Requires the thread lock on entry, drops on exit. */ -int +void setrunnable(struct thread *td, int srqflags) { THREAD_LOCK_ASSERT(td, MA_OWNED); @@ -586,8 +586,6 @@ default: panic("setrunnable: state 0x%x", TD_GET_STATE(td)); } - - return (0); } /* diff --git a/sys/kern/kern_thread.c b/sys/kern/kern_thread.c --- a/sys/kern/kern_thread.c +++ b/sys/kern/kern_thread.c @@ -145,7 +145,7 @@ static void thread_reap_all(void); static void thread_reap_task_cb(void *, int); static void thread_reap_callout_cb(void *); -static int thread_unsuspend_one(struct thread *td, struct proc *p, +static void thread_unsuspend_one(struct thread *td, struct proc *p, bool boundary); static void thread_free_batched(struct thread *td); @@ -935,7 +935,6 @@ struct thread *td; struct thread *td2; struct proc *p; - int wakeup_swapper; td = curthread; p = td->td_proc; @@ -981,10 +980,8 @@ if (P_SHOULDSTOP(p) == P_STOPPED_SINGLE) { if (p->p_numthreads == p->p_suspcount) { thread_lock(p->p_singlethread); - wakeup_swapper = thread_unsuspend_one( - p->p_singlethread, p, false); - if (wakeup_swapper) - kick_proc0(); + thread_unsuspend_one(p->p_singlethread, + p, false); } } @@ -1137,17 +1134,13 @@ return (mode == SINGLE_ALLPROC ? 0 : 1); } -static int +static void weed_inhib(int mode, struct thread *td2, struct proc *p) { - int wakeup_swapper; - PROC_LOCK_ASSERT(p, MA_OWNED); PROC_SLOCK_ASSERT(p, MA_OWNED); THREAD_LOCK_ASSERT(td2, MA_OWNED); - wakeup_swapper = 0; - /* * Since the thread lock is dropped by the scheduler we have * to retry to check for races. @@ -1156,26 +1149,26 @@ switch (mode) { case SINGLE_EXIT: if (TD_IS_SUSPENDED(td2)) { - wakeup_swapper |= thread_unsuspend_one(td2, p, true); + thread_unsuspend_one(td2, p, true); thread_lock(td2); goto restart; } if (TD_CAN_ABORT(td2)) { - wakeup_swapper |= sleepq_abort(td2, EINTR); - return (wakeup_swapper); + sleepq_abort(td2, EINTR); + return; } break; case SINGLE_BOUNDARY: case SINGLE_NO_EXIT: if (TD_IS_SUSPENDED(td2) && (td2->td_flags & TDF_BOUNDARY) == 0) { - wakeup_swapper |= thread_unsuspend_one(td2, p, false); + thread_unsuspend_one(td2, p, false); thread_lock(td2); goto restart; } if (TD_CAN_ABORT(td2)) { - wakeup_swapper |= sleepq_abort(td2, ERESTART); - return (wakeup_swapper); + sleepq_abort(td2, ERESTART); + return; } break; case SINGLE_ALLPROC: @@ -1189,21 +1182,20 @@ */ if (TD_IS_SUSPENDED(td2) && (td2->td_flags & TDF_ALLPROCSUSP) == 0) { - wakeup_swapper |= thread_unsuspend_one(td2, p, false); + thread_unsuspend_one(td2, p, false); thread_lock(td2); goto restart; } if (TD_CAN_ABORT(td2)) { td2->td_flags |= TDF_ALLPROCSUSP; - wakeup_swapper |= sleepq_abort(td2, ERESTART); - return (wakeup_swapper); + sleepq_abort(td2, ERESTART); + return; } break; default: break; } thread_unlock(td2); - return (wakeup_swapper); } /* @@ -1224,7 +1216,7 @@ { struct thread *td; struct thread *td2; - int remaining, wakeup_swapper; + int remaining; td = curthread; KASSERT(mode == SINGLE_EXIT || mode == SINGLE_BOUNDARY || @@ -1279,14 +1271,13 @@ while (remaining != remain_for_mode(mode)) { if (P_SHOULDSTOP(p) != P_STOPPED_SINGLE) goto stopme; - wakeup_swapper = 0; FOREACH_THREAD_IN_PROC(p, td2) { if (td2 == td) continue; thread_lock(td2); ast_sched_locked(td2, TDA_SUSPEND); if (TD_IS_INHIBITED(td2)) { - wakeup_swapper |= weed_inhib(mode, td2, p); + weed_inhib(mode, td2, p); #ifdef SMP } else if (TD_IS_RUNNING(td2)) { forward_signal(td2); @@ -1295,8 +1286,6 @@ } else thread_unlock(td2); } - if (wakeup_swapper) - kick_proc0(); remaining = calc_remaining(p, mode); /* @@ -1412,7 +1401,6 @@ { struct thread *td; struct proc *p; - int wakeup_swapper; td = curthread; p = td->td_proc; @@ -1475,10 +1463,8 @@ if (P_SHOULDSTOP(p) == P_STOPPED_SINGLE) { if (p->p_numthreads == p->p_suspcount + 1) { thread_lock(p->p_singlethread); - wakeup_swapper = thread_unsuspend_one( - p->p_singlethread, p, false); - if (wakeup_swapper) - kick_proc0(); + thread_unsuspend_one(p->p_singlethread, p, + false); } } PROC_UNLOCK(p); @@ -1585,7 +1571,7 @@ sched_sleep(td, 0); } -static int +static void thread_unsuspend_one(struct thread *td, struct proc *p, bool boundary) { @@ -1601,7 +1587,7 @@ p->p_boundary_count--; } } - return (setrunnable(td, 0)); + setrunnable(td, 0); } void @@ -1625,8 +1611,7 @@ MPASS(p->p_suspcount > 0); p->p_suspcount--; PROC_SUNLOCK(p); - if (setrunnable(td, 0)) - kick_proc0(); + setrunnable(td, 0); } /* @@ -1636,17 +1621,14 @@ thread_unsuspend(struct proc *p) { struct thread *td; - int wakeup_swapper; PROC_LOCK_ASSERT(p, MA_OWNED); PROC_SLOCK_ASSERT(p, MA_OWNED); - wakeup_swapper = 0; if (!P_SHOULDSTOP(p)) { FOREACH_THREAD_IN_PROC(p, td) { thread_lock(td); if (TD_IS_SUSPENDED(td)) - wakeup_swapper |= thread_unsuspend_one(td, p, - true); + thread_unsuspend_one(td, p, true); else thread_unlock(td); } @@ -1659,12 +1641,9 @@ */ if (p->p_singlethread->td_proc == p) { thread_lock(p->p_singlethread); - wakeup_swapper = thread_unsuspend_one( - p->p_singlethread, p, false); + thread_unsuspend_one(p->p_singlethread, p, false); } } - if (wakeup_swapper) - kick_proc0(); } /* @@ -1674,7 +1653,6 @@ thread_single_end(struct proc *p, int mode) { struct thread *td; - int wakeup_swapper; KASSERT(mode == SINGLE_EXIT || mode == SINGLE_BOUNDARY || mode == SINGLE_ALLPROC || mode == SINGLE_NO_EXIT, @@ -1693,7 +1671,7 @@ P_TOTAL_STOP); PROC_SLOCK(p); p->p_singlethread = NULL; - wakeup_swapper = 0; + /* * If there are other threads they may now run, * unless of course there is a blanket 'stop order' @@ -1703,18 +1681,15 @@ if (p->p_numthreads != remain_for_mode(mode) && !P_SHOULDSTOP(p)) { FOREACH_THREAD_IN_PROC(p, td) { thread_lock(td); - if (TD_IS_SUSPENDED(td)) { - wakeup_swapper |= thread_unsuspend_one(td, p, - true); - } else + if (TD_IS_SUSPENDED(td)) + thread_unsuspend_one(td, p, true); + else thread_unlock(td); } } KASSERT(mode != SINGLE_BOUNDARY || p->p_boundary_count == 0, ("inconsistent boundary count %d", p->p_boundary_count)); PROC_SUNLOCK(p); - if (wakeup_swapper) - kick_proc0(); wakeup(&p->p_flag); } diff --git a/sys/kern/subr_sleepqueue.c b/sys/kern/subr_sleepqueue.c --- a/sys/kern/subr_sleepqueue.c +++ b/sys/kern/subr_sleepqueue.c @@ -169,7 +169,7 @@ static void sleepq_dtor(void *mem, int size, void *arg); #endif static int sleepq_init(void *mem, int size, int flags); -static int sleepq_resume_thread(struct sleepqueue *sq, struct thread *td, +static void sleepq_resume_thread(struct sleepqueue *sq, struct thread *td, int pri, int srqflags); static void sleepq_remove_thread(struct sleepqueue *sq, struct thread *td); static void sleepq_switch(const void *wchan, int pri); @@ -731,13 +731,12 @@ } /* - * Removes a thread from a sleep queue and makes it - * runnable. + * Removes a thread from a sleep queue and makes it runnable. * * Requires the sc chain locked on entry. If SRQ_HOLD is specified it will * be locked on return. Returns without the thread lock held. */ -static int +static void sleepq_resume_thread(struct sleepqueue *sq, struct thread *td, int pri, int srqflags) { @@ -788,12 +787,11 @@ if (TD_IS_SLEEPING(td)) { MPASS(!drop); TD_CLR_SLEEPING(td); - return (setrunnable(td, srqflags)); + setrunnable(td, srqflags); + } else { + MPASS(drop); + thread_unlock(td); } - MPASS(drop); - thread_unlock(td); - - return (0); } static void @@ -929,7 +927,6 @@ struct sleepqueue *sq; struct threadqueue *head; struct thread *td, *besttd; - int wakeup_swapper; CTR2(KTR_PROC, "sleepq_signal(%p, %d)", wchan, flags); KASSERT(wchan != NULL, ("%s: invalid NULL wait channel", __func__)); @@ -972,9 +969,9 @@ } } MPASS(besttd != NULL); - wakeup_swapper = sleepq_resume_thread(sq, besttd, pri, + sleepq_resume_thread(sq, besttd, pri, (flags & SLEEPQ_DROP) ? 0 : SRQ_HOLD); - return (wakeup_swapper); + return (0); } static bool @@ -1012,7 +1009,6 @@ bool (*matches)(struct thread *), int pri) { struct thread *td, *tdn; - int wakeup_swapper; /* * The last thread will be given ownership of sq and may @@ -1020,14 +1016,12 @@ * so we must cache the "next" queue item at the beginning * of the final iteration. */ - wakeup_swapper = 0; TAILQ_FOREACH_SAFE(td, &sq->sq_blocked[queue], td_slpq, tdn) { if (matches(td)) - wakeup_swapper |= sleepq_resume_thread(sq, td, pri, - SRQ_HOLD); + sleepq_resume_thread(sq, td, pri, SRQ_HOLD); } - return (wakeup_swapper); + return (0); } /* @@ -1041,7 +1035,6 @@ struct sleepqueue *sq; struct thread *td; const void *wchan; - int wakeup_swapper; td = arg; CTR3(KTR_PROC, "sleepq_timeout: thread %p (pid %ld, %s)", @@ -1064,9 +1057,7 @@ sq = sleepq_lookup(wchan); MPASS(sq != NULL); td->td_flags |= TDF_TIMEOUT; - wakeup_swapper = sleepq_resume_thread(sq, td, 0, 0); - if (wakeup_swapper) - kick_proc0(); + sleepq_resume_thread(sq, td, 0, 0); return; } else if (TD_ON_SLEEPQ(td)) { /* @@ -1089,7 +1080,6 @@ { struct sleepqueue_chain *sc; struct sleepqueue *sq; - int wakeup_swapper; /* * Look up the sleep queue for this wait channel, then re-check @@ -1114,9 +1104,7 @@ sq = sleepq_lookup(wchan); MPASS(sq != NULL); MPASS(td->td_wchan == wchan); - wakeup_swapper = sleepq_resume_thread(sq, td, 0, 0); - if (wakeup_swapper) - kick_proc0(); + sleepq_resume_thread(sq, td, 0, 0); } /* @@ -1165,7 +1153,8 @@ MPASS(sq != NULL); /* Thread is asleep on sleep queue sq, so wake it up. */ - return (sleepq_resume_thread(sq, td, 0, 0)); + sleepq_resume_thread(sq, td, 0, 0); + return (0); } void @@ -1173,25 +1162,19 @@ { struct sleepqueue_chain *sc; struct sleepqueue *sq, *sq1; - int i, wakeup_swapper; + int i; - wakeup_swapper = 0; for (sc = &sleepq_chains[0]; sc < sleepq_chains + SC_TABLESIZE; ++sc) { if (LIST_EMPTY(&sc->sc_queues)) { continue; } mtx_lock_spin(&sc->sc_lock); LIST_FOREACH_SAFE(sq, &sc->sc_queues, sq_hash, sq1) { - for (i = 0; i < NR_SLEEPQS; ++i) { - wakeup_swapper |= sleepq_remove_matching(sq, i, - matches, 0); - } + for (i = 0; i < NR_SLEEPQS; ++i) + sleepq_remove_matching(sq, i, matches, 0); } mtx_unlock_spin(&sc->sc_lock); } - if (wakeup_swapper) { - kick_proc0(); - } } /* diff --git a/sys/sys/proc.h b/sys/sys/proc.h --- a/sys/sys/proc.h +++ b/sys/sys/proc.h @@ -1200,7 +1200,7 @@ int securelevel_gt(struct ucred *cr, int level); void sess_hold(struct session *); void sess_release(struct session *); -int setrunnable(struct thread *, int); +void setrunnable(struct thread *, int); void setsugid(struct proc *p); bool should_yield(void); int sigonstack(size_t sp);