/* $NetBSD: pthread_sig.c,v 1.23 2003/10/16 13:38:28 yamt Exp $ */ /*- * Copyright (c) 2001 The NetBSD Foundation, Inc. * All rights reserved. * * This code is derived from software contributed to The NetBSD Foundation * by Nathan J. Williams. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software * must display the following acknowledgement: * This product includes software developed by the NetBSD * Foundation, Inc. and its contributors. * 4. Neither the name of The NetBSD Foundation nor the names of its * contributors may be used to endorse or promote products derived * from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. */ #include __RCSID("$NetBSD: pthread_sig.c,v 1.23 2003/10/16 13:38:28 yamt Exp $"); /* We're interposing a specific version of the signal interface. */ #define __LIBC12_SOURCE__ #define __PTHREAD_SIGNAL_PRIVATE #include #include #include #include #include #include /* for memcpy() */ #include #include #include #include #include "pthread.h" #include "pthread_int.h" #ifdef PTHREAD_SIG_DEBUG #define SDPRINTF(x) DPRINTF(x) #else #define SDPRINTF(x) #endif extern int pthread__started; extern pthread_spin_t pthread__runqueue_lock; extern struct pthread_queue_t pthread__runqueue; extern pthread_spin_t pthread__allqueue_lock; extern struct pthread_queue_t pthread__allqueue; static pthread_spin_t pt_sigacts_lock; static struct sigaction pt_sigacts[_NSIG]; static pthread_spin_t pt_process_siglock; static sigset_t pt_process_sigmask; static sigset_t pt_process_siglist; /* Queue of threads that are waiting in sigsuspend(). */ static struct pthread_queue_t pt_sigsuspended; static pthread_spin_t pt_sigsuspended_lock; /* * Nothing actually signals or waits on this lock, but the sleepobj * needs to point to something. */ static pthread_cond_t pt_sigsuspended_cond = PTHREAD_COND_INITIALIZER; /* Queue of threads that are waiting in sigtimedwait(). */ static struct pthread_queue_t pt_sigwaiting; static pthread_spin_t pt_sigwaiting_lock; static pthread_t pt_sigwmaster; static pthread_cond_t pt_sigwaiting_cond = PTHREAD_COND_INITIALIZER; static void pthread__make_siginfo(siginfo_t *, int); static void pthread__kill(pthread_t, pthread_t, siginfo_t *); static void pthread__kill_self(pthread_t, siginfo_t *); static void pthread__signal_tramp(void (*)(int, siginfo_t *, void *), siginfo_t *, ucontext_t *); static int firstsig(const sigset_t *); int _sys_execve(const char *, char *const [], char *const []); __strong_alias(__libc_thr_sigsetmask,pthread_sigmask) __strong_alias(__exeve,execve) void pthread__signal_init(void) { SDPRINTF(("(signal_init) setting process sigmask\n")); __sigprocmask14(0, NULL, &pt_process_sigmask); PTQ_INIT(&pt_sigsuspended); PTQ_INIT(&pt_sigwaiting); } static void pthread__make_siginfo(siginfo_t *si, int sig) { (void)memset(si, 0, sizeof(*si)); si->si_signo = sig; si->si_code = SI_USER; /* * XXX: these could be cached, but beware of setuid(). */ si->si_uid = getuid(); si->si_pid = getpid(); } int pthread_kill(pthread_t thread, int sig) { pthread_t self; void (*handler)(int); self = pthread__self(); SDPRINTF(("(pthread_kill %p) kill %p sig %d\n", self, thread, sig)); if ((sig < 0) || (sig >= _NSIG)) return EINVAL; if (pthread__find(self, thread) != 0) return ESRCH; /* * We only let the thread handle this signal if the action for * the signal is an explicit handler. Otherwise we feed it to * the kernel so that it can affect the whole process. */ pthread_spinlock(self, &pt_sigacts_lock); handler = pt_sigacts[sig].sa_handler; pthread_spinunlock(self, &pt_sigacts_lock); if (handler == SIG_IGN) { SDPRINTF(("(pthread_kill %p) do nothing\n", self, thread, sig)); /* Do nothing */ } else if ((sig == SIGKILL) || (sig == SIGSTOP) || (handler == SIG_DFL)) { /* Let the kernel do the work */ SDPRINTF(("(pthread_kill %p) kernel kill\n", self, thread, sig)); kill(getpid(), sig); } else { siginfo_t si; pthread__make_siginfo(&si, sig); pthread_spinlock(self, &thread->pt_siglock); pthread__kill(self, thread, &si); pthread_spinunlock(self, &thread->pt_siglock); } return 0; } /* * Interpositioning is our friend. We need to intercept sigaction(), * sigsuspend() and sigtimedwait(). */ int __sigaction14(int sig, const struct sigaction *act, struct sigaction *oact) { struct sigaction realact; sigset_t oldmask; pthread_t self; int retval; if ((sig <= 0) || (sig >= _NSIG)) return EINVAL; self = pthread__self(); if (act != NULL) { /* Save the information for our internal dispatch. */ pthread_spinlock(self, &pt_sigacts_lock); oldmask = pt_sigacts[sig].sa_mask; pt_sigacts[sig] = *act; pthread_spinunlock(self, &pt_sigacts_lock); /* * We want to handle all signals ourself, and not have * the kernel mask them. Therefore, we clear the * sa_mask field before passing this to the kernel. We * do not set SA_NODEFER, which seems like it might be * appropriate, because that would permit a continuous * stream of signals to exhaust the supply of upcalls. */ realact = *act; __sigemptyset14(&realact.sa_mask); act = &realact; } retval = __libc_sigaction14(sig, act, oact); if (oact && (retval == 0)) oact->sa_mask = oldmask; return retval; } int _sys___sigsuspend14(const sigset_t *); int __sigsuspend14(const sigset_t *sigmask) { pthread_t self; sigset_t oldmask; /* if threading not started yet, just do the syscall */ if (__predict_false(pthread__started == 0)) return (_sys___sigsuspend14(sigmask)); self = pthread__self(); pthread_spinlock(self, &pt_sigsuspended_lock); pthread_spinlock(self, &self->pt_statelock); if (self->pt_cancel) { pthread_spinunlock(self, &self->pt_statelock); pthread_spinunlock(self, &pt_sigsuspended_lock); pthread_exit(PTHREAD_CANCELED); } pthread_sigmask(SIG_SETMASK, sigmask, &oldmask); self->pt_state = PT_STATE_BLOCKED_QUEUE; self->pt_sleepobj = &pt_sigsuspended_cond; self->pt_sleepq = &pt_sigsuspended; self->pt_sleeplock = &pt_sigsuspended_lock; pthread_spinunlock(self, &self->pt_statelock); PTQ_INSERT_TAIL(&pt_sigsuspended, self, pt_sleep); pthread__block(self, &pt_sigsuspended_lock); pthread__testcancel(self); pthread_sigmask(SIG_SETMASK, &oldmask, NULL); errno = EINTR; return -1; } /* * Interpositioned sigtimedwait(2), need to multiplex all * eventual callers to a single kernel lwp. */ int _sigtimedwait(const sigset_t * __restrict, siginfo_t * __restrict, const struct timespec * __restrict); static void pthread_sigtimedwait__callback(void *arg) { pthread__sched(pthread__self(), (pthread_t) arg); } int sigtimedwait(const sigset_t * __restrict set, siginfo_t * __restrict info, const struct timespec * __restrict timeout) { pthread_t self; int error = 0; pthread_t target; sigset_t wset; struct timespec timo; /* if threading not started yet, just do the syscall */ if (__predict_false(pthread__started == 0)) return (_sigtimedwait(set, info, timeout)); self = pthread__self(); pthread__testcancel(self); /* also call syscall if timeout is zero (i.e. polling) */ if (timeout && timeout->tv_sec == 0 && timeout->tv_nsec == 0) { error = _sigtimedwait(set, info, timeout); pthread__testcancel(self); return (error); } if (timeout) { if ((u_int) timeout->tv_nsec >= 1000000000) return (EINVAL); timo = *timeout; } pthread_spinlock(self, &pt_sigwaiting_lock); /* * If there is already master thread running, arrange things * to accomodate for eventual extra signals to wait for, * and join the sigwaiting list. */ if (pt_sigwmaster) { struct pt_alarm_t timoalarm; struct timespec etimo; /* * Get current time. We need it if we would become master. */ if (timeout) { clock_gettime(CLOCK_MONOTONIC, &etimo); timespecadd(&etimo, timeout, &etimo); } /* * Check if this thread's wait set is different to master set. */ wset = *set; __sigminusset(pt_sigwmaster->pt_sigwait, &wset); if (firstsig(&wset)) { /* * Some new signal in set, wakeup master. It will * rebuild its wait set. */ _lwp_wakeup(pt_sigwmaster->pt_blockedlwp); } /* Save our wait set and info pointer */ wset = *set; self->pt_sigwait = &wset; self->pt_wsig = info; /* zero to recognize when we get passed the signal from master */ info->si_signo = 0; if (timeout) { pthread__alarm_add(self, &timoalarm, &etimo, pthread_sigtimedwait__callback, self); } block: pthread_spinlock(self, &self->pt_statelock); self->pt_state = PT_STATE_BLOCKED_QUEUE; self->pt_sleepobj = &pt_sigwaiting_cond; self->pt_sleepq = &pt_sigwaiting; self->pt_sleeplock = &pt_sigwaiting_lock; pthread_spinunlock(self, &self->pt_statelock); PTQ_INSERT_TAIL(&pt_sigwaiting, self, pt_sleep); pthread__block(self, &pt_sigwaiting_lock); /* check if we got a signal we waited for */ if (info->si_signo) { /* got the signal from master */ pthread__testcancel(self); return (0); } /* need the lock from now on */ pthread_spinlock(self, &pt_sigwaiting_lock); /* * If alarm fired, remove us from queue, adjust master * wait set and return with EAGAIN. */ if (timeout) { if (pthread__alarm_fired(&timoalarm)) { PTQ_REMOVE(&pt_sigwaiting, self, pt_sleep); /* * Signal master. It will rebuild it's wait set. */ _lwp_wakeup(pt_sigwmaster->pt_blockedlwp); pthread_spinunlock(self, &pt_sigwaiting_lock); errno = EAGAIN; return (-1); } pthread__alarm_del(self, &timoalarm); } /* * May have been woken up to deliver signal - check if we are * the master and reblock if appropriate. */ if (pt_sigwmaster != self) goto block; /* not signal nor alarm, must have been upgraded to master */ pthread__assert(pt_sigwmaster == self); /* update timeout before upgrading to master */ if (timeout) { struct timespec tnow; clock_gettime(CLOCK_MONOTONIC, &tnow); /* compute difference to end time */ timespecsub(&tnow, &etimo, &tnow); /* substract the difference from timeout */ timespecsub(&timo, &tnow, &timo); } } /* MASTER */ self->pt_sigwait = &wset; self->pt_wsig = NULL; /* Master thread loop */ pt_sigwmaster = self; for(;;) { /* Build our wait set */ wset = *set; if (!PTQ_EMPTY(&pt_sigwaiting)) { PTQ_FOREACH(target, &pt_sigwaiting, pt_sleep) __sigplusset(target->pt_sigwait, &wset); } pthread_spinunlock(self, &pt_sigwaiting_lock); /* * We are either the only one, or wait set was setup already. * Just do the syscall now. */ error = __sigtimedwait(&wset, info, (timeout) ? &timo : NULL); pthread_spinlock(self, &pt_sigwaiting_lock); if ((error && (errno != ECANCELED || self->pt_cancel)) || (!error && __sigismember14(set, info->si_signo)) ) { /* * Normal function return. Clear pt_sigwmaster, * and if wait queue is nonempty, make first waiter * new master. */ pt_sigwmaster = NULL; if (!PTQ_EMPTY(&pt_sigwaiting)) { pt_sigwmaster = PTQ_FIRST(&pt_sigwaiting); PTQ_REMOVE(&pt_sigwaiting, pt_sigwmaster, pt_sleep); pthread__sched(self, pt_sigwmaster); } pthread_spinunlock(self, &pt_sigwaiting_lock); pthread__testcancel(self); return (error); } if (!error) { /* * Got a signal, but not from _our_ wait set. * Scan the queue of sigwaiters and wakeup * the first thread waiting for this signal. */ PTQ_FOREACH(target, &pt_sigwaiting, pt_sleep) { if (__sigismember14(target->pt_sigwait, info->si_signo)) { pthread__assert(target->pt_state == PT_STATE_BLOCKED_QUEUE); /* copy to waiter siginfo */ memcpy(target->pt_wsig, info, sizeof(*info)); PTQ_REMOVE(&pt_sigwaiting, target, pt_sleep); pthread__sched(self, target); break; } } if (!target) { /* * Didn't find anyone waiting on this signal. * Deliver signal normally. This might * happen if a thread times out, but * 'their' signal arrives before the master * thread would be scheduled after _lwp_wakeup(). */ pthread__signal(self, NULL, info); } else { /* * Signal waiter removed, adjust our wait set. */ wset = *set; PTQ_FOREACH(target, &pt_sigwaiting, pt_sleep) __sigplusset(target->pt_sigwait, &wset); } } else { /* * ECANCELED - new sigwaiter arrived and added to master * wait set, or some sigwaiter exited due to timeout * and removed from master wait set. All the work * was done already, so just update our timeout * and go back to syscall. */ } /* Timeout was adjusted by the syscall, just call again. */ } /* NOTREACHED */ return (0); } /* * firstsig is stolen from kern_sig.c * XXX this is an abstraction violation; except for this, all of * the knowledge about the composition of sigset_t's was encapsulated * in signal.h. * Putting this function in signal.h caused problems with other parts of the * kernel that #included but didn't have a prototype for ffs. */ static int firstsig(const sigset_t *ss) { int sig; sig = ffs((int)ss->__bits[0]); if (sig != 0) return (sig); #if _NSIG > 33 sig = ffs((int)ss->__bits[1]); if (sig != 0) return (sig + 32); #endif #if _NSIG > 65 sig = ffs((int)ss->__bits[2]); if (sig != 0) return (sig + 64); #endif #if _NSIG > 97 sig = ffs((int)ss->__bits[3]); if (sig != 0) return (sig + 96); #endif return (0); } int pthread_sigmask(int how, const sigset_t *set, sigset_t *oset) { sigset_t tmp, takelist; pthread_t self; int i; siginfo_t si; self = pthread__self(); /* * While other threads may read a process's sigmask, * they won't write it, so we don't need to lock our reads of it. */ if (oset != NULL) *oset = self->pt_sigmask; if (set == NULL) return 0; pthread_spinlock(self, &self->pt_siglock); if (how == SIG_BLOCK) { __sigplusset(set, &self->pt_sigmask); /* * Blocking of signals that are now * blocked by all threads will be done * lazily, at signal delivery time. */ pthread_spinunlock(self, &self->pt_siglock); return 0; } else if (how == SIG_UNBLOCK) __sigminusset(set, &self->pt_sigmask); else if (how == SIG_SETMASK) self->pt_sigmask = *set; else { pthread_spinunlock(self, &self->pt_siglock); return EINVAL; } /* See if there are any signals to take */ __sigemptyset14(&takelist); while ((i = firstsig(&self->pt_siglist)) != 0) { if (!__sigismember14(&self->pt_sigmask, i)) { __sigaddset14(&takelist, i); __sigdelset14(&self->pt_siglist, i); } } pthread_spinlock(self, &pt_process_siglock); while ((i = firstsig(&pt_process_siglist)) != 0) { if (!__sigismember14(&self->pt_sigmask, i)) { __sigaddset14(&takelist, i); __sigdelset14(&pt_process_siglist, i); } } /* Unblock any signals that were blocked process-wide before this. */ tmp = pt_process_sigmask; __sigandset(&self->pt_sigmask, &tmp); if (!__sigsetequal(&tmp, &pt_process_sigmask)) { pt_process_sigmask = tmp; __sigprocmask14(SIG_SETMASK, &pt_process_sigmask, NULL); } pthread_spinunlock(self, &pt_process_siglock); pthread__make_siginfo(&si, 0); while ((i = firstsig(&takelist)) != 0) { /* Take the signal */ SDPRINTF(("(pt_sigmask %p) taking unblocked signal %d\n", self, i)); si.si_signo = i; pthread__kill_self(self, &si); __sigdelset14(&takelist, i); } pthread_spinunlock(self, &self->pt_siglock); return 0; } /* * Dispatch a signal to thread t, if it is non-null, and to any * willing thread, if t is null. */ void pthread__signal(pthread_t self, pthread_t t, siginfo_t *si) { pthread_t target, good, okay; if (t) { target = t; pthread_spinlock(self, &target->pt_siglock); } else { /* * Pick a thread that doesn't have the signal blocked * and can be reasonably forced to run. */ okay = good = NULL; pthread_spinlock(self, &pthread__allqueue_lock); PTQ_FOREACH(target, &pthread__allqueue, pt_allq) { /* * Changing to PT_STATE_ZOMBIE is protected by * the pthread__allqueue lock, so we can just * test for it here. */ if ((target->pt_state == PT_STATE_ZOMBIE) || (target->pt_type != PT_THREAD_NORMAL)) continue; pthread_spinlock(self, &target->pt_siglock); SDPRINTF(( "(pt_signal %p) target %p: state %d, mask %08x\n", self, target, target->pt_state, target->pt_sigmask.__bits[0])); if (!__sigismember14(&target->pt_sigmask, si->si_signo)) { if (target->pt_state != PT_STATE_BLOCKED_SYS) { good = target; /* Leave target locked */ break; } else if (okay == NULL) { okay = target; /* Leave target locked */ continue; } } pthread_spinunlock(self, &target->pt_siglock); } pthread_spinunlock(self, &pthread__allqueue_lock); if (good) { target = good; if (okay) pthread_spinunlock(self, &okay->pt_siglock); } else { target = okay; } if (target == NULL) { /* * They all have it blocked. Note that in our * process-wide signal mask, and stash the signal * for later unblocking. */ pthread_spinlock(self, &pt_process_siglock); __sigaddset14(&pt_process_sigmask, si->si_signo); SDPRINTF(("(pt_signal %p) lazily setting proc sigmask to " "%08x\n", self, pt_process_sigmask.__bits[0])); __sigprocmask14(SIG_SETMASK, &pt_process_sigmask, NULL); __sigaddset14(&pt_process_siglist, si->si_signo); pthread_spinunlock(self, &pt_process_siglock); return; } } /* * We now have a signal and a victim. * The victim's pt_siglock is locked. */ /* * Reset the process signal mask so we can take another * signal. We will not exhaust our supply of upcalls; if * another signal is delivered after this, the resolve_locks * dance will permit us to finish and recycle before the next * upcall reaches this point. */ pthread_spinlock(self, &pt_process_siglock); SDPRINTF(("(pt_signal %p) setting proc sigmask to " "%08x\n", self, pt_process_sigmask.__bits[0])); __sigprocmask14(SIG_SETMASK, &pt_process_sigmask, NULL); pthread_spinunlock(self, &pt_process_siglock); pthread__kill(self, target, si); pthread_spinunlock(self, &target->pt_siglock); } /* * Call the signal handler in the context of this thread. Not quite as * suicidal as it sounds. * Must be called with target's siglock held. */ static void pthread__kill_self(pthread_t self, siginfo_t *si) { sigset_t oldmask; struct sigaction act; pthread_spinlock(self, &pt_sigacts_lock); act = pt_sigacts[si->si_signo]; pthread_spinunlock(self, &pt_sigacts_lock); SDPRINTF(("(pthread__kill_self %p) sig %d\n", self, si->si_signo)); oldmask = self->pt_sigmask; __sigplusset(&self->pt_sigmask, &act.sa_mask); if ((act.sa_flags & SA_NODEFER) == 0) __sigaddset14(&self->pt_sigmask, si->si_signo); pthread_spinunlock(self, &self->pt_siglock); #ifdef __HAVE_SIGINFO { ucontext_t uc; (*act.sa_sigaction)(si->si_signo, si, &uc); } #else { struct sigcontext xxxsc; void (*handler)(int, int, struct sigcontext *); handler = (void (*)(int, int, struct sigcontext *)) act.sa_handler; (*handler)(si->si_signo, si->si_trap, &xxxsc); } #endif pthread_spinlock(self, &self->pt_siglock); self->pt_sigmask = oldmask; } /* Must be called with target's siglock held */ static void pthread__kill(pthread_t self, pthread_t target, siginfo_t *si) { SDPRINTF(("(pthread__kill %p) target %p sig %d code %d\n", self, target, si->si_signo)); if (__sigismember14(&target->pt_sigmask, si->si_signo)) { /* Record the signal for later delivery. */ __sigaddset14(&target->pt_siglist, si->si_signo); return; } if (self == target) { pthread__kill_self(self, si); return; } /* * Ensure the victim is not running. * In a MP world, it could be on another processor somewhere. * * XXX As long as this is uniprocessor, encountering a running * target process is a bug. */ pthread__assert(target->pt_state != PT_STATE_RUNNING); /* * Holding the state lock blocks out cancellation and any other * attempts to set this thread up to take a signal. */ pthread_spinlock(self, &target->pt_statelock); switch (target->pt_state) { case PT_STATE_RUNNABLE: pthread_spinlock(self, &pthread__runqueue_lock); PTQ_REMOVE(&pthread__runqueue, target, pt_runq); pthread_spinunlock(self, &pthread__runqueue_lock); break; case PT_STATE_BLOCKED_QUEUE: pthread_spinlock(self, target->pt_sleeplock); PTQ_REMOVE(target->pt_sleepq, target, pt_sleep); pthread_spinunlock(self, target->pt_sleeplock); break; case PT_STATE_BLOCKED_SYS: /* * The target is not on a queue at all, and won't run * again for a while. Try to wake it from its torpor, then * mark the signal for later processing. */ __sigaddset14(&target->pt_sigblocked, si->si_signo); __sigaddset14(&target->pt_sigmask, si->si_signo); pthread_spinlock(self, &target->pt_flaglock); target->pt_flags |= PT_FLAG_SIGDEFERRED; pthread_spinunlock(self, &target->pt_flaglock); pthread_spinunlock(self, &target->pt_statelock); _lwp_wakeup(target->pt_blockedlwp); return; default: ; } pthread__deliver_signal(self, target, si); pthread__sched(self, target); pthread_spinunlock(self, &target->pt_statelock); } /* Must be called with target's siglock held */ void pthread__deliver_signal(pthread_t self, pthread_t target, siginfo_t *si) { sigset_t oldmask, *maskp; ucontext_t *uc, *olduc; struct sigaction act; siginfo_t *siginfop; pthread_spinlock(self, &pt_sigacts_lock); act = pt_sigacts[si->si_signo]; pthread_spinunlock(self, &pt_sigacts_lock); /* * Prevent anyone else from considering this thread for handling * more instances of this signal. */ oldmask = target->pt_sigmask; __sigplusset(&target->pt_sigmask, &act.sa_mask); __sigaddset14(&target->pt_sigmask, si->si_signo); if (target->pt_trapuc) { olduc = target->pt_trapuc; target->pt_trapuc = NULL; } else olduc = target->pt_uc; /* * We'd like to just pass oldmask and si to the * pthread__signal_tramp(), but makecontext() can't reasonably * pass structures, just word-size things or smaller. We also * don't want to involve malloc() here, inside the upcall * handler. So we borrow a bit of space from the target's * stack, which we were adjusting anyway. */ maskp = (sigset_t *)(void *)((char *)(void *)olduc - STACKSPACE - sizeof(sigset_t)); *maskp = oldmask; siginfop = (siginfo_t *)(void *)((char *)(void *)maskp - sizeof(*siginfop)); *siginfop = *si; /* * XXX We are blatantly ignoring SIGALTSTACK. It would screw * with our notion of stack->thread mappings. */ uc = (ucontext_t *)(void *)((char *)(void *)siginfop - sizeof(ucontext_t)); #ifdef _UC_UCONTEXT_ALIGN uc = (ucontext_t *)((uintptr_t)uc & _UC_UCONTEXT_ALIGN); #endif _INITCONTEXT_U(uc); uc->uc_stack.ss_sp = maskp; uc->uc_stack.ss_size = 0; uc->uc_link = NULL; SDPRINTF(("(makecontext %p): target %p: sig: %d uc: %p oldmask: %08x\n", self, target, si->si_signo, target->pt_uc, maskp->__bits[0])); makecontext(uc, pthread__signal_tramp, 3, act.sa_handler, siginfop, olduc); target->pt_uc = uc; } void pthread__signal_deferred(pthread_t self, pthread_t t) { int i; siginfo_t si; pthread_spinlock(self, &t->pt_siglock); pthread__make_siginfo(&si, 0); while ((i = firstsig(&t->pt_sigblocked)) != 0) { __sigdelset14(&t->pt_sigblocked, i); si.si_signo = i; pthread__deliver_signal(self, t, &si); } t->pt_flags &= ~PT_FLAG_SIGDEFERRED; pthread_spinunlock(self, &t->pt_siglock); } static void pthread__signal_tramp(void (*handler)(int, siginfo_t *, void *), siginfo_t *info, ucontext_t *uc) { SDPRINTF(("(tramp %p) sig %d uc %p oldmask %08x\n", pthread__self(), info->si_signo, uc, uc->uc_sigmask.__bits[0])); #ifdef __HAVE_SIGINFO (*handler)(info->si_signo, info, uc); #else { struct pthread__sigcontext psc; /* * XXX we don't support siginfo here yet. * Note that uc_stack.ss_sp holds the old sigmask */ sigset_t *maskp = uc->uc_stack.ss_sp; PTHREAD_UCONTEXT_TO_SIGCONTEXT(maskp, uc, &psc); ((void *(*)(int, int, struct sigcontext *))handler) (info->si_signo, info->si_trap, &psc.psc_context); PTHREAD_SIGCONTEXT_TO_UCONTEXT(&psc, uc); } #endif /* * We've finished the handler, so this thread can restore the * original signal mask. Note that traditional BSD behavior * allows for the handler to change the signal mask; we handle * that here. */ pthread_sigmask(SIG_SETMASK, &uc->uc_sigmask, NULL); /* * Jump back to what we were doing before we were interrupted * by a signal. */ _setcontext_u(uc); /*NOTREACHED*/ pthread__abort(); } /* * The execve() system call and the libc exec*() calls that use it are * specified to propagate the signal mask of the current thread to the * initial thread of the new process image. Since thread signal masks * are maintained in userlevel, this wrapper is necessary to give the * kernel the correct value. */ int execve(const char *path, char *const argv[], char *const envp[]) { pthread_t self; int ret; self = pthread__self(); /* * Don't acquire pt_process_siglock, even though it seems like * the right thing to do. The most common reason to be here is * that we're on the child side of a fork() or vfork() * call. In either case, another thread could have held * pt_process_siglock at the moment of forking, and acquiring * it here would cause us to deadlock. Additionally, in the * case of vfork(), acquiring the lock here would cause it to * be locked in the parent's address space and cause a * deadlock there the next time a signal routine is called. * * The remaining case is where a live multithreaded program * calls exec*() from one of several threads with no explicit * synchronization. It may get the wrong process sigmask in * the new process image if another thread executes a signal * routine between the sigprocmask and the _sys_execve() * call. I don't have much sympathy for such a program. */ __sigprocmask14(SIG_SETMASK, &self->pt_sigmask, NULL); ret = _sys_execve(path, argv, envp); /* Normally, we shouldn't get here; this is an error condition. */ __sigprocmask14(SIG_SETMASK, &pt_process_sigmask, NULL); return ret; }