1024 lines
23 KiB
C
1024 lines
23 KiB
C
/* $NetBSD: pthread_dbg.c,v 1.35 2007/03/14 21:09:01 skrll Exp $ */
|
|
|
|
/*-
|
|
* Copyright (c) 2002 Wasabi Systems, Inc.
|
|
* All rights reserved.
|
|
*
|
|
* Written by Nathan J. Williams for Wasabi Systems, Inc.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
* must display the following acknowledgement:
|
|
* This product includes software developed for the NetBSD Project by
|
|
* Wasabi Systems, Inc.
|
|
* 4. The name of Wasabi Systems, Inc. may not be used to endorse
|
|
* or promote products derived from this software without specific
|
|
* prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY WASABI SYSTEMS, INC. ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
|
|
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
|
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL WASABI SYSTEMS, INC
|
|
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__RCSID("$NetBSD: pthread_dbg.c,v 1.35 2007/03/14 21:09:01 skrll Exp $");
|
|
|
|
#define __EXPOSE_STACK 1
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/types.h>
|
|
|
|
#include <stddef.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
#include <errno.h>
|
|
#include <unistd.h>
|
|
#include <lwp.h>
|
|
|
|
#include <machine/reg.h>
|
|
|
|
#include <pthread.h>
|
|
#include <pthread_int.h>
|
|
#include <pthread_dbg.h>
|
|
#include <pthread_dbg_int.h>
|
|
|
|
#ifndef PT_FIXEDSTACKSIZE_LG
|
|
#undef PT_STACKMASK
|
|
#define PT_STACKMASK (proc->stackmask)
|
|
#endif
|
|
|
|
/* Compensate for debuggers that want a zero ID to be a sentinel */
|
|
#define TN_OFFSET 1
|
|
|
|
static int td__getthread(td_proc_t *proc, caddr_t addr, td_thread_t **threadp);
|
|
static int td__getsync(td_proc_t *proc, caddr_t addr, td_sync_t **syncp);
|
|
static int td__getstacksize(td_proc_t *proc);
|
|
|
|
int
|
|
td_open(struct td_proc_callbacks_t *cb, void *arg, td_proc_t **procp)
|
|
{
|
|
td_proc_t *proc;
|
|
caddr_t addr;
|
|
int dbg;
|
|
int val;
|
|
|
|
proc = malloc(sizeof(*proc));
|
|
if (proc == NULL)
|
|
return TD_ERR_NOMEM;
|
|
proc->cb = cb;
|
|
proc->arg = arg;
|
|
|
|
val = LOOKUP(proc, "pthread__dbg", &addr);
|
|
if (val != 0) {
|
|
if (val == TD_ERR_NOSYM)
|
|
val = TD_ERR_NOLIB;
|
|
goto error;
|
|
}
|
|
proc->dbgaddr = addr;
|
|
|
|
val = LOOKUP(proc, "pthread__allqueue", &addr);
|
|
if (val != 0) {
|
|
if (val == TD_ERR_NOSYM)
|
|
val = TD_ERR_NOLIB;
|
|
goto error;
|
|
}
|
|
proc->allqaddr = addr;
|
|
|
|
val = LOOKUP(proc, "pthread__tsd_alloc", &addr);
|
|
if (val != 0) {
|
|
if (val == TD_ERR_NOSYM)
|
|
val = TD_ERR_NOLIB;
|
|
goto error;
|
|
}
|
|
proc->tsdallocaddr = addr;
|
|
|
|
val = LOOKUP(proc, "pthread__tsd_destructors", &addr);
|
|
if (val != 0) {
|
|
if (val == TD_ERR_NOSYM)
|
|
val = TD_ERR_NOLIB;
|
|
goto error;
|
|
}
|
|
proc->tsddestaddr = addr;
|
|
|
|
val = READ(proc, proc->dbgaddr, &dbg, sizeof(int));
|
|
if (val != 0)
|
|
goto error;
|
|
|
|
if (dbg != 0) {
|
|
/* Another instance of libpthread_dbg is already attached. */
|
|
val = TD_ERR_INUSE;
|
|
goto error;
|
|
}
|
|
|
|
val = LOOKUP(proc, "pthread_stacksize_lg", &addr);
|
|
if (val == 0)
|
|
proc->stacksizeaddr = addr;
|
|
else
|
|
proc->stacksizeaddr = NULL;
|
|
proc->stacksizelg = -1;
|
|
proc->stacksize = 0;
|
|
proc->stackmask = 0;
|
|
|
|
proc->regbuf = NULL;
|
|
proc->fpregbuf = NULL;
|
|
|
|
dbg = getpid();
|
|
/*
|
|
* If this fails it probably means we're debugging a core file and
|
|
* can't write to it.
|
|
* If it's something else we'll lose the next time we hit WRITE,
|
|
* but not before, and that's OK.
|
|
*/
|
|
WRITE(proc, proc->dbgaddr, &dbg, sizeof(int));
|
|
|
|
PTQ_INIT(&proc->threads);
|
|
PTQ_INIT(&proc->syncs);
|
|
|
|
*procp = proc;
|
|
|
|
return 0;
|
|
|
|
error:
|
|
free(proc);
|
|
return val;
|
|
}
|
|
|
|
int
|
|
td_close(td_proc_t *proc)
|
|
{
|
|
int dbg;
|
|
td_thread_t *t, *next;
|
|
td_sync_t *s, *nexts;
|
|
|
|
dbg = 0;
|
|
/*
|
|
* Error returns from this write are mot really a problem;
|
|
* the process doesn't exist any more.
|
|
*/
|
|
WRITE(proc, proc->dbgaddr, &dbg, sizeof(int));
|
|
|
|
/* Deallocate the list of thread structures */
|
|
for (t = PTQ_FIRST(&proc->threads); t; t = next) {
|
|
next = PTQ_NEXT(t, list);
|
|
PTQ_REMOVE(&proc->threads, t, list);
|
|
free(t);
|
|
}
|
|
/* Deallocate the list of sync objects */
|
|
for (s = PTQ_FIRST(&proc->syncs); s; s = nexts) {
|
|
nexts = PTQ_NEXT(s, list);
|
|
PTQ_REMOVE(&proc->syncs, s, list);
|
|
free(s);
|
|
}
|
|
|
|
if (proc->regbuf != NULL) {
|
|
free(proc->regbuf);
|
|
free(proc->fpregbuf);
|
|
}
|
|
|
|
free(proc);
|
|
return 0;
|
|
}
|
|
|
|
|
|
int
|
|
td_thr_iter(td_proc_t *proc, int (*call)(td_thread_t *, void *), void *callarg)
|
|
{
|
|
int val;
|
|
caddr_t next;
|
|
struct pthread_queue_t allq;
|
|
td_thread_t *thread;
|
|
|
|
val = READ(proc, proc->allqaddr, &allq, sizeof(allq));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
next = (void *)allq.ptqh_first;
|
|
while (next != NULL) {
|
|
val = td__getthread(proc, next, &thread);
|
|
if (val != 0)
|
|
return val;
|
|
val = (*call)(thread, callarg);
|
|
if (val != 0)
|
|
return 0;
|
|
|
|
val = READ(proc,
|
|
next + offsetof(struct __pthread_st, pt_allq.ptqe_next),
|
|
&next, sizeof(next));
|
|
if (val != 0)
|
|
return val;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
td_thr_info(td_thread_t *thread, td_thread_info_t *info)
|
|
{
|
|
int tmp, val;
|
|
struct pthread_queue_t queue;
|
|
|
|
val = READ(thread->proc, thread->addr, &tmp, sizeof(tmp));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
if (tmp != PT_MAGIC)
|
|
return TD_ERR_BADTHREAD;
|
|
|
|
info->thread_addr = thread->addr;
|
|
if ((val = READ(thread->proc,
|
|
thread->addr + offsetof(struct __pthread_st, pt_state),
|
|
&tmp, sizeof(tmp))) != 0)
|
|
return val;
|
|
switch (tmp) {
|
|
case PT_STATE_RUNNING:
|
|
info->thread_state = TD_STATE_RUNNING;
|
|
break;
|
|
#ifdef XXXLWP
|
|
case PT_STATE_RUNNABLE:
|
|
info->thread_state = TD_STATE_RUNNABLE;
|
|
break;
|
|
case _PT_STATE_BLOCKED_SYS:
|
|
info->thread_state = TD_STATE_BLOCKED;
|
|
break;
|
|
case PT_STATE_BLOCKED_QUEUE:
|
|
info->thread_state = TD_STATE_SLEEPING;
|
|
break;
|
|
case PT_STATE_SUSPENDED:
|
|
info->thread_state = TD_STATE_SUSPENDED;
|
|
break;
|
|
#endif
|
|
case PT_STATE_ZOMBIE:
|
|
info->thread_state = TD_STATE_ZOMBIE;
|
|
break;
|
|
default:
|
|
info->thread_state = TD_STATE_UNKNOWN;
|
|
}
|
|
|
|
info->thread_type = TD_TYPE_USER;
|
|
|
|
if ((val = READ(thread->proc,
|
|
thread->addr + offsetof(struct __pthread_st, pt_stack),
|
|
&info->thread_stack, sizeof(stack_t))) != 0)
|
|
return val;
|
|
|
|
if ((val = READ(thread->proc,
|
|
thread->addr + offsetof(struct __pthread_st, pt_joiners),
|
|
&queue, sizeof(queue))) != 0)
|
|
return val;
|
|
|
|
if (PTQ_EMPTY(&queue))
|
|
info->thread_hasjoiners = 0;
|
|
else
|
|
info->thread_hasjoiners = 1;
|
|
|
|
if ((val = READ(thread->proc,
|
|
thread->addr + offsetof(struct __pthread_st, pt_errno),
|
|
&info->thread_errno, sizeof(info->thread_errno))) != 0)
|
|
return val;
|
|
|
|
if ((val = READ(thread->proc,
|
|
thread->addr + offsetof(struct __pthread_st, pt_num),
|
|
&info->thread_id, sizeof(info->thread_id))) != 0)
|
|
return val;
|
|
|
|
info->thread_id += TN_OFFSET;
|
|
|
|
#ifdef XXXLWP
|
|
if ((val = READ(thread->proc,
|
|
thread->addr + offsetof(struct __pthread_st, pt_sigmask),
|
|
&info->thread_sigmask, sizeof(info->thread_sigmask))) != 0)
|
|
return val;
|
|
|
|
if ((val = READ(thread->proc,
|
|
thread->addr + offsetof(struct __pthread_st, pt_siglist),
|
|
&info->thread_sigpending, sizeof(info->thread_sigpending))) != 0)
|
|
return val;
|
|
#endif
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
td_thr_getname(td_thread_t *thread, char *name, int len)
|
|
{
|
|
int val, tmp;
|
|
caddr_t nameaddr;
|
|
|
|
|
|
val = READ(thread->proc, thread->addr, &tmp, sizeof(tmp));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
if (tmp != PT_MAGIC)
|
|
return TD_ERR_BADTHREAD;
|
|
|
|
if ((val = READ(thread->proc,
|
|
thread->addr + offsetof(struct __pthread_st, pt_name),
|
|
&nameaddr, sizeof(nameaddr))) != 0)
|
|
return val;
|
|
|
|
if (nameaddr == 0)
|
|
name[0] = '\0';
|
|
else if ((val = READ(thread->proc, nameaddr,
|
|
name, (size_t)MIN(PTHREAD_MAX_NAMELEN_NP, len))) != 0)
|
|
return val;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
td_thr_getregs(td_thread_t *thread, int regset, void *buf)
|
|
{
|
|
int tmp, val;
|
|
|
|
if ((val = READ(thread->proc,
|
|
thread->addr + offsetof(struct __pthread_st, pt_state),
|
|
&tmp, sizeof(tmp))) != 0)
|
|
return val;
|
|
|
|
switch (tmp) {
|
|
case PT_STATE_RUNNING:
|
|
/*
|
|
* The register state of the thread is live in the
|
|
* inferior process's register state.
|
|
*/
|
|
val = GETREGS(thread->proc, regset, thread->lwp, buf);
|
|
if (val != 0)
|
|
return val;
|
|
break;
|
|
case PT_STATE_ZOMBIE:
|
|
default:
|
|
return TD_ERR_BADTHREAD;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
td_thr_setregs(td_thread_t *thread, int regset, void *buf)
|
|
{
|
|
int val, tmp;
|
|
|
|
if ((val = READ(thread->proc,
|
|
thread->addr + offsetof(struct __pthread_st, pt_state),
|
|
&tmp, sizeof(tmp))) != 0)
|
|
return val;
|
|
|
|
switch (tmp) {
|
|
case PT_STATE_RUNNING:
|
|
/*
|
|
* The register state of the thread is live in the
|
|
* inferior process's register state.
|
|
*/
|
|
val = SETREGS(thread->proc, regset, thread->lwp, buf);
|
|
if (val != 0)
|
|
return val;
|
|
break;
|
|
case PT_STATE_ZOMBIE:
|
|
default:
|
|
return TD_ERR_BADTHREAD;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
td_thr_join_iter(td_thread_t *thread, int (*call)(td_thread_t *, void *),
|
|
void *arg)
|
|
{
|
|
int val;
|
|
caddr_t next;
|
|
td_thread_t *thread2;
|
|
struct pthread_queue_t queue;
|
|
|
|
if ((val = READ(thread->proc,
|
|
thread->addr + offsetof(struct __pthread_st, pt_joiners),
|
|
&queue, sizeof(queue))) != 0)
|
|
return val;
|
|
|
|
next = (void *)queue.ptqh_first;
|
|
while (next != NULL) {
|
|
val = td__getthread(thread->proc, next, &thread2);
|
|
if (val != 0)
|
|
return val;
|
|
val = (*call)(thread2, arg);
|
|
if (val != 0)
|
|
return 0;
|
|
|
|
val = READ(thread->proc,
|
|
next + offsetof(struct __pthread_st, pt_sleep.ptqe_next),
|
|
&next, sizeof(next));
|
|
if (val != 0)
|
|
return val;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
td_sync_info(td_sync_t *s, td_sync_info_t *info)
|
|
{
|
|
int val, magic, n;
|
|
struct pthread_queue_t queue;
|
|
pthread_spin_t slock;
|
|
pthread_t taddr;
|
|
td_proc_t *proc = s->proc;
|
|
|
|
val = READ(proc, s->addr, &magic, sizeof(magic));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
info->sync_type = TD_SYNC_UNKNOWN;
|
|
info->sync_size = 0;
|
|
info->sync_haswaiters = 0;
|
|
switch (magic) {
|
|
case _PT_MUTEX_MAGIC:
|
|
info->sync_type = TD_SYNC_MUTEX;
|
|
info->sync_size = sizeof(struct __pthread_mutex_st);
|
|
if ((val = READ(s->proc,
|
|
s->addr + offsetof(struct __pthread_mutex_st, ptm_blocked),
|
|
&queue, sizeof(queue))) != 0)
|
|
return val;
|
|
|
|
if (!PTQ_EMPTY(&queue))
|
|
info->sync_haswaiters = 1;
|
|
/*
|
|
* The cast to (void *) is to explicitly throw away the
|
|
* volatile qualifier on pthread_spin_t,
|
|
* from __cpu_simple_lock_t.
|
|
*/
|
|
if ((val = READ(proc,
|
|
s->addr + offsetof(struct __pthread_mutex_st, ptm_lock),
|
|
(void *)&slock, sizeof(slock))) != 0)
|
|
return val;
|
|
if (slock == __SIMPLELOCK_LOCKED) {
|
|
info->sync_data.mutex.locked = 1;
|
|
if ((val = READ(proc,
|
|
s->addr + offsetof(struct __pthread_mutex_st,
|
|
ptm_owner),
|
|
&taddr, sizeof(taddr))) != 0)
|
|
return val;
|
|
if ((val = td__getstacksize(proc)) != 0)
|
|
return val;
|
|
taddr = pthread__id(taddr);
|
|
td__getthread(proc, (void *)taddr,
|
|
&info->sync_data.mutex.owner);
|
|
} else
|
|
info->sync_data.mutex.locked = 0;
|
|
break;
|
|
case _PT_COND_MAGIC:
|
|
info->sync_type = TD_SYNC_COND;
|
|
info->sync_size = sizeof(struct __pthread_cond_st);
|
|
if ((val = READ(proc,
|
|
s->addr + offsetof(struct __pthread_cond_st, ptc_waiters),
|
|
&queue, sizeof(queue))) != 0)
|
|
return val;
|
|
if (!PTQ_EMPTY(&queue))
|
|
info->sync_haswaiters = 1;
|
|
break;
|
|
case _PT_SPINLOCK_MAGIC:
|
|
info->sync_type = TD_SYNC_SPIN;
|
|
info->sync_size = sizeof(struct __pthread_spinlock_st);
|
|
if ((val = READ(proc,
|
|
s->addr + offsetof(struct __pthread_spinlock_st, pts_spin),
|
|
(void *)&slock, sizeof(slock))) != 0)
|
|
return val;
|
|
if (slock == __SIMPLELOCK_LOCKED)
|
|
info->sync_data.spin.locked = 1;
|
|
else
|
|
info->sync_data.spin.locked = 0;
|
|
break;
|
|
case PT_MAGIC:
|
|
info->sync_type = TD_SYNC_JOIN;
|
|
info->sync_size = sizeof(struct __pthread_st);
|
|
td__getthread(proc, s->addr,
|
|
&info->sync_data.join.thread);
|
|
if ((val = READ(proc,
|
|
s->addr + offsetof(struct __pthread_st, pt_joiners),
|
|
&queue, sizeof(queue))) != 0)
|
|
return val;
|
|
|
|
if (!PTQ_EMPTY(&queue))
|
|
info->sync_haswaiters = 1;
|
|
break;
|
|
case (int)_PT_RWLOCK_MAGIC:
|
|
info->sync_type = TD_SYNC_RWLOCK;
|
|
info->sync_size = sizeof(struct __pthread_rwlock_st);
|
|
if ((val = READ(proc,
|
|
s->addr + offsetof(struct __pthread_rwlock_st, ptr_rblocked),
|
|
&queue, sizeof(queue))) != 0)
|
|
return val;
|
|
if (!PTQ_EMPTY(&queue))
|
|
info->sync_haswaiters = 1;
|
|
|
|
if ((val = READ(proc,
|
|
s->addr + offsetof(struct __pthread_rwlock_st, ptr_wblocked),
|
|
&queue, sizeof(queue))) != 0)
|
|
return val;
|
|
if (!PTQ_EMPTY(&queue))
|
|
info->sync_haswaiters = 1;
|
|
|
|
|
|
info->sync_data.rwlock.locked = 0;
|
|
if ((val = READ(proc,
|
|
s->addr + offsetof(struct __pthread_rwlock_st, ptr_nreaders),
|
|
&n, sizeof(n))) != 0)
|
|
return val;
|
|
info->sync_data.rwlock.readlocks = n;
|
|
if (n > 0)
|
|
info->sync_data.rwlock.locked = 1;
|
|
|
|
if ((val = READ(proc,
|
|
s->addr + offsetof(struct __pthread_rwlock_st, ptr_writer),
|
|
&taddr, sizeof(taddr))) != 0)
|
|
return val;
|
|
if (taddr != 0) {
|
|
info->sync_data.rwlock.locked = 1;
|
|
td__getthread(proc, (void *)taddr,
|
|
&info->sync_data.rwlock.writeowner);
|
|
}
|
|
/*FALLTHROUGH*/
|
|
default:
|
|
return (0);
|
|
}
|
|
|
|
info->sync_addr = s->addr;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
int
|
|
td_sync_waiters_iter(td_sync_t *s, int (*call)(td_thread_t *, void *),
|
|
void *arg)
|
|
{
|
|
int val, magic;
|
|
caddr_t next;
|
|
struct pthread_queue_t queue;
|
|
td_thread_t *thread;
|
|
|
|
val = READ(s->proc, s->addr, &magic, sizeof(magic));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
switch (magic) {
|
|
case _PT_MUTEX_MAGIC:
|
|
if ((val = READ(s->proc,
|
|
s->addr + offsetof(struct __pthread_mutex_st, ptm_blocked),
|
|
&queue, sizeof(queue))) != 0)
|
|
return val;
|
|
break;
|
|
case _PT_COND_MAGIC:
|
|
if ((val = READ(s->proc,
|
|
s->addr + offsetof(struct __pthread_cond_st, ptc_waiters),
|
|
&queue, sizeof(queue))) != 0)
|
|
return val;
|
|
break;
|
|
case PT_MAGIC:
|
|
/* Redundant with join_iter, but what the hell... */
|
|
if ((val = READ(s->proc,
|
|
s->addr + offsetof(struct __pthread_st, pt_joiners),
|
|
&queue, sizeof(queue))) != 0)
|
|
return val;
|
|
break;
|
|
default:
|
|
return (0);
|
|
}
|
|
|
|
next = (void *)queue.ptqh_first;
|
|
while (next != NULL) {
|
|
val = td__getthread(s->proc, next, &thread);
|
|
if (val != 0)
|
|
return val;
|
|
val = (*call)(thread, arg);
|
|
if (val != 0)
|
|
return 0;
|
|
|
|
val = READ(s->proc,
|
|
next + offsetof(struct __pthread_st, pt_sleep.ptqe_next),
|
|
&next, sizeof(next));
|
|
if (val != 0)
|
|
return val;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
|
|
int
|
|
td_map_addr2sync(td_proc_t *proc, caddr_t addr, td_sync_t **syncp)
|
|
{
|
|
int magic, val;
|
|
|
|
val = READ(proc, addr, &magic, sizeof(magic));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
if ((magic != _PT_MUTEX_MAGIC) &&
|
|
(magic != _PT_COND_MAGIC) &&
|
|
(magic != _PT_SPINLOCK_MAGIC) &&
|
|
(magic != _PT_RWLOCK_MAGIC) &&
|
|
(magic != PT_MAGIC))
|
|
return TD_ERR_NOOBJ;
|
|
|
|
val = td__getsync(proc, addr, syncp);
|
|
if (val != 0)
|
|
return val;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
int
|
|
td_map_pth2thr(td_proc_t *proc, pthread_t thread, td_thread_t **threadp)
|
|
{
|
|
int magic, val;
|
|
|
|
val = READ(proc, (void *)thread, &magic, sizeof(magic));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
if (magic != PT_MAGIC)
|
|
return TD_ERR_NOOBJ;
|
|
|
|
val = td__getthread(proc, (void *)thread, threadp);
|
|
if (val != 0)
|
|
return val;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
td_map_id2thr(td_proc_t *proc, int threadid, td_thread_t **threadp)
|
|
{
|
|
int val, num;
|
|
caddr_t next;
|
|
struct pthread_queue_t allq;
|
|
td_thread_t *thread;
|
|
|
|
|
|
val = READ(proc, proc->allqaddr, &allq, sizeof(allq));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
/* Correct for offset */
|
|
threadid -= TN_OFFSET;
|
|
next = (void *)allq.ptqh_first;
|
|
while (next != NULL) {
|
|
val = READ(proc,
|
|
next + offsetof(struct __pthread_st, pt_num),
|
|
&num, sizeof(num));
|
|
|
|
if (num == threadid)
|
|
break;
|
|
|
|
val = READ(proc,
|
|
next + offsetof(struct __pthread_st, pt_allq.ptqe_next),
|
|
&next, sizeof(next));
|
|
if (val != 0)
|
|
return val;
|
|
}
|
|
|
|
if (next == 0) {
|
|
/* A matching thread was not found. */
|
|
return TD_ERR_NOOBJ;
|
|
}
|
|
|
|
val = td__getthread(proc, next, &thread);
|
|
if (val != 0)
|
|
return val;
|
|
*threadp = thread;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Return the thread handle of the thread running on the given LWP */
|
|
int
|
|
td_map_lwp2thr(td_proc_t *proc, int lwp, td_thread_t **threadp)
|
|
{
|
|
int val, magic;
|
|
struct reg gregs;
|
|
ucontext_t uc;
|
|
void *th;
|
|
|
|
val = GETREGS(proc, 0, lwp, &gregs);
|
|
if (val != 0)
|
|
return val;
|
|
|
|
PTHREAD_REG_TO_UCONTEXT(&uc, &gregs);
|
|
|
|
val = td__getstacksize(proc);
|
|
if (val != 0)
|
|
return val;
|
|
|
|
th = pthread__id(pthread__uc_sp(&uc));
|
|
|
|
val = READ(proc, th, &magic, sizeof(magic));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
if (magic != PT_MAGIC)
|
|
return TD_ERR_NOOBJ;
|
|
|
|
val = td__getthread(proc, th, threadp);
|
|
if (val != 0)
|
|
return val;
|
|
|
|
(*threadp)->lwp = lwp;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
td_map_lwps(td_proc_t *proc)
|
|
{
|
|
int i, val, nlwps;
|
|
td_thread_t *thread;
|
|
|
|
val = READ(proc, proc->maxlwpsaddr, &nlwps, sizeof(nlwps));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
if (nlwps < 1)
|
|
nlwps = 1; /* always at least one LWP */
|
|
|
|
PTQ_FOREACH(thread, &proc->threads, list) {
|
|
thread->lwp = -1;
|
|
}
|
|
|
|
for (i = 1; i <= nlwps; i++) {
|
|
/*
|
|
* Errors are deliberately ignored for the call to
|
|
* td_map_lwp2thr(); it is entirely likely that not
|
|
* all LWPs in the range 1..nlwps exist, and that's
|
|
* not a problem.
|
|
*/
|
|
td_map_lwp2thr(proc, i, &thread);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
td_tsd_iter(td_proc_t *proc,
|
|
int (*call)(pthread_key_t, void (*)(void *), void *), void *arg)
|
|
{
|
|
int val;
|
|
int i, allocated;
|
|
void (*destructor)(void *);
|
|
|
|
for (i = 0; i < PTHREAD_KEYS_MAX; i++) {
|
|
val = READ(proc, proc->tsdallocaddr + i * sizeof(int),
|
|
&allocated, sizeof(allocated));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
if (allocated) {
|
|
val = READ(proc, proc->tsddestaddr +
|
|
i * sizeof(destructor),
|
|
&destructor, sizeof(destructor));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
val = (call)(i, destructor, arg);
|
|
if (val != 0)
|
|
return val;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Get the synchronization object that the thread is sleeping on */
|
|
int
|
|
td_thr_sleepinfo(td_thread_t *thread, td_sync_t **s)
|
|
{
|
|
int val;
|
|
caddr_t addr;
|
|
|
|
if ((val = READ(thread->proc,
|
|
thread->addr + offsetof(struct __pthread_st, pt_sleepobj),
|
|
&addr, sizeof(addr))) != 0)
|
|
return val;
|
|
|
|
td__getsync(thread->proc, addr, s);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#define DPTQ_REMOVE(head, elm, field) do { \
|
|
int _val; \
|
|
PTQ_ENTRY(__pthread_st) _qent; \
|
|
\
|
|
_val = READ(thread->proc, \
|
|
(elm) + offsetof(struct __pthread_st, field), \
|
|
&_qent, sizeof(_qent)); \
|
|
if (_val != 0) \
|
|
return _val; \
|
|
if (_qent.ptqe_next != NULL) { \
|
|
_val = WRITE(thread->proc, \
|
|
(caddr_t)(void *)_qent.ptqe_next + \
|
|
offsetof(struct __pthread_st, field.ptqe_prev), \
|
|
&_qent.ptqe_prev, sizeof(_qent.ptqe_prev)); \
|
|
if (_val != 0) \
|
|
return _val; \
|
|
} else { \
|
|
_val = WRITE(thread->proc, (head) + \
|
|
offsetof(struct pthread_queue_t, ptqh_last), \
|
|
&_qent.ptqe_prev, sizeof(_qent.ptqe_prev)); \
|
|
if (_val != 0) \
|
|
return _val; \
|
|
} \
|
|
_val = WRITE(thread->proc, (caddr_t)(void *)_qent.ptqe_prev, \
|
|
&_qent.ptqe_next, sizeof(_qent.ptqe_next)); \
|
|
if (_val != 0) \
|
|
return _val; \
|
|
} while (/*CONSTCOND*/0)
|
|
|
|
#define DPTQ_INSERT_TAIL(head, elm, field) do { \
|
|
int _val; \
|
|
struct pthread_queue_t _qhead; \
|
|
PTQ_ENTRY(__pthread_st) _qent; \
|
|
\
|
|
/* if ((head)->ptqh_last == NULL) */ \
|
|
/* (head)->ptqh_last = &(head)->ptqh_first; */ \
|
|
_val = READ(thread->proc, (head), &_qhead, sizeof(_qhead)); \
|
|
\
|
|
if (_val != 0) \
|
|
return _val; \
|
|
if (_qhead.ptqh_last == NULL) \
|
|
_qhead.ptqh_last = (void *)(head); \
|
|
\
|
|
/* (elm)->field.ptqe_prev = (head)->ptqh_last; */ \
|
|
_qent.ptqe_prev = _qhead.ptqh_last; \
|
|
\
|
|
/* *(head)->ptqh_last = (elm); */ \
|
|
_qent.ptqe_next = (void *)elm; \
|
|
_val = WRITE(thread->proc, (caddr_t)(void *)_qhead.ptqh_last, \
|
|
&_qent.ptqe_next, sizeof(_qent.ptqe_next)); \
|
|
if (_val != 0) \
|
|
return _val; \
|
|
\
|
|
/* (elm)->field.ptqe_next = NULL; */ \
|
|
_qent.ptqe_next = NULL; \
|
|
\
|
|
/* (head)->ptqh_last = &(elm)->field.ptqe_next; */ \
|
|
_qhead.ptqh_last = (void *) ((elm) + \
|
|
offsetof(struct __pthread_st, field.ptqe_next)); \
|
|
\
|
|
_val = WRITE(thread->proc, (elm) + \
|
|
offsetof(struct __pthread_st, field), \
|
|
&_qent, sizeof(_qent)); \
|
|
if (_val != 0) \
|
|
return _val; \
|
|
_val = WRITE(thread->proc, \
|
|
(head) + offsetof(struct pthread_queue_t, ptqh_last), \
|
|
&_qhead.ptqh_last, sizeof(_qhead.ptqh_last)); \
|
|
if (_val != 0) \
|
|
return _val; \
|
|
} while (/*CONSTCOND*/0)
|
|
|
|
|
|
/* Suspend a thread from running */
|
|
int
|
|
td_thr_suspend(td_thread_t *thread)
|
|
{
|
|
int tmp, val;
|
|
|
|
/* validate the thread */
|
|
val = READ(thread->proc, thread->addr, &tmp, sizeof(tmp));
|
|
if (val != 0)
|
|
return val;
|
|
if (tmp != PT_MAGIC)
|
|
return TD_ERR_BADTHREAD;
|
|
|
|
/* XXXLWP get lid, suspend the sucker */;
|
|
return 0;
|
|
}
|
|
|
|
/* Restore a suspended thread to its previous state */
|
|
int
|
|
td_thr_resume(td_thread_t *thread)
|
|
{
|
|
int tmp, val;
|
|
|
|
/* validate the thread */
|
|
val = READ(thread->proc, thread->addr, &tmp, sizeof(tmp));
|
|
if (val != 0)
|
|
return val;
|
|
if (tmp != PT_MAGIC)
|
|
return TD_ERR_BADTHREAD;
|
|
|
|
/* XXXLWP get lid */
|
|
val = READ(thread->proc, thread->addr +
|
|
offsetof(struct __pthread_st, pt_flags),
|
|
&tmp, sizeof(tmp));
|
|
if (val != 0)
|
|
return val;
|
|
|
|
/* XXXLWP suspend the sucker */;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
static int
|
|
td__getthread(td_proc_t *proc, caddr_t addr, td_thread_t **threadp)
|
|
{
|
|
td_thread_t *thread;
|
|
|
|
/*
|
|
* Check if we've allocated a descriptor for this thread.
|
|
* Sadly, this makes iterating over a set of threads O(N^2)
|
|
* in the number of threads. More sophisticated data structures
|
|
* can wait.
|
|
*/
|
|
PTQ_FOREACH(thread, &proc->threads, list) {
|
|
if (thread->addr == addr)
|
|
break;
|
|
}
|
|
if (thread == NULL) {
|
|
thread = malloc(sizeof(*thread));
|
|
if (thread == NULL)
|
|
return TD_ERR_NOMEM;
|
|
thread->proc = proc;
|
|
thread->addr = addr;
|
|
thread->lwp = 0;
|
|
PTQ_INSERT_HEAD(&proc->threads, thread, list);
|
|
}
|
|
|
|
*threadp = thread;
|
|
return 0;
|
|
}
|
|
|
|
|
|
static int
|
|
td__getsync(td_proc_t *proc, caddr_t addr, td_sync_t **syncp)
|
|
{
|
|
td_sync_t *s;
|
|
|
|
/* Check if we've allocated a descriptor for this object. */
|
|
PTQ_FOREACH(s, &proc->syncs, list) {
|
|
if (s->addr == addr)
|
|
break;
|
|
}
|
|
/* Allocate a fresh one */
|
|
if (s == NULL) {
|
|
s = malloc(sizeof(*s));
|
|
if (s == NULL)
|
|
return TD_ERR_NOMEM;
|
|
s->proc = proc;
|
|
s->addr = addr;
|
|
PTQ_INSERT_HEAD(&proc->syncs, s, list);
|
|
}
|
|
|
|
*syncp = s;
|
|
return 0;
|
|
}
|
|
|
|
|
|
int
|
|
td_thr_tsd(td_thread_t *thread, pthread_key_t key, void **value)
|
|
{
|
|
int val;
|
|
|
|
val = READ(thread->proc, thread->addr +
|
|
offsetof(struct __pthread_st, pt_specific) +
|
|
key * sizeof(void *), value, sizeof(*value));
|
|
|
|
return val;
|
|
}
|
|
|
|
|
|
static int
|
|
td__getstacksize(td_proc_t *proc)
|
|
{
|
|
int lg, val;
|
|
|
|
if (proc->stacksizeaddr == NULL)
|
|
return 0;
|
|
|
|
val = READ(proc, proc->stacksizeaddr, &lg, sizeof(int));
|
|
if (val != 0)
|
|
return 0;
|
|
if (lg != proc->stacksizelg) {
|
|
proc->stacksizelg = lg;
|
|
proc->stacksize = (1 << lg);
|
|
proc->stackmask = proc->stacksize - 1;
|
|
}
|
|
return 0;
|
|
}
|