2002-10-23 21:31:10 +04:00
|
|
|
/* Threading routines */
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
/*
|
|
|
|
** Copyright 2001-2002, Travis Geiselbrecht. All rights reserved.
|
|
|
|
** Distributed under the terms of the NewOS License.
|
|
|
|
*/
|
|
|
|
|
2003-05-03 20:20:38 +04:00
|
|
|
#include <OS.h>
|
|
|
|
#include <Errors.h>
|
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
#include <kernel.h>
|
|
|
|
#include <debug.h>
|
|
|
|
#include <console.h>
|
|
|
|
#include <thread.h>
|
|
|
|
#include <arch/thread.h>
|
|
|
|
#include <khash.h>
|
|
|
|
#include <int.h>
|
|
|
|
#include <smp.h>
|
|
|
|
#include <timer.h>
|
|
|
|
#include <cpu.h>
|
|
|
|
#include <arch/int.h>
|
|
|
|
#include <arch/cpu.h>
|
|
|
|
#include <arch/vm.h>
|
|
|
|
#include <sem.h>
|
|
|
|
#include <port.h>
|
|
|
|
#include <vfs.h>
|
|
|
|
#include <elf.h>
|
2002-10-30 02:07:06 +03:00
|
|
|
#include <malloc.h>
|
2002-07-09 16:24:59 +04:00
|
|
|
#include <user_runtime.h>
|
2003-05-03 20:20:38 +04:00
|
|
|
#include <boot/stage2.h>
|
2002-07-09 16:24:59 +04:00
|
|
|
#include <string.h>
|
2003-01-12 19:30:09 +03:00
|
|
|
#include <kimage.h>
|
2002-07-09 16:24:59 +04:00
|
|
|
#include <stdio.h>
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include <resource.h>
|
2002-07-12 02:21:56 +04:00
|
|
|
#include <kerrors.h>
|
2002-07-18 18:07:56 +04:00
|
|
|
#include <syscalls.h>
|
2003-01-27 06:17:36 +03:00
|
|
|
#include <ksignal.h>
|
2003-01-07 12:48:01 +03:00
|
|
|
#include <tls.h>
|
2002-07-18 18:07:56 +04:00
|
|
|
|
2002-11-29 11:42:52 +03:00
|
|
|
#define TRACE_THREAD 0
|
|
|
|
#if TRACE_THREAD
|
|
|
|
# define TRACE(x) dprintf x
|
|
|
|
#else
|
|
|
|
# define TRACE(x) ;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
#define THREAD_MAX_MESSAGE_SIZE 65536
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
struct thread_key {
|
|
|
|
thread_id id;
|
|
|
|
};
|
|
|
|
|
|
|
|
// global
|
2002-10-26 20:13:36 +04:00
|
|
|
spinlock thread_spinlock = 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// thread list
|
|
|
|
static struct thread *idle_threads[MAX_BOOT_CPUS];
|
|
|
|
static void *thread_hash = NULL;
|
2002-08-05 00:10:06 +04:00
|
|
|
static thread_id next_thread_id = 1;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
static sem_id gSnoozeSem = -1;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// death stacks - used temporarily as a thread cleans itself up
|
|
|
|
struct death_stack {
|
|
|
|
region_id rid;
|
|
|
|
addr address;
|
|
|
|
bool in_use;
|
|
|
|
};
|
|
|
|
static struct death_stack *death_stacks;
|
|
|
|
static unsigned int num_death_stacks;
|
|
|
|
static unsigned int volatile death_stack_bitmap;
|
|
|
|
static sem_id death_stack_sem;
|
|
|
|
|
2002-08-16 17:14:29 +04:00
|
|
|
// The dead queue is used as a pool from which to retrieve and reuse previously
|
|
|
|
// allocated thread structs when creating a new thread. It should be gone once
|
|
|
|
// the slab allocator is in.
|
2002-08-04 03:39:50 +04:00
|
|
|
struct thread_queue dead_q;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-10-26 16:58:52 +04:00
|
|
|
static void thread_kthread_entry(void);
|
2002-07-09 16:24:59 +04:00
|
|
|
static void thread_kthread_exit(void);
|
|
|
|
|
2002-08-16 17:14:29 +04:00
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
static void
|
|
|
|
insert_thread_into_team(struct team *p, struct thread *t)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-08-03 04:41:27 +04:00
|
|
|
t->team_next = p->thread_list;
|
2002-07-09 16:24:59 +04:00
|
|
|
p->thread_list = t;
|
|
|
|
p->num_threads++;
|
2002-10-05 23:38:42 +04:00
|
|
|
if (p->num_threads == 1) {
|
2002-07-09 16:24:59 +04:00
|
|
|
// this was the first thread
|
|
|
|
p->main_thread = t;
|
|
|
|
}
|
2002-08-03 04:41:27 +04:00
|
|
|
t->team = p;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
|
|
|
|
static void
|
|
|
|
remove_thread_from_team(struct team *p, struct thread *t)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread *temp, *last = NULL;
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
for (temp = p->thread_list; temp != NULL; temp = temp->team_next) {
|
|
|
|
if (temp == t) {
|
|
|
|
if (last == NULL)
|
2002-08-03 04:41:27 +04:00
|
|
|
p->thread_list = temp->team_next;
|
2002-10-05 23:38:42 +04:00
|
|
|
else
|
2002-08-03 04:41:27 +04:00
|
|
|
last->team_next = temp->team_next;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
p->num_threads--;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
last = temp;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
|
|
|
|
static int
|
|
|
|
thread_struct_compare(void *_t, const void *_key)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread *t = _t;
|
|
|
|
const struct thread_key *key = _key;
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
if (t->id == key->id)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
return 1;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-11-29 11:42:52 +03:00
|
|
|
static uint32
|
|
|
|
thread_struct_hash(void *_t, const void *_key, uint32 range)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread *t = _t;
|
|
|
|
const struct thread_key *key = _key;
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
if (t != NULL)
|
2002-11-29 11:42:52 +03:00
|
|
|
return t->id % range;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-11-29 11:42:52 +03:00
|
|
|
return key->id % range;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
|
|
|
|
static struct thread *
|
|
|
|
create_thread_struct(const char *name)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread *t;
|
2002-12-04 19:40:07 +03:00
|
|
|
cpu_status state;
|
2002-08-19 12:28:39 +04:00
|
|
|
char temp[64];
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-07-25 05:05:51 +04:00
|
|
|
state = disable_interrupts();
|
2002-07-09 16:24:59 +04:00
|
|
|
GRAB_THREAD_LOCK();
|
|
|
|
t = thread_dequeue(&dead_q);
|
|
|
|
RELEASE_THREAD_LOCK();
|
2002-07-25 05:05:51 +04:00
|
|
|
restore_interrupts(state);
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
if (t == NULL) {
|
2002-10-30 02:07:06 +03:00
|
|
|
t = (struct thread *)malloc(sizeof(struct thread));
|
2002-10-05 23:38:42 +04:00
|
|
|
if (t == NULL)
|
2002-07-09 16:24:59 +04:00
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
|
|
|
|
strncpy(&t->name[0], name, SYS_MAX_OS_NAME_LEN-1);
|
|
|
|
t->name[SYS_MAX_OS_NAME_LEN-1] = 0;
|
|
|
|
|
|
|
|
t->id = atomic_add(&next_thread_id, 1);
|
2002-08-03 04:41:27 +04:00
|
|
|
t->team = NULL;
|
2002-07-09 16:24:59 +04:00
|
|
|
t->cpu = NULL;
|
|
|
|
t->sem_blocking = -1;
|
2002-12-04 19:40:07 +03:00
|
|
|
t->fault_handler = 0;
|
2002-12-03 17:11:35 +03:00
|
|
|
t->page_faults_allowed = 1;
|
2002-07-09 16:24:59 +04:00
|
|
|
t->kernel_stack_region_id = -1;
|
|
|
|
t->kernel_stack_base = 0;
|
|
|
|
t->user_stack_region_id = -1;
|
|
|
|
t->user_stack_base = 0;
|
2003-01-07 12:48:01 +03:00
|
|
|
t->user_local_storage = 0;
|
|
|
|
t->kernel_errno = 0;
|
2002-08-03 04:41:27 +04:00
|
|
|
t->team_next = NULL;
|
2003-01-27 06:17:36 +03:00
|
|
|
t->queue_next = NULL;
|
2002-07-09 16:24:59 +04:00
|
|
|
t->priority = -1;
|
2003-04-18 13:38:28 +04:00
|
|
|
t->args1 = NULL; t->args2 = NULL;
|
2002-10-23 21:31:10 +04:00
|
|
|
t->sig_pending = 0;
|
|
|
|
t->sig_block_mask = 0;
|
|
|
|
memset(t->sig_action, 0, 32 * sizeof(struct sigaction));
|
2002-07-09 16:24:59 +04:00
|
|
|
t->in_kernel = true;
|
|
|
|
t->user_time = 0;
|
|
|
|
t->kernel_time = 0;
|
2002-10-23 21:31:10 +04:00
|
|
|
t->last_time = 0;
|
2003-01-30 00:59:37 +03:00
|
|
|
t->last_time_type = KERNEL_TIME;
|
2002-10-26 05:11:15 +04:00
|
|
|
t->return_code = 0;
|
|
|
|
t->return_flags = 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
sprintf(temp, "thread_0x%lx_retcode_sem", t->id);
|
2002-08-19 12:28:39 +04:00
|
|
|
t->return_code_sem = create_sem(0, temp);
|
2002-10-05 23:38:42 +04:00
|
|
|
if (t->return_code_sem < 0)
|
2002-08-19 12:28:39 +04:00
|
|
|
goto err1;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
sprintf(temp, "%s data write sem", t->name);
|
|
|
|
t->msg.write_sem = create_sem(1, temp);
|
|
|
|
if (t->msg.write_sem < 0)
|
2002-07-09 16:24:59 +04:00
|
|
|
goto err2;
|
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
sprintf(temp, "%s data read sem", t->name);
|
|
|
|
t->msg.read_sem = create_sem(0, temp);
|
|
|
|
if (t->msg.read_sem < 0)
|
|
|
|
goto err3;
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
if (arch_thread_init_thread_struct(t) < 0)
|
2002-08-19 12:28:39 +04:00
|
|
|
goto err4;
|
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
return t;
|
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
err4:
|
2002-10-26 05:11:15 +04:00
|
|
|
delete_sem_etc(t->msg.read_sem, -1, false);
|
2002-08-19 12:28:39 +04:00
|
|
|
err3:
|
2002-10-26 05:11:15 +04:00
|
|
|
delete_sem_etc(t->msg.write_sem, -1, false);
|
2002-07-09 16:24:59 +04:00
|
|
|
err2:
|
2002-10-26 05:11:15 +04:00
|
|
|
delete_sem_etc(t->return_code_sem, -1, false);
|
2002-07-09 16:24:59 +04:00
|
|
|
err1:
|
2002-10-30 02:07:06 +03:00
|
|
|
free(t);
|
2002-07-09 16:24:59 +04:00
|
|
|
err:
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
|
|
|
|
static void
|
|
|
|
delete_thread_struct(struct thread *t)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-08-19 12:28:39 +04:00
|
|
|
if (t->return_code_sem >= 0)
|
2002-10-26 05:11:15 +04:00
|
|
|
delete_sem_etc(t->return_code_sem, -1, false);
|
2002-08-19 12:28:39 +04:00
|
|
|
if (t->msg.write_sem >= 0)
|
2002-10-26 05:11:15 +04:00
|
|
|
delete_sem_etc(t->msg.write_sem, -1, false);
|
2002-08-19 12:28:39 +04:00
|
|
|
if (t->msg.read_sem >= 0)
|
2002-10-26 05:11:15 +04:00
|
|
|
delete_sem_etc(t->msg.read_sem, -1, false);
|
2002-10-30 02:07:06 +03:00
|
|
|
free(t);
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-07 12:48:01 +03:00
|
|
|
/** Initializes the thread and jumps to its userspace entry point.
|
|
|
|
* This function is called at creation time of every user thread,
|
|
|
|
* but not for the team's main threads.
|
|
|
|
*/
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
static int
|
|
|
|
_create_user_thread_kentry(void)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2003-01-07 12:48:01 +03:00
|
|
|
struct thread *thread = thread_get_current_thread();
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// a signal may have been delivered here
|
2002-10-23 21:31:10 +04:00
|
|
|
// thread_atkernel_exit();
|
2003-01-30 00:59:37 +03:00
|
|
|
// start tracking kernel & user time
|
2003-01-07 12:48:01 +03:00
|
|
|
thread->last_time = system_time();
|
2003-01-30 00:59:37 +03:00
|
|
|
thread->last_time_type = KERNEL_TIME;
|
2003-01-07 12:48:01 +03:00
|
|
|
thread->in_kernel = false;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// jump to the entry point in user space
|
2003-04-18 13:38:28 +04:00
|
|
|
arch_thread_enter_uspace(thread, (addr)thread->entry, thread->args1, thread->args2);
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// never get here
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-07 12:48:01 +03:00
|
|
|
/** Initializes the thread and calls it kernel space entry point.
|
|
|
|
*/
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
static int
|
|
|
|
_create_kernel_thread_kentry(void)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2003-01-07 12:48:01 +03:00
|
|
|
struct thread *thread = thread_get_current_thread();
|
2002-07-09 16:24:59 +04:00
|
|
|
int (*func)(void *args);
|
|
|
|
|
2003-01-30 00:59:37 +03:00
|
|
|
// start tracking kernel & user time
|
2003-01-07 12:48:01 +03:00
|
|
|
thread->last_time = system_time();
|
2003-01-30 00:59:37 +03:00
|
|
|
thread->last_time_type = KERNEL_TIME;
|
2003-01-07 12:48:01 +03:00
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
// call the entry function with the appropriate args
|
2003-01-07 12:48:01 +03:00
|
|
|
func = (void *)thread->entry;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2003-04-18 13:38:28 +04:00
|
|
|
return func(thread->args1);
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
static thread_id
|
2003-04-18 13:38:28 +04:00
|
|
|
create_thread(const char *name, team_id teamID, thread_func entry, void *args1, void *args2, int32 priority, bool kernel)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread *t;
|
2003-01-07 12:48:01 +03:00
|
|
|
struct team *team;
|
2002-12-04 19:40:07 +03:00
|
|
|
cpu_status state;
|
2002-07-09 16:24:59 +04:00
|
|
|
char stack_name[64];
|
|
|
|
bool abort = false;
|
2003-01-07 12:48:01 +03:00
|
|
|
addr mainThreadStackBase = USER_STACK_REGION + USER_STACK_REGION_SIZE;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
t = create_thread_struct(name);
|
2002-09-24 04:08:50 +04:00
|
|
|
if (t == NULL)
|
2002-07-14 14:10:22 +04:00
|
|
|
return ENOMEM;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-08-05 00:10:06 +04:00
|
|
|
t->priority = priority == -1 ? B_NORMAL_PRIORITY : priority;
|
2002-10-05 23:38:42 +04:00
|
|
|
// ToDo: revisit this one - the thread might go to early to B_THREAD_SUSPENDED
|
2002-08-05 00:10:06 +04:00
|
|
|
// t->state = THREAD_STATE_BIRTH;
|
|
|
|
t->state = B_THREAD_SUSPENDED; // Is this right?
|
|
|
|
t->next_state = B_THREAD_SUSPENDED;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-07-25 05:05:51 +04:00
|
|
|
state = disable_interrupts();
|
2002-07-09 16:24:59 +04:00
|
|
|
GRAB_THREAD_LOCK();
|
|
|
|
|
|
|
|
// insert into global list
|
|
|
|
hash_insert(thread_hash, t);
|
|
|
|
RELEASE_THREAD_LOCK();
|
|
|
|
|
2002-08-03 04:41:27 +04:00
|
|
|
GRAB_TEAM_LOCK();
|
|
|
|
// look at the team, make sure it's not being deleted
|
2003-01-07 12:48:01 +03:00
|
|
|
team = team_get_team_struct_locked(teamID);
|
|
|
|
if (team != NULL && team->state != TEAM_STATE_DEATH)
|
|
|
|
insert_thread_into_team(team, t);
|
2002-10-05 23:38:42 +04:00
|
|
|
else
|
2002-07-09 16:24:59 +04:00
|
|
|
abort = true;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-07 12:48:01 +03:00
|
|
|
if (!kernel && team->main_thread)
|
|
|
|
mainThreadStackBase = team->main_thread->user_stack_base;
|
|
|
|
|
2002-08-03 04:41:27 +04:00
|
|
|
RELEASE_TEAM_LOCK();
|
2002-09-24 04:08:50 +04:00
|
|
|
if (abort) {
|
2002-07-09 16:24:59 +04:00
|
|
|
GRAB_THREAD_LOCK();
|
|
|
|
hash_remove(thread_hash, t);
|
|
|
|
RELEASE_THREAD_LOCK();
|
|
|
|
}
|
2002-07-25 05:05:51 +04:00
|
|
|
restore_interrupts(state);
|
2002-09-24 04:08:50 +04:00
|
|
|
if (abort) {
|
2002-07-09 16:24:59 +04:00
|
|
|
delete_thread_struct(t);
|
|
|
|
return ERR_TASK_PROC_DELETED;
|
|
|
|
}
|
|
|
|
|
|
|
|
sprintf(stack_name, "%s_kstack", name);
|
2003-08-19 18:28:11 +04:00
|
|
|
t->kernel_stack_region_id = create_area(stack_name, (void **)&t->kernel_stack_base,
|
|
|
|
B_ANY_KERNEL_ADDRESS, KSTACK_SIZE, B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA);
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
if (t->kernel_stack_region_id < 0)
|
2002-07-09 16:24:59 +04:00
|
|
|
panic("_create_thread: error creating kernel stack!\n");
|
|
|
|
|
2003-04-18 13:38:28 +04:00
|
|
|
t->args1 = args1;
|
|
|
|
t->args2 = args2;
|
2002-07-09 16:24:59 +04:00
|
|
|
t->entry = entry;
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
if (kernel) {
|
2002-07-09 16:24:59 +04:00
|
|
|
// this sets up an initial kthread stack that runs the entry
|
2003-01-07 12:48:01 +03:00
|
|
|
|
|
|
|
// Note: whatever function wants to set up a user stack later for this thread
|
|
|
|
// must initialize the TLS for it
|
|
|
|
arch_thread_init_kthread_stack(t, &_create_kernel_thread_kentry, &thread_kthread_entry, &thread_kthread_exit);
|
2002-07-09 16:24:59 +04:00
|
|
|
} else {
|
|
|
|
// create user stack
|
2003-01-06 11:10:03 +03:00
|
|
|
|
|
|
|
// ToDo: make this better. For now just keep trying to create a stack
|
|
|
|
// until we find a spot.
|
2003-01-07 12:48:01 +03:00
|
|
|
// -> implement B_BASE_ADDRESS for create_area() and use that one!
|
|
|
|
// (we can then also eliminate the mainThreadStackBase variable)
|
|
|
|
|
|
|
|
// ToDo: should we move the stack creation to _create_user_thread_kentry()?
|
|
|
|
|
|
|
|
// the stack will be between USER_STACK_REGION and the main thread stack area
|
|
|
|
// (the user stack of the main thread is created in team_create_team())
|
|
|
|
t->user_stack_base = USER_STACK_REGION;
|
2003-01-06 11:10:03 +03:00
|
|
|
|
2003-01-07 12:48:01 +03:00
|
|
|
while (t->user_stack_base < mainThreadStackBase) {
|
|
|
|
sprintf(stack_name, "%s_stack%ld", team->name, t->id);
|
2003-08-19 21:38:13 +04:00
|
|
|
t->user_stack_region_id = create_area_etc(team, stack_name,
|
|
|
|
(void **)&t->user_stack_base, B_EXACT_ADDRESS,
|
|
|
|
STACK_SIZE + TLS_SIZE, B_NO_LOCK, B_READ_AREA | B_WRITE_AREA);
|
2003-01-07 12:48:01 +03:00
|
|
|
if (t->user_stack_region_id >= 0)
|
2002-07-09 16:24:59 +04:00
|
|
|
break;
|
2003-01-07 12:48:01 +03:00
|
|
|
|
|
|
|
t->user_stack_base += STACK_SIZE + TLS_SIZE;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
2002-09-24 04:08:50 +04:00
|
|
|
if (t->user_stack_region_id < 0)
|
2002-07-09 16:24:59 +04:00
|
|
|
panic("_create_thread: unable to create user stack!\n");
|
|
|
|
|
2003-01-07 12:48:01 +03:00
|
|
|
// now that the TLS area is allocated, initialize TLS
|
|
|
|
arch_thread_init_tls(t);
|
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
// copy the user entry over to the args field in the thread struct
|
|
|
|
// the function this will call will immediately switch the thread into
|
|
|
|
// user space.
|
2003-01-07 12:48:01 +03:00
|
|
|
arch_thread_init_kthread_stack(t, &_create_user_thread_kentry, &thread_kthread_entry, &thread_kthread_exit);
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-08-05 00:10:06 +04:00
|
|
|
t->state = B_THREAD_SUSPENDED;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
return t->id;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
static const char *
|
|
|
|
state_to_text(int state)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-09-24 04:08:50 +04:00
|
|
|
switch (state) {
|
2002-08-05 00:10:06 +04:00
|
|
|
case B_THREAD_READY:
|
2002-07-09 16:24:59 +04:00
|
|
|
return "READY";
|
2002-08-05 00:10:06 +04:00
|
|
|
case B_THREAD_RUNNING:
|
2002-07-09 16:24:59 +04:00
|
|
|
return "RUNNING";
|
2002-08-05 00:10:06 +04:00
|
|
|
case B_THREAD_WAITING:
|
2002-07-09 16:24:59 +04:00
|
|
|
return "WAITING";
|
2002-08-05 00:10:06 +04:00
|
|
|
case B_THREAD_SUSPENDED:
|
2002-07-09 16:24:59 +04:00
|
|
|
return "SUSPEND";
|
|
|
|
case THREAD_STATE_FREE_ON_RESCHED:
|
|
|
|
return "DEATH";
|
|
|
|
default:
|
|
|
|
return "UNKNOWN";
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
static struct thread *last_thread_dumped = NULL;
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
static void
|
|
|
|
_dump_thread_info(struct thread *t)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
dprintf("THREAD: %p\n", t);
|
2002-09-24 04:08:50 +04:00
|
|
|
dprintf("id: 0x%lx\n", t->id);
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("name: '%s'\n", t->name);
|
2002-08-03 04:41:27 +04:00
|
|
|
dprintf("all_next: %p\nteam_next: %p\nq_next: %p\n",
|
2003-01-27 06:17:36 +03:00
|
|
|
t->all_next, t->team_next, t->queue_next);
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("priority: 0x%x\n", t->priority);
|
|
|
|
dprintf("state: %s\n", state_to_text(t->state));
|
|
|
|
dprintf("next_state: %s\n", state_to_text(t->next_state));
|
|
|
|
dprintf("cpu: %p ", t->cpu);
|
2002-09-24 04:08:50 +04:00
|
|
|
if (t->cpu)
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("(%d)\n", t->cpu->info.cpu_num);
|
|
|
|
else
|
|
|
|
dprintf("\n");
|
2002-10-23 21:31:10 +04:00
|
|
|
dprintf("sig_pending: 0x%lx\n", t->sig_pending);
|
2003-01-08 12:27:55 +03:00
|
|
|
dprintf("in_kernel: %d\n", t->in_kernel);
|
|
|
|
dprintf("sem_blocking: 0x%lx\n", t->sem_blocking);
|
|
|
|
dprintf("sem_count: 0x%x\n", t->sem_count);
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("sem_deleted_retcode: 0x%x\n", t->sem_deleted_retcode);
|
2003-01-08 12:27:55 +03:00
|
|
|
dprintf("sem_errcode: 0x%x\n", t->sem_errcode);
|
|
|
|
dprintf("sem_flags: 0x%x\n", t->sem_flags);
|
|
|
|
dprintf("fault_handler: %p\n", (void *)t->fault_handler);
|
2003-04-18 13:38:28 +04:00
|
|
|
dprintf("args: %p %p\n", t->args1, t->args2);
|
2003-01-08 12:27:55 +03:00
|
|
|
dprintf("entry: %p\n", (void *)t->entry);
|
|
|
|
dprintf("team: %p\n", t->team);
|
2002-09-24 04:08:50 +04:00
|
|
|
dprintf("return_code_sem: 0x%lx\n", t->return_code_sem);
|
|
|
|
dprintf("kernel_stack_region_id: 0x%lx\n", t->kernel_stack_region_id);
|
2003-01-08 12:27:55 +03:00
|
|
|
dprintf("kernel_stack_base: %p\n", (void *)t->kernel_stack_base);
|
2002-09-24 04:08:50 +04:00
|
|
|
dprintf("user_stack_region_id: 0x%lx\n", t->user_stack_region_id);
|
2003-01-08 12:27:55 +03:00
|
|
|
dprintf("user_stack_base: %p\n", (void *)t->user_stack_base);
|
|
|
|
dprintf("user_local_storage: %p\n", (void *)t->user_local_storage);
|
|
|
|
dprintf("kernel_errno: %d\n", t->kernel_errno);
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("kernel_time: %Ld\n", t->kernel_time);
|
|
|
|
dprintf("user_time: %Ld\n", t->user_time);
|
|
|
|
dprintf("architecture dependant section:\n");
|
|
|
|
arch_thread_dump_info(&t->arch_info);
|
|
|
|
|
|
|
|
last_thread_dumped = t;
|
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
static int
|
|
|
|
dump_thread_info(int argc, char **argv)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread *t;
|
|
|
|
int id = -1;
|
|
|
|
struct hash_iterator i;
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
if (argc < 2) {
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("thread: not enough arguments\n");
|
2002-07-18 02:07:37 +04:00
|
|
|
return 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
// if the argument looks like a hex number, treat it as such
|
2002-10-31 17:32:56 +03:00
|
|
|
if (strlen(argv[1]) > 2 && argv[1][0] == '0' && argv[1][1] == 'x')
|
|
|
|
id = strtoul(argv[1], NULL, 16);
|
|
|
|
else
|
|
|
|
id = atoi(argv[1]);
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// walk through the thread list, trying to match name or id
|
|
|
|
hash_open(thread_hash, &i);
|
2002-09-24 04:08:50 +04:00
|
|
|
while ((t = hash_next(thread_hash, &i)) != NULL) {
|
|
|
|
if ((t->name && strcmp(argv[1], t->name) == 0) || t->id == id) {
|
2002-07-09 16:24:59 +04:00
|
|
|
_dump_thread_info(t);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
hash_close(thread_hash, &i, false);
|
2002-07-18 02:07:37 +04:00
|
|
|
return 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
static int
|
|
|
|
dump_thread_list(int argc, char **argv)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread *t;
|
|
|
|
struct hash_iterator i;
|
|
|
|
|
|
|
|
hash_open(thread_hash, &i);
|
2002-09-24 04:08:50 +04:00
|
|
|
while ((t = hash_next(thread_hash, &i)) != NULL) {
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("%p", t);
|
2002-09-24 04:08:50 +04:00
|
|
|
if (t->name != NULL)
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("\t%32s", t->name);
|
|
|
|
else
|
|
|
|
dprintf("\t%32s", "<NULL>");
|
2002-09-24 04:08:50 +04:00
|
|
|
dprintf("\t0x%lx", t->id);
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("\t%16s", state_to_text(t->state));
|
2002-09-24 04:08:50 +04:00
|
|
|
if (t->cpu)
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("\t%d", t->cpu->info.cpu_num);
|
|
|
|
else
|
|
|
|
dprintf("\tNOCPU");
|
|
|
|
dprintf("\t0x%lx\n", t->kernel_stack_base);
|
|
|
|
}
|
|
|
|
hash_close(thread_hash, &i, false);
|
2002-07-18 02:07:37 +04:00
|
|
|
return 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
static int
|
|
|
|
dump_next_thread_in_q(int argc, char **argv)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread *t = last_thread_dumped;
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
if (t == NULL) {
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("no thread previously dumped. Examine a thread first.\n");
|
2002-07-18 02:07:37 +04:00
|
|
|
return 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
dprintf("next thread in queue after thread @ %p\n", t);
|
2003-01-27 06:17:36 +03:00
|
|
|
if (t->queue_next != NULL)
|
|
|
|
_dump_thread_info(t->queue_next);
|
2002-10-05 23:38:42 +04:00
|
|
|
else
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("NULL\n");
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-07-18 02:07:37 +04:00
|
|
|
return 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
static int
|
|
|
|
dump_next_thread_in_all_list(int argc, char **argv)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread *t = last_thread_dumped;
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
if (t == NULL) {
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("no thread previously dumped. Examine a thread first.\n");
|
2002-07-18 02:07:37 +04:00
|
|
|
return 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
dprintf("next thread in global list after thread @ %p\n", t);
|
2002-10-05 23:38:42 +04:00
|
|
|
if (t->all_next != NULL)
|
2002-07-09 16:24:59 +04:00
|
|
|
_dump_thread_info(t->all_next);
|
2002-10-05 23:38:42 +04:00
|
|
|
else
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("NULL\n");
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-07-18 02:07:37 +04:00
|
|
|
return 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
static int
|
|
|
|
dump_next_thread_in_team(int argc, char **argv)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread *t = last_thread_dumped;
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
if (t == NULL) {
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("no thread previously dumped. Examine a thread first.\n");
|
2002-07-18 02:07:37 +04:00
|
|
|
return 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-08-03 04:41:27 +04:00
|
|
|
dprintf("next thread in team after thread @ %p\n", t);
|
2002-10-05 23:38:42 +04:00
|
|
|
if (t->team_next != NULL)
|
2002-08-03 04:41:27 +04:00
|
|
|
_dump_thread_info(t->team_next);
|
2002-10-05 23:38:42 +04:00
|
|
|
else
|
2002-07-09 16:24:59 +04:00
|
|
|
dprintf("NULL\n");
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-07-18 02:07:37 +04:00
|
|
|
return 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
static int
|
|
|
|
get_death_stack(void)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-12-04 19:40:07 +03:00
|
|
|
cpu_status state;
|
2002-07-09 16:24:59 +04:00
|
|
|
unsigned int bit;
|
2002-12-04 19:40:07 +03:00
|
|
|
int i;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
acquire_sem(death_stack_sem);
|
|
|
|
|
|
|
|
// grap the thread lock, find a free spot and release
|
2002-07-25 05:05:51 +04:00
|
|
|
state = disable_interrupts();
|
2002-07-09 16:24:59 +04:00
|
|
|
GRAB_THREAD_LOCK();
|
|
|
|
bit = death_stack_bitmap;
|
|
|
|
bit = (~bit)&~((~bit)-1);
|
|
|
|
death_stack_bitmap |= bit;
|
|
|
|
RELEASE_THREAD_LOCK();
|
|
|
|
|
|
|
|
// sanity checks
|
2002-09-24 04:08:50 +04:00
|
|
|
if (!bit)
|
2002-07-09 16:24:59 +04:00
|
|
|
panic("get_death_stack: couldn't find free stack!\n");
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
if (bit & (bit - 1))
|
|
|
|
panic("get_death_stack: impossible bitmap result!\n");
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// bit to number
|
2002-09-24 04:08:50 +04:00
|
|
|
for (i = -1; bit; i++) {
|
2002-07-09 16:24:59 +04:00
|
|
|
bit >>= 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
// dprintf("get_death_stack: returning 0x%lx\n", death_stacks[i].address);
|
|
|
|
|
|
|
|
return i;
|
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
static void
|
|
|
|
put_death_stack_and_reschedule(unsigned int index)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
// dprintf("put_death_stack...: passed %d\n", index);
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
if (index >= num_death_stacks)
|
2002-07-09 16:24:59 +04:00
|
|
|
panic("put_death_stack: passed invalid stack index %d\n", index);
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
if (!(death_stack_bitmap & (1 << index)))
|
2002-07-09 16:24:59 +04:00
|
|
|
panic("put_death_stack: passed invalid stack index %d\n", index);
|
|
|
|
|
2002-07-25 05:05:51 +04:00
|
|
|
disable_interrupts();
|
2002-07-09 16:24:59 +04:00
|
|
|
GRAB_THREAD_LOCK();
|
|
|
|
|
|
|
|
death_stack_bitmap &= ~(1 << index);
|
|
|
|
|
|
|
|
release_sem_etc(death_stack_sem, 1, B_DO_NOT_RESCHEDULE);
|
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
scheduler_reschedule();
|
2002-07-13 00:24:28 +04:00
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
// this function gets run by a new thread before anything else
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
static void
|
2002-10-26 16:58:52 +04:00
|
|
|
thread_kthread_entry(void)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
// simulates the thread spinlock release that would occur if the thread had been
|
|
|
|
// rescheded from. The resched didn't happen because the thread is new.
|
|
|
|
RELEASE_THREAD_LOCK();
|
2002-07-25 05:05:51 +04:00
|
|
|
enable_interrupts(); // this essentially simulates a return-from-interrupt
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
// used to pass messages between thread_exit and thread_exit2
|
2002-09-24 04:08:50 +04:00
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
struct thread_exit_args {
|
|
|
|
struct thread *t;
|
|
|
|
region_id old_kernel_stack;
|
|
|
|
int int_state;
|
|
|
|
unsigned int death_stack;
|
2002-08-01 00:08:27 +04:00
|
|
|
sem_id death_sem;
|
2002-07-09 16:24:59 +04:00
|
|
|
};
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
static void
|
|
|
|
thread_exit2(void *_args)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread_exit_args args;
|
|
|
|
|
|
|
|
// copy the arguments over, since the source is probably on the kernel stack we're about to delete
|
|
|
|
memcpy(&args, _args, sizeof(struct thread_exit_args));
|
|
|
|
|
|
|
|
// restore the interrupts
|
2002-11-29 11:42:52 +03:00
|
|
|
enable_interrupts();
|
|
|
|
// ToDo: was:
|
|
|
|
// restore_interrupts(args.int_state);
|
|
|
|
// we probably don't want to let the interrupts disabled at this point??
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-11-29 11:42:52 +03:00
|
|
|
TRACE(("thread_exit2, running on death stack 0x%lx\n", args.t->kernel_stack_base));
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// delete the old kernel stack region
|
2002-11-29 11:42:52 +03:00
|
|
|
TRACE(("thread_exit2: deleting old kernel stack id 0x%x for thread 0x%x\n", args.old_kernel_stack, args.t->id));
|
2003-08-19 18:28:11 +04:00
|
|
|
delete_area(args.old_kernel_stack);
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// remove this thread from all of the global lists
|
2002-11-29 11:42:52 +03:00
|
|
|
TRACE(("thread_exit2: removing thread 0x%x from global lists\n", args.t->id));
|
|
|
|
|
2002-07-25 05:05:51 +04:00
|
|
|
disable_interrupts();
|
2002-08-03 04:41:27 +04:00
|
|
|
GRAB_TEAM_LOCK();
|
2002-08-04 03:39:50 +04:00
|
|
|
remove_thread_from_team(team_get_kernel_team(), args.t);
|
2002-08-03 04:41:27 +04:00
|
|
|
RELEASE_TEAM_LOCK();
|
2002-07-09 16:24:59 +04:00
|
|
|
GRAB_THREAD_LOCK();
|
|
|
|
hash_remove(thread_hash, args.t);
|
|
|
|
RELEASE_THREAD_LOCK();
|
|
|
|
|
2002-11-29 11:42:52 +03:00
|
|
|
// ToDo: is this correct at this point?
|
2002-12-04 19:40:07 +03:00
|
|
|
//restore_interrupts(args.int_state);
|
2002-11-29 11:42:52 +03:00
|
|
|
|
|
|
|
TRACE(("thread_exit2: done removing thread from lists\n"));
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-08-01 00:08:27 +04:00
|
|
|
if (args.death_sem >= 0)
|
|
|
|
release_sem_etc(args.death_sem, 1, B_DO_NOT_RESCHEDULE);
|
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
// set the next state to be gone. Will return the thread structure to a ready pool upon reschedule
|
|
|
|
args.t->next_state = THREAD_STATE_FREE_ON_RESCHED;
|
|
|
|
|
|
|
|
// return the death stack and reschedule one last time
|
|
|
|
put_death_stack_and_reschedule(args.death_stack);
|
|
|
|
// never get to here
|
|
|
|
panic("thread_exit2: made it where it shouldn't have!\n");
|
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
void
|
2002-10-26 05:11:15 +04:00
|
|
|
thread_exit(void)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-12-04 19:40:07 +03:00
|
|
|
cpu_status state;
|
2002-07-09 16:24:59 +04:00
|
|
|
struct thread *t = thread_get_current_thread();
|
2003-01-07 12:48:01 +03:00
|
|
|
struct team *team = t->team;
|
2002-08-03 04:41:27 +04:00
|
|
|
bool delete_team = false;
|
2002-07-09 16:24:59 +04:00
|
|
|
unsigned int death_stack;
|
2002-08-01 00:08:27 +04:00
|
|
|
sem_id cached_death_sem;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-11-29 11:42:52 +03:00
|
|
|
if (!are_interrupts_enabled())
|
|
|
|
dprintf("thread_exit() called with interrupts disabled!\n");
|
|
|
|
|
|
|
|
TRACE(("thread 0x%lx exiting %s w/return code 0x%x\n", t->id,
|
2002-10-26 05:11:15 +04:00
|
|
|
t->return_flags & THREAD_RETURN_INTERRUPTED ? "due to signal" : "normally",
|
2002-11-29 11:42:52 +03:00
|
|
|
(int)t->return_code));
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// boost our priority to get this over with
|
2002-10-31 16:20:00 +03:00
|
|
|
t->priority = B_FIRST_REAL_TIME_PRIORITY;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-10-26 02:36:08 +04:00
|
|
|
// Cancel previously installed alarm timer, if any
|
|
|
|
cancel_timer(&t->alarm);
|
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
// delete the user stack region first
|
2003-01-07 12:48:01 +03:00
|
|
|
if (team->_aspace_id >= 0 && t->user_stack_region_id >= 0) {
|
2002-07-09 16:24:59 +04:00
|
|
|
region_id rid = t->user_stack_region_id;
|
|
|
|
t->user_stack_region_id = -1;
|
2003-08-19 21:38:13 +04:00
|
|
|
delete_area_etc(team, rid);
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2003-01-07 12:48:01 +03:00
|
|
|
if (team != team_get_kernel_team()) {
|
2002-08-03 04:41:27 +04:00
|
|
|
// remove this thread from the current team and add it to the kernel
|
|
|
|
// put the thread into the kernel team until it dies
|
2002-07-25 05:05:51 +04:00
|
|
|
state = disable_interrupts();
|
2002-08-03 04:41:27 +04:00
|
|
|
GRAB_TEAM_LOCK();
|
2003-01-07 12:48:01 +03:00
|
|
|
remove_thread_from_team(team, t);
|
2002-08-04 03:39:50 +04:00
|
|
|
insert_thread_into_team(team_get_kernel_team(), t);
|
2003-01-07 12:48:01 +03:00
|
|
|
if (team->main_thread == t) {
|
2002-08-03 04:41:27 +04:00
|
|
|
// this was main thread in this team
|
|
|
|
delete_team = true;
|
2003-01-07 12:48:01 +03:00
|
|
|
team_remove_team_from_hash(team);
|
|
|
|
team->state = TEAM_STATE_DEATH;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
2002-08-03 04:41:27 +04:00
|
|
|
RELEASE_TEAM_LOCK();
|
2002-07-09 16:24:59 +04:00
|
|
|
// swap address spaces, to make sure we're running on the kernel's pgdir
|
2002-08-04 03:39:50 +04:00
|
|
|
vm_aspace_swap(team_get_kernel_team()->kaspace);
|
2002-07-25 05:05:51 +04:00
|
|
|
restore_interrupts(state);
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// dprintf("thread_exit: thread 0x%x now a kernel thread!\n", t->id);
|
|
|
|
}
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-07 12:48:01 +03:00
|
|
|
cached_death_sem = team->death_sem;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-08-03 04:41:27 +04:00
|
|
|
// delete the team
|
2002-09-24 04:08:50 +04:00
|
|
|
if (delete_team) {
|
2003-01-07 12:48:01 +03:00
|
|
|
if (team->num_threads > 0) {
|
2002-08-03 04:41:27 +04:00
|
|
|
// there are other threads still in this team,
|
2002-07-09 16:24:59 +04:00
|
|
|
// cycle through and signal kill on each of the threads
|
|
|
|
// XXX this can be optimized. There's got to be a better solution.
|
|
|
|
struct thread *temp_thread;
|
2002-08-01 00:08:27 +04:00
|
|
|
char death_sem_name[SYS_MAX_OS_NAME_LEN];
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2003-01-07 12:48:01 +03:00
|
|
|
sprintf(death_sem_name, "team %ld death sem", team->id);
|
|
|
|
team->death_sem = create_sem(0, death_sem_name);
|
|
|
|
if (team->death_sem < 0)
|
|
|
|
panic("thread_exit: cannot init death sem for team %ld\n", team->id);
|
2002-08-01 00:08:27 +04:00
|
|
|
|
2002-07-25 05:05:51 +04:00
|
|
|
state = disable_interrupts();
|
2002-08-03 04:41:27 +04:00
|
|
|
GRAB_TEAM_LOCK();
|
2002-07-09 16:24:59 +04:00
|
|
|
// we can safely walk the list because of the lock. no new threads can be created
|
2002-08-03 04:41:27 +04:00
|
|
|
// because of the TEAM_STATE_DEATH flag on the team
|
2003-01-07 12:48:01 +03:00
|
|
|
temp_thread = team->thread_list;
|
2002-07-09 16:24:59 +04:00
|
|
|
while(temp_thread) {
|
2002-08-03 04:41:27 +04:00
|
|
|
struct thread *next = temp_thread->team_next;
|
2002-07-09 16:24:59 +04:00
|
|
|
thread_kill_thread_nowait(temp_thread->id);
|
|
|
|
temp_thread = next;
|
|
|
|
}
|
2002-08-03 04:41:27 +04:00
|
|
|
RELEASE_TEAM_LOCK();
|
2002-07-25 05:05:51 +04:00
|
|
|
restore_interrupts(state);
|
2002-10-23 21:31:10 +04:00
|
|
|
// wait until all threads in team are dead.
|
2003-01-07 12:48:01 +03:00
|
|
|
acquire_sem_etc(team->death_sem, team->num_threads, 0, 0);
|
|
|
|
delete_sem(team->death_sem);
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
2002-08-01 00:08:27 +04:00
|
|
|
cached_death_sem = -1;
|
2003-01-12 19:30:09 +03:00
|
|
|
|
|
|
|
// free team resources
|
2003-01-07 12:48:01 +03:00
|
|
|
vm_put_aspace(team->aspace);
|
|
|
|
vm_delete_aspace(team->_aspace_id);
|
|
|
|
delete_owned_ports(team->id);
|
|
|
|
sem_delete_owned_sems(team->id);
|
2003-01-12 19:30:09 +03:00
|
|
|
remove_images(team);
|
2003-01-07 12:48:01 +03:00
|
|
|
vfs_free_io_context(team->io_context);
|
2003-01-12 19:30:09 +03:00
|
|
|
|
2003-01-07 12:48:01 +03:00
|
|
|
free(team);
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
// delete the sem that others will use to wait on us and get the retcode
|
|
|
|
{
|
|
|
|
sem_id s = t->return_code_sem;
|
|
|
|
|
|
|
|
t->return_code_sem = -1;
|
2002-10-26 05:11:15 +04:00
|
|
|
delete_sem_etc(s, t->return_code, t->return_flags & THREAD_RETURN_INTERRUPTED ? true : false);
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
death_stack = get_death_stack();
|
|
|
|
{
|
|
|
|
struct thread_exit_args args;
|
|
|
|
|
|
|
|
args.t = t;
|
|
|
|
args.old_kernel_stack = t->kernel_stack_region_id;
|
|
|
|
args.death_stack = death_stack;
|
2002-08-01 00:08:27 +04:00
|
|
|
args.death_sem = cached_death_sem;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// disable the interrupts. Must remain disabled until the kernel stack pointer can be officially switched
|
2002-07-25 05:05:51 +04:00
|
|
|
args.int_state = disable_interrupts();
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// set the new kernel stack officially to the death stack, wont be really switched until
|
|
|
|
// the next function is called. This bookkeeping must be done now before a context switch
|
|
|
|
// happens, or the processor will interrupt to the old stack
|
|
|
|
t->kernel_stack_region_id = death_stacks[death_stack].rid;
|
|
|
|
t->kernel_stack_base = death_stacks[death_stack].address;
|
|
|
|
|
|
|
|
// we will continue in thread_exit2(), on the new stack
|
|
|
|
arch_thread_switch_kstack_and_call(t, t->kernel_stack_base + KSTACK_SIZE, thread_exit2, &args);
|
|
|
|
}
|
|
|
|
|
|
|
|
panic("never can get here\n");
|
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
int
|
|
|
|
thread_kill_thread(thread_id id)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2003-01-27 06:17:36 +03:00
|
|
|
status_t status = send_signal_etc(id, SIGKILLTHR, B_DO_NOT_RESCHEDULE);
|
|
|
|
if (status < 0)
|
|
|
|
return status;
|
|
|
|
|
2002-10-23 21:31:10 +04:00
|
|
|
if (id != thread_get_current_thread()->id)
|
2003-01-27 06:17:36 +03:00
|
|
|
wait_for_thread(id, NULL);
|
|
|
|
|
|
|
|
return status;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
int
|
|
|
|
thread_kill_thread_nowait(thread_id id)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-10-23 21:31:10 +04:00
|
|
|
return send_signal_etc(id, SIGKILLTHR, B_DO_NOT_RESCHEDULE);
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
static void
|
|
|
|
thread_kthread_exit(void)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-10-26 05:11:15 +04:00
|
|
|
struct thread *t = thread_get_current_thread();
|
|
|
|
|
|
|
|
t->return_flags = THREAD_RETURN_EXIT;
|
|
|
|
thread_exit();
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
struct thread *
|
|
|
|
thread_get_thread_struct(thread_id id)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread *t;
|
2002-12-04 19:40:07 +03:00
|
|
|
cpu_status state;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-07-25 05:05:51 +04:00
|
|
|
state = disable_interrupts();
|
2002-07-09 16:24:59 +04:00
|
|
|
GRAB_THREAD_LOCK();
|
|
|
|
|
|
|
|
t = thread_get_thread_struct_locked(id);
|
|
|
|
|
|
|
|
RELEASE_THREAD_LOCK();
|
2002-07-25 05:05:51 +04:00
|
|
|
restore_interrupts(state);
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
return t;
|
|
|
|
}
|
|
|
|
|
2002-10-23 21:31:10 +04:00
|
|
|
|
|
|
|
struct thread *
|
|
|
|
thread_get_thread_struct_locked(thread_id id)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
|
|
|
struct thread_key key;
|
|
|
|
|
|
|
|
key.id = id;
|
|
|
|
|
|
|
|
return hash_lookup(thread_hash, &key);
|
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
// called in the int handler code when a thread enters the kernel for any reason
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
void
|
|
|
|
thread_atkernel_entry(void)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-12-04 19:40:07 +03:00
|
|
|
cpu_status state;
|
2002-07-09 16:24:59 +04:00
|
|
|
struct thread *t;
|
|
|
|
bigtime_t now;
|
|
|
|
|
|
|
|
// dprintf("thread_atkernel_entry: entry thread 0x%x\n", t->id);
|
|
|
|
|
|
|
|
t = thread_get_current_thread();
|
|
|
|
|
2002-07-25 05:05:51 +04:00
|
|
|
state = disable_interrupts();
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
// track user time
|
|
|
|
now = system_time();
|
|
|
|
t->user_time += now - t->last_time;
|
|
|
|
t->last_time = now;
|
2003-01-30 00:59:37 +03:00
|
|
|
t->last_time_type = KERNEL_TIME;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
t->in_kernel = true;
|
|
|
|
|
2002-07-25 05:05:51 +04:00
|
|
|
restore_interrupts(state);
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
|
2002-07-09 16:24:59 +04:00
|
|
|
// called when a thread exits kernel space to user space
|
2002-09-24 04:08:50 +04:00
|
|
|
|
|
|
|
void
|
|
|
|
thread_atkernel_exit(void)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-12-04 19:40:07 +03:00
|
|
|
cpu_status state;
|
|
|
|
int global_resched;
|
2002-07-09 16:24:59 +04:00
|
|
|
struct thread *t;
|
|
|
|
bigtime_t now;
|
|
|
|
|
|
|
|
// dprintf("thread_atkernel_exit: entry\n");
|
|
|
|
|
|
|
|
t = thread_get_current_thread();
|
|
|
|
|
2002-07-25 05:05:51 +04:00
|
|
|
state = disable_interrupts();
|
2002-07-09 16:24:59 +04:00
|
|
|
GRAB_THREAD_LOCK();
|
|
|
|
|
2002-10-28 23:38:16 +03:00
|
|
|
global_resched = handle_signals(t, state);
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
t->in_kernel = false;
|
|
|
|
|
|
|
|
// track kernel time
|
|
|
|
now = system_time();
|
|
|
|
t->kernel_time += now - t->last_time;
|
|
|
|
t->last_time = now;
|
2003-01-30 00:59:37 +03:00
|
|
|
t->last_time_type = USER_TIME;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-10-28 23:38:16 +03:00
|
|
|
RELEASE_THREAD_LOCK();
|
2002-07-25 05:05:51 +04:00
|
|
|
restore_interrupts(state);
|
2003-01-27 06:17:36 +03:00
|
|
|
|
2002-10-28 23:38:16 +03:00
|
|
|
if (global_resched)
|
|
|
|
smp_send_broadcast_ici(SMP_MSG_RESCHEDULE, 0, 0, 0, NULL, SMP_MSG_FLAG_SYNC);
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-08-04 06:04:37 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
// #pragma mark -
|
|
|
|
// private kernel exported functions
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
|
|
|
|
/** insert a thread onto the tail of a queue
|
|
|
|
*/
|
|
|
|
|
|
|
|
void
|
|
|
|
thread_enqueue(struct thread *thread, struct thread_queue *queue)
|
|
|
|
{
|
|
|
|
thread->queue_next = NULL;
|
|
|
|
if (queue->head == NULL) {
|
|
|
|
queue->head = thread;
|
|
|
|
queue->tail = thread;
|
|
|
|
} else {
|
|
|
|
queue->tail->queue_next = thread;
|
|
|
|
queue->tail = thread;
|
2002-08-19 12:28:39 +04:00
|
|
|
}
|
2003-01-27 06:17:36 +03:00
|
|
|
}
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
struct thread *
|
|
|
|
thread_lookat_queue(struct thread_queue *queue)
|
|
|
|
{
|
|
|
|
return queue->head;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
struct thread *
|
|
|
|
thread_dequeue(struct thread_queue *queue)
|
|
|
|
{
|
|
|
|
struct thread *thread = queue->head;
|
|
|
|
|
|
|
|
if (thread != NULL) {
|
|
|
|
queue->head = thread->queue_next;
|
|
|
|
if (queue->tail == thread)
|
|
|
|
queue->tail = NULL;
|
|
|
|
}
|
|
|
|
return thread;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
struct thread *
|
|
|
|
thread_dequeue_id(struct thread_queue *q, thread_id thr_id)
|
|
|
|
{
|
|
|
|
struct thread *t;
|
|
|
|
struct thread *last = NULL;
|
|
|
|
|
|
|
|
t = q->head;
|
|
|
|
while (t != NULL) {
|
|
|
|
if (t->id == thr_id) {
|
|
|
|
if (last == NULL)
|
|
|
|
q->head = t->queue_next;
|
|
|
|
else
|
|
|
|
last->queue_next = t->queue_next;
|
|
|
|
|
|
|
|
if (q->tail == t)
|
|
|
|
q->tail = last;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
last = t;
|
|
|
|
t = t->queue_next;
|
|
|
|
}
|
|
|
|
return t;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int
|
|
|
|
thread_init(kernel_args *ka)
|
|
|
|
{
|
|
|
|
struct thread *t;
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
// dprintf("thread_init: entry\n");
|
|
|
|
|
|
|
|
// create the thread hash table
|
|
|
|
thread_hash = hash_init(15, (addr)&t->all_next - (addr)t,
|
|
|
|
&thread_struct_compare, &thread_struct_hash);
|
|
|
|
|
|
|
|
// zero out the dead thread structure q
|
|
|
|
memset(&dead_q, 0, sizeof(dead_q));
|
|
|
|
|
|
|
|
// allocate snooze sem
|
|
|
|
gSnoozeSem = create_sem(0, "snooze sem");
|
|
|
|
if (gSnoozeSem < 0) {
|
|
|
|
panic("error creating snooze sem\n");
|
|
|
|
return gSnoozeSem;
|
|
|
|
}
|
|
|
|
|
|
|
|
// create an idle thread for each cpu
|
|
|
|
for (i = 0; i < ka->num_cpus; i++) {
|
|
|
|
char temp[64];
|
|
|
|
vm_region *region;
|
|
|
|
|
|
|
|
sprintf(temp, "idle_thread%d", i);
|
|
|
|
t = create_thread_struct(temp);
|
|
|
|
if (t == NULL) {
|
|
|
|
panic("error creating idle thread struct\n");
|
|
|
|
return ENOMEM;
|
|
|
|
}
|
|
|
|
t->team = team_get_kernel_team();
|
|
|
|
t->priority = B_IDLE_PRIORITY;
|
|
|
|
t->state = B_THREAD_RUNNING;
|
|
|
|
t->next_state = B_THREAD_READY;
|
|
|
|
sprintf(temp, "idle_thread%d_kstack", i);
|
|
|
|
t->kernel_stack_region_id = vm_find_region_by_name(vm_get_kernel_aspace_id(), temp);
|
|
|
|
region = vm_get_region_by_id(t->kernel_stack_region_id);
|
|
|
|
if (!region)
|
|
|
|
panic("error finding idle kstack region\n");
|
|
|
|
|
|
|
|
t->kernel_stack_base = region->base;
|
|
|
|
vm_put_region(region);
|
|
|
|
hash_insert(thread_hash, t);
|
|
|
|
insert_thread_into_team(t->team, t);
|
|
|
|
idle_threads[i] = t;
|
|
|
|
if (i == 0)
|
|
|
|
arch_thread_set_current_thread(t);
|
|
|
|
t->cpu = &cpu[i];
|
|
|
|
}
|
|
|
|
|
|
|
|
// create a set of death stacks
|
|
|
|
num_death_stacks = smp_get_num_cpus();
|
|
|
|
if (num_death_stacks > 8*sizeof(death_stack_bitmap)) {
|
|
|
|
/*
|
|
|
|
* clamp values for really beefy machines
|
|
|
|
*/
|
|
|
|
num_death_stacks = 8*sizeof(death_stack_bitmap);
|
|
|
|
}
|
|
|
|
death_stack_bitmap = 0;
|
|
|
|
death_stacks = (struct death_stack *)malloc(num_death_stacks * sizeof(struct death_stack));
|
|
|
|
if (death_stacks == NULL) {
|
|
|
|
panic("error creating death stacks\n");
|
|
|
|
return ENOMEM;
|
|
|
|
}
|
|
|
|
{
|
|
|
|
char temp[64];
|
|
|
|
|
|
|
|
for (i = 0; i < num_death_stacks; i++) {
|
|
|
|
sprintf(temp, "death_stack%d", i);
|
2003-08-19 18:28:11 +04:00
|
|
|
death_stacks[i].rid = create_area(temp, (void **)&death_stacks[i].address,
|
|
|
|
B_ANY_KERNEL_ADDRESS, KSTACK_SIZE, B_FULL_LOCK, B_KERNEL_READ_AREA | B_KERNEL_WRITE_AREA);
|
2003-01-27 06:17:36 +03:00
|
|
|
if (death_stacks[i].rid < 0) {
|
|
|
|
panic("error creating death stacks\n");
|
|
|
|
return death_stacks[i].rid;
|
|
|
|
}
|
|
|
|
death_stacks[i].in_use = false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
death_stack_sem = create_sem(num_death_stacks, "death_stack_noavail_sem");
|
|
|
|
|
|
|
|
// set up some debugger commands
|
|
|
|
add_debugger_command("threads", &dump_thread_list, "list all threads");
|
|
|
|
add_debugger_command("thread", &dump_thread_info, "list info about a particular thread");
|
|
|
|
add_debugger_command("next_q", &dump_next_thread_in_q, "dump the next thread in the queue of last thread viewed");
|
|
|
|
add_debugger_command("next_all", &dump_next_thread_in_all_list, "dump the next thread in the global list of the last thread viewed");
|
|
|
|
add_debugger_command("next_team", &dump_next_thread_in_team, "dump the next thread in the team of the last thread viewed");
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int
|
|
|
|
thread_init_percpu(int cpu_num)
|
|
|
|
{
|
|
|
|
arch_thread_set_current_thread(idle_threads[cpu_num]);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
thread_id
|
|
|
|
spawn_kernel_thread_etc(thread_func function, const char *name, int32 priority, void *arg, team_id team)
|
|
|
|
{
|
2003-04-18 13:38:28 +04:00
|
|
|
return create_thread(name, team, function, arg, NULL, priority, true);
|
2003-01-27 06:17:36 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// #pragma mark -
|
|
|
|
// public kernel exported functions
|
|
|
|
|
|
|
|
|
|
|
|
status_t
|
|
|
|
send_data(thread_id tid, int32 code, const void *buffer, size_t buffer_size)
|
|
|
|
{
|
|
|
|
struct thread *target;
|
|
|
|
sem_id cached_sem;
|
|
|
|
cpu_status state;
|
|
|
|
status_t rv;
|
|
|
|
cbuf *data;
|
|
|
|
|
|
|
|
state = disable_interrupts();
|
|
|
|
GRAB_THREAD_LOCK();
|
|
|
|
target = thread_get_thread_struct_locked(tid);
|
|
|
|
if (!target) {
|
|
|
|
RELEASE_THREAD_LOCK();
|
|
|
|
restore_interrupts(state);
|
|
|
|
return B_BAD_THREAD_ID;
|
|
|
|
}
|
|
|
|
cached_sem = target->msg.write_sem;
|
|
|
|
RELEASE_THREAD_LOCK();
|
|
|
|
restore_interrupts(state);
|
|
|
|
|
|
|
|
if (buffer_size > THREAD_MAX_MESSAGE_SIZE)
|
|
|
|
return B_NO_MEMORY;
|
|
|
|
|
|
|
|
rv = acquire_sem_etc(cached_sem, 1, B_CAN_INTERRUPT, 0);
|
2002-08-19 12:28:39 +04:00
|
|
|
if (rv == B_INTERRUPTED)
|
|
|
|
// We got interrupted by a signal
|
|
|
|
return rv;
|
|
|
|
if (rv != B_OK)
|
|
|
|
// Any other acquisition problems may be due to thread deletion
|
|
|
|
return B_BAD_THREAD_ID;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
if (buffer_size > 0) {
|
|
|
|
data = cbuf_get_chain(buffer_size);
|
|
|
|
if (!data)
|
|
|
|
return B_NO_MEMORY;
|
|
|
|
rv = cbuf_user_memcpy_to_chain(data, 0, buffer, buffer_size);
|
|
|
|
if (rv < 0) {
|
|
|
|
cbuf_free_chain(data);
|
|
|
|
return B_NO_MEMORY;
|
|
|
|
}
|
|
|
|
} else
|
|
|
|
data = NULL;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
state = disable_interrupts();
|
|
|
|
GRAB_THREAD_LOCK();
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
// The target thread could have been deleted at this point
|
|
|
|
target = thread_get_thread_struct_locked(tid);
|
|
|
|
if (!target) {
|
|
|
|
RELEASE_THREAD_LOCK();
|
|
|
|
restore_interrupts(state);
|
|
|
|
cbuf_free_chain(data);
|
|
|
|
return B_BAD_THREAD_ID;
|
|
|
|
}
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
// Save message informations
|
|
|
|
target->msg.sender = thread_get_current_thread()->id;
|
|
|
|
target->msg.code = code;
|
|
|
|
target->msg.size = buffer_size;
|
|
|
|
target->msg.buffer = data;
|
|
|
|
cached_sem = target->msg.read_sem;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
RELEASE_THREAD_LOCK();
|
|
|
|
restore_interrupts(state);
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
release_sem(cached_sem);
|
2002-08-19 12:28:39 +04:00
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
return B_OK;
|
2002-08-19 12:28:39 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
status_t
|
|
|
|
receive_data(thread_id *sender, void *buffer, size_t buffer_size)
|
|
|
|
{
|
|
|
|
struct thread *t = thread_get_current_thread();
|
|
|
|
status_t rv;
|
|
|
|
size_t size;
|
|
|
|
int32 code;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
acquire_sem(t->msg.read_sem);
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
size = min(buffer_size, t->msg.size);
|
|
|
|
rv = cbuf_user_memcpy_from_chain(buffer, t->msg.buffer, 0, size);
|
|
|
|
if (rv < 0) {
|
|
|
|
cbuf_free_chain(t->msg.buffer);
|
|
|
|
release_sem(t->msg.write_sem);
|
|
|
|
return rv;
|
|
|
|
}
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
*sender = t->msg.sender;
|
|
|
|
code = t->msg.code;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
cbuf_free_chain(t->msg.buffer);
|
|
|
|
release_sem(t->msg.write_sem);
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
return code;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
bool
|
|
|
|
has_data(thread_id thread)
|
|
|
|
{
|
|
|
|
int32 count;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-19 12:28:39 +04:00
|
|
|
if (get_sem_count(thread_get_current_thread()->msg.read_sem, &count) != B_OK)
|
|
|
|
return false;
|
2003-01-27 06:17:36 +03:00
|
|
|
|
2002-09-24 04:08:50 +04:00
|
|
|
return count == 0 ? false : true;
|
2002-08-19 12:28:39 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2002-08-04 06:04:37 +04:00
|
|
|
status_t
|
|
|
|
_get_thread_info(thread_id id, thread_info *info, size_t size)
|
|
|
|
{
|
2002-12-04 19:40:07 +03:00
|
|
|
cpu_status state;
|
2002-08-04 06:04:37 +04:00
|
|
|
status_t rc = B_OK;
|
|
|
|
struct thread *t;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-04 06:04:37 +04:00
|
|
|
state = disable_interrupts();
|
|
|
|
GRAB_THREAD_LOCK();
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-04 06:04:37 +04:00
|
|
|
t = thread_get_thread_struct_locked(id);
|
|
|
|
if (!t) {
|
|
|
|
rc = B_BAD_VALUE;
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
info->thread = t->id;
|
|
|
|
info->team = t->team->id;
|
|
|
|
strncpy(info->name, t->name, B_OS_NAME_LENGTH);
|
|
|
|
info->name[B_OS_NAME_LENGTH - 1] = '\0';
|
2002-08-05 00:10:06 +04:00
|
|
|
if (t->state == B_THREAD_WAITING) {
|
2003-01-27 06:17:36 +03:00
|
|
|
if (t->sem_blocking == gSnoozeSem)
|
2002-08-05 00:10:06 +04:00
|
|
|
info->state = B_THREAD_ASLEEP;
|
2002-08-19 13:10:09 +04:00
|
|
|
else if (t->sem_blocking == t->msg.read_sem)
|
|
|
|
info->state = B_THREAD_RECEIVING;
|
2002-08-05 00:10:06 +04:00
|
|
|
else
|
|
|
|
info->state = B_THREAD_WAITING;
|
|
|
|
} else
|
|
|
|
info->state = t->state;
|
2002-08-04 06:04:37 +04:00
|
|
|
info->priority = t->priority;
|
|
|
|
info->sem = t->sem_blocking;
|
|
|
|
info->user_time = t->user_time;
|
|
|
|
info->kernel_time = t->kernel_time;
|
|
|
|
info->stack_base = (void *)t->user_stack_base;
|
|
|
|
info->stack_end = (void *)(t->user_stack_base + STACK_SIZE);
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-04 06:04:37 +04:00
|
|
|
err:
|
|
|
|
RELEASE_THREAD_LOCK();
|
|
|
|
restore_interrupts(state);
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
status_t
|
|
|
|
_get_next_thread_info(team_id tid, int32 *cookie, thread_info *info, size_t size)
|
|
|
|
{
|
2002-12-04 19:40:07 +03:00
|
|
|
cpu_status state;
|
2002-08-04 06:04:37 +04:00
|
|
|
int slot;
|
|
|
|
status_t rc = B_BAD_VALUE;
|
|
|
|
struct team *team;
|
|
|
|
struct thread *t = NULL;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-04 06:04:37 +04:00
|
|
|
if (tid == 0)
|
|
|
|
tid = team_get_current_team_id();
|
|
|
|
team = team_get_team_struct(tid);
|
|
|
|
if (!team)
|
|
|
|
return B_BAD_VALUE;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-04 06:04:37 +04:00
|
|
|
state = disable_interrupts();
|
|
|
|
GRAB_THREAD_LOCK();
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-04 06:04:37 +04:00
|
|
|
if (*cookie == 0)
|
|
|
|
slot = 0;
|
|
|
|
else {
|
|
|
|
slot = *cookie;
|
|
|
|
if (slot >= next_thread_id)
|
|
|
|
goto err;
|
|
|
|
}
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-05 00:10:06 +04:00
|
|
|
while ((slot < next_thread_id) && (!(t = thread_get_thread_struct_locked(slot)) || (t->team->id != tid)))
|
2002-08-04 06:04:37 +04:00
|
|
|
slot++;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-05 00:10:06 +04:00
|
|
|
if ((t) && (t->team->id == tid)) {
|
2002-08-04 06:04:37 +04:00
|
|
|
info->thread = t->id;
|
|
|
|
info->team = t->team->id;
|
|
|
|
strncpy(info->name, t->name, B_OS_NAME_LENGTH);
|
|
|
|
info->name[B_OS_NAME_LENGTH - 1] = '\0';
|
2002-08-05 00:10:06 +04:00
|
|
|
if (t->state == B_THREAD_WAITING) {
|
2003-01-27 06:17:36 +03:00
|
|
|
if (t->sem_blocking == gSnoozeSem)
|
2002-08-05 00:10:06 +04:00
|
|
|
info->state = B_THREAD_ASLEEP;
|
2002-08-19 13:10:09 +04:00
|
|
|
else if (t->sem_blocking == t->msg.read_sem)
|
|
|
|
info->state = B_THREAD_RECEIVING;
|
2002-08-05 00:10:06 +04:00
|
|
|
else
|
|
|
|
info->state = B_THREAD_WAITING;
|
|
|
|
} else
|
|
|
|
info->state = t->state;
|
2002-08-04 06:04:37 +04:00
|
|
|
info->priority = t->priority;
|
|
|
|
info->sem = t->sem_blocking;
|
|
|
|
info->user_time = t->user_time;
|
|
|
|
info->kernel_time = t->kernel_time;
|
|
|
|
info->stack_base = (void *)t->user_stack_base;
|
|
|
|
info->stack_end = (void *)(t->user_stack_base + STACK_SIZE);
|
|
|
|
slot++;
|
|
|
|
*cookie = slot;
|
|
|
|
rc = B_OK;
|
|
|
|
}
|
|
|
|
err:
|
|
|
|
RELEASE_THREAD_LOCK();
|
|
|
|
restore_interrupts(state);
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2002-08-04 06:04:37 +04:00
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
status_t
|
|
|
|
set_thread_priority(thread_id id, int32 priority)
|
|
|
|
{
|
|
|
|
struct thread *thread;
|
|
|
|
int32 oldPriority;
|
|
|
|
|
|
|
|
// make sure the passed in priority is within bounds
|
|
|
|
if (priority > B_MAX_PRIORITY)
|
|
|
|
priority = B_MAX_PRIORITY;
|
|
|
|
if (priority < B_MIN_PRIORITY)
|
|
|
|
priority = B_MIN_PRIORITY;
|
|
|
|
|
|
|
|
thread = thread_get_current_thread();
|
|
|
|
if (thread->id == id) {
|
|
|
|
// it's ourself, so we know we aren't in the run queue, and we can manipulate
|
|
|
|
// our structure directly
|
|
|
|
oldPriority = thread->priority;
|
|
|
|
// note that this might not return the correct value if we are preempted
|
|
|
|
// here, and another thread changes our priority before the next line is
|
|
|
|
// executed
|
|
|
|
thread->priority = priority;
|
|
|
|
} else {
|
|
|
|
cpu_status state = disable_interrupts();
|
|
|
|
GRAB_THREAD_LOCK();
|
|
|
|
|
|
|
|
thread = thread_get_thread_struct_locked(id);
|
|
|
|
if (thread) {
|
|
|
|
oldPriority = thread->priority;
|
|
|
|
if (thread->state == B_THREAD_READY && thread->priority != priority) {
|
|
|
|
// if the thread is in the run queue, we reinsert it at a new position
|
|
|
|
// ToDo: this doesn't seem to be necessary: if a thread is already in
|
|
|
|
// the run queue, running it once doesn't hurt, and if the priority
|
|
|
|
// is now very low, it probably wouldn't have been selected to be in
|
|
|
|
// the run queue at all right now, anyway.
|
|
|
|
scheduler_remove_from_run_queue(thread);
|
|
|
|
thread->priority = priority;
|
|
|
|
scheduler_enqueue_in_run_queue(thread);
|
|
|
|
} else
|
|
|
|
thread->priority = priority;
|
|
|
|
} else
|
|
|
|
oldPriority = B_BAD_THREAD_ID;
|
|
|
|
|
|
|
|
RELEASE_THREAD_LOCK();
|
|
|
|
restore_interrupts(state);
|
|
|
|
}
|
|
|
|
|
|
|
|
return oldPriority;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
status_t
|
|
|
|
snooze_etc(bigtime_t timeout, int timebase, uint32 flags)
|
|
|
|
{
|
|
|
|
status_t status;
|
|
|
|
|
|
|
|
if (timebase != B_SYSTEM_TIMEBASE)
|
|
|
|
return B_BAD_VALUE;
|
|
|
|
|
|
|
|
status = acquire_sem_etc(gSnoozeSem, 1, B_ABSOLUTE_TIMEOUT | flags, timeout);
|
|
|
|
if (status == B_TIMED_OUT || status == B_WOULD_BLOCK)
|
|
|
|
return B_OK;
|
|
|
|
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/** snooze() for internal kernel use only; doesn't interrupt on signals.
|
|
|
|
*/
|
|
|
|
|
|
|
|
status_t
|
|
|
|
snooze(bigtime_t timeout)
|
|
|
|
{
|
|
|
|
return snooze_etc(timeout, B_SYSTEM_TIMEBASE, B_RELATIVE_TIMEOUT);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/** snooze_until() for internal kernel use only; doesn't interrupt on signals.
|
|
|
|
*/
|
|
|
|
|
|
|
|
status_t
|
|
|
|
snooze_until(bigtime_t timeout, int timebase)
|
|
|
|
{
|
|
|
|
return snooze_etc(timeout, timebase, B_ABSOLUTE_TIMEOUT);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
status_t
|
|
|
|
wait_for_thread(thread_id id, status_t *_returnCode)
|
|
|
|
{
|
|
|
|
sem_id sem;
|
|
|
|
cpu_status state;
|
|
|
|
struct thread *t;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
rc = send_signal_etc(id, SIGCONT, 0);
|
|
|
|
if (rc != B_OK)
|
|
|
|
return rc;
|
|
|
|
|
|
|
|
state = disable_interrupts();
|
|
|
|
GRAB_THREAD_LOCK();
|
|
|
|
|
|
|
|
t = thread_get_thread_struct_locked(id);
|
|
|
|
sem = t != NULL ? t->return_code_sem : B_BAD_THREAD_ID;
|
|
|
|
|
|
|
|
RELEASE_THREAD_LOCK();
|
|
|
|
restore_interrupts(state);
|
|
|
|
|
|
|
|
rc = acquire_sem(sem);
|
|
|
|
|
|
|
|
/* This thread died the way it should, dont ripple a non-error up */
|
|
|
|
if (rc == B_BAD_SEM_ID) {
|
|
|
|
rc = B_NO_ERROR;
|
|
|
|
|
|
|
|
if (_returnCode) {
|
|
|
|
t = thread_get_current_thread();
|
|
|
|
dprintf("thread_wait_on_thread: thread %ld got return code 0x%x\n",
|
|
|
|
t->id, t->sem_deleted_retcode);
|
|
|
|
*_returnCode = t->sem_deleted_retcode;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
status_t
|
|
|
|
suspend_thread(thread_id id)
|
|
|
|
{
|
|
|
|
return send_signal_etc(id, SIGSTOP, B_DO_NOT_RESCHEDULE);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
status_t
|
|
|
|
resume_thread(thread_id id)
|
|
|
|
{
|
|
|
|
return send_signal_etc(id, SIGCONT, B_DO_NOT_RESCHEDULE);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
thread_id
|
|
|
|
spawn_kernel_thread(thread_func function, const char *name, int32 priority, void *arg)
|
|
|
|
{
|
2003-04-18 13:38:28 +04:00
|
|
|
return create_thread(name, team_get_kernel_team()->id, function, arg, NULL, priority, true);
|
2003-01-27 06:17:36 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
int
|
|
|
|
getrlimit(int resource, struct rlimit * rlp)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-10-05 23:38:42 +04:00
|
|
|
if (!rlp)
|
|
|
|
return -1;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
switch (resource) {
|
|
|
|
case RLIMIT_NOFILE:
|
|
|
|
return vfs_getrlimit(resource, rlp);
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
default:
|
|
|
|
return -1;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
return 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
|
|
|
|
int
|
|
|
|
setrlimit(int resource, const struct rlimit * rlp)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-10-05 23:38:42 +04:00
|
|
|
if (!rlp)
|
2002-07-09 16:24:59 +04:00
|
|
|
return -1;
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
switch (resource) {
|
2002-07-09 16:24:59 +04:00
|
|
|
case RLIMIT_NOFILE:
|
2002-10-05 23:38:42 +04:00
|
|
|
return vfs_setrlimit(resource, rlp);
|
2002-07-09 16:24:59 +04:00
|
|
|
|
|
|
|
default:
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
// #pragma mark -
|
|
|
|
// Calls from userland (with extra address checks)
|
|
|
|
|
2002-11-18 05:04:57 +03:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
void
|
|
|
|
user_exit_thread(status_t return_value)
|
2002-12-07 03:25:50 +03:00
|
|
|
{
|
2003-01-27 06:17:36 +03:00
|
|
|
struct thread *thread = thread_get_current_thread();
|
|
|
|
|
|
|
|
thread->return_code = return_value;
|
|
|
|
thread->return_flags = THREAD_RETURN_EXIT;
|
2002-12-07 03:25:50 +03:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
send_signal_etc(thread->id, SIGKILLTHR, B_DO_NOT_RESCHEDULE);
|
2002-12-07 03:25:50 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
status_t
|
|
|
|
user_resume_thread(thread_id thread)
|
|
|
|
{
|
|
|
|
return resume_thread(thread);
|
|
|
|
}
|
|
|
|
|
2002-12-07 03:25:50 +03:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
status_t
|
|
|
|
user_suspend_thread(thread_id thread)
|
|
|
|
{
|
|
|
|
return suspend_thread(thread);
|
|
|
|
}
|
2002-12-07 03:25:50 +03:00
|
|
|
|
2002-11-18 05:04:57 +03:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
int32
|
|
|
|
user_set_thread_priority(thread_id thread, int32 newPriority)
|
2002-11-18 05:04:57 +03:00
|
|
|
{
|
2003-01-27 06:17:36 +03:00
|
|
|
return set_thread_priority(thread, newPriority);
|
2002-11-18 05:04:57 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
thread_id
|
2003-04-18 13:38:28 +04:00
|
|
|
user_spawn_thread(thread_func entry, const char *userName, int32 priority, void *data1, void *data2)
|
2002-11-18 05:04:57 +03:00
|
|
|
{
|
2003-01-27 06:17:36 +03:00
|
|
|
char name[SYS_MAX_OS_NAME_LEN];
|
2002-11-18 05:04:57 +03:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
if (!CHECK_USER_ADDRESS(entry) || entry == NULL
|
|
|
|
|| !CHECK_USER_ADDRESS(userName)
|
|
|
|
|| user_strlcpy(name, userName, SYS_MAX_OS_NAME_LEN) < B_OK)
|
|
|
|
return B_BAD_ADDRESS;
|
2002-11-18 05:04:57 +03:00
|
|
|
|
2003-04-18 13:38:28 +04:00
|
|
|
return create_thread(name, thread_get_current_thread()->team->id, entry, data1, data2, priority, false);
|
2003-01-27 06:17:36 +03:00
|
|
|
}
|
2002-11-18 05:04:57 +03:00
|
|
|
|
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
status_t
|
|
|
|
user_snooze_etc(bigtime_t timeout, int timebase, uint32 flags)
|
|
|
|
{
|
|
|
|
return snooze_etc(timeout, timebase, flags | B_CAN_INTERRUPT);
|
|
|
|
}
|
2002-10-05 23:38:42 +04:00
|
|
|
|
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
status_t
|
|
|
|
user_get_thread_info(thread_id id, thread_info *userInfo)
|
2002-10-05 23:38:42 +04:00
|
|
|
{
|
2003-01-27 06:17:36 +03:00
|
|
|
thread_info info;
|
|
|
|
status_t status;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
if (!CHECK_USER_ADDRESS(userInfo))
|
|
|
|
return B_BAD_ADDRESS;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
status = _get_thread_info(id, &info, sizeof(thread_info));
|
|
|
|
|
|
|
|
if (status >= B_OK && user_memcpy(userInfo, &info, sizeof(thread_info)) < B_OK)
|
|
|
|
return B_BAD_ADDRESS;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
return status;
|
2002-10-05 23:38:42 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
status_t
|
2003-01-27 06:17:36 +03:00
|
|
|
user_get_next_thread_info(team_id team, int32 *userCookie, thread_info *userInfo)
|
2002-10-05 23:38:42 +04:00
|
|
|
{
|
2003-01-27 06:17:36 +03:00
|
|
|
status_t status;
|
|
|
|
thread_info info;
|
|
|
|
int32 cookie;
|
|
|
|
|
|
|
|
if (!CHECK_USER_ADDRESS(userCookie) || !CHECK_USER_ADDRESS(userInfo)
|
|
|
|
|| user_memcpy(&cookie, userCookie, sizeof(int32)) < B_OK)
|
|
|
|
return B_BAD_ADDRESS;
|
|
|
|
|
|
|
|
status = _get_next_thread_info(team, &cookie, &info, sizeof(thread_info));
|
|
|
|
if (status < B_OK)
|
|
|
|
return status;
|
|
|
|
|
|
|
|
if (user_memcpy(userCookie, &cookie, sizeof(int32)) < B_OK
|
|
|
|
|| user_memcpy(userInfo, &info, sizeof(thread_info)) < B_OK)
|
|
|
|
return B_BAD_ADDRESS;
|
|
|
|
|
|
|
|
return status;
|
2002-10-05 23:38:42 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
status_t
|
2003-01-27 06:17:36 +03:00
|
|
|
user_wait_for_thread(thread_id id, status_t *userReturnCode)
|
2002-10-05 23:38:42 +04:00
|
|
|
{
|
2003-01-27 06:17:36 +03:00
|
|
|
status_t returnCode;
|
|
|
|
status_t status;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
if (!CHECK_USER_ADDRESS(userReturnCode))
|
|
|
|
return B_BAD_ADDRESS;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
status = wait_for_thread(id, &returnCode);
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
if (userReturnCode && user_memcpy(userReturnCode, &returnCode, sizeof(status_t)) < B_OK)
|
|
|
|
return B_BAD_ADDRESS;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
return status;
|
|
|
|
}
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
|
|
|
|
bool
|
|
|
|
user_has_data(thread_id thread)
|
|
|
|
{
|
|
|
|
return has_data(thread);
|
2002-10-05 23:38:42 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
status_t
|
2003-01-27 06:17:36 +03:00
|
|
|
user_send_data(thread_id thread, int32 code, const void *buffer, size_t bufferSize)
|
2002-10-05 23:38:42 +04:00
|
|
|
{
|
2003-01-27 06:17:36 +03:00
|
|
|
if (!CHECK_USER_ADDRESS(buffer))
|
2002-10-05 23:38:42 +04:00
|
|
|
return B_BAD_ADDRESS;
|
2003-01-27 06:17:36 +03:00
|
|
|
|
|
|
|
return send_data(thread, code, buffer, bufferSize);
|
|
|
|
// supports userland buffers
|
2002-10-05 23:38:42 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
status_t
|
2003-01-27 06:17:36 +03:00
|
|
|
user_receive_data(thread_id *userSender, void *buffer, size_t bufferSize)
|
2002-10-05 23:38:42 +04:00
|
|
|
{
|
2003-01-27 06:17:36 +03:00
|
|
|
thread_id sender;
|
2002-10-05 23:38:42 +04:00
|
|
|
status_t code;
|
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
if (!CHECK_USER_ADDRESS(userSender)
|
|
|
|
|| !CHECK_USER_ADDRESS(buffer))
|
2002-10-05 23:38:42 +04:00
|
|
|
return B_BAD_ADDRESS;
|
2003-01-27 06:17:36 +03:00
|
|
|
|
|
|
|
code = receive_data(&sender, buffer, bufferSize);
|
|
|
|
// supports userland buffers
|
|
|
|
|
|
|
|
if (user_memcpy(userSender, &sender, sizeof(thread_id)) < B_OK)
|
2002-10-05 23:38:42 +04:00
|
|
|
return B_BAD_ADDRESS;
|
2003-01-27 06:17:36 +03:00
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
return code;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int
|
2003-01-27 06:17:36 +03:00
|
|
|
user_getrlimit(int resource, struct rlimit *urlp)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-10-05 23:38:42 +04:00
|
|
|
struct rlimit rl;
|
|
|
|
int ret;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
if (urlp == NULL)
|
2002-07-14 14:10:22 +04:00
|
|
|
return EINVAL;
|
2002-10-05 23:38:42 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
if (!CHECK_USER_ADDRESS(urlp))
|
|
|
|
return B_BAD_ADDRESS;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
ret = getrlimit(resource, &rl);
|
|
|
|
|
|
|
|
if (ret == 0) {
|
|
|
|
ret = user_memcpy(urlp, &rl, sizeof(struct rlimit));
|
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
return 0;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
return ret;
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
|
|
|
|
int
|
2003-01-27 06:17:36 +03:00
|
|
|
user_setrlimit(int resource, const struct rlimit *urlp)
|
2002-07-09 16:24:59 +04:00
|
|
|
{
|
2002-10-05 23:38:42 +04:00
|
|
|
struct rlimit rl;
|
|
|
|
int err;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
if (urlp == NULL)
|
|
|
|
return EINVAL;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2003-01-27 06:17:36 +03:00
|
|
|
if (!CHECK_USER_ADDRESS(urlp))
|
|
|
|
return B_BAD_ADDRESS;
|
2002-07-09 16:24:59 +04:00
|
|
|
|
2002-10-05 23:38:42 +04:00
|
|
|
err = user_memcpy(&rl, urlp, sizeof(struct rlimit));
|
|
|
|
if (err < 0)
|
|
|
|
return err;
|
|
|
|
|
|
|
|
return setrlimit(resource, &rl);
|
2002-07-09 16:24:59 +04:00
|
|
|
}
|
2002-10-05 23:38:42 +04:00
|
|
|
|