in fork1(), only add make the new proc visible (by giving it a pid

and adding it to allproc) after it's fully initialized.
this prevents the scheduler from coming in via a clock interrupt
and tripping over a partially-initialized proc.
This commit is contained in:
chs 2000-11-08 05:16:23 +00:00
parent b7192d086c
commit a284236c04
1 changed files with 90 additions and 92 deletions

View File

@ -1,4 +1,4 @@
/* $NetBSD: kern_fork.c,v 1.75 2000/11/07 12:41:52 jdolecek Exp $ */
/* $NetBSD: kern_fork.c,v 1.76 2000/11/08 05:16:23 chs Exp $ */
/*
* Copyright (c) 1982, 1986, 1989, 1991, 1993
@ -106,9 +106,8 @@ fork1(struct proc *p1, int flags, int exitsig, void *stack, size_t stacksize,
void (*func)(void *), void *arg, register_t *retval,
struct proc **rnewprocp)
{
struct proc *p2;
struct proc *p2, *tp;
uid_t uid;
struct proc *newproc;
int count, s;
vaddr_t uaddr;
static int nextpid, pidchecked = 0;
@ -126,6 +125,7 @@ fork1(struct proc *p1, int flags, int exitsig, void *stack, size_t stacksize,
tablefull("proc", "increase kern.maxproc or NPROC");
return (EAGAIN);
}
nprocs++;
/*
* Increment the count of procs running with this uid. Don't allow
@ -135,6 +135,7 @@ fork1(struct proc *p1, int flags, int exitsig, void *stack, size_t stacksize,
if (__predict_false(uid != 0 && count >
p1->p_rlimit[RLIMIT_NPROC].rlim_cur)) {
(void)chgproccnt(uid, -1);
nprocs--;
return (EAGAIN);
}
@ -147,6 +148,7 @@ fork1(struct proc *p1, int flags, int exitsig, void *stack, size_t stacksize,
uaddr = uvm_km_valloc(kernel_map, USPACE);
if (__predict_false(uaddr == 0)) {
(void)chgproccnt(uid, -1);
nprocs--;
return (ENOMEM);
}
@ -156,94 +158,7 @@ fork1(struct proc *p1, int flags, int exitsig, void *stack, size_t stacksize,
*/
/* Allocate new proc. */
newproc = pool_get(&proc_pool, PR_WAITOK);
/*
* BEGIN PID ALLOCATION.
*/
s = proclist_lock_write();
/*
* Find an unused process ID. We remember a range of unused IDs
* ready to use (from nextpid+1 through pidchecked-1).
*/
nextpid++;
retry:
/*
* If the process ID prototype has wrapped around,
* restart somewhat above 0, as the low-numbered procs
* tend to include daemons that don't exit.
*/
if (nextpid >= PID_MAX) {
nextpid = 100;
pidchecked = 0;
}
if (nextpid >= pidchecked) {
const struct proclist_desc *pd;
pidchecked = PID_MAX;
/*
* Scan the process lists to check whether this pid
* is in use. Remember the lowest pid that's greater
* than nextpid, so we can avoid checking for a while.
*/
pd = proclists;
again:
for (p2 = LIST_FIRST(pd->pd_list); p2 != 0;
p2 = LIST_NEXT(p2, p_list)) {
while (p2->p_pid == nextpid ||
p2->p_pgrp->pg_id == nextpid ||
p2->p_session->s_sid == nextpid) {
nextpid++;
if (nextpid >= pidchecked)
goto retry;
}
if (p2->p_pid > nextpid && pidchecked > p2->p_pid)
pidchecked = p2->p_pid;
if (p2->p_pgrp->pg_id > nextpid &&
pidchecked > p2->p_pgrp->pg_id)
pidchecked = p2->p_pgrp->pg_id;
if (p2->p_session->s_sid > nextpid &&
pidchecked > p2->p_session->s_sid)
pidchecked = p2->p_session->s_sid;
}
/*
* If there's another list, scan it. If we have checked
* them all, we've found one!
*/
pd++;
if (pd->pd_list != NULL)
goto again;
}
nprocs++;
p2 = newproc;
/* Record the pid we've allocated. */
p2->p_pid = nextpid;
/* Record the signal to be delivered to the parent on exit. */
p2->p_exitsig = exitsig;
/*
* Put the proc on allproc before unlocking PID allocation
* so that waiters won't grab it as soon as we unlock.
*/
p2->p_stat = SIDL; /* protect against others */
p2->p_forw = p2->p_back = NULL; /* shouldn't be necessary */
LIST_INSERT_HEAD(&allproc, p2, p_list);
LIST_INSERT_HEAD(PIDHASH(p2->p_pid), p2, p_hash);
/*
* END PID ALLOCATION.
*/
proclist_unlock_write(s);
p2 = pool_get(&proc_pool, PR_WAITOK);
/*
* Make a proc table entry for the new process.
@ -274,7 +189,7 @@ again:
/*
* Duplicate sub-structures as needed.
* Increase reference counts on shared objects.
* The p_stats and p_sigacts substructs are set in vm_fork.
* The p_stats and p_sigacts substructs are set in uvm_fork().
*/
p2->p_flag = P_INMEM | (p1->p_flag & P_SUGID);
p2->p_emul = p1->p_emul;
@ -368,6 +283,89 @@ again:
(func != NULL) ? func : child_return,
(arg != NULL) ? arg : p2);
/*
* BEGIN PID ALLOCATION.
*/
s = proclist_lock_write();
/*
* Find an unused process ID. We remember a range of unused IDs
* ready to use (from nextpid+1 through pidchecked-1).
*/
nextpid++;
retry:
/*
* If the process ID prototype has wrapped around,
* restart somewhat above 0, as the low-numbered procs
* tend to include daemons that don't exit.
*/
if (nextpid >= PID_MAX) {
nextpid = 100;
pidchecked = 0;
}
if (nextpid >= pidchecked) {
const struct proclist_desc *pd;
pidchecked = PID_MAX;
/*
* Scan the process lists to check whether this pid
* is in use. Remember the lowest pid that's greater
* than nextpid, so we can avoid checking for a while.
*/
pd = proclists;
again:
LIST_FOREACH(tp, pd->pd_list, p_list) {
while (tp->p_pid == nextpid ||
tp->p_pgrp->pg_id == nextpid ||
tp->p_session->s_sid == nextpid) {
nextpid++;
if (nextpid >= pidchecked)
goto retry;
}
if (tp->p_pid > nextpid && pidchecked > tp->p_pid)
pidchecked = tp->p_pid;
if (tp->p_pgrp->pg_id > nextpid &&
pidchecked > tp->p_pgrp->pg_id)
pidchecked = tp->p_pgrp->pg_id;
if (tp->p_session->s_sid > nextpid &&
pidchecked > tp->p_session->s_sid)
pidchecked = tp->p_session->s_sid;
}
/*
* If there's another list, scan it. If we have checked
* them all, we've found one!
*/
pd++;
if (pd->pd_list != NULL)
goto again;
}
/* Record the pid we've allocated. */
p2->p_pid = nextpid;
/* Record the signal to be delivered to the parent on exit. */
p2->p_exitsig = exitsig;
/*
* Put the proc on allproc before unlocking PID allocation
* so that waiters won't grab it as soon as we unlock.
*/
p2->p_stat = SIDL; /* protect against others */
p2->p_forw = p2->p_back = NULL; /* shouldn't be necessary */
LIST_INSERT_HEAD(&allproc, p2, p_list);
LIST_INSERT_HEAD(PIDHASH(p2->p_pid), p2, p_hash);
/*
* END PID ALLOCATION.
*/
proclist_unlock_write(s);
/*
* Make child runnable, set start time, and add to run queue.
*/