in fork1(), only add make the new proc visible (by giving it a pid
and adding it to allproc) after it's fully initialized. this prevents the scheduler from coming in via a clock interrupt and tripping over a partially-initialized proc.
This commit is contained in:
parent
b7192d086c
commit
a284236c04
|
@ -1,4 +1,4 @@
|
||||||
/* $NetBSD: kern_fork.c,v 1.75 2000/11/07 12:41:52 jdolecek Exp $ */
|
/* $NetBSD: kern_fork.c,v 1.76 2000/11/08 05:16:23 chs Exp $ */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Copyright (c) 1982, 1986, 1989, 1991, 1993
|
* Copyright (c) 1982, 1986, 1989, 1991, 1993
|
||||||
|
@ -106,9 +106,8 @@ fork1(struct proc *p1, int flags, int exitsig, void *stack, size_t stacksize,
|
||||||
void (*func)(void *), void *arg, register_t *retval,
|
void (*func)(void *), void *arg, register_t *retval,
|
||||||
struct proc **rnewprocp)
|
struct proc **rnewprocp)
|
||||||
{
|
{
|
||||||
struct proc *p2;
|
struct proc *p2, *tp;
|
||||||
uid_t uid;
|
uid_t uid;
|
||||||
struct proc *newproc;
|
|
||||||
int count, s;
|
int count, s;
|
||||||
vaddr_t uaddr;
|
vaddr_t uaddr;
|
||||||
static int nextpid, pidchecked = 0;
|
static int nextpid, pidchecked = 0;
|
||||||
|
@ -126,6 +125,7 @@ fork1(struct proc *p1, int flags, int exitsig, void *stack, size_t stacksize,
|
||||||
tablefull("proc", "increase kern.maxproc or NPROC");
|
tablefull("proc", "increase kern.maxproc or NPROC");
|
||||||
return (EAGAIN);
|
return (EAGAIN);
|
||||||
}
|
}
|
||||||
|
nprocs++;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Increment the count of procs running with this uid. Don't allow
|
* Increment the count of procs running with this uid. Don't allow
|
||||||
|
@ -135,6 +135,7 @@ fork1(struct proc *p1, int flags, int exitsig, void *stack, size_t stacksize,
|
||||||
if (__predict_false(uid != 0 && count >
|
if (__predict_false(uid != 0 && count >
|
||||||
p1->p_rlimit[RLIMIT_NPROC].rlim_cur)) {
|
p1->p_rlimit[RLIMIT_NPROC].rlim_cur)) {
|
||||||
(void)chgproccnt(uid, -1);
|
(void)chgproccnt(uid, -1);
|
||||||
|
nprocs--;
|
||||||
return (EAGAIN);
|
return (EAGAIN);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -147,6 +148,7 @@ fork1(struct proc *p1, int flags, int exitsig, void *stack, size_t stacksize,
|
||||||
uaddr = uvm_km_valloc(kernel_map, USPACE);
|
uaddr = uvm_km_valloc(kernel_map, USPACE);
|
||||||
if (__predict_false(uaddr == 0)) {
|
if (__predict_false(uaddr == 0)) {
|
||||||
(void)chgproccnt(uid, -1);
|
(void)chgproccnt(uid, -1);
|
||||||
|
nprocs--;
|
||||||
return (ENOMEM);
|
return (ENOMEM);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -156,94 +158,7 @@ fork1(struct proc *p1, int flags, int exitsig, void *stack, size_t stacksize,
|
||||||
*/
|
*/
|
||||||
|
|
||||||
/* Allocate new proc. */
|
/* Allocate new proc. */
|
||||||
newproc = pool_get(&proc_pool, PR_WAITOK);
|
p2 = pool_get(&proc_pool, PR_WAITOK);
|
||||||
|
|
||||||
/*
|
|
||||||
* BEGIN PID ALLOCATION.
|
|
||||||
*/
|
|
||||||
s = proclist_lock_write();
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Find an unused process ID. We remember a range of unused IDs
|
|
||||||
* ready to use (from nextpid+1 through pidchecked-1).
|
|
||||||
*/
|
|
||||||
nextpid++;
|
|
||||||
retry:
|
|
||||||
/*
|
|
||||||
* If the process ID prototype has wrapped around,
|
|
||||||
* restart somewhat above 0, as the low-numbered procs
|
|
||||||
* tend to include daemons that don't exit.
|
|
||||||
*/
|
|
||||||
if (nextpid >= PID_MAX) {
|
|
||||||
nextpid = 100;
|
|
||||||
pidchecked = 0;
|
|
||||||
}
|
|
||||||
if (nextpid >= pidchecked) {
|
|
||||||
const struct proclist_desc *pd;
|
|
||||||
|
|
||||||
pidchecked = PID_MAX;
|
|
||||||
/*
|
|
||||||
* Scan the process lists to check whether this pid
|
|
||||||
* is in use. Remember the lowest pid that's greater
|
|
||||||
* than nextpid, so we can avoid checking for a while.
|
|
||||||
*/
|
|
||||||
pd = proclists;
|
|
||||||
again:
|
|
||||||
for (p2 = LIST_FIRST(pd->pd_list); p2 != 0;
|
|
||||||
p2 = LIST_NEXT(p2, p_list)) {
|
|
||||||
while (p2->p_pid == nextpid ||
|
|
||||||
p2->p_pgrp->pg_id == nextpid ||
|
|
||||||
p2->p_session->s_sid == nextpid) {
|
|
||||||
nextpid++;
|
|
||||||
if (nextpid >= pidchecked)
|
|
||||||
goto retry;
|
|
||||||
}
|
|
||||||
if (p2->p_pid > nextpid && pidchecked > p2->p_pid)
|
|
||||||
pidchecked = p2->p_pid;
|
|
||||||
|
|
||||||
if (p2->p_pgrp->pg_id > nextpid &&
|
|
||||||
pidchecked > p2->p_pgrp->pg_id)
|
|
||||||
pidchecked = p2->p_pgrp->pg_id;
|
|
||||||
|
|
||||||
if (p2->p_session->s_sid > nextpid &&
|
|
||||||
pidchecked > p2->p_session->s_sid)
|
|
||||||
pidchecked = p2->p_session->s_sid;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* If there's another list, scan it. If we have checked
|
|
||||||
* them all, we've found one!
|
|
||||||
*/
|
|
||||||
pd++;
|
|
||||||
if (pd->pd_list != NULL)
|
|
||||||
goto again;
|
|
||||||
}
|
|
||||||
|
|
||||||
nprocs++;
|
|
||||||
p2 = newproc;
|
|
||||||
|
|
||||||
/* Record the pid we've allocated. */
|
|
||||||
p2->p_pid = nextpid;
|
|
||||||
|
|
||||||
/* Record the signal to be delivered to the parent on exit. */
|
|
||||||
p2->p_exitsig = exitsig;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Put the proc on allproc before unlocking PID allocation
|
|
||||||
* so that waiters won't grab it as soon as we unlock.
|
|
||||||
*/
|
|
||||||
|
|
||||||
p2->p_stat = SIDL; /* protect against others */
|
|
||||||
p2->p_forw = p2->p_back = NULL; /* shouldn't be necessary */
|
|
||||||
|
|
||||||
LIST_INSERT_HEAD(&allproc, p2, p_list);
|
|
||||||
|
|
||||||
LIST_INSERT_HEAD(PIDHASH(p2->p_pid), p2, p_hash);
|
|
||||||
|
|
||||||
/*
|
|
||||||
* END PID ALLOCATION.
|
|
||||||
*/
|
|
||||||
proclist_unlock_write(s);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Make a proc table entry for the new process.
|
* Make a proc table entry for the new process.
|
||||||
|
@ -274,7 +189,7 @@ again:
|
||||||
/*
|
/*
|
||||||
* Duplicate sub-structures as needed.
|
* Duplicate sub-structures as needed.
|
||||||
* Increase reference counts on shared objects.
|
* Increase reference counts on shared objects.
|
||||||
* The p_stats and p_sigacts substructs are set in vm_fork.
|
* The p_stats and p_sigacts substructs are set in uvm_fork().
|
||||||
*/
|
*/
|
||||||
p2->p_flag = P_INMEM | (p1->p_flag & P_SUGID);
|
p2->p_flag = P_INMEM | (p1->p_flag & P_SUGID);
|
||||||
p2->p_emul = p1->p_emul;
|
p2->p_emul = p1->p_emul;
|
||||||
|
@ -368,6 +283,89 @@ again:
|
||||||
(func != NULL) ? func : child_return,
|
(func != NULL) ? func : child_return,
|
||||||
(arg != NULL) ? arg : p2);
|
(arg != NULL) ? arg : p2);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* BEGIN PID ALLOCATION.
|
||||||
|
*/
|
||||||
|
s = proclist_lock_write();
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Find an unused process ID. We remember a range of unused IDs
|
||||||
|
* ready to use (from nextpid+1 through pidchecked-1).
|
||||||
|
*/
|
||||||
|
nextpid++;
|
||||||
|
retry:
|
||||||
|
/*
|
||||||
|
* If the process ID prototype has wrapped around,
|
||||||
|
* restart somewhat above 0, as the low-numbered procs
|
||||||
|
* tend to include daemons that don't exit.
|
||||||
|
*/
|
||||||
|
if (nextpid >= PID_MAX) {
|
||||||
|
nextpid = 100;
|
||||||
|
pidchecked = 0;
|
||||||
|
}
|
||||||
|
if (nextpid >= pidchecked) {
|
||||||
|
const struct proclist_desc *pd;
|
||||||
|
|
||||||
|
pidchecked = PID_MAX;
|
||||||
|
/*
|
||||||
|
* Scan the process lists to check whether this pid
|
||||||
|
* is in use. Remember the lowest pid that's greater
|
||||||
|
* than nextpid, so we can avoid checking for a while.
|
||||||
|
*/
|
||||||
|
pd = proclists;
|
||||||
|
again:
|
||||||
|
LIST_FOREACH(tp, pd->pd_list, p_list) {
|
||||||
|
while (tp->p_pid == nextpid ||
|
||||||
|
tp->p_pgrp->pg_id == nextpid ||
|
||||||
|
tp->p_session->s_sid == nextpid) {
|
||||||
|
nextpid++;
|
||||||
|
if (nextpid >= pidchecked)
|
||||||
|
goto retry;
|
||||||
|
}
|
||||||
|
if (tp->p_pid > nextpid && pidchecked > tp->p_pid)
|
||||||
|
pidchecked = tp->p_pid;
|
||||||
|
|
||||||
|
if (tp->p_pgrp->pg_id > nextpid &&
|
||||||
|
pidchecked > tp->p_pgrp->pg_id)
|
||||||
|
pidchecked = tp->p_pgrp->pg_id;
|
||||||
|
|
||||||
|
if (tp->p_session->s_sid > nextpid &&
|
||||||
|
pidchecked > tp->p_session->s_sid)
|
||||||
|
pidchecked = tp->p_session->s_sid;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* If there's another list, scan it. If we have checked
|
||||||
|
* them all, we've found one!
|
||||||
|
*/
|
||||||
|
pd++;
|
||||||
|
if (pd->pd_list != NULL)
|
||||||
|
goto again;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Record the pid we've allocated. */
|
||||||
|
p2->p_pid = nextpid;
|
||||||
|
|
||||||
|
/* Record the signal to be delivered to the parent on exit. */
|
||||||
|
p2->p_exitsig = exitsig;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Put the proc on allproc before unlocking PID allocation
|
||||||
|
* so that waiters won't grab it as soon as we unlock.
|
||||||
|
*/
|
||||||
|
|
||||||
|
p2->p_stat = SIDL; /* protect against others */
|
||||||
|
p2->p_forw = p2->p_back = NULL; /* shouldn't be necessary */
|
||||||
|
|
||||||
|
LIST_INSERT_HEAD(&allproc, p2, p_list);
|
||||||
|
|
||||||
|
LIST_INSERT_HEAD(PIDHASH(p2->p_pid), p2, p_hash);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* END PID ALLOCATION.
|
||||||
|
*/
|
||||||
|
proclist_unlock_write(s);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Make child runnable, set start time, and add to run queue.
|
* Make child runnable, set start time, and add to run queue.
|
||||||
*/
|
*/
|
||||||
|
|
Loading…
Reference in New Issue