2010-06-22 22:29:01 +04:00
|
|
|
/* $NetBSD: subr_xcall.c,v 1.12 2010/06/22 18:29:01 rmind Exp $ */
|
2007-10-08 19:12:05 +04:00
|
|
|
|
|
|
|
/*-
|
2010-06-22 22:29:01 +04:00
|
|
|
* Copyright (c) 2007-2010 The NetBSD Foundation, Inc.
|
2007-10-08 19:12:05 +04:00
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* This code is derived from software contributed to The NetBSD Foundation
|
2010-06-22 22:29:01 +04:00
|
|
|
* by Andrew Doran and Mindaugas Rasiukevicius.
|
2007-10-08 19:12:05 +04:00
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
|
|
|
|
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
|
|
|
|
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
|
|
|
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
|
|
|
|
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Cross call support
|
|
|
|
*
|
|
|
|
* Background
|
|
|
|
*
|
|
|
|
* Sometimes it is necessary to modify hardware state that is tied
|
|
|
|
* directly to individual CPUs (such as a CPU's local timer), and
|
|
|
|
* these updates can not be done remotely by another CPU. The LWP
|
|
|
|
* requesting the update may be unable to guarantee that it will be
|
|
|
|
* running on the CPU where the update must occur, when the update
|
|
|
|
* occurs.
|
|
|
|
*
|
|
|
|
* Additionally, it's sometimes necessary to modify per-CPU software
|
|
|
|
* state from a remote CPU. Where these update operations are so
|
|
|
|
* rare or the access to the per-CPU data so frequent that the cost
|
|
|
|
* of using locking or atomic operations to provide coherency is
|
2007-10-27 05:22:53 +04:00
|
|
|
* prohibitive, another way must be found.
|
2007-10-08 19:12:05 +04:00
|
|
|
*
|
|
|
|
* Cross calls help to solve these types of problem by allowing
|
|
|
|
* any CPU in the system to request that an arbitrary function be
|
|
|
|
* executed on any other CPU.
|
|
|
|
*
|
|
|
|
* Implementation
|
|
|
|
*
|
|
|
|
* A slow mechanism for making 'low priority' cross calls is
|
|
|
|
* provided. The function to be executed runs on the remote CPU
|
|
|
|
* within a bound kthread. No queueing is provided, and the
|
|
|
|
* implementation uses global state. The function being called may
|
|
|
|
* block briefly on locks, but in doing so must be careful to not
|
|
|
|
* interfere with other cross calls in the system. The function is
|
|
|
|
* called with thread context and not from a soft interrupt, so it
|
|
|
|
* can ensure that it is not interrupting other code running on the
|
|
|
|
* CPU, and so has exclusive access to the CPU. Since this facility
|
|
|
|
* is heavyweight, it's expected that it will not be used often.
|
|
|
|
*
|
2007-10-27 05:22:53 +04:00
|
|
|
* Cross calls must not allocate memory, as the pagedaemon uses
|
|
|
|
* them (and memory allocation may need to wait on the pagedaemon).
|
|
|
|
*
|
2010-06-22 22:29:01 +04:00
|
|
|
* A low-overhead mechanism for high priority calls (XC_HIGHPRI) is
|
|
|
|
* also provided. The function to be executed runs on a software
|
|
|
|
* interrupt context, at SOFTINT_CLOCK level, and is expected to be
|
|
|
|
* very lightweight, e.g. avoid blocking.
|
2007-10-08 19:12:05 +04:00
|
|
|
*/
|
|
|
|
|
|
|
|
#include <sys/cdefs.h>
|
2010-06-22 22:29:01 +04:00
|
|
|
__KERNEL_RCSID(0, "$NetBSD: subr_xcall.c,v 1.12 2010/06/22 18:29:01 rmind Exp $");
|
2007-10-08 19:12:05 +04:00
|
|
|
|
|
|
|
#include <sys/types.h>
|
|
|
|
#include <sys/param.h>
|
|
|
|
#include <sys/xcall.h>
|
|
|
|
#include <sys/mutex.h>
|
|
|
|
#include <sys/condvar.h>
|
|
|
|
#include <sys/evcnt.h>
|
|
|
|
#include <sys/kthread.h>
|
2007-10-08 21:26:40 +04:00
|
|
|
#include <sys/cpu.h>
|
2007-10-08 19:12:05 +04:00
|
|
|
|
2010-06-22 22:29:01 +04:00
|
|
|
/* Cross-call state box. */
|
|
|
|
typedef struct {
|
|
|
|
kmutex_t xc_lock;
|
|
|
|
kcondvar_t xc_busy;
|
|
|
|
xcfunc_t xc_func;
|
|
|
|
void * xc_arg1;
|
|
|
|
void * xc_arg2;
|
|
|
|
uint64_t xc_headp;
|
|
|
|
uint64_t xc_donep;
|
|
|
|
} xc_state_t;
|
|
|
|
|
|
|
|
/* Bit indicating high (1) or low (0) priority. */
|
|
|
|
#define XC_PRI_BIT (1ULL << 63)
|
|
|
|
|
|
|
|
/* Low priority xcall structures. */
|
|
|
|
static xc_state_t xc_low_pri;
|
|
|
|
static uint64_t xc_tailp;
|
|
|
|
|
|
|
|
/* High priority xcall structures. */
|
|
|
|
static xc_state_t xc_high_pri;
|
|
|
|
static void * xc_sih;
|
2007-10-08 19:12:05 +04:00
|
|
|
|
2010-06-22 22:29:01 +04:00
|
|
|
/* Event counters. */
|
2007-10-08 19:12:05 +04:00
|
|
|
static struct evcnt xc_unicast_ev;
|
|
|
|
static struct evcnt xc_broadcast_ev;
|
2010-06-22 22:29:01 +04:00
|
|
|
|
|
|
|
static void xc_init(void);
|
|
|
|
static void xc_thread(void *);
|
|
|
|
static void xc_highpri_intr(void *);
|
|
|
|
|
|
|
|
static inline uint64_t xc_highpri(xcfunc_t, void *, void *, struct cpu_info *);
|
|
|
|
static inline uint64_t xc_lowpri(xcfunc_t, void *, void *, struct cpu_info *);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* xc_init:
|
|
|
|
*
|
|
|
|
* Initialize low and high priority cross-call structures.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
xc_init(void)
|
|
|
|
{
|
|
|
|
xc_state_t *xclo = &xc_low_pri, *xchi = &xc_high_pri;
|
|
|
|
|
|
|
|
memset(xclo, 0, sizeof(xc_state_t));
|
|
|
|
mutex_init(&xclo->xc_lock, MUTEX_DEFAULT, IPL_NONE);
|
|
|
|
cv_init(&xclo->xc_busy, "xclocv");
|
|
|
|
xc_tailp = 0;
|
|
|
|
|
|
|
|
memset(xchi, 0, sizeof(xc_state_t));
|
|
|
|
mutex_init(&xchi->xc_lock, MUTEX_DEFAULT, IPL_SOFTCLOCK);
|
|
|
|
cv_init(&xchi->xc_busy, "xchicv");
|
|
|
|
xc_sih = softint_establish(SOFTINT_CLOCK | SOFTINT_MPSAFE,
|
|
|
|
xc_highpri_intr, NULL);
|
|
|
|
KASSERT(xc_sih != NULL);
|
|
|
|
|
|
|
|
evcnt_attach_dynamic(&xc_unicast_ev, EVCNT_TYPE_MISC, NULL,
|
|
|
|
"crosscall", "unicast");
|
|
|
|
evcnt_attach_dynamic(&xc_broadcast_ev, EVCNT_TYPE_MISC, NULL,
|
|
|
|
"crosscall", "broadcast");
|
|
|
|
}
|
2007-10-08 19:12:05 +04:00
|
|
|
|
|
|
|
/*
|
|
|
|
* xc_init_cpu:
|
|
|
|
*
|
|
|
|
* Initialize the cross-call subsystem. Called once for each CPU
|
|
|
|
* in the system as they are attached.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
xc_init_cpu(struct cpu_info *ci)
|
|
|
|
{
|
2009-11-30 18:37:56 +03:00
|
|
|
static bool again = false;
|
2007-10-08 19:12:05 +04:00
|
|
|
int error;
|
|
|
|
|
|
|
|
if (!again) {
|
|
|
|
/* Autoconfiguration will prevent re-entry. */
|
2010-06-22 22:29:01 +04:00
|
|
|
xc_init();
|
2007-10-08 19:12:05 +04:00
|
|
|
again = true;
|
|
|
|
}
|
|
|
|
cv_init(&ci->ci_data.cpu_xcall, "xcall");
|
|
|
|
error = kthread_create(PRI_XCALL, KTHREAD_MPSAFE, ci, xc_thread,
|
2008-03-11 01:20:14 +03:00
|
|
|
NULL, NULL, "xcall/%u", ci->ci_index);
|
2010-06-22 22:29:01 +04:00
|
|
|
KASSERT(error == 0);
|
2007-10-08 19:12:05 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2008-04-14 04:18:43 +04:00
|
|
|
* xc_broadcast:
|
2007-10-08 19:12:05 +04:00
|
|
|
*
|
|
|
|
* Trigger a call on all CPUs in the system.
|
|
|
|
*/
|
|
|
|
uint64_t
|
|
|
|
xc_broadcast(u_int flags, xcfunc_t func, void *arg1, void *arg2)
|
|
|
|
{
|
|
|
|
|
2010-06-22 22:29:01 +04:00
|
|
|
KASSERT(!cpu_intr_p() && !cpu_softintr_p());
|
|
|
|
|
2007-10-08 19:12:05 +04:00
|
|
|
if ((flags & XC_HIGHPRI) != 0) {
|
2010-06-22 22:29:01 +04:00
|
|
|
return xc_highpri(func, arg1, arg2, NULL);
|
2007-10-08 19:12:05 +04:00
|
|
|
} else {
|
2010-06-22 22:29:01 +04:00
|
|
|
return xc_lowpri(func, arg1, arg2, NULL);
|
2007-10-08 19:12:05 +04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* xc_unicast:
|
|
|
|
*
|
|
|
|
* Trigger a call on one CPU.
|
|
|
|
*/
|
|
|
|
uint64_t
|
|
|
|
xc_unicast(u_int flags, xcfunc_t func, void *arg1, void *arg2,
|
|
|
|
struct cpu_info *ci)
|
|
|
|
{
|
|
|
|
|
2010-06-22 22:29:01 +04:00
|
|
|
KASSERT(ci != NULL);
|
|
|
|
KASSERT(!cpu_intr_p() && !cpu_softintr_p());
|
|
|
|
|
2007-10-08 19:12:05 +04:00
|
|
|
if ((flags & XC_HIGHPRI) != 0) {
|
2010-06-22 22:29:01 +04:00
|
|
|
return xc_highpri(func, arg1, arg2, ci);
|
2007-10-08 19:12:05 +04:00
|
|
|
} else {
|
2010-06-22 22:29:01 +04:00
|
|
|
return xc_lowpri(func, arg1, arg2, ci);
|
2007-10-08 19:12:05 +04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2010-06-22 22:29:01 +04:00
|
|
|
/*
|
|
|
|
* xc_wait:
|
|
|
|
*
|
|
|
|
* Wait for a cross call to complete.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
xc_wait(uint64_t where)
|
|
|
|
{
|
|
|
|
xc_state_t *xc;
|
|
|
|
|
|
|
|
KASSERT(!cpu_intr_p() && !cpu_softintr_p());
|
|
|
|
|
|
|
|
/* Determine whether it is high or low priority cross-call. */
|
|
|
|
if ((where & XC_PRI_BIT) != 0) {
|
|
|
|
xc = &xc_high_pri;
|
|
|
|
where &= ~XC_PRI_BIT;
|
|
|
|
} else {
|
|
|
|
xc = &xc_low_pri;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Fast path, if already done. */
|
|
|
|
if (xc->xc_donep >= where) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Slow path: block until awoken. */
|
|
|
|
mutex_enter(&xc->xc_lock);
|
|
|
|
while (xc->xc_donep < where) {
|
|
|
|
cv_wait(&xc->xc_busy, &xc->xc_lock);
|
|
|
|
}
|
|
|
|
mutex_exit(&xc->xc_lock);
|
|
|
|
}
|
|
|
|
|
2007-10-08 19:12:05 +04:00
|
|
|
/*
|
|
|
|
* xc_lowpri:
|
|
|
|
*
|
|
|
|
* Trigger a low priority call on one or more CPUs.
|
|
|
|
*/
|
2010-06-22 22:29:01 +04:00
|
|
|
static inline uint64_t
|
|
|
|
xc_lowpri(xcfunc_t func, void *arg1, void *arg2, struct cpu_info *ci)
|
2007-10-08 19:12:05 +04:00
|
|
|
{
|
2010-06-22 22:29:01 +04:00
|
|
|
xc_state_t *xc = &xc_low_pri;
|
2007-10-08 19:12:05 +04:00
|
|
|
CPU_INFO_ITERATOR cii;
|
2009-03-05 16:18:51 +03:00
|
|
|
uint64_t where;
|
2007-10-08 19:12:05 +04:00
|
|
|
|
2010-06-22 22:29:01 +04:00
|
|
|
mutex_enter(&xc->xc_lock);
|
|
|
|
while (xc->xc_headp != xc_tailp) {
|
|
|
|
cv_wait(&xc->xc_busy, &xc->xc_lock);
|
|
|
|
}
|
|
|
|
xc->xc_arg1 = arg1;
|
|
|
|
xc->xc_arg2 = arg2;
|
|
|
|
xc->xc_func = func;
|
2007-10-08 19:12:05 +04:00
|
|
|
if (ci == NULL) {
|
|
|
|
xc_broadcast_ev.ev_count++;
|
|
|
|
for (CPU_INFO_FOREACH(cii, ci)) {
|
2008-04-24 17:56:30 +04:00
|
|
|
if ((ci->ci_schedstate.spc_flags & SPCF_RUNNING) == 0)
|
|
|
|
continue;
|
2010-06-22 22:29:01 +04:00
|
|
|
xc->xc_headp += 1;
|
2007-10-08 19:12:05 +04:00
|
|
|
ci->ci_data.cpu_xcall_pending = true;
|
|
|
|
cv_signal(&ci->ci_data.cpu_xcall);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
xc_unicast_ev.ev_count++;
|
2010-06-22 22:29:01 +04:00
|
|
|
xc->xc_headp += 1;
|
2007-10-08 19:12:05 +04:00
|
|
|
ci->ci_data.cpu_xcall_pending = true;
|
|
|
|
cv_signal(&ci->ci_data.cpu_xcall);
|
|
|
|
}
|
2010-06-22 22:29:01 +04:00
|
|
|
KASSERT(xc_tailp < xc->xc_headp);
|
|
|
|
where = xc->xc_headp;
|
|
|
|
mutex_exit(&xc->xc_lock);
|
2007-10-08 19:12:05 +04:00
|
|
|
|
2010-06-22 22:29:01 +04:00
|
|
|
/* Return a low priority ticket. */
|
|
|
|
KASSERT((where & XC_PRI_BIT) == 0);
|
2007-10-08 19:12:05 +04:00
|
|
|
return where;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* xc_thread:
|
|
|
|
*
|
|
|
|
* One thread per-CPU to dispatch low priority calls.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
xc_thread(void *cookie)
|
|
|
|
{
|
2010-06-22 22:29:01 +04:00
|
|
|
struct cpu_info *ci = curcpu();
|
|
|
|
xc_state_t *xc = &xc_low_pri;
|
2007-10-08 19:12:05 +04:00
|
|
|
void *arg1, *arg2;
|
|
|
|
xcfunc_t func;
|
|
|
|
|
2010-06-22 22:29:01 +04:00
|
|
|
mutex_enter(&xc->xc_lock);
|
2007-10-08 19:12:05 +04:00
|
|
|
for (;;) {
|
|
|
|
while (!ci->ci_data.cpu_xcall_pending) {
|
2010-06-22 22:29:01 +04:00
|
|
|
if (xc->xc_headp == xc_tailp) {
|
|
|
|
cv_broadcast(&xc->xc_busy);
|
|
|
|
}
|
|
|
|
cv_wait(&ci->ci_data.cpu_xcall, &xc->xc_lock);
|
2007-10-08 19:12:05 +04:00
|
|
|
KASSERT(ci == curcpu());
|
|
|
|
}
|
|
|
|
ci->ci_data.cpu_xcall_pending = false;
|
2010-06-22 22:29:01 +04:00
|
|
|
func = xc->xc_func;
|
|
|
|
arg1 = xc->xc_arg1;
|
|
|
|
arg2 = xc->xc_arg2;
|
2007-10-08 19:12:05 +04:00
|
|
|
xc_tailp++;
|
2010-06-22 22:29:01 +04:00
|
|
|
mutex_exit(&xc->xc_lock);
|
2007-10-08 19:12:05 +04:00
|
|
|
|
2010-06-22 22:29:01 +04:00
|
|
|
KASSERT(func != NULL);
|
2007-10-08 19:12:05 +04:00
|
|
|
(*func)(arg1, arg2);
|
|
|
|
|
2010-06-22 22:29:01 +04:00
|
|
|
mutex_enter(&xc->xc_lock);
|
|
|
|
xc->xc_donep++;
|
2007-10-08 19:12:05 +04:00
|
|
|
}
|
|
|
|
/* NOTREACHED */
|
|
|
|
}
|
2010-06-22 22:29:01 +04:00
|
|
|
|
|
|
|
/*
|
|
|
|
* xc_ipi_handler:
|
|
|
|
*
|
|
|
|
* Handler of cross-call IPI.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
xc_ipi_handler(void)
|
|
|
|
{
|
|
|
|
/* Executes xc_highpri_intr() via software interrupt. */
|
|
|
|
softint_schedule(xc_sih);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* xc_highpri_intr:
|
|
|
|
*
|
|
|
|
* A software interrupt handler for high priority calls.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
xc_highpri_intr(void *dummy)
|
|
|
|
{
|
|
|
|
xc_state_t *xc = &xc_high_pri;
|
|
|
|
void *arg1, *arg2;
|
|
|
|
xcfunc_t func;
|
|
|
|
|
|
|
|
KASSERT(cpu_softintr_p());
|
|
|
|
/*
|
|
|
|
* Lock-less fetch of function and its arguments.
|
|
|
|
* Safe since it cannot change at this point.
|
|
|
|
*/
|
|
|
|
KASSERT(xc->xc_donep < xc->xc_headp);
|
|
|
|
func = xc->xc_func;
|
|
|
|
arg1 = xc->xc_arg1;
|
|
|
|
arg2 = xc->xc_arg2;
|
|
|
|
|
|
|
|
KASSERT(func != NULL);
|
|
|
|
(*func)(arg1, arg2);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Note the request as done, and if we have reached the head,
|
|
|
|
* cross-call has been processed - notify waiters, if any.
|
|
|
|
*/
|
|
|
|
mutex_enter(&xc->xc_lock);
|
|
|
|
if (++xc->xc_donep == xc->xc_headp) {
|
|
|
|
cv_broadcast(&xc->xc_busy);
|
|
|
|
}
|
|
|
|
mutex_exit(&xc->xc_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* xc_highpri:
|
|
|
|
*
|
|
|
|
* Trigger a high priority call on one or more CPUs.
|
|
|
|
*/
|
|
|
|
static inline uint64_t
|
|
|
|
xc_highpri(xcfunc_t func, void *arg1, void *arg2, struct cpu_info *ci)
|
|
|
|
{
|
|
|
|
xc_state_t *xc = &xc_high_pri;
|
|
|
|
uint64_t where;
|
|
|
|
|
|
|
|
mutex_enter(&xc->xc_lock);
|
|
|
|
while (xc->xc_headp != xc->xc_donep) {
|
|
|
|
cv_wait(&xc->xc_busy, &xc->xc_lock);
|
|
|
|
}
|
|
|
|
xc->xc_func = func;
|
|
|
|
xc->xc_arg1 = arg1;
|
|
|
|
xc->xc_arg2 = arg2;
|
|
|
|
xc->xc_headp += (ci ? 1 : ncpu);
|
|
|
|
where = xc->xc_headp;
|
|
|
|
mutex_exit(&xc->xc_lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Send the IPI once lock is released.
|
|
|
|
* Note: it will handle the local CPU case.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifdef MULTIPROCESSOR
|
|
|
|
kpreempt_disable();
|
|
|
|
if (curcpu() == ci) {
|
|
|
|
/* Unicast: local CPU. */
|
|
|
|
xc_ipi_handler();
|
|
|
|
} else if (ci) {
|
|
|
|
/* Unicast: remote CPU. */
|
|
|
|
xc_send_ipi(ci);
|
|
|
|
} else {
|
|
|
|
/* Broadcast: all, including local. */
|
|
|
|
xc_send_ipi(NULL);
|
|
|
|
xc_ipi_handler();
|
|
|
|
}
|
|
|
|
kpreempt_enable();
|
|
|
|
#else
|
|
|
|
KASSERT(curcpu() == ci);
|
|
|
|
xc_ipi_handler();
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* Indicate a high priority ticket. */
|
|
|
|
return (where | XC_PRI_BIT);
|
|
|
|
}
|