852 lines
24 KiB
C
852 lines
24 KiB
C
/* $NetBSD: puffs_msgif.c,v 1.21 2007/03/20 10:21:58 pooka Exp $ */
|
|
|
|
/*
|
|
* Copyright (c) 2005, 2006, 2007 Antti Kantee. All Rights Reserved.
|
|
*
|
|
* Development of this software was supported by the
|
|
* Google Summer of Code program and the Ulla Tuominen Foundation.
|
|
* The Google SoC project was mentored by Bill Studenmund.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. The name of the company nor the name of the author may be used to
|
|
* endorse or promote products derived from this software without specific
|
|
* prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS
|
|
* OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
|
* WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
* DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
|
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__KERNEL_RCSID(0, "$NetBSD: puffs_msgif.c,v 1.21 2007/03/20 10:21:58 pooka Exp $");
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/fstrans.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/mount.h>
|
|
#include <sys/vnode.h>
|
|
#include <sys/lock.h>
|
|
|
|
#include <fs/puffs/puffs_msgif.h>
|
|
#include <fs/puffs/puffs_sys.h>
|
|
|
|
|
|
/*
|
|
* kernel-user-kernel waitqueues
|
|
*/
|
|
|
|
static int touser(struct puffs_mount *, struct puffs_park *, uint64_t,
|
|
struct vnode *, struct vnode *);
|
|
|
|
uint64_t
|
|
puffs_getreqid(struct puffs_mount *pmp)
|
|
{
|
|
uint64_t rv;
|
|
|
|
simple_lock(&pmp->pmp_lock);
|
|
rv = pmp->pmp_nextreq++;
|
|
simple_unlock(&pmp->pmp_lock);
|
|
|
|
return rv;
|
|
}
|
|
|
|
/* vfs request */
|
|
int
|
|
puffs_vfstouser(struct puffs_mount *pmp, int optype, void *kbuf, size_t buflen)
|
|
{
|
|
struct puffs_park park;
|
|
|
|
park.park_preq = kbuf;
|
|
|
|
park.park_preq->preq_opclass = PUFFSOP_VFS;
|
|
park.park_preq->preq_optype = optype;
|
|
|
|
park.park_maxlen = park.park_copylen = buflen;
|
|
park.park_flags = 0;
|
|
|
|
return touser(pmp, &park, puffs_getreqid(pmp), NULL, NULL);
|
|
}
|
|
|
|
void
|
|
puffs_suspendtouser(struct puffs_mount *pmp, int status)
|
|
{
|
|
struct puffs_vfsreq_suspend *pvfsr_susp;
|
|
struct puffs_park *ppark;
|
|
|
|
pvfsr_susp = malloc(sizeof(struct puffs_vfsreq_suspend),
|
|
M_PUFFS, M_WAITOK | M_ZERO);
|
|
ppark = malloc(sizeof(struct puffs_park), M_PUFFS, M_WAITOK | M_ZERO);
|
|
|
|
pvfsr_susp->pvfsr_status = status;
|
|
ppark->park_preq = (struct puffs_req *)pvfsr_susp;
|
|
|
|
ppark->park_preq->preq_opclass = PUFFSOP_VFS | PUFFSOPFLAG_FAF;
|
|
ppark->park_preq->preq_optype = PUFFS_VFS_SUSPEND;
|
|
|
|
ppark->park_maxlen = ppark->park_copylen
|
|
= sizeof(struct puffs_vfsreq_suspend);
|
|
ppark->park_flags = 0;
|
|
|
|
(void)touser(pmp, ppark, 0, NULL, NULL);
|
|
}
|
|
|
|
/*
|
|
* vnode level request
|
|
*/
|
|
int
|
|
puffs_vntouser(struct puffs_mount *pmp, int optype,
|
|
void *kbuf, size_t buflen, void *cookie,
|
|
struct vnode *vp1, struct vnode *vp2)
|
|
{
|
|
struct puffs_park park;
|
|
|
|
park.park_preq = kbuf;
|
|
|
|
park.park_preq->preq_opclass = PUFFSOP_VN;
|
|
park.park_preq->preq_optype = optype;
|
|
park.park_preq->preq_cookie = cookie;
|
|
|
|
park.park_maxlen = park.park_copylen = buflen;
|
|
park.park_flags = 0;
|
|
|
|
return touser(pmp, &park, puffs_getreqid(pmp), vp1, vp2);
|
|
}
|
|
|
|
/*
|
|
* vnode level request, caller-controller req id
|
|
*/
|
|
int
|
|
puffs_vntouser_req(struct puffs_mount *pmp, int optype,
|
|
void *kbuf, size_t buflen, void *cookie, uint64_t reqid,
|
|
struct vnode *vp1, struct vnode *vp2)
|
|
{
|
|
struct puffs_park park;
|
|
|
|
park.park_preq = kbuf;
|
|
|
|
park.park_preq->preq_opclass = PUFFSOP_VN;
|
|
park.park_preq->preq_optype = optype;
|
|
park.park_preq->preq_cookie = cookie;
|
|
|
|
park.park_maxlen = park.park_copylen = buflen;
|
|
park.park_flags = 0;
|
|
|
|
return touser(pmp, &park, reqid, vp1, vp2);
|
|
}
|
|
|
|
/*
|
|
* vnode level request, copy routines can adjust "kernbuf".
|
|
*/
|
|
int
|
|
puffs_vntouser_adjbuf(struct puffs_mount *pmp, int optype,
|
|
void **kbuf, size_t *buflen, size_t maxdelta,
|
|
void *cookie, struct vnode *vp1, struct vnode *vp2)
|
|
{
|
|
struct puffs_park park;
|
|
int error;
|
|
|
|
park.park_preq = *kbuf;
|
|
|
|
park.park_preq->preq_opclass = PUFFSOP_VN;
|
|
park.park_preq->preq_optype = optype;
|
|
park.park_preq->preq_cookie = cookie;
|
|
|
|
park.park_copylen = *buflen;
|
|
park.park_maxlen = maxdelta + *buflen;
|
|
park.park_flags = PUFFS_PARKFLAG_ADJUSTABLE;
|
|
|
|
error = touser(pmp, &park, puffs_getreqid(pmp), vp1, vp2);
|
|
|
|
*kbuf = park.park_preq;
|
|
*buflen = park.park_copylen;
|
|
|
|
return error;
|
|
}
|
|
|
|
/*
|
|
* File server interaction is async from caller perspective.
|
|
* biodone(bp)J is signalled in putop.
|
|
*/
|
|
void
|
|
puffs_vntouser_bioread_async(struct puffs_mount *pmp, void *cookie,
|
|
size_t tomove, off_t offset, struct buf *bp,
|
|
struct vnode *vp1, struct vnode *vp2)
|
|
{
|
|
struct puffs_park *ppark;
|
|
struct puffs_vnreq_read *read_argp;
|
|
|
|
MALLOC(ppark, struct puffs_park *, sizeof(struct puffs_park),
|
|
M_PUFFS, M_WAITOK);
|
|
MALLOC(read_argp, struct puffs_vnreq_read *,
|
|
sizeof(struct puffs_vnreq_read), M_PUFFS, M_WAITOK | M_ZERO);
|
|
|
|
read_argp->pvnr_ioflag = 0;
|
|
read_argp->pvnr_resid = tomove;
|
|
read_argp->pvnr_offset = offset;
|
|
puffs_credcvt(&read_argp->pvnr_cred, FSCRED);
|
|
|
|
ppark->park_preq = (void *)read_argp;
|
|
ppark->park_preq->preq_opclass = PUFFSOP_VN;
|
|
ppark->park_preq->preq_optype = PUFFS_VN_READ;
|
|
ppark->park_preq->preq_cookie = cookie;
|
|
|
|
ppark->park_copylen = sizeof(struct puffs_vnreq_read);
|
|
ppark->park_maxlen = sizeof(struct puffs_vnreq_read) + tomove;
|
|
ppark->park_bp = bp;
|
|
ppark->park_flags
|
|
= PUFFS_PARKFLAG_ADJUSTABLE | PUFFS_PARKFLAG_ASYNCBIOREAD;
|
|
|
|
(void)touser(pmp, ppark, puffs_getreqid(pmp), vp1, vp2);
|
|
}
|
|
|
|
/*
|
|
* Notice: kbuf will be free'd later. I must be allocated from the
|
|
* kernel heap and it's ownership is shifted to this function from
|
|
* now on, i.e. the caller is not allowed to use it anymore!
|
|
*/
|
|
void
|
|
puffs_vntouser_faf(struct puffs_mount *pmp, int optype,
|
|
void *kbuf, size_t buflen, void *cookie)
|
|
{
|
|
struct puffs_park *ppark;
|
|
|
|
/* XXX: is it allowable to sleep here? */
|
|
ppark = malloc(sizeof(struct puffs_park), M_PUFFS, M_NOWAIT | M_ZERO);
|
|
if (ppark == NULL)
|
|
return; /* 2bad */
|
|
|
|
ppark->park_preq = kbuf;
|
|
|
|
ppark->park_preq->preq_opclass = PUFFSOP_VN | PUFFSOPFLAG_FAF;
|
|
ppark->park_preq->preq_optype = optype;
|
|
ppark->park_preq->preq_cookie = cookie;
|
|
|
|
ppark->park_maxlen = ppark->park_copylen = buflen;
|
|
ppark->park_flags = 0;
|
|
|
|
(void)touser(pmp, ppark, 0, NULL, NULL);
|
|
}
|
|
|
|
void
|
|
puffs_cacheop(struct puffs_mount *pmp, struct puffs_park *ppark,
|
|
struct puffs_cacheinfo *pcinfo, size_t pcilen, void *cookie)
|
|
{
|
|
|
|
ppark->park_preq = (struct puffs_req *)pcinfo;
|
|
ppark->park_preq->preq_opclass = PUFFSOP_CACHE | PUFFSOPFLAG_FAF;
|
|
ppark->park_preq->preq_optype = PCACHE_TYPE_WRITE; /* XXX */
|
|
ppark->park_preq->preq_cookie = cookie;
|
|
|
|
ppark->park_maxlen = ppark->park_copylen = pcilen;
|
|
|
|
(void)touser(pmp, ppark, 0, NULL, NULL);
|
|
}
|
|
|
|
/*
|
|
* Wait for the userspace ping-pong game in calling process context.
|
|
*
|
|
* This unlocks vnodes if they are supplied. vp1 is the vnode
|
|
* before in the locking order, i.e. the one which must be locked
|
|
* before accessing vp2. This is done here so that operations are
|
|
* already ordered in the queue when vnodes are unlocked (I'm not
|
|
* sure if that's really necessary, but it can't hurt). Okok, maybe
|
|
* there's a slight ugly-factor also, but let's not worry about that.
|
|
*/
|
|
static int
|
|
touser(struct puffs_mount *pmp, struct puffs_park *ppark, uint64_t reqid,
|
|
struct vnode *vp1, struct vnode *vp2)
|
|
{
|
|
struct lwp *l = curlwp;
|
|
struct mount *mp;
|
|
struct puffs_req *preq;
|
|
int rv = 0;
|
|
|
|
mp = PMPTOMP(pmp);
|
|
preq = ppark->park_preq;
|
|
preq->preq_id = ppark->park_id = reqid;
|
|
preq->preq_buflen = ALIGN(ppark->park_maxlen);
|
|
|
|
/*
|
|
* To support PCATCH, yet another movie: check if there are signals
|
|
* pending and we are issueing a non-FAF. If so, return an error
|
|
* directly UNLESS we are issueing INACTIVE. In that case, convert
|
|
* it to a FAF, fire off to the file server and return an error.
|
|
* Yes, this is bordering disgusting. Barfbags are on me.
|
|
*/
|
|
if (PUFFSOP_WANTREPLY(ppark->park_preq->preq_opclass)
|
|
&& (ppark->park_flags & PUFFS_PARKFLAG_ASYNCBIOREAD) == 0
|
|
&& (l->l_flag & LW_PENDSIG) != 0 && sigispending(l, 0)) {
|
|
if (PUFFSOP_OPCLASS(preq->preq_opclass) == PUFFSOP_VN
|
|
&& preq->preq_optype == PUFFS_VN_INACTIVE) {
|
|
struct puffs_park *newpark;
|
|
|
|
newpark = puffs_reqtofaf(ppark);
|
|
DPRINTF(("puffs touser: converted to FAF, old %p, "
|
|
"new %p\n", ppark, newpark));
|
|
ppark = newpark;
|
|
rv = EINTR;
|
|
} else {
|
|
return EINTR;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* test for suspension lock.
|
|
*
|
|
* Note that we *DO NOT* keep the lock, since that might block
|
|
* lock acquiring PLUS it would give userlandia control over
|
|
* the lock. The operation queue enforces a strict ordering:
|
|
* when the fs server gets in the op stream, it knows things
|
|
* are in order. The kernel locks can't guarantee that for
|
|
* userspace, in any case.
|
|
*
|
|
* BUT: this presents a problem for ops which have a consistency
|
|
* clause based on more than one operation. Unfortunately such
|
|
* operations (read, write) do not reliably work yet.
|
|
*
|
|
* Ya, Ya, it's wrong wong wrong, me be fixink this someday.
|
|
*
|
|
* XXX: and there is one more problem. We sometimes need to
|
|
* take a lazy lock in case the fs is suspending and we are
|
|
* executing as the fs server context. This might happen
|
|
* e.g. in the case that the user server triggers a reclaim
|
|
* in the kernel while the fs is suspending. It's not a very
|
|
* likely event, but it needs to be fixed some day.
|
|
*/
|
|
fstrans_start(mp, FSTRANS_NORMAL);
|
|
simple_lock(&pmp->pmp_lock);
|
|
fstrans_done(mp);
|
|
|
|
if (pmp->pmp_status != PUFFSTAT_RUNNING) {
|
|
simple_unlock(&pmp->pmp_lock);
|
|
return ENXIO;
|
|
}
|
|
|
|
TAILQ_INSERT_TAIL(&pmp->pmp_req_touser, ppark, park_entries);
|
|
pmp->pmp_req_touser_waiters++;
|
|
|
|
/*
|
|
* Don't do unlock-relock dance yet. There are a couple of
|
|
* unsolved issues with it. If we don't unlock, we can have
|
|
* processes wanting vn_lock in case userspace hangs. But
|
|
* that can be "solved" by killing the userspace process. It
|
|
* would of course be nicer to have antilocking in the userspace
|
|
* interface protocol itself.. your patience will be rewarded.
|
|
*/
|
|
#if 0
|
|
/* unlock */
|
|
if (vp2)
|
|
VOP_UNLOCK(vp2, 0);
|
|
if (vp1)
|
|
VOP_UNLOCK(vp1, 0);
|
|
#endif
|
|
|
|
/*
|
|
* XXX: does releasing the lock here cause trouble? Can't hold
|
|
* it, because otherwise the below would cause locking against
|
|
* oneself-problems in the kqueue stuff. yes, it is a
|
|
* theoretical race, so it must be solved
|
|
*/
|
|
simple_unlock(&pmp->pmp_lock);
|
|
|
|
DPRINTF(("touser: req %" PRIu64 ", preq: %p, park: %p, "
|
|
"c/t: 0x%x/0x%x, f: 0x%x\n", preq->preq_id, preq, ppark,
|
|
preq->preq_opclass, preq->preq_optype, ppark->park_flags));
|
|
|
|
wakeup(&pmp->pmp_req_touser);
|
|
selnotify(pmp->pmp_sel, 0);
|
|
|
|
if (PUFFSOP_WANTREPLY(ppark->park_preq->preq_opclass)
|
|
&& (ppark->park_flags & PUFFS_PARKFLAG_ASYNCBIOREAD) == 0) {
|
|
struct puffs_park *valetpark = NULL;
|
|
int error;
|
|
|
|
error = ltsleep(ppark, PUSER | PCATCH, "puffs1", 0, NULL);
|
|
rv = ppark->park_preq->preq_rv;
|
|
|
|
/*
|
|
* Ok, so it gets a bit tricky around here once again.
|
|
* We want to give interruptibility to the sleep to work
|
|
* around all kinds of locking-against-oneself problems
|
|
* and the file system recursing into itself and so forth.
|
|
* So if we break out of the ltsleep() for anything except
|
|
* natural causes, we need to caution ourselves.
|
|
*
|
|
* The stages at which we can break out are:
|
|
* 1) operation waiting to be fetched by file server
|
|
* 2) operation being copied to userspace, not on either queue
|
|
* 3) file server operating on .. err .. operation
|
|
* 4) putop: locate the correct park structure from the queue
|
|
* 5) putop: copy response from userspace
|
|
* 6) putop: wakeup waiter
|
|
*
|
|
* If we are still at stage 1, no problem, just remove
|
|
* ourselves from the queue to userspace. If we are at
|
|
* the stage before 4 has completed, replace the park structure
|
|
* with a park structure indicating that the caller is
|
|
* no more and no proper reply is required. If the server
|
|
* is already copying data from userspace to the kernel,
|
|
* wait for it to finish and return the real return value to
|
|
* the caller.
|
|
*/
|
|
checkagain:
|
|
if (valetpark) {
|
|
FREE(valetpark, M_PUFFS);
|
|
valetpark = NULL;
|
|
}
|
|
|
|
if (error) {
|
|
DPRINTF(("puffs touser: got %d from ltsleep, "
|
|
"(unlocked) flags 0x%x (park %p)\n",
|
|
error, ppark->park_flags, ppark));
|
|
rv = error;
|
|
|
|
MALLOC(valetpark, struct puffs_park *,
|
|
sizeof(struct puffs_park), M_PUFFS,
|
|
M_ZERO | M_WAITOK);
|
|
|
|
simple_lock(&pmp->pmp_lock);
|
|
|
|
/*
|
|
* The order here for the clauses, per description
|
|
* in comment above, is:
|
|
* 1, after 6, after 4, 2-3.
|
|
*/
|
|
if ((ppark->park_flags&PUFFS_PARKFLAG_PROCESSING)==0) {
|
|
TAILQ_REMOVE(&pmp->pmp_req_touser, ppark,
|
|
park_entries);
|
|
simple_unlock(&pmp->pmp_lock);
|
|
FREE(valetpark, M_PUFFS);
|
|
DPRINTF(("puffs touser: park %p removed "
|
|
"from queue one\n", ppark));
|
|
} else if
|
|
(ppark->park_flags & PUFFS_PARKFLAG_RECVREPLY) {
|
|
if (ppark->park_flags & PUFFS_PARKFLAG_DONE) {
|
|
rv = ppark->park_preq->preq_rv;
|
|
simple_unlock(&pmp->pmp_lock);
|
|
FREE(valetpark, M_PUFFS);
|
|
} else {
|
|
error = ltsleep(ppark,
|
|
PUSER | PCATCH | PNORELOCK,
|
|
"puffsre1", 0, &pmp->pmp_lock);
|
|
goto checkagain;
|
|
}
|
|
} else {
|
|
valetpark->park_flags
|
|
= PUFFS_PARKFLAG_WAITERGONE;
|
|
ppark->park_flags |= PUFFS_PARKFLAG_WAITERGONE;
|
|
valetpark->park_id = ppark->park_id;
|
|
|
|
if (ppark->park_flags & PUFFS_PARKFLAG_RQUEUE) {
|
|
TAILQ_INSERT_BEFORE(ppark, valetpark,
|
|
park_entries);
|
|
TAILQ_REMOVE(&pmp->pmp_req_replywait,
|
|
ppark, park_entries);
|
|
} else {
|
|
TAILQ_INSERT_TAIL(
|
|
&pmp->pmp_req_replywait,
|
|
valetpark, park_entries);
|
|
}
|
|
|
|
simple_unlock(&pmp->pmp_lock);
|
|
DPRINTF(("puffs touser: replaced park %p "
|
|
"with valet park %p\n", ppark, valetpark));
|
|
}
|
|
}
|
|
|
|
/*
|
|
* retake the lock and release. This makes sure (haha,
|
|
* I'm humorous) that we don't process the same vnode in
|
|
* multiple threads due to the locks hacks we have in
|
|
* puffs_lock(). In reality this is well protected by
|
|
* the biglock, but once that's gone, well, hopefully
|
|
* this will be fixed for real. (and when you read this
|
|
* comment in 2017 and subsequently barf, my condolences ;).
|
|
*/
|
|
if (rv == 0 && !fstrans_is_owner(mp)) {
|
|
fstrans_start(mp, FSTRANS_NORMAL);
|
|
fstrans_done(mp);
|
|
}
|
|
}
|
|
|
|
#if 0
|
|
/* relock */
|
|
if (vp1)
|
|
KASSERT(vn_lock(vp1, LK_EXCLUSIVE | LK_RETRY) == 0);
|
|
if (vp2)
|
|
KASSERT(vn_lock(vp2, LK_EXCLUSIVE | LK_RETRY) == 0);
|
|
#endif
|
|
|
|
simple_lock(&pmp->pmp_lock);
|
|
if (--pmp->pmp_req_touser_waiters == 0)
|
|
wakeup(&pmp->pmp_req_touser_waiters);
|
|
simple_unlock(&pmp->pmp_lock);
|
|
|
|
return rv;
|
|
}
|
|
|
|
|
|
/*
|
|
* getop: scan through queued requests until:
|
|
* 1) max number of requests satisfied
|
|
* OR
|
|
* 2) buffer runs out of space
|
|
* OR
|
|
* 3) nonblocking is set AND there are no operations available
|
|
* OR
|
|
* 4) at least one operation was transferred AND there are no more waiting
|
|
*/
|
|
int
|
|
puffs_getop(struct puffs_mount *pmp, struct puffs_reqh_get *phg, int nonblock)
|
|
{
|
|
struct puffs_park *park;
|
|
struct puffs_req *preq;
|
|
uint8_t *bufpos;
|
|
int error, donesome;
|
|
|
|
donesome = error = 0;
|
|
bufpos = phg->phg_buf;
|
|
|
|
simple_lock(&pmp->pmp_lock);
|
|
while (phg->phg_nops == 0 || donesome != phg->phg_nops) {
|
|
again:
|
|
if (pmp->pmp_status != PUFFSTAT_RUNNING) {
|
|
/* if we got some, they don't really matter anymore */
|
|
error = ENXIO;
|
|
goto out;
|
|
}
|
|
if (TAILQ_EMPTY(&pmp->pmp_req_touser)) {
|
|
if (donesome)
|
|
goto out;
|
|
|
|
if (nonblock) {
|
|
error = EWOULDBLOCK;
|
|
goto out;
|
|
}
|
|
|
|
error = ltsleep(&pmp->pmp_req_touser, PUSER | PCATCH,
|
|
"puffs2", 0, &pmp->pmp_lock);
|
|
if (error)
|
|
goto out;
|
|
else
|
|
goto again;
|
|
}
|
|
|
|
park = TAILQ_FIRST(&pmp->pmp_req_touser);
|
|
preq = park->park_preq;
|
|
|
|
if (phg->phg_buflen < preq->preq_buflen) {
|
|
if (!donesome)
|
|
error = E2BIG;
|
|
goto out;
|
|
}
|
|
TAILQ_REMOVE(&pmp->pmp_req_touser, park, park_entries);
|
|
park->park_flags |= PUFFS_PARKFLAG_PROCESSING;
|
|
simple_unlock(&pmp->pmp_lock);
|
|
|
|
DPRINTF(("puffsgetop: get op %" PRIu64 " (%d.), from %p "
|
|
"len %zu (buflen %zu), target %p\n", preq->preq_id,
|
|
donesome, preq, park->park_copylen, preq->preq_buflen,
|
|
bufpos));
|
|
|
|
if ((error = copyout(preq, bufpos, park->park_copylen)) != 0) {
|
|
DPRINTF((" FAILED %d\n", error));
|
|
/*
|
|
* ok, user server is probably trying to cheat.
|
|
* stuff op back & return error to user
|
|
*/
|
|
simple_lock(&pmp->pmp_lock);
|
|
TAILQ_INSERT_HEAD(&pmp->pmp_req_touser, park,
|
|
park_entries);
|
|
|
|
if (donesome)
|
|
error = 0;
|
|
goto out;
|
|
}
|
|
bufpos += preq->preq_buflen;
|
|
phg->phg_buflen -= preq->preq_buflen;
|
|
donesome++;
|
|
|
|
simple_lock(&pmp->pmp_lock);
|
|
if (PUFFSOP_WANTREPLY(preq->preq_opclass)) {
|
|
if ((park->park_flags & PUFFS_PARKFLAG_WAITERGONE)==0) {
|
|
TAILQ_INSERT_TAIL(&pmp->pmp_req_replywait, park,
|
|
park_entries);
|
|
park->park_flags |= PUFFS_PARKFLAG_RQUEUE;
|
|
}
|
|
} else {
|
|
simple_unlock(&pmp->pmp_lock);
|
|
free(preq, M_PUFFS);
|
|
free(park, M_PUFFS);
|
|
simple_lock(&pmp->pmp_lock);
|
|
}
|
|
}
|
|
|
|
out:
|
|
phg->phg_more = pmp->pmp_req_touser_waiters;
|
|
simple_unlock(&pmp->pmp_lock);
|
|
|
|
phg->phg_nops = donesome;
|
|
|
|
return error;
|
|
}
|
|
|
|
/*
|
|
* urgh, too complex, be very very careful while editing, i.e. NEEDS CLEANUP
|
|
*/
|
|
int
|
|
puffs_putop(struct puffs_mount *pmp, struct puffs_reqh_put *php)
|
|
{
|
|
struct puffs_park *park;
|
|
struct puffs_req tmpreq;
|
|
struct puffs_req *nextpreq;
|
|
struct buf *bp;
|
|
void *userbuf;
|
|
uint64_t id;
|
|
size_t reqlen;
|
|
int donesome, error, wgone;
|
|
|
|
donesome = error = wgone = 0;
|
|
|
|
id = php->php_id;
|
|
userbuf = php->php_buf;
|
|
reqlen = php->php_buflen;
|
|
|
|
simple_lock(&pmp->pmp_lock);
|
|
while (donesome != php->php_nops) {
|
|
#ifdef PUFFSDEBUG
|
|
simple_unlock(&pmp->pmp_lock);
|
|
DPRINTF(("puffsputop: searching for %" PRIu64 ", ubuf: %p, "
|
|
"len %zu\n", id, userbuf, reqlen));
|
|
simple_lock(&pmp->pmp_lock);
|
|
#endif
|
|
TAILQ_FOREACH(park, &pmp->pmp_req_replywait, park_entries) {
|
|
if (park->park_id == id)
|
|
break;
|
|
}
|
|
|
|
if (park == NULL) {
|
|
error = EINVAL;
|
|
break;
|
|
}
|
|
TAILQ_REMOVE(&pmp->pmp_req_replywait, park, park_entries);
|
|
park->park_flags |= PUFFS_PARKFLAG_RECVREPLY;
|
|
simple_unlock(&pmp->pmp_lock);
|
|
|
|
/*
|
|
* If the caller has gone south, go to next, collect
|
|
* $200 and free the structure there instead of wakeup.
|
|
* We also need to copyin the
|
|
*/
|
|
if (park->park_flags & PUFFS_PARKFLAG_WAITERGONE) {
|
|
DPRINTF(("puffs_putop: bad service - waiter gone for "
|
|
"park %p\n", park));
|
|
wgone = 1;
|
|
error = copyin(userbuf, &tmpreq,
|
|
sizeof(struct puffs_req));
|
|
if (error)
|
|
goto loopout;
|
|
nextpreq = &tmpreq;
|
|
goto next;
|
|
}
|
|
|
|
if (park->park_flags & PUFFS_PARKFLAG_ADJUSTABLE) {
|
|
/* sanitycheck size of incoming transmission. */
|
|
if (reqlen > pmp->pmp_req_maxsize) {
|
|
DPRINTF(("puffsputop: outrageous user buf "
|
|
"size: %zu\n", reqlen));
|
|
error = EINVAL;
|
|
goto loopout;
|
|
}
|
|
|
|
if (reqlen > park->park_copylen) {
|
|
if (reqlen > park->park_maxlen) {
|
|
DPRINTF(("puffsputop: adj copysize "
|
|
"> max size, %zu vs %zu\n",
|
|
reqlen, park->park_maxlen));
|
|
error = EINVAL;
|
|
goto loopout;
|
|
}
|
|
free(park->park_preq, M_PUFFS);
|
|
park->park_preq = malloc(reqlen,
|
|
M_PUFFS, M_WAITOK);
|
|
|
|
park->park_copylen = reqlen;
|
|
DPRINTF(("puffsputop: adjbuf, new addr %p, "
|
|
"len %zu\n", park->park_preq, reqlen));
|
|
}
|
|
} else {
|
|
if (reqlen == 0 || reqlen > park->park_copylen) {
|
|
reqlen = park->park_copylen;
|
|
DPRINTF(("puffsputop: kernel bufsize override: "
|
|
"%zu\n", reqlen));
|
|
}
|
|
}
|
|
|
|
DPRINTF(("puffsputpop: copyin from %p to %p, len %zu\n",
|
|
userbuf, park->park_preq, reqlen));
|
|
error = copyin(userbuf, park->park_preq, reqlen);
|
|
if (error)
|
|
goto loopout;
|
|
nextpreq = park->park_preq;
|
|
bp = park->park_bp;
|
|
|
|
if (park->park_flags & PUFFS_PARKFLAG_ASYNCBIOREAD) {
|
|
struct puffs_vnreq_read *read_argp;
|
|
size_t moved;
|
|
|
|
bp->b_error = park->park_preq->preq_rv;
|
|
|
|
DPRINTF(("puffs_putop: async bioread for park %p, "
|
|
"bp %p, error %d\n", park, bp, bp->b_error));
|
|
|
|
if (bp->b_error == 0) {
|
|
read_argp = (void *)park->park_preq;
|
|
moved = park->park_maxlen
|
|
- sizeof(struct puffs_vnreq_read)
|
|
- read_argp->pvnr_resid;
|
|
memcpy(bp->b_data, read_argp->pvnr_data, moved);
|
|
bp->b_resid = bp->b_bcount - moved;
|
|
biodone(bp);
|
|
}
|
|
}
|
|
|
|
next:
|
|
/* all's well, prepare for next op */
|
|
id = nextpreq->preq_id;
|
|
reqlen = nextpreq->preq_buflen;
|
|
userbuf = nextpreq->preq_nextbuf;
|
|
donesome++;
|
|
|
|
loopout:
|
|
if (error && park->park_preq) {
|
|
park->park_preq->preq_rv = error;
|
|
if (park->park_flags & PUFFS_PARKFLAG_ASYNCBIOREAD) {
|
|
bp = park->park_bp;
|
|
bp->b_error = error;
|
|
bp->b_flags |= B_ERROR;
|
|
biodone(bp);
|
|
}
|
|
}
|
|
|
|
if (wgone) {
|
|
FREE(park, M_PUFFS);
|
|
simple_lock(&pmp->pmp_lock);
|
|
} else if (park->park_flags & PUFFS_PARKFLAG_ASYNCBIOREAD) {
|
|
free(park->park_preq, M_PUFFS);
|
|
FREE(park, M_PUFFS);
|
|
simple_lock(&pmp->pmp_lock);
|
|
} else {
|
|
DPRINTF(("puffs_putop: flagging done for park %p\n",
|
|
park));
|
|
simple_lock(&pmp->pmp_lock);
|
|
park->park_flags |= PUFFS_PARKFLAG_DONE;
|
|
wakeup(park);
|
|
}
|
|
|
|
if (error)
|
|
break;
|
|
wgone = 0;
|
|
}
|
|
|
|
simple_unlock(&pmp->pmp_lock);
|
|
php->php_nops -= donesome;
|
|
|
|
return error;
|
|
}
|
|
|
|
/* this is probably going to die away at some point? */
|
|
/*
|
|
* XXX: currently bitrotted
|
|
*/
|
|
#if 0
|
|
static int
|
|
puffssizeop(struct puffs_mount *pmp, struct puffs_sizeop *psop_user)
|
|
{
|
|
struct puffs_sizepark *pspark;
|
|
void *kernbuf;
|
|
size_t copylen;
|
|
int error;
|
|
|
|
/* locate correct op */
|
|
simple_lock(&pmp->pmp_lock);
|
|
TAILQ_FOREACH(pspark, &pmp->pmp_req_sizepark, pkso_entries) {
|
|
if (pspark->pkso_reqid == psop_user->pso_reqid) {
|
|
TAILQ_REMOVE(&pmp->pmp_req_sizepark, pspark,
|
|
pkso_entries);
|
|
break;
|
|
}
|
|
}
|
|
simple_unlock(&pmp->pmp_lock);
|
|
|
|
if (pspark == NULL)
|
|
return EINVAL;
|
|
|
|
error = 0;
|
|
copylen = MIN(pspark->pkso_bufsize, psop_user->pso_bufsize);
|
|
|
|
/*
|
|
* XXX: uvm stuff to avoid bouncy-bouncy copying?
|
|
*/
|
|
if (PUFFS_SIZEOP_UIO(pspark->pkso_reqtype)) {
|
|
kernbuf = malloc(copylen, M_PUFFS, M_WAITOK | M_ZERO);
|
|
if (pspark->pkso_reqtype == PUFFS_SIZEOPREQ_UIO_IN) {
|
|
error = copyin(psop_user->pso_userbuf,
|
|
kernbuf, copylen);
|
|
if (error) {
|
|
printf("psop ERROR1 %d\n", error);
|
|
goto escape;
|
|
}
|
|
}
|
|
error = uiomove(kernbuf, copylen, pspark->pkso_uio);
|
|
if (error) {
|
|
printf("uiomove from kernel %p, len %d failed: %d\n",
|
|
kernbuf, (int)copylen, error);
|
|
goto escape;
|
|
}
|
|
|
|
if (pspark->pkso_reqtype == PUFFS_SIZEOPREQ_UIO_OUT) {
|
|
error = copyout(kernbuf,
|
|
psop_user->pso_userbuf, copylen);
|
|
if (error) {
|
|
printf("psop ERROR2 %d\n", error);
|
|
goto escape;
|
|
}
|
|
}
|
|
escape:
|
|
free(kernbuf, M_PUFFS);
|
|
} else if (PUFFS_SIZEOP_BUF(pspark->pkso_reqtype)) {
|
|
copylen = MAX(pspark->pkso_bufsize, psop_user->pso_bufsize);
|
|
if (pspark->pkso_reqtype == PUFFS_SIZEOPREQ_BUF_IN) {
|
|
error = copyin(psop_user->pso_userbuf,
|
|
pspark->pkso_copybuf, copylen);
|
|
} else {
|
|
error = copyout(pspark->pkso_copybuf,
|
|
psop_user->pso_userbuf, copylen);
|
|
}
|
|
}
|
|
#ifdef DIAGNOSTIC
|
|
else
|
|
panic("puffssizeop: invalid reqtype %d\n",
|
|
pspark->pkso_reqtype);
|
|
#endif /* DIAGNOSTIC */
|
|
|
|
return error;
|
|
}
|
|
#endif
|