e90eb67076
* replace some b_flags magic with BUF_ISREAD/WRITE()
359 lines
8.6 KiB
C
359 lines
8.6 KiB
C
/* $NetBSD: specfs.c,v 1.27 2008/09/11 13:42:35 pooka Exp $ */
|
|
|
|
/*
|
|
* Copyright (c) 2007 Antti Kantee. All Rights Reserved.
|
|
*
|
|
* Development of this software was supported by Google Summer of Code.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS
|
|
* OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
|
* WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
* DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
|
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/types.h>
|
|
#include <sys/vnode.h>
|
|
#include <sys/vnode_if.h>
|
|
#include <sys/fcntl.h>
|
|
#include <sys/disklabel.h>
|
|
#include <sys/stat.h>
|
|
|
|
#include <miscfs/genfs/genfs.h>
|
|
#include <miscfs/specfs/specdev.h>
|
|
|
|
#include <uvm/uvm_extern.h>
|
|
|
|
#include <rump/rumpuser.h>
|
|
|
|
#include "rump_private.h"
|
|
|
|
/* We have special special ops */
|
|
static int rump_specopen(void *);
|
|
static int rump_specioctl(void *);
|
|
static int rump_specclose(void *);
|
|
static int rump_specfsync(void *);
|
|
static int rump_specbmap(void *);
|
|
static int rump_specputpages(void *);
|
|
static int rump_specstrategy(void *);
|
|
static int rump_specsimpleul(void *);
|
|
|
|
kmutex_t specfs_lock;
|
|
|
|
int (**spec_vnodeop_p)(void *);
|
|
const struct vnodeopv_entry_desc rumpspec_vnodeop_entries[] = {
|
|
{ &vop_default_desc, vn_default_error },
|
|
{ &vop_bwrite_desc, vn_bwrite }, /* bwrite */
|
|
{ &vop_lock_desc, genfs_lock }, /* lock */
|
|
{ &vop_unlock_desc, genfs_unlock }, /* unlock */
|
|
{ &vop_open_desc, rump_specopen }, /* open */
|
|
{ &vop_close_desc, rump_specclose }, /* close */
|
|
{ &vop_ioctl_desc, rump_specioctl }, /* ioctl */
|
|
{ &vop_fsync_desc, rump_specfsync }, /* fsync */
|
|
{ &vop_bmap_desc, rump_specbmap }, /* bmap */
|
|
{ &vop_putpages_desc, rump_specputpages }, /* putpages */
|
|
{ &vop_strategy_desc, rump_specstrategy }, /* strategy */
|
|
{ &vop_getpages_desc, rump_specsimpleul }, /* getpages */
|
|
{ &vop_putpages_desc, rump_specsimpleul }, /* putpages */
|
|
{ NULL, NULL }
|
|
};
|
|
const struct vnodeopv_desc spec_vnodeop_opv_desc =
|
|
{ &spec_vnodeop_p, rumpspec_vnodeop_entries };
|
|
|
|
static int
|
|
rump_specopen(void *v)
|
|
{
|
|
struct vop_open_args /* {
|
|
struct vnode *a_vp;
|
|
int a_mode;
|
|
kauth_cred_t a_cred;
|
|
struct lwp *a_l;
|
|
} */ *ap = v;
|
|
struct vnode *vp = ap->a_vp;
|
|
struct rump_specpriv *sp = vp->v_data;
|
|
struct stat sb;
|
|
int fd, error;
|
|
|
|
fd = rumpuser_open(sp->rsp_path, OFLAGS(ap->a_mode), &error);
|
|
if (fd == -1)
|
|
return error;
|
|
|
|
/* XXX uh */
|
|
if (rumpuser_ioctl(fd, DIOCGDINFO, &sp->rsp_dl, &error) == -1) {
|
|
memset(&sp->rsp_dl, 0, sizeof(sp->rsp_dl));
|
|
|
|
if (rumpuser_stat(sp->rsp_path, &sb, &error) == -1) {
|
|
int dummy;
|
|
|
|
rumpuser_close(fd, &dummy);
|
|
return error;
|
|
}
|
|
sp->rsp_pi.p_size = sb.st_size >> DEV_BSHIFT;
|
|
sp->rsp_dl.d_secsize = DEV_BSIZE;
|
|
sp->rsp_curpi = &sp->rsp_pi;
|
|
} else {
|
|
sp->rsp_curpi = &sp->rsp_dl.d_partitions[0]; /* XXX */
|
|
}
|
|
|
|
sp->rsp_fd = fd;
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
rump_specioctl(void *v)
|
|
{
|
|
struct vop_ioctl_args /* {
|
|
struct vnode *a_vp;
|
|
u_long a_command;
|
|
void *a_data;
|
|
int a_fflag;
|
|
kauth_cred_t a_cred;
|
|
struct lwp *a_l;
|
|
} */ *ap = v;
|
|
struct vnode *vp = ap->a_vp;
|
|
struct rump_specpriv *sp = vp->v_data;
|
|
int rv, error;
|
|
|
|
if (ap->a_command == DIOCGPART) {
|
|
struct partinfo *pi = (struct partinfo *)ap->a_data;
|
|
|
|
pi->part = sp->rsp_curpi;
|
|
pi->disklab = &sp->rsp_dl;
|
|
|
|
return 0;
|
|
}
|
|
|
|
rv = rumpuser_ioctl(sp->rsp_fd, ap->a_command, ap->a_data, &error);
|
|
if (rv == -1)
|
|
return error;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
rump_specclose(void *v)
|
|
{
|
|
struct vop_close_args /* {
|
|
struct vnode *a_vp;
|
|
int a_fflag;
|
|
kauth_cred_t a_cred;
|
|
struct lwp *a_l;
|
|
} */ *ap = v;
|
|
struct vnode *vp = ap->a_vp;
|
|
struct rump_specpriv *sp = vp->v_data;
|
|
int error;
|
|
|
|
rumpuser_close(sp->rsp_fd, &error);
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
rump_specfsync(void *v)
|
|
{
|
|
struct vop_fsync_args /* {
|
|
struct vnode *a_vp;
|
|
kauth_cred_t a_cred;
|
|
int a_flags;
|
|
off_t a_offlo;
|
|
off_t a_offhi;
|
|
struct lwp *a_l;
|
|
} */ *ap = v;
|
|
struct vnode *vp = ap->a_vp;
|
|
struct mount *mp;
|
|
int error;
|
|
|
|
KASSERT(vp->v_type == VBLK);
|
|
if ((mp = vp->v_specmountpoint) != NULL) {
|
|
error = VFS_FSYNC(mp, vp, ap->a_flags | FSYNC_VFS);
|
|
if (error != EOPNOTSUPP)
|
|
return error;
|
|
}
|
|
vflushbuf(vp, (ap->a_flags & FSYNC_WAIT) != 0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
rump_specputpages(void *v)
|
|
{
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
rump_specbmap(void *v)
|
|
{
|
|
struct vop_bmap_args /* {
|
|
struct vnode *a_vp;
|
|
daddr_t a_bn;
|
|
struct vnode **a_vpp;
|
|
daddr_t *a_bnp;
|
|
int *a_runp;
|
|
} */ *ap = v;
|
|
|
|
if (ap->a_vpp != NULL)
|
|
*ap->a_vpp = ap->a_vp;
|
|
if (ap->a_bnp != NULL)
|
|
*ap->a_bnp = ap->a_bn;
|
|
if (ap->a_runp != NULL)
|
|
*ap->a_runp = (MAXBSIZE >> DEV_BSHIFT) -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
rump_specstrategy(void *v)
|
|
{
|
|
struct vop_strategy_args /* {
|
|
struct vnode *a_vp;
|
|
struct buf *a_bp;
|
|
} */ *ap = v;
|
|
struct vnode *vp = ap->a_vp;
|
|
struct buf *bp = ap->a_bp;
|
|
struct rump_specpriv *sp;
|
|
int async;
|
|
off_t off;
|
|
|
|
KASSERT(vp->v_type == VBLK);
|
|
sp = vp->v_data;
|
|
|
|
off = bp->b_blkno << DEV_BSHIFT;
|
|
DPRINTF(("specstrategy: 0x%x bytes %s off 0x%" PRIx64
|
|
" (0x%" PRIx64 " - 0x%" PRIx64")\n",
|
|
bp->b_bcount, BUF_ISREAD(bp) "READ" : "WRITE",
|
|
off, off, (off + bp->b_bcount)));
|
|
|
|
/*
|
|
* Do I/O. We have different paths for async and sync I/O.
|
|
* Async I/O is done by passing a request to rumpuser where
|
|
* it is executed. The rumpuser routine then calls
|
|
* biodone() to signal any waiters in the kernel. I/O's are
|
|
* executed in series. Technically executing them in parallel
|
|
* would produce better results, but then we'd need either
|
|
* more threads or posix aio. Maybe worth investigating
|
|
* this later.
|
|
*
|
|
* Synchronous I/O is done directly in the context mainly to
|
|
* avoid unnecessary scheduling with the I/O thread.
|
|
*/
|
|
async = bp->b_flags & B_ASYNC;
|
|
if (async && rump_threads) {
|
|
struct rumpuser_aio *rua;
|
|
|
|
rua = kmem_alloc(sizeof(struct rumpuser_aio), KM_SLEEP);
|
|
rua->rua_fd = sp->rsp_fd;
|
|
rua->rua_data = bp->b_data;
|
|
rua->rua_dlen = bp->b_bcount;
|
|
rua->rua_off = off;
|
|
rua->rua_bp = bp;
|
|
rua->rua_op = BUF_ISREAD(bp);
|
|
|
|
rumpuser_mutex_enter(&rua_mtx);
|
|
|
|
/*
|
|
* Check if our buffer is full. Doing it this way
|
|
* throttles the I/O a bit if we have a massive
|
|
* async I/O burst.
|
|
*
|
|
* XXX: this actually leads to deadlocks with spl()
|
|
* (caller maybe be at splbio() legally for async I/O),
|
|
* so for now set N_AIOS high and FIXXXME some day.
|
|
*/
|
|
if ((rua_head+1) % N_AIOS == rua_tail) {
|
|
kmem_free(rua, sizeof(*rua));
|
|
rumpuser_mutex_exit(&rua_mtx);
|
|
goto syncfallback;
|
|
}
|
|
|
|
/* insert into queue & signal */
|
|
rua_aios[rua_head] = rua;
|
|
rua_head = (rua_head+1) % (N_AIOS-1);
|
|
rumpuser_cv_signal(&rua_cv);
|
|
rumpuser_mutex_exit(&rua_mtx);
|
|
} else {
|
|
syncfallback:
|
|
if (BUF_ISREAD(bp)) {
|
|
rumpuser_read_bio(sp->rsp_fd, bp->b_data,
|
|
bp->b_bcount, off, bp);
|
|
} else {
|
|
rumpuser_write_bio(sp->rsp_fd, bp->b_data,
|
|
bp->b_bcount, off, bp);
|
|
}
|
|
if (!async) {
|
|
int error;
|
|
|
|
if (BUF_ISWRITE(bp))
|
|
rumpuser_fsync(sp->rsp_fd, &error);
|
|
biowait(bp);
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
rump_specsimpleul(void *v)
|
|
{
|
|
struct vop_generic_args *ap = v;
|
|
struct vnode *vp;
|
|
int offset;
|
|
|
|
offset = ap->a_desc->vdesc_vp_offsets[0];
|
|
KASSERT(offset != VDESC_NO_OFFSET);
|
|
|
|
vp = *VOPARG_OFFSETTO(struct vnode **, offset, ap);
|
|
mutex_exit(&vp->v_interlock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void
|
|
spec_node_init(struct vnode *nvp, dev_t nvp_rdev)
|
|
{
|
|
specdev_t *sd;
|
|
|
|
sd = kmem_zalloc(sizeof(specdev_t), KM_SLEEP);
|
|
sd->sd_rdev = nvp_rdev;
|
|
sd->sd_refcnt = 1;
|
|
nvp->v_specnode = kmem_alloc(sizeof(specnode_t), KM_SLEEP);
|
|
nvp->v_specnode->sn_dev = sd;
|
|
nvp->v_rdev = nvp_rdev;
|
|
}
|
|
|
|
void
|
|
spec_node_destroy(vnode_t *vp)
|
|
{
|
|
specnode_t *sn;
|
|
specdev_t *sd;
|
|
|
|
sn = vp->v_specnode;
|
|
sd = sn->sn_dev;
|
|
|
|
KASSERT(sd->sd_refcnt == 1);
|
|
kmem_free(sd, sizeof(*sd));
|
|
kmem_free(sn, sizeof(*sn));
|
|
}
|
|
|
|
void
|
|
spec_node_revoke(vnode_t *vp)
|
|
{
|
|
|
|
panic("spec_node_revoke: should not be called");
|
|
}
|