/* $NetBSD: dmover_io.c,v 1.2 2002/08/02 00:31:35 thorpej Exp $ */ /* * Copyright (c) 2002 Wasabi Systems, Inc. * All rights reserved. * * Written by Jason R. Thorpe for Wasabi Systems, Inc. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software * must display the following acknowledgement: * This product includes software developed for the NetBSD Project by * Wasabi Systems, Inc. * 4. The name of Wasabi Systems, Inc. may not be used to endorse * or promote products derived from this software without specific prior * written permission. * * THIS SOFTWARE IS PROVIDED BY WASABI SYSTEMS, INC. ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL WASABI SYSTEMS, INC * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. */ /* * dmover_io.c: Support for user-space access to dmover-api * * This interface is quite simple: * * 1. The user opens /dev/dmover, which is a cloning device. This * allocates internal state for the session. * * 2. The user does a DMIO_SETFUNC to select the data movement * function. This actually creates the dmover session. * * 3. The user writes request messages to its dmover handle. * * 4. The user reads request responses from its dmover handle. * * 5. The user closes the file descriptor and the session is * torn down. */ #include __KERNEL_RCSID(0, "$NetBSD: dmover_io.c,v 1.2 2002/08/02 00:31:35 thorpej Exp $"); #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include struct dmio_usrreq_state { TAILQ_ENTRY(dmio_usrreq_state) dus_q; struct uio dus_uio_out; struct uio *dus_uio_in; struct dmover_request *dus_req; uint32_t dus_id; }; struct dmio_state { struct dmover_session *ds_session; TAILQ_HEAD(, dmio_usrreq_state) ds_pending; TAILQ_HEAD(, dmio_usrreq_state) ds_complete; struct selinfo ds_selq; __volatile int ds_flags; u_int ds_nreqs; struct simplelock ds_slock; }; #define DMIO_STATE_SEL 0x0001 #define DMIO_STATE_DEAD 0x0002 #define DMIO_STATE_LARVAL 0x0004 #define DMIO_STATE_READ_WAIT 0x0008 #define DMIO_STATE_WRITE_WAIT 0x0010 #define DMIO_NREQS_MAX 64 /* XXX pulled out of a hat */ struct pool dmio_state_pool; struct pool dmio_usrreq_state_pool; void dmoverioattach(int); cdev_decl(dmoverio); /* * dmoverioattach: * * Pseudo-device attach routine. */ void dmoverioattach(int count) { pool_init(&dmio_state_pool, sizeof(struct dmio_state), 0, 0, 0, "dmiostate", NULL); pool_init(&dmio_usrreq_state_pool, sizeof(struct dmio_usrreq_state), 0, 0, 0, "dmiourstate", NULL); } /* * dmio_usrreq_init: * * Build a request structure. */ static int dmio_usrreq_init(struct file *fp, struct dmio_usrreq_state *dus, struct dmio_usrreq *req, struct dmover_request *dreq) { struct dmio_state *ds = (struct dmio_state *) fp->f_data; struct dmover_session *dses = ds->ds_session; struct uio *uio_out = &dus->dus_uio_out; struct uio *uio_in; dmio_buffer inbuf; size_t len; u_int i, j; int error; /* XXX How should malloc interact w/ FNONBLOCK? */ if (req->req_outbuf.dmbuf_iovcnt > IOV_MAX) return (EINVAL); len = sizeof(struct iovec) * req->req_outbuf.dmbuf_iovcnt; if (len == 0) return (EINVAL); uio_out->uio_iov = malloc(len, M_TEMP, M_WAITOK); error = copyin(req->req_outbuf.dmbuf_iov, uio_out->uio_iov, len); if (error) { free(uio_out->uio_iov, M_TEMP); return (error); } for (j = 0, len = 0; j < req->req_outbuf.dmbuf_iovcnt; j++) { len += uio_out->uio_iov[j].iov_len; if (len > SSIZE_MAX) { free(uio_out->uio_iov, M_TEMP); return (error); } } uio_out->uio_iovcnt = req->req_outbuf.dmbuf_iovcnt; uio_out->uio_resid = len; uio_out->uio_rw = UIO_READ; uio_out->uio_segflg = UIO_USERSPACE; uio_out->uio_procp = curproc; dreq->dreq_outbuf_type = DMOVER_BUF_UIO; dreq->dreq_outbuf.dmbuf_uio = uio_out; if (dses->dses_ninputs == 0) { /* No inputs; copy the immediate. */ memcpy(dreq->dreq_immediate, req->req_immediate, sizeof(dreq->dreq_immediate)); return (0); } dreq->dreq_inbuf_type = DMOVER_BUF_UIO; dus->dus_uio_in = malloc(sizeof(struct uio) * dses->dses_ninputs, M_TEMP, M_WAITOK); memset(dus->dus_uio_in, 0, sizeof(struct uio) * dses->dses_ninputs); for (i = 0; i < dses->dses_ninputs; i++) { uio_in = &dus->dus_uio_in[i]; error = copyin(&req->req_inbuf[i], &inbuf, sizeof(inbuf)); if (error) goto bad; if (inbuf.dmbuf_iovcnt > IOV_MAX) { error = EINVAL; goto bad; } len = sizeof(struct iovec) * inbuf.dmbuf_iovcnt; if (len == 0) { error = EINVAL; goto bad; } uio_in->uio_iov = malloc(len, M_TEMP, M_WAITOK); error = copyin(inbuf.dmbuf_iov, uio_in->uio_iov, len); if (error) { free(uio_in->uio_iov, M_TEMP); goto bad; } for (j = 0, len = 0; j < req->req_outbuf.dmbuf_iovcnt; j++) { len += uio_in->uio_iov[j].iov_len; if (len > SSIZE_MAX) { free(uio_in->uio_iov, M_TEMP); error = EINVAL; goto bad; } } if (len != uio_out->uio_resid) { free(uio_in->uio_iov, M_TEMP); error = EINVAL; goto bad; } uio_in->uio_iovcnt = inbuf.dmbuf_iovcnt; uio_in->uio_resid = len; uio_in->uio_rw = UIO_WRITE; uio_in->uio_segflg = UIO_USERSPACE; uio_in->uio_procp = curproc; dreq->dreq_inbuf[i].dmbuf_uio = uio_in; } return (0); bad: if (i > 0) { for (--i; i >= 0; i--) { uio_in = &dus->dus_uio_in[i]; free(uio_in->uio_iov, M_TEMP); } } free(dus->dus_uio_in, M_TEMP); free(uio_out->uio_iov, M_TEMP); return (error); } /* * dmio_usrreq_fini: * * Tear down a request. Must be called at splsoftclock(). */ static void dmio_usrreq_fini(struct dmio_state *ds, struct dmio_usrreq_state *dus) { struct dmover_session *dses = ds->ds_session; struct uio *uio_out = &dus->dus_uio_out; struct uio *uio_in; int i; free(uio_out->uio_iov, M_TEMP); if (dses->dses_ninputs == 0) { pool_put(&dmio_usrreq_state_pool, dus); return; } for (i = 0; i < dses->dses_ninputs; i++) { uio_in = &dus->dus_uio_in[i]; free(uio_in->uio_iov, M_TEMP); } free(dus->dus_uio_in, M_TEMP); pool_put(&dmio_usrreq_state_pool, dus); } /* * dmio_read: * * Read file op. */ static int dmio_read(struct file *fp, off_t *offp, struct uio *uio, struct ucred *cred, int flags) { struct dmio_state *ds = (struct dmio_state *) fp->f_data; struct dmio_usrreq_state *dus; struct dmover_request *dreq; struct dmio_usrresp resp; int s, error = 0, progress = 0; if ((uio->uio_resid % sizeof(resp)) != 0) return (EINVAL); if (ds->ds_session == NULL) return (ENXIO); s = splsoftclock(); simple_lock(&ds->ds_slock); while (uio->uio_resid != 0) { for (;;) { dus = TAILQ_FIRST(&ds->ds_complete); if (dus == NULL) { if (fp->f_flag & FNONBLOCK) { error = progress ? 0 : EWOULDBLOCK; goto out; } error = ltsleep(&ds->ds_complete, PRIBIO | PCATCH, "dmvrrd", 0, &ds->ds_slock); if (error) goto out; continue; } /* Have a completed request. */ TAILQ_REMOVE(&ds->ds_complete, dus, dus_q); ds->ds_nreqs--; if (ds->ds_flags & DMIO_STATE_WRITE_WAIT) { ds->ds_flags &= ~DMIO_STATE_WRITE_WAIT; wakeup(&ds->ds_nreqs); } if (ds->ds_flags & DMIO_STATE_SEL) { ds->ds_flags &= ~DMIO_STATE_SEL; selwakeup(&ds->ds_selq); } break; } simple_unlock(&ds->ds_slock); dreq = dus->dus_req; resp.resp_id = dus->dus_id; resp.resp_error = (dreq->dreq_flags & DMOVER_REQ_ERROR) ? dreq->dreq_error : 0; dmio_usrreq_fini(ds, dus); splx(s); progress = 1; dmover_request_free(dreq); error = uiomove(&resp, sizeof(resp), uio); if (error) return (error); s = splsoftclock(); simple_lock(&ds->ds_slock); } out: simple_unlock(&ds->ds_slock); splx(s); return (error); } /* * dmio_usrreq_done: * * Dmover completion callback. */ static void dmio_usrreq_done(struct dmover_request *dreq) { struct dmio_usrreq_state *dus = dreq->dreq_cookie; struct dmio_state *ds = dreq->dreq_session->dses_cookie; /* We're already at splsoftclock(). */ simple_lock(&ds->ds_slock); TAILQ_REMOVE(&ds->ds_pending, dus, dus_q); if (ds->ds_flags & DMIO_STATE_DEAD) { ds->ds_nreqs--; dmio_usrreq_fini(ds, dus); dmover_request_free(dreq); if (ds->ds_nreqs == 0) { simple_unlock(&ds->ds_slock); pool_put(&dmio_state_pool, ds); return; } } else { TAILQ_INSERT_TAIL(&ds->ds_complete, dus, dus_q); if (ds->ds_flags & DMIO_STATE_READ_WAIT) { ds->ds_flags &= ~DMIO_STATE_READ_WAIT; wakeup(&ds->ds_complete); } if (ds->ds_flags & DMIO_STATE_SEL) { ds->ds_flags &= ~DMIO_STATE_SEL; selwakeup(&ds->ds_selq); } } simple_unlock(&ds->ds_slock); } /* * dmio_write: * * Write file op. */ static int dmio_write(struct file *fp, off_t *offp, struct uio *uio, struct ucred *cred, int flags) { struct dmio_state *ds = (struct dmio_state *) fp->f_data; struct dmio_usrreq_state *dus; struct dmover_request *dreq; struct dmio_usrreq req; int error = 0, s, progress = 0; if ((uio->uio_resid % sizeof(req)) != 0) return (EINVAL); if (ds->ds_session == NULL) return (ENXIO); s = splsoftclock(); simple_lock(&ds->ds_slock); while (uio->uio_resid != 0) { if (ds->ds_nreqs == DMIO_NREQS_MAX) { if (fp->f_flag & FNONBLOCK) { error = progress ? 0 : EWOULDBLOCK; break; } ds->ds_flags |= DMIO_STATE_WRITE_WAIT; error = ltsleep(&ds->ds_nreqs, PRIBIO | PCATCH, "dmiowr", 0, &ds->ds_slock); if (error) break; continue; } ds->ds_nreqs++; simple_unlock(&ds->ds_slock); splx(s); progress = 1; error = uiomove(&req, sizeof(req), uio); if (error) { s = splsoftclock(); simple_lock(&ds->ds_slock); ds->ds_nreqs--; break; } /* XXX How should this interact with FNONBLOCK? */ dreq = dmover_request_alloc(ds->ds_session, NULL); if (dreq == NULL) { /* XXX */ s = splsoftclock(); simple_lock(&ds->ds_slock); ds->ds_nreqs--; error = ENOMEM; break; } s = splsoftclock(); dus = pool_get(&dmio_usrreq_state_pool, PR_WAITOK); splx(s); error = dmio_usrreq_init(fp, dus, &req, dreq); if (error) { dmover_request_free(dreq); s = splsoftclock(); pool_put(&dmio_usrreq_state_pool, dus); simple_lock(&ds->ds_slock); break; } dreq->dreq_callback = dmio_usrreq_done; dreq->dreq_cookie = dus; dus->dus_req = dreq; dus->dus_id = req.req_id; s = splsoftclock(); simple_lock(&ds->ds_slock); TAILQ_INSERT_TAIL(&ds->ds_pending, dus, dus_q); simple_unlock(&ds->ds_slock); splx(s); dmover_process(dreq); s = splsoftclock(); simple_lock(&ds->ds_slock); } simple_unlock(&ds->ds_slock); splx(s); return (error); } /* * dmio_ioctl: * * Ioctl file op. */ static int dmio_ioctl(struct file *fp, u_long cmd, caddr_t data, struct proc *p) { struct dmio_state *ds = (struct dmio_state *) fp->f_data; int error, s; switch (cmd) { case FIONBIO: case FIOASYNC: return (0); case DMIO_SETFUNC: { struct dmio_setfunc *dsf = (void *) data; struct dmover_session *dses; s = splsoftclock(); simple_lock(&ds->ds_slock); if (ds->ds_session != NULL || (ds->ds_flags & DMIO_STATE_LARVAL) != 0) { simple_unlock(&ds->ds_slock); splx(s); return (EBUSY); } ds->ds_flags |= DMIO_STATE_LARVAL; simple_unlock(&ds->ds_slock); splx(s); dsf->dsf_name[DMIO_MAX_FUNCNAME - 1] = '\0'; error = dmover_session_create(dsf->dsf_name, &dses); s = splsoftclock(); simple_lock(&ds->ds_slock); if (error == 0) { dses->dses_cookie = ds; ds->ds_session = dses; } ds->ds_flags &= ~DMIO_STATE_LARVAL; simple_unlock(&ds->ds_slock); splx(s); break; } default: error = ENOTTY; } return (error); } /* * dmio_fcntl: * * Fcntl file op. */ static int dmio_fcntl(struct file *fp, u_int cmd, caddr_t data, struct proc *p) { if (cmd == FNONBLOCK || cmd == FASYNC) return (0); return (EOPNOTSUPP); } /* * dmio_poll: * * Poll file op. */ static int dmio_poll(struct file *fp, int events, struct proc *p) { struct dmio_state *ds = (struct dmio_state *) fp->f_data; int s, revents = 0; if ((events & (POLLIN | POLLRDNORM | POLLOUT | POLLWRNORM)) == 0) return (revents); s = splsoftclock(); simple_lock(&ds->ds_slock); if (ds->ds_flags & DMIO_STATE_DEAD) { /* EOF */ revents |= events & (POLLIN | POLLRDNORM | POLLOUT | POLLWRNORM); goto out; } /* We can read if there are completed requests. */ if (events & (POLLIN | POLLRDNORM)) if (TAILQ_EMPTY(&ds->ds_complete) == 0) revents |= events & (POLLIN | POLLRDNORM); /* * We can write if there is there are fewer then DMIO_NREQS_MAX * are already in the queue. */ if (events & (POLLOUT | POLLWRNORM)) if (ds->ds_nreqs < DMIO_NREQS_MAX) revents |= events & (POLLOUT | POLLWRNORM); if (revents == 0) { selrecord(p, &ds->ds_selq); ds->ds_flags |= DMIO_STATE_SEL; } out: simple_unlock(&ds->ds_slock); splx(s); return (revents); } /* * dmio_stat: * * Stat file op. */ static int dmio_stat(struct file *fp, struct stat *sb, struct proc *p) { return (EOPNOTSUPP); } /* * dmio_close: * * Close file op. */ static int dmio_close(struct file *fp, struct proc *p) { struct dmio_state *ds = (struct dmio_state *) fp->f_data; struct dmio_usrreq_state *dus; struct dmover_session *dses; int s; s = splsoftclock(); simple_lock(&ds->ds_slock); ds->ds_flags |= DMIO_STATE_DEAD; /* Garbage-collect all the responses on the queue. */ while ((dus = TAILQ_FIRST(&ds->ds_complete)) != NULL) { TAILQ_REMOVE(&ds->ds_complete, dus, dus_q); ds->ds_nreqs--; dmover_request_free(dus->dus_req); dmio_usrreq_fini(ds, dus); } /* * If there are any requests pending, we have to wait for * them. Don't free the dmio_state in this case. */ if (ds->ds_nreqs == 0) { dses = ds->ds_session; simple_unlock(&ds->ds_slock); pool_put(&dmio_state_pool, ds); } else { dses = NULL; simple_unlock(&ds->ds_slock); } splx(s); fp->f_data = NULL; if (dses != NULL) dmover_session_destroy(dses); return (0); } static struct fileops dmio_fileops = { dmio_read, dmio_write, dmio_ioctl, dmio_fcntl, dmio_poll, dmio_stat, dmio_close, }; /* * dmoverioopen: * * Device switch open routine. */ int dmoverioopen(dev_t dev, int flag, int mode, struct proc *p) { struct dmio_state *ds; struct file *fp; int error, fd, s; /* falloc() will use the descriptor for us. */ if ((error = falloc(p, &fp, &fd)) != 0) return (error); s = splsoftclock(); ds = pool_get(&dmio_state_pool, PR_WAITOK); splx(s); memset(ds, 0, sizeof(*ds)); TAILQ_INIT(&ds->ds_pending); TAILQ_INIT(&ds->ds_complete); fp->f_flag = FREAD | FWRITE; fp->f_type = DTYPE_MISC; fp->f_ops = &dmio_fileops; fp->f_data = (caddr_t) ds; p->p_dupfd = fd; FILE_SET_MATURE(fp); FILE_UNUSE(fp, p); return (ENXIO); }