NetBSD/sys/dev/ic/dpt.c

1102 lines
29 KiB
C

/* $NetBSD: dpt.c,v 1.22 2000/06/13 13:36:42 ad Exp $ */
/*-
* Copyright (c) 1997, 1998, 1999, 2000 The NetBSD Foundation, Inc.
* All rights reserved.
*
* This code is derived from software contributed to The NetBSD Foundation
* by Andrew Doran, Charles M. Hannum and by Jason R. Thorpe of the Numerical
* Aerospace Simulation Facility, NASA Ames Research Center.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. All advertising materials mentioning features or use of this software
* must display the following acknowledgement:
* This product includes software developed by the NetBSD
* Foundation, Inc. and its contributors.
* 4. Neither the name of The NetBSD Foundation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/
/*
* Portions of this code fall under the following copyright:
*
* Originally written by Julian Elischer (julian@tfs.com)
* for TRW Financial Systems for use under the MACH(2.5) operating system.
*
* TRW Financial Systems, in accordance with their agreement with Carnegie
* Mellon University, makes this software available to CMU to distribute
* or use in any manner that they see fit as long as this message is kept with
* the software. For this reason TFS also grants any other persons or
* organisations permission to use or modify this software.
*
* TFS supplies this software to be publicly redistributed
* on the understanding that TFS is not responsible for the correct
* functioning of this software in any circumstances.
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: dpt.c,v 1.22 2000/06/13 13:36:42 ad Exp $");
#include <sys/param.h>
#include <sys/systm.h>
#include <sys/kernel.h>
#include <sys/device.h>
#include <sys/queue.h>
#include <sys/proc.h>
#include <sys/buf.h>
#include <sys/endian.h>
#include <machine/bswap.h>
#include <machine/bus.h>
#include <dev/scsipi/scsi_all.h>
#include <dev/scsipi/scsipi_all.h>
#include <dev/scsipi/scsiconf.h>
#include <dev/ic/dptreg.h>
#include <dev/ic/dptvar.h>
/* A default for our link struct */
static struct scsipi_device dpt_dev = {
NULL, /* Use default error handler */
NULL, /* have a queue, served by this */
NULL, /* have no async handler */
NULL, /* Use default 'done' routine */
};
static char *dpt_cname[] = {
"PM3334", "SmartRAID IV",
"PM3332", "SmartRAID IV",
"PM2144", "SmartCache IV",
"PM2044", "SmartCache IV",
"PM2142", "SmartCache IV",
"PM2042", "SmartCache IV",
"PM2041", "SmartCache IV",
"PM3224", "SmartRAID III",
"PM3222", "SmartRAID III",
"PM3021", "SmartRAID III",
"PM2124", "SmartCache III",
"PM2024", "SmartCache III",
"PM2122", "SmartCache III",
"PM2022", "SmartCache III",
"PM2021", "SmartCache III",
"SK2012", "SmartCache Plus",
"SK2011", "SmartCache Plus",
NULL, "unknown adapter, please report using send-pr(1)",
};
TAILQ_HEAD(, dpt_softc) dpt_hba; /* tailq of HBA softc's */
void *dpt_sdh; /* shutdown hook */
/*
* Handle an interrupt from the HBA.
*/
int
dpt_intr(xxx_sc)
void *xxx_sc;
{
struct dpt_softc *sc;
struct dpt_ccb *ccb;
struct eata_sp *sp;
volatile int junk;
sc = xxx_sc;
sp = sc->sc_stp;
#ifdef DEBUG
if ((dpt_inb(sc, HA_AUX_STATUS) & HA_AUX_INTR) == 0) {
printf("%s: spurious intr\n", sc->sc_dv.dv_xname);
return (1);
}
#endif
for (;;) {
/*
* HBA might have interrupted while we were dealing with the
* last completed command, since we ACK before we deal; keep
* polling.
*/
if ((dpt_inb(sc, HA_AUX_STATUS) & HA_AUX_INTR) == 0)
break;
bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap, sc->sc_stpoff,
sizeof(struct eata_sp), BUS_DMASYNC_POSTREAD);
/* Might have looped before HBA can reset HBA_AUX_INTR */
if (sp->sp_ccbid == -1) {
DELAY(50);
if ((dpt_inb(sc, HA_AUX_STATUS) & HA_AUX_INTR) == 0)
return (0);
printf("%s: no status\n", sc->sc_dv.dv_xname);
/* Re-sync DMA map */
bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap,
sc->sc_stpoff, sizeof(struct eata_sp),
BUS_DMASYNC_POSTREAD);
}
/* Make sure CCB ID from status packet is realistic */
if (sp->sp_ccbid >= 0 && sp->sp_ccbid < sc->sc_nccbs) {
/* Sync up DMA map and cache cmd status */
ccb = sc->sc_ccbs + sp->sp_ccbid;
bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap,
CCB_OFF(sc, ccb), sizeof(struct dpt_ccb),
BUS_DMASYNC_POSTWRITE);
ccb->ccb_hba_status = sp->sp_hba_status & 0x7F;
ccb->ccb_scsi_status = sp->sp_scsi_status;
/*
* Ack the interrupt and process the CCB. If this
* is a private CCB it's up to dpt_poll() to notice.
*/
sp->sp_ccbid = -1;
ccb->ccb_flg |= CCB_INTR;
junk = dpt_inb(sc, HA_STATUS);
if ((ccb->ccb_flg & CCB_PRIVATE) == 0)
dpt_done_ccb(sc, ccb);
} else {
printf("%s: bogus status (returned CCB id %d)\n",
sc->sc_dv.dv_xname, sp->sp_ccbid);
/* Ack the interrupt */
sp->sp_ccbid = -1;
junk = dpt_inb(sc, HA_STATUS);
}
}
return (1);
}
/*
* Initialize and attach the HBA. This is the entry point from bus
* specific probe-and-attach code.
*/
void
dpt_init(sc, intrstr)
struct dpt_softc *sc;
const char *intrstr;
{
struct eata_inquiry_data *ei;
int i, j, error, rseg, maxchannel, maxtarget;
bus_dma_segment_t seg;
struct eata_cfg *ec;
struct scsipi_link *link;
char model[16];
ec = &sc->sc_ec;
/* Allocate the CCB/status packet/scratch DMA map and load */
sc->sc_nccbs =
min(be16toh(*(int16_t *)ec->ec_queuedepth), DPT_MAX_CCBS);
sc->sc_stpoff = sc->sc_nccbs * sizeof(struct dpt_ccb);
sc->sc_scroff = sc->sc_stpoff + sizeof(struct eata_sp);
sc->sc_scrlen = DPT_SCRATCH_SIZE;
sc->sc_dmamapsize = sc->sc_nccbs * sizeof(struct dpt_ccb) +
sc->sc_scrlen + sizeof(struct eata_sp);
if ((error = bus_dmamem_alloc(sc->sc_dmat, sc->sc_dmamapsize, NBPG, 0,
&seg, 1, &rseg, BUS_DMA_NOWAIT)) != 0) {
printf("%s: unable to allocate CCBs, error = %d\n",
sc->sc_dv.dv_xname, error);
return;
}
if ((error = bus_dmamem_map(sc->sc_dmat, &seg, rseg, sc->sc_dmamapsize,
(caddr_t *)&sc->sc_ccbs, BUS_DMA_NOWAIT|BUS_DMA_COHERENT)) != 0) {
printf("%s: unable to map CCBs, error = %d\n",
sc->sc_dv.dv_xname, error);
return;
}
if ((error = bus_dmamap_create(sc->sc_dmat, sc->sc_dmamapsize,
sc->sc_dmamapsize, 1, 0, BUS_DMA_NOWAIT, &sc->sc_dmamap)) != 0) {
printf("%s: unable to create CCB DMA map, error = %d\n",
sc->sc_dv.dv_xname, error);
return;
}
if ((error = bus_dmamap_load(sc->sc_dmat, sc->sc_dmamap,
sc->sc_ccbs, sc->sc_dmamapsize, NULL, BUS_DMA_NOWAIT)) != 0) {
printf("%s: unable to load CCB DMA map, error = %d\n",
sc->sc_dv.dv_xname, error);
return;
}
sc->sc_stp = (struct eata_sp *)((caddr_t)sc->sc_ccbs + sc->sc_stpoff);
sc->sc_stppa = sc->sc_dmamap->dm_segs[0].ds_addr + sc->sc_stpoff;
sc->sc_scr = (caddr_t)sc->sc_ccbs + sc->sc_scroff;
sc->sc_scrpa = sc->sc_dmamap->dm_segs[0].ds_addr + sc->sc_scroff;
sc->sc_stp->sp_ccbid = -1;
/* Initialize the CCBs */
TAILQ_INIT(&sc->sc_free_ccb);
i = dpt_create_ccbs(sc, sc->sc_ccbs, sc->sc_nccbs);
if (i == 0) {
printf("%s: unable to create CCBs\n", sc->sc_dv.dv_xname);
return;
} else if (i != sc->sc_nccbs) {
printf("%s: %d/%d CCBs created!\n", sc->sc_dv.dv_xname, i,
sc->sc_nccbs);
sc->sc_nccbs = i;
}
/* Set shutdownhook before we start any device activity */
if (dpt_sdh == NULL) {
TAILQ_INIT(&dpt_hba);
dpt_sdh = shutdownhook_establish(dpt_shutdown, NULL);
}
/* Get the page 0 inquiry data from the HBA */
dpt_hba_inquire(sc, &ei);
/*
* dpt0 at pci0 dev 12 function 0: DPT SmartRAID III (PM3224A/9X-R)
* dpt0: interrupting at irq 10
* dpt0: 64 queued commands, 1 channel(s), adapter on ID(s) 7
*/
for (i = 0; ei->ei_vendor[i] != ' ' && i < 8; i++)
;
ei->ei_vendor[i] = '\0';
for (i = 0; ei->ei_model[i] != ' ' && i < 7; i++)
model[i] = ei->ei_model[i];
for (j = 0; ei->ei_suffix[j] != ' ' && j < 7; j++)
model[i++] = ei->ei_model[i];
model[i] = '\0';
/* Find the cannonical name for the board */
for (i = 0; dpt_cname[i] != NULL; i += 2)
if (memcmp(ei->ei_model, dpt_cname[i], 6) == 0)
break;
printf("%s %s (%s)\n", ei->ei_vendor, dpt_cname[i + 1], model);
if (intrstr != NULL)
printf("%s: interrupting at %s\n", sc->sc_dv.dv_xname, intrstr);
maxchannel = (ec->ec_feat3 & EC_F3_MAX_CHANNEL_MASK) >>
EC_F3_MAX_CHANNEL_SHIFT;
maxtarget = (ec->ec_feat3 & EC_F3_MAX_TARGET_MASK) >>
EC_F3_MAX_TARGET_SHIFT;
printf("%s: %d queued commands, %d channel(s), adapter on ID(s)",
sc->sc_dv.dv_xname, sc->sc_nccbs, maxchannel + 1);
for (i = 0; i <= maxchannel; i++) {
sc->sc_hbaid[i] = ec->ec_hba[3 - i];
printf(" %d", sc->sc_hbaid[i]);
}
printf("\n");
/* Reset the SCSI bus */
if (dpt_cmd(sc, NULL, 0, CP_IMMEDIATE, CPI_BUS_RESET))
panic("%s: dpt_cmd failed", sc->sc_dv.dv_xname);
/* Fill in the adapter, each link and attach in turn */
sc->sc_adapter.scsipi_cmd = dpt_scsi_cmd;
sc->sc_adapter.scsipi_minphys = dpt_minphys;
for (i = 0; i <= maxchannel; i++) {
link = &sc->sc_link[i];
link->scsipi_scsi.channel = i;
link->scsipi_scsi.adapter_target = sc->sc_hbaid[i];
link->scsipi_scsi.max_lun = ec->ec_maxlun;
link->scsipi_scsi.max_target = maxtarget;
link->type = BUS_SCSI;
link->device = &dpt_dev;
link->adapter = &sc->sc_adapter;
link->adapter_softc = sc;
link->openings = sc->sc_nccbs; /* XXX */
config_found(&sc->sc_dv, link, scsiprint);
}
TAILQ_INSERT_TAIL(&dpt_hba, sc, sc_chain);
}
/*
* Our 'shutdownhook' to cleanly shut down the HBA. The HBA must flush
* all data from it's cache and mark array groups as clean.
*/
void
dpt_shutdown(xxx_sc)
void *xxx_sc;
{
struct dpt_softc *sc;
printf("shutting down dpt devices...");
for (sc = TAILQ_FIRST(&dpt_hba); sc != NULL;
sc = TAILQ_NEXT(sc, sc_chain))
dpt_cmd(sc, NULL, 0, CP_IMMEDIATE, CPI_POWEROFF_WARN);
DELAY(5000*1000);
printf(" done\n");
}
/*
* Send an EATA command to the HBA.
*/
int
dpt_cmd(sc, cp, addr, eatacmd, icmd)
struct dpt_softc *sc;
struct eata_cp *cp;
u_int32_t addr;
int eatacmd, icmd;
{
int i;
for (i = 20000; i; i--) {
if ((dpt_inb(sc, HA_AUX_STATUS) & HA_AUX_BUSY) == 0)
break;
DELAY(50);
}
/* Not the most graceful way to handle this */
if (i == 0) {
printf("%s: HBA timeout on EATA command issue; aborting\n",
sc->sc_dv.dv_xname);
return (-1);
}
if (cp == NULL)
addr = 0;
dpt_outl(sc, HA_DMA_BASE, (u_int32_t)addr);
if (eatacmd == CP_IMMEDIATE) {
if (cp == NULL) {
/* XXX should really pass meaningful values */
dpt_outb(sc, HA_ICMD_CODE2, 0);
dpt_outb(sc, HA_ICMD_CODE1, 0);
}
dpt_outb(sc, HA_ICMD, icmd);
}
dpt_outb(sc, HA_COMMAND, eatacmd);
return (0);
}
/*
* Wait for the HBA status register to reach a specific state.
*/
int
dpt_wait(sc, mask, state, ms)
struct dpt_softc *sc;
u_int8_t mask, state;
int ms;
{
for (ms *= 10; ms; ms--) {
if ((dpt_inb(sc, HA_STATUS) & mask) == state)
return (0);
DELAY(100);
}
return (-1);
}
/*
* Wait for the specified CCB to finish. This is used when we may not be
* able to sleep and/or interrupts are disabled (eg autoconfiguration).
* The timeout value from the CCB is used. This should only be used for
* CCB_PRIVATE requests; otherwise the CCB will get recycled before we get
* a look at it.
*/
int
dpt_poll(sc, ccb)
struct dpt_softc *sc;
struct dpt_ccb *ccb;
{
int i;
#ifdef DEBUG
if ((ccb->ccb_flg & CCB_PRIVATE) == 0)
panic("dpt_poll: called for non-CCB_PRIVATE request\n");
#endif
if ((ccb->ccb_flg & CCB_INTR) != 0)
return (0);
for (i = ccb->ccb_timeout * 20; i; i--) {
if ((dpt_inb(sc, HA_AUX_STATUS) & HA_AUX_INTR) != 0) {
dpt_intr(sc);
if ((ccb->ccb_flg & CCB_INTR) != 0)
return (0);
}
DELAY(50);
}
return (-1);
}
/*
* Read the EATA configuration from the HBA and perform some sanity checks.
*/
int
dpt_readcfg(sc)
struct dpt_softc *sc;
{
struct eata_cfg *ec;
int i, j, stat;
u_int16_t *p;
ec = &sc->sc_ec;
/* Older firmware may puke if we talk to it too soon after reset */
dpt_outb(sc, HA_COMMAND, CP_RESET);
DELAY(750000);
for (i = 1000; i; i--) {
if ((dpt_inb(sc, HA_STATUS) & HA_ST_READY) != 0)
break;
DELAY(2000);
}
if (i == 0) {
printf("%s: HBA not ready after reset (hba status:%02x)\n",
sc->sc_dv.dv_xname, dpt_inb(sc, HA_STATUS));
return (-1);
}
while((((stat = dpt_inb(sc, HA_STATUS))
!= (HA_ST_READY|HA_ST_SEEK_COMPLETE))
&& (stat != (HA_ST_READY|HA_ST_SEEK_COMPLETE|HA_ST_ERROR))
&& (stat != (HA_ST_READY|HA_ST_SEEK_COMPLETE|HA_ST_ERROR|HA_ST_DRQ)))
&& dpt_wait(sc, HA_ST_BUSY, 0, 2000)) {
/* RAID drives still spinning up? */
if((dpt_inb(sc, HA_ERROR) != 'D')
|| (dpt_inb(sc, HA_ERROR + 1) != 'P')
|| (dpt_inb(sc, HA_ERROR + 2) != 'T')) {
printf("%s: HBA not ready\n", sc->sc_dv.dv_xname);
return (-1);
}
}
/*
* Issue the read-config command and wait for the data to appear.
* XXX we shouldn't be doing this with PIO, but it makes it a lot
* easier as no DMA setup is required.
*/
dpt_outb(sc, HA_COMMAND, CP_PIO_GETCFG);
memset(ec, 0, sizeof(*ec));
i = ((int)&((struct eata_cfg *)0)->ec_cfglen +
sizeof(ec->ec_cfglen)) >> 1;
p = (u_int16_t *)ec;
if (dpt_wait(sc, 0xFF, HA_ST_DATA_RDY, 2000)) {
printf("%s: cfg data didn't appear (hba status:%02x)\n",
sc->sc_dv.dv_xname, dpt_inb(sc, HA_STATUS));
return (-1);
}
/* Begin reading */
while (i--)
*p++ = dpt_inw(sc, HA_DATA);
if ((i = ec->ec_cfglen) > (sizeof(struct eata_cfg)
- (int)(&(((struct eata_cfg *)0L)->ec_cfglen))
- sizeof(ec->ec_cfglen)))
i = sizeof(struct eata_cfg)
- (int)(&(((struct eata_cfg *)0L)->ec_cfglen))
- sizeof(ec->ec_cfglen);
j = i + (int)(&(((struct eata_cfg *)0L)->ec_cfglen)) +
sizeof(ec->ec_cfglen);
i >>= 1;
while (i--)
*p++ = dpt_inw(sc, HA_DATA);
/* Flush until we have read 512 bytes. */
i = (512 - j + 1) >> 1;
while (i--)
dpt_inw(sc, HA_DATA);
/* Defaults for older Firmware */
if (p <= (u_short *)&ec->ec_hba[DPT_MAX_CHANNELS - 1])
ec->ec_hba[DPT_MAX_CHANNELS - 1] = 7;
if ((dpt_inb(sc, HA_STATUS) & HA_ST_ERROR) != 0) {
printf("%s: HBA error\n", sc->sc_dv.dv_xname);
return (-1);
}
if (memcmp(ec->ec_eatasig, "EATA", 4) != 0) {
printf("%s: EATA signature mismatch\n", sc->sc_dv.dv_xname);
return (-1);
}
if ((ec->ec_feat0 & EC_F0_HBA_VALID) == 0) {
printf("%s: ec_hba field invalid\n", sc->sc_dv.dv_xname);
return (-1);
}
if ((ec->ec_feat0 & EC_F0_DMA_SUPPORTED) == 0) {
printf("%s: DMA not supported\n", sc->sc_dv.dv_xname);
return (-1);
}
return (0);
}
/*
* Adjust the size of each I/O before it passes to the SCSI layer.
*/
void
dpt_minphys(bp)
struct buf *bp;
{
if (bp->b_bcount > DPT_MAX_XFER)
bp->b_bcount = DPT_MAX_XFER;
minphys(bp);
}
/*
* Put a CCB onto the freelist.
*/
void
dpt_free_ccb(sc, ccb)
struct dpt_softc *sc;
struct dpt_ccb *ccb;
{
int s;
s = splbio();
ccb->ccb_flg = 0;
TAILQ_INSERT_HEAD(&sc->sc_free_ccb, ccb, ccb_chain);
/* Wake anybody waiting for a free ccb */
if (ccb->ccb_chain.tqe_next == 0)
wakeup(&sc->sc_free_ccb);
splx(s);
}
/*
* Initialize the specified CCB.
*/
int
dpt_init_ccb(sc, ccb)
struct dpt_softc *sc;
struct dpt_ccb *ccb;
{
int error;
/* Create the DMA map for this CCB's data */
error = bus_dmamap_create(sc->sc_dmat, DPT_MAX_XFER, DPT_SG_SIZE,
DPT_MAX_XFER, 0, BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW,
&ccb->ccb_dmamap_xfer);
if (error) {
printf("%s: can't create ccb dmamap (%d)\n",
sc->sc_dv.dv_xname, error);
return (error);
}
ccb->ccb_flg = 0;
ccb->ccb_ccbpa = sc->sc_dmamap->dm_segs[0].ds_addr +
CCB_OFF(sc, ccb);
return (0);
}
/*
* Create a set of CCBs and add them to the free list.
*/
int
dpt_create_ccbs(sc, ccbstore, count)
struct dpt_softc *sc;
struct dpt_ccb *ccbstore;
int count;
{
struct dpt_ccb *ccb;
int i, error;
memset(ccbstore, 0, sizeof(struct dpt_ccb) * count);
for (i = 0, ccb = ccbstore; i < count; i++, ccb++) {
if ((error = dpt_init_ccb(sc, ccb)) != 0) {
printf("%s: unable to init ccb, error = %d\n",
sc->sc_dv.dv_xname, error);
break;
}
ccb->ccb_id = i;
TAILQ_INSERT_TAIL(&sc->sc_free_ccb, ccb, ccb_chain);
}
return (i);
}
/*
* Get a free ccb. If there are none, see if we can allocate a new one. If
* none are available right now and we are permitted to sleep, then wait
* until one becomes free, otherwise return an error.
*/
struct dpt_ccb *
dpt_alloc_ccb(sc, flg)
struct dpt_softc *sc;
int flg;
{
struct dpt_ccb *ccb;
int s;
s = splbio();
for (;;) {
if ((ccb = TAILQ_FIRST(&sc->sc_free_ccb)) != NULL) {
TAILQ_REMOVE(&sc->sc_free_ccb, ccb, ccb_chain);
break;
}
if ((flg & XS_CTL_NOSLEEP) != 0) {
ccb = NULL;
break;
}
tsleep(&sc->sc_free_ccb, PRIBIO, "dptccb", 0);
}
splx(s);
return (ccb);
}
/*
* We have a CCB which has been processed by the HBA, now we look to see how
* the operation went. CCBs marked with CCB_PRIVATE are not automatically
* passed here by dpt_intr().
*/
void
dpt_done_ccb(sc, ccb)
struct dpt_softc *sc;
struct dpt_ccb *ccb;
{
struct scsipi_sense_data *s1, *s2;
struct scsipi_xfer *xs;
bus_dma_tag_t dmat;
dmat = sc->sc_dmat;
xs = ccb->ccb_xs;
SC_DEBUG(xs->sc_link, SDEV_DB2, ("dpt_done_ccb\n"));
/*
* If we were a data transfer, unload the map that described the
* data buffer.
*/
if (xs->datalen != 0) {
bus_dmamap_sync(dmat, ccb->ccb_dmamap_xfer, 0,
ccb->ccb_dmamap_xfer->dm_mapsize,
(xs->xs_control & XS_CTL_DATA_IN) ? BUS_DMASYNC_POSTREAD :
BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(dmat, ccb->ccb_dmamap_xfer);
}
if (xs->error == XS_NOERROR) {
if (ccb->ccb_hba_status != SP_HBA_NO_ERROR) {
switch (ccb->ccb_hba_status) {
case SP_HBA_ERROR_SEL_TO:
xs->error = XS_SELTIMEOUT;
break;
case SP_HBA_ERROR_RESET:
xs->error = XS_RESET;
break;
default: /* Other scsi protocol messes */
printf("%s: HBA status %x\n",
sc->sc_dv.dv_xname, ccb->ccb_hba_status);
xs->error = XS_DRIVER_STUFFUP;
}
} else if (ccb->ccb_scsi_status != SCSI_OK) {
switch (ccb->ccb_scsi_status) {
case SCSI_CHECK:
s1 = &ccb->ccb_sense;
s2 = &xs->sense.scsi_sense;
*s2 = *s1;
xs->error = XS_SENSE;
break;
case SCSI_BUSY:
xs->error = XS_BUSY;
break;
default:
printf("%s: SCSI status %x\n",
sc->sc_dv.dv_xname, ccb->ccb_scsi_status);
xs->error = XS_DRIVER_STUFFUP;
}
} else
xs->resid = 0;
xs->status = ccb->ccb_scsi_status;
}
/* Free up the CCB and mark the command as done */
dpt_free_ccb(sc, ccb);
xs->xs_status |= XS_STS_DONE;
scsipi_done(xs);
/*
* If there are entries in the software queue, try to run the first
* one. We should be more or less guaranteed to succeed, since we
* just freed an CCB. NOTE: dpt_scsi_cmd() relies on our calling it
* with the first entry in the queue.
*/
if ((xs = TAILQ_FIRST(&sc->sc_queue)) != NULL)
dpt_scsi_cmd(xs);
}
/*
* Start a SCSI command.
*/
int
dpt_scsi_cmd(xs)
struct scsipi_xfer *xs;
{
int error, i, flags, s, fromqueue, dontqueue, nowait;
struct scsipi_link *sc_link;
struct dpt_softc *sc;
struct dpt_ccb *ccb;
struct eata_sg *sg;
struct eata_cp *cp;
bus_dma_tag_t dmat;
bus_dmamap_t xfer;
sc_link = xs->sc_link;
flags = xs->xs_control;
sc = sc_link->adapter_softc;
dmat = sc->sc_dmat;
fromqueue = 0;
dontqueue = 0;
nowait = 0;
SC_DEBUG(sc_link, SDEV_DB2, ("dpt_scsi_cmd\n"));
/* Protect the queue */
s = splbio();
/*
* If we're running the queue from dpt_done_ccb(), we've been called
* with the first queue entry as our argument.
*/
if (xs == TAILQ_FIRST(&sc->sc_queue)) {
TAILQ_REMOVE(&sc->sc_queue, xs, adapter_q);
fromqueue = 1;
nowait = 1;
} else {
/* Cmds must be no more than 12 bytes for us */
if (xs->cmdlen > 12) {
splx(s);
xs->error = XS_DRIVER_STUFFUP;
return (COMPLETE);
}
/*
* XXX we can't reset devices just yet. Apparently some
* older firmware revisions don't even support it.
*/
if ((flags & XS_CTL_RESET) != 0) {
xs->error = XS_DRIVER_STUFFUP;
return (COMPLETE);
}
/* Polled requests can't be queued for later */
dontqueue = flags & XS_CTL_POLL;
/* If there are jobs in the queue, run them first */
if (TAILQ_FIRST(&sc->sc_queue) != NULL) {
/*
* If we can't queue we abort, since we must
* preserve the queue order.
*/
if (dontqueue) {
splx(s);
xs->error = XS_DRIVER_STUFFUP;
return (TRY_AGAIN_LATER);
}
/* Swap with the first queue entry. */
TAILQ_INSERT_TAIL(&sc->sc_queue, xs, adapter_q);
xs = TAILQ_FIRST(&sc->sc_queue);
TAILQ_REMOVE(&sc->sc_queue, xs, adapter_q);
fromqueue = 1;
}
}
/* Get a CCB */
if (nowait)
flags |= XS_CTL_NOSLEEP;
if ((ccb = dpt_alloc_ccb(sc, flags)) == NULL) {
/* If we can't queue, we lose */
if (dontqueue) {
splx(s);
xs->error = XS_DRIVER_STUFFUP;
return (TRY_AGAIN_LATER);
}
/*
* Stuff request into the queue, in front if we came off
* it in the first place.
*/
if (fromqueue)
TAILQ_INSERT_HEAD(&sc->sc_queue, xs, adapter_q);
else
TAILQ_INSERT_TAIL(&sc->sc_queue, xs, adapter_q);
splx(s);
return (SUCCESSFULLY_QUEUED);
}
splx(s);
ccb->ccb_xs = xs;
ccb->ccb_timeout = xs->timeout;
cp = &ccb->ccb_eata_cp;
memcpy(&cp->cp_cdb_cmd, xs->cmd, xs->cmdlen);
cp->cp_ccbid = ccb->ccb_id;
cp->cp_senselen = sizeof(ccb->ccb_sense);
cp->cp_stataddr = htobe32(sc->sc_stppa);
cp->cp_ctl0 = CP_C0_AUTO_SENSE;
cp->cp_ctl1 = 0;
cp->cp_ctl2 = 0;
cp->cp_ctl3 = sc_link->scsipi_scsi.target << CP_C3_ID_SHIFT;
cp->cp_ctl3 |= sc_link->scsipi_scsi.channel << CP_C3_CHANNEL_SHIFT;
cp->cp_ctl4 = sc_link->scsipi_scsi.lun << CP_C4_LUN_SHIFT;
cp->cp_ctl4 |= CP_C4_DIS_PRI | CP_C4_IDENTIFY;
if ((flags & XS_CTL_DATA_IN) != 0)
cp->cp_ctl0 |= CP_C0_DATA_IN;
if ((flags & XS_CTL_DATA_OUT) != 0)
cp->cp_ctl0 |= CP_C0_DATA_OUT;
if (sc->sc_hbaid[sc_link->scsipi_scsi.channel] ==
sc_link->scsipi_scsi.target)
cp->cp_ctl0 |= CP_C0_INTERPRET;
/* Synchronous xfers musn't write-back through the cache */
if (xs->bp != NULL && (xs->bp->b_flags & (B_ASYNC | B_READ)) == 0)
cp->cp_ctl2 |= CP_C2_NO_CACHE;
cp->cp_senseaddr = htobe32(sc->sc_dmamap->dm_segs[0].ds_addr +
CCB_OFF(sc, ccb) + offsetof(struct dpt_ccb, ccb_sense));
if (xs->datalen != 0) {
xfer = ccb->ccb_dmamap_xfer;
#ifdef TFS
if ((flags & XS_CTL_DATA_UIO) != 0) {
error = bus_dmamap_load_uio(dmat, xfer,
(struct uio *)xs->data, (flags & XS_CTL_NOSLEEP) ?
BUS_DMA_NOWAIT : BUS_DMA_WAITOK);
} else
#endif /* TFS */
{
error = bus_dmamap_load(dmat, xfer, xs->data,
xs->datalen, NULL, (flags & XS_CTL_NOSLEEP) ?
BUS_DMA_NOWAIT : BUS_DMA_WAITOK);
}
if (error) {
printf("%s: dpt_scsi_cmd: ", sc->sc_dv.dv_xname);
if (error == EFBIG)
printf("more than %d dma segs\n", DPT_SG_SIZE);
else
printf("error %d loading dma map\n", error);
xs->error = XS_DRIVER_STUFFUP;
dpt_free_ccb(sc, ccb);
return (COMPLETE);
}
bus_dmamap_sync(dmat, xfer, 0, xfer->dm_mapsize,
(flags & XS_CTL_DATA_IN) ? BUS_DMASYNC_PREREAD :
BUS_DMASYNC_PREWRITE);
/* Don't bother using scatter/gather for just 1 segment */
if (xfer->dm_nsegs == 1) {
cp->cp_dataaddr = htobe32(xfer->dm_segs[0].ds_addr);
cp->cp_datalen = htobe32(xfer->dm_segs[0].ds_len);
} else {
/*
* Load the hardware scatter/gather map with the
* contents of the DMA map.
*/
sg = ccb->ccb_sg;
for (i = 0; i < xfer->dm_nsegs; i++, sg++) {
sg->sg_addr = htobe32(xfer->dm_segs[i].ds_addr);
sg->sg_len = htobe32(xfer->dm_segs[i].ds_len);
}
cp->cp_dataaddr = htobe32(CCB_OFF(sc, ccb) +
sc->sc_dmamap->dm_segs[0].ds_addr +
offsetof(struct dpt_ccb, ccb_sg));
cp->cp_datalen = htobe32(i * sizeof(struct eata_sg));
cp->cp_ctl0 |= CP_C0_SCATTER;
}
} else {
cp->cp_dataaddr = 0;
cp->cp_datalen = 0;
}
/* Sync up CCB and status packet */
bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap, CCB_OFF(sc, ccb),
sizeof(struct dpt_ccb), BUS_DMASYNC_PREWRITE);
bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap, sc->sc_stpoff,
sizeof(struct eata_sp), BUS_DMASYNC_PREREAD);
/*
* Start the command. If we are polling on completion, mark it
* private so that dpt_intr/dpt_done_ccb don't recycle the CCB
* without us noticing.
*/
if (dontqueue != 0)
ccb->ccb_flg |= CCB_PRIVATE;
if (dpt_cmd(sc, &ccb->ccb_eata_cp, ccb->ccb_ccbpa, CP_DMA_CMD, 0)) {
printf("%s: dpt_cmd failed\n", sc->sc_dv.dv_xname);
xs->error = XS_DRIVER_STUFFUP;
dpt_free_ccb(sc, ccb);
return (TRY_AGAIN_LATER);
}
if (dontqueue == 0)
return (SUCCESSFULLY_QUEUED);
/* Don't wait longer than this single command wants to wait */
if (dpt_poll(sc, ccb)) {
dpt_timeout(ccb);
/* Wait for abort to complete */
if (dpt_poll(sc, ccb))
dpt_timeout(ccb);
}
dpt_done_ccb(sc, ccb);
return (COMPLETE);
}
/*
* Specified CCB has timed out, abort it.
*/
void
dpt_timeout(arg)
void *arg;
{
struct scsipi_link *sc_link;
struct scsipi_xfer *xs;
struct dpt_softc *sc;
struct dpt_ccb *ccb;
int s;
ccb = arg;
xs = ccb->ccb_xs;
sc_link = xs->sc_link;
sc = sc_link->adapter_softc;
scsi_print_addr(sc_link);
printf("timed out (status:%02x aux status:%02x)",
dpt_inb(sc, HA_STATUS), dpt_inb(sc, HA_AUX_STATUS));
s = splbio();
if ((ccb->ccb_flg & CCB_ABORT) != 0) {
/* Abort timed out, reset the HBA */
printf(" AGAIN, resetting HBA\n");
dpt_outb(sc, HA_COMMAND, CP_RESET);
DELAY(750000);
} else {
/* Abort the operation that has timed out */
printf("\n");
ccb->ccb_xs->error = XS_TIMEOUT;
ccb->ccb_timeout = DPT_ABORT_TIMEOUT;
ccb->ccb_flg |= CCB_ABORT;
/* Start the abort */
if (dpt_cmd(sc, &ccb->ccb_eata_cp, ccb->ccb_ccbpa,
CP_IMMEDIATE, CPI_SPEC_ABORT))
printf("%s: dpt_cmd failed\n", sc->sc_dv.dv_xname);
}
splx(s);
}
/*
* Get inquiry data from the adapter.
*/
void
dpt_hba_inquire(sc, ei)
struct dpt_softc *sc;
struct eata_inquiry_data **ei;
{
struct dpt_ccb *ccb;
struct eata_cp *cp;
bus_dma_tag_t dmat;
*ei = (struct eata_inquiry_data *)sc->sc_scr;
dmat = sc->sc_dmat;
/* Get a CCB and mark as private */
if ((ccb = dpt_alloc_ccb(sc, 0)) == NULL)
panic("%s: no CCB for inquiry", sc->sc_dv.dv_xname);
ccb->ccb_flg |= CCB_PRIVATE;
ccb->ccb_timeout = 200;
/* Put all the arguments into the CCB */
cp = &ccb->ccb_eata_cp;
cp->cp_ccbid = ccb->ccb_id;
cp->cp_senselen = sizeof(ccb->ccb_sense);
cp->cp_senseaddr = 0;
cp->cp_stataddr = htobe32(sc->sc_stppa);
cp->cp_dataaddr = htobe32(sc->sc_scrpa);
cp->cp_datalen = htobe32(sizeof(struct eata_inquiry_data));
cp->cp_ctl0 = CP_C0_DATA_IN | CP_C0_INTERPRET;
cp->cp_ctl1 = 0;
cp->cp_ctl2 = 0;
cp->cp_ctl3 = sc->sc_hbaid[0] << CP_C3_ID_SHIFT;
cp->cp_ctl4 = CP_C4_DIS_PRI | CP_C4_IDENTIFY;
/* Put together the SCSI inquiry command */
memset(&cp->cp_cdb_cmd, 0, 12); /* XXX */
cp->cp_cdb_cmd = INQUIRY;
cp->cp_cdb_len = sizeof(struct eata_inquiry_data);
/* Sync up CCB, status packet and scratch area */
bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap, CCB_OFF(sc, ccb),
sizeof(struct dpt_ccb), BUS_DMASYNC_PREWRITE);
bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap, sc->sc_stpoff,
sizeof(struct eata_sp), BUS_DMASYNC_PREREAD);
bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap, sc->sc_scroff,
sizeof(struct eata_inquiry_data), BUS_DMASYNC_PREREAD);
/* Start the command and poll on completion */
if (dpt_cmd(sc, &ccb->ccb_eata_cp, ccb->ccb_ccbpa, CP_DMA_CMD, 0))
panic("%s: dpt_cmd failed", sc->sc_dv.dv_xname);
if (dpt_poll(sc, ccb))
panic("%s: inquiry timed out", sc->sc_dv.dv_xname);
if (ccb->ccb_hba_status != SP_HBA_NO_ERROR ||
ccb->ccb_scsi_status != SCSI_OK)
panic("%s: inquiry failed (hba:%02x scsi:%02x)",
sc->sc_dv.dv_xname, ccb->ccb_hba_status,
ccb->ccb_scsi_status);
/* Sync up the DMA map and free CCB, returning */
bus_dmamap_sync(sc->sc_dmat, sc->sc_dmamap, sc->sc_scroff,
sizeof(struct eata_inquiry_data), BUS_DMASYNC_POSTREAD);
dpt_free_ccb(sc, ccb);
}