1951 lines
47 KiB
C
1951 lines
47 KiB
C
/* $NetBSD: bha.c,v 1.75 2012/10/27 17:18:19 chs Exp $ */
|
|
|
|
/*-
|
|
* Copyright (c) 1997, 1998, 1999 The NetBSD Foundation, Inc.
|
|
* All rights reserved.
|
|
*
|
|
* This code is derived from software contributed to The NetBSD Foundation
|
|
* by Charles M. Hannum and by Jason R. Thorpe of the Numerical Aerospace
|
|
* Simulation Facility, NASA Ames Research Center.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
|
|
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
|
|
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
|
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
|
|
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
/*
|
|
* Originally written by Julian Elischer (julian@tfs.com)
|
|
* for TRW Financial Systems for use under the MACH(2.5) operating system.
|
|
*
|
|
* TRW Financial Systems, in accordance with their agreement with Carnegie
|
|
* Mellon University, makes this software available to CMU to distribute
|
|
* or use in any manner that they see fit as long as this message is kept with
|
|
* the software. For this reason TFS also grants any other persons or
|
|
* organisations permission to use or modify this software.
|
|
*
|
|
* TFS supplies this software to be publicly redistributed
|
|
* on the understanding that TFS is not responsible for the correct
|
|
* functioning of this software in any circumstances.
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__KERNEL_RCSID(0, "$NetBSD: bha.c,v 1.75 2012/10/27 17:18:19 chs Exp $");
|
|
|
|
#include "opt_ddb.h"
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/callout.h>
|
|
#include <sys/kernel.h>
|
|
#include <sys/errno.h>
|
|
#include <sys/ioctl.h>
|
|
#include <sys/device.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/buf.h>
|
|
#include <sys/proc.h>
|
|
|
|
#include <sys/bus.h>
|
|
#include <sys/intr.h>
|
|
|
|
#include <dev/scsipi/scsi_all.h>
|
|
#include <dev/scsipi/scsipi_all.h>
|
|
#include <dev/scsipi/scsiconf.h>
|
|
|
|
#include <dev/ic/bhareg.h>
|
|
#include <dev/ic/bhavar.h>
|
|
|
|
#ifndef DDB
|
|
#define Debugger() panic("should call debugger here (bha.c)")
|
|
#endif /* ! DDB */
|
|
|
|
#define BHA_MAXXFER ((BHA_NSEG - 1) << PGSHIFT)
|
|
|
|
#ifdef BHADEBUG
|
|
int bha_debug = 0;
|
|
#endif /* BHADEBUG */
|
|
|
|
static int bha_cmd(bus_space_tag_t, bus_space_handle_t, const char *, int,
|
|
u_char *, int, u_char *);
|
|
|
|
static void bha_scsipi_request(struct scsipi_channel *,
|
|
scsipi_adapter_req_t, void *);
|
|
static void bha_minphys(struct buf *);
|
|
|
|
static void bha_get_xfer_mode(struct bha_softc *,
|
|
struct scsipi_xfer_mode *);
|
|
|
|
static void bha_done(struct bha_softc *, struct bha_ccb *);
|
|
static int bha_poll(struct bha_softc *, struct scsipi_xfer *, int);
|
|
static void bha_timeout(void *arg);
|
|
|
|
static int bha_init(struct bha_softc *);
|
|
|
|
static int bha_create_mailbox(struct bha_softc *);
|
|
static void bha_collect_mbo(struct bha_softc *);
|
|
|
|
static void bha_queue_ccb(struct bha_softc *, struct bha_ccb *);
|
|
static void bha_start_ccbs(struct bha_softc *);
|
|
static void bha_finish_ccbs(struct bha_softc *);
|
|
|
|
static struct bha_ccb *bha_ccb_phys_kv(struct bha_softc *, bus_addr_t);
|
|
static void bha_create_ccbs(struct bha_softc *, int);
|
|
static int bha_init_ccb(struct bha_softc *, struct bha_ccb *);
|
|
static struct bha_ccb *bha_get_ccb(struct bha_softc *);
|
|
static void bha_free_ccb(struct bha_softc *, struct bha_ccb *);
|
|
|
|
#define BHA_RESET_TIMEOUT 2000 /* time to wait for reset (mSec) */
|
|
#define BHA_ABORT_TIMEOUT 2000 /* time to wait for abort (mSec) */
|
|
|
|
/*
|
|
* Number of CCBs in an allocation group; must be computed at run-time.
|
|
*/
|
|
static int bha_ccbs_per_group;
|
|
|
|
static inline struct bha_mbx_out *
|
|
bha_nextmbo(struct bha_softc *sc, struct bha_mbx_out *mbo)
|
|
{
|
|
|
|
if (mbo == &sc->sc_mbo[sc->sc_mbox_count - 1])
|
|
return (&sc->sc_mbo[0]);
|
|
return (mbo + 1);
|
|
}
|
|
|
|
static inline struct bha_mbx_in *
|
|
bha_nextmbi(struct bha_softc *sc, struct bha_mbx_in *mbi)
|
|
{
|
|
if (mbi == &sc->sc_mbi[sc->sc_mbox_count - 1])
|
|
return (&sc->sc_mbi[0]);
|
|
return (mbi + 1);
|
|
}
|
|
|
|
/*
|
|
* bha_attach:
|
|
*
|
|
* Finish attaching a Buslogic controller, and configure children.
|
|
*/
|
|
void
|
|
bha_attach(struct bha_softc *sc)
|
|
{
|
|
struct scsipi_adapter *adapt = &sc->sc_adapter;
|
|
struct scsipi_channel *chan = &sc->sc_channel;
|
|
int initial_ccbs;
|
|
|
|
/*
|
|
* Initialize the number of CCBs per group.
|
|
*/
|
|
if (bha_ccbs_per_group == 0)
|
|
bha_ccbs_per_group = BHA_CCBS_PER_GROUP;
|
|
|
|
initial_ccbs = bha_info(sc);
|
|
if (initial_ccbs == 0) {
|
|
aprint_error_dev(sc->sc_dev, "unable to get adapter info\n");
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Fill in the scsipi_adapter.
|
|
*/
|
|
memset(adapt, 0, sizeof(*adapt));
|
|
adapt->adapt_dev = sc->sc_dev;
|
|
adapt->adapt_nchannels = 1;
|
|
/* adapt_openings initialized below */
|
|
adapt->adapt_max_periph = sc->sc_mbox_count;
|
|
adapt->adapt_request = bha_scsipi_request;
|
|
adapt->adapt_minphys = bha_minphys;
|
|
|
|
/*
|
|
* Fill in the scsipi_channel.
|
|
*/
|
|
memset(chan, 0, sizeof(*chan));
|
|
chan->chan_adapter = adapt;
|
|
chan->chan_bustype = &scsi_bustype;
|
|
chan->chan_channel = 0;
|
|
chan->chan_flags = SCSIPI_CHAN_CANGROW;
|
|
chan->chan_ntargets = (sc->sc_flags & BHAF_WIDE) ? 16 : 8;
|
|
chan->chan_nluns = (sc->sc_flags & BHAF_WIDE_LUN) ? 32 : 8;
|
|
chan->chan_id = sc->sc_scsi_id;
|
|
|
|
TAILQ_INIT(&sc->sc_free_ccb);
|
|
TAILQ_INIT(&sc->sc_waiting_ccb);
|
|
TAILQ_INIT(&sc->sc_allocating_ccbs);
|
|
|
|
if (bha_create_mailbox(sc) != 0)
|
|
return;
|
|
|
|
bha_create_ccbs(sc, initial_ccbs);
|
|
if (sc->sc_cur_ccbs < 2) {
|
|
aprint_error_dev(sc->sc_dev, "not enough CCBs to run\n");
|
|
return;
|
|
}
|
|
|
|
adapt->adapt_openings = sc->sc_cur_ccbs;
|
|
|
|
if (bha_init(sc) != 0)
|
|
return;
|
|
|
|
(void) config_found(sc->sc_dev, &sc->sc_channel, scsiprint);
|
|
}
|
|
|
|
/*
|
|
* bha_intr:
|
|
*
|
|
* Interrupt service routine.
|
|
*/
|
|
int
|
|
bha_intr(void *arg)
|
|
{
|
|
struct bha_softc *sc = arg;
|
|
bus_space_tag_t iot = sc->sc_iot;
|
|
bus_space_handle_t ioh = sc->sc_ioh;
|
|
u_char sts;
|
|
|
|
#ifdef BHADEBUG
|
|
printf("%s: bha_intr ", device_xname(sc->sc_dev));
|
|
#endif /* BHADEBUG */
|
|
|
|
/*
|
|
* First acknowledge the interrupt, Then if it's not telling about
|
|
* a completed operation just return.
|
|
*/
|
|
sts = bus_space_read_1(iot, ioh, BHA_INTR_PORT);
|
|
if ((sts & BHA_INTR_ANYINTR) == 0)
|
|
return (0);
|
|
bus_space_write_1(iot, ioh, BHA_CTRL_PORT, BHA_CTRL_IRST);
|
|
|
|
#ifdef BHADIAG
|
|
/* Make sure we clear CCB_SENDING before finishing a CCB. */
|
|
bha_collect_mbo(sc);
|
|
#endif
|
|
|
|
/* Mail box out empty? */
|
|
if (sts & BHA_INTR_MBOA) {
|
|
struct bha_toggle toggle;
|
|
|
|
toggle.cmd.opcode = BHA_MBO_INTR_EN;
|
|
toggle.cmd.enable = 0;
|
|
bha_cmd(iot, ioh, device_xname(sc->sc_dev),
|
|
sizeof(toggle.cmd), (u_char *)&toggle.cmd,
|
|
0, (u_char *)0);
|
|
bha_start_ccbs(sc);
|
|
}
|
|
|
|
/* Mail box in full? */
|
|
if (sts & BHA_INTR_MBIF)
|
|
bha_finish_ccbs(sc);
|
|
|
|
return (1);
|
|
}
|
|
|
|
/*****************************************************************************
|
|
* SCSI interface routines
|
|
*****************************************************************************/
|
|
|
|
/*
|
|
* bha_scsipi_request:
|
|
*
|
|
* Perform a request for the SCSIPI layer.
|
|
*/
|
|
static void
|
|
bha_scsipi_request(struct scsipi_channel *chan, scsipi_adapter_req_t req,
|
|
void *arg)
|
|
{
|
|
struct scsipi_adapter *adapt = chan->chan_adapter;
|
|
struct bha_softc *sc = device_private(adapt->adapt_dev);
|
|
struct scsipi_xfer *xs;
|
|
struct scsipi_periph *periph;
|
|
bus_dma_tag_t dmat = sc->sc_dmat;
|
|
struct bha_ccb *ccb;
|
|
int error, seg, flags, s;
|
|
|
|
switch (req) {
|
|
case ADAPTER_REQ_RUN_XFER:
|
|
xs = arg;
|
|
periph = xs->xs_periph;
|
|
flags = xs->xs_control;
|
|
|
|
SC_DEBUG(periph, SCSIPI_DB2, ("bha_scsipi_request\n"));
|
|
|
|
/* Get a CCB to use. */
|
|
ccb = bha_get_ccb(sc);
|
|
#ifdef DIAGNOSTIC
|
|
/*
|
|
* This should never happen as we track the resources
|
|
* in the mid-layer.
|
|
*/
|
|
if (ccb == NULL) {
|
|
scsipi_printaddr(periph);
|
|
printf("unable to allocate ccb\n");
|
|
panic("bha_scsipi_request");
|
|
}
|
|
#endif
|
|
|
|
ccb->xs = xs;
|
|
ccb->timeout = xs->timeout;
|
|
|
|
/*
|
|
* Put all the arguments for the xfer in the ccb
|
|
*/
|
|
if (flags & XS_CTL_RESET) {
|
|
ccb->opcode = BHA_RESET_CCB;
|
|
ccb->scsi_cmd_length = 0;
|
|
} else {
|
|
/* can't use S/G if zero length */
|
|
if (xs->cmdlen > sizeof(ccb->scsi_cmd)) {
|
|
printf("%s: cmdlen %d too large for CCB\n",
|
|
device_xname(sc->sc_dev), xs->cmdlen);
|
|
xs->error = XS_DRIVER_STUFFUP;
|
|
goto out_bad;
|
|
}
|
|
ccb->opcode = (xs->datalen ? BHA_INIT_SCAT_GATH_CCB
|
|
: BHA_INITIATOR_CCB);
|
|
memcpy(&ccb->scsi_cmd, xs->cmd,
|
|
ccb->scsi_cmd_length = xs->cmdlen);
|
|
}
|
|
|
|
if (xs->datalen) {
|
|
/*
|
|
* Map the DMA transfer.
|
|
*/
|
|
#ifdef TFS
|
|
if (flags & XS_CTL_DATA_UIO) {
|
|
error = bus_dmamap_load_uio(dmat,
|
|
ccb->dmamap_xfer, (struct uio *)xs->data,
|
|
((flags & XS_CTL_NOSLEEP) ? BUS_DMA_NOWAIT :
|
|
BUS_DMA_WAITOK) | BUS_DMA_STREAMING |
|
|
((flags & XS_CTL_DATA_IN) ? BUS_DMA_READ :
|
|
BUS_DMA_WRITE));
|
|
} else
|
|
#endif /* TFS */
|
|
{
|
|
error = bus_dmamap_load(dmat,
|
|
ccb->dmamap_xfer, xs->data, xs->datalen,
|
|
NULL,
|
|
((flags & XS_CTL_NOSLEEP) ? BUS_DMA_NOWAIT :
|
|
BUS_DMA_WAITOK) | BUS_DMA_STREAMING |
|
|
((flags & XS_CTL_DATA_IN) ? BUS_DMA_READ :
|
|
BUS_DMA_WRITE));
|
|
}
|
|
|
|
switch (error) {
|
|
case 0:
|
|
break;
|
|
|
|
case ENOMEM:
|
|
case EAGAIN:
|
|
xs->error = XS_RESOURCE_SHORTAGE;
|
|
goto out_bad;
|
|
|
|
default:
|
|
xs->error = XS_DRIVER_STUFFUP;
|
|
aprint_error_dev(sc->sc_dev, "error %d loading DMA map\n", error);
|
|
out_bad:
|
|
bha_free_ccb(sc, ccb);
|
|
scsipi_done(xs);
|
|
return;
|
|
}
|
|
|
|
bus_dmamap_sync(dmat, ccb->dmamap_xfer, 0,
|
|
ccb->dmamap_xfer->dm_mapsize,
|
|
(flags & XS_CTL_DATA_IN) ? BUS_DMASYNC_PREREAD :
|
|
BUS_DMASYNC_PREWRITE);
|
|
|
|
/*
|
|
* Load the hardware scatter/gather map with the
|
|
* contents of the DMA map.
|
|
*/
|
|
for (seg = 0; seg < ccb->dmamap_xfer->dm_nsegs; seg++) {
|
|
ltophys(ccb->dmamap_xfer->dm_segs[seg].ds_addr,
|
|
ccb->scat_gath[seg].seg_addr);
|
|
ltophys(ccb->dmamap_xfer->dm_segs[seg].ds_len,
|
|
ccb->scat_gath[seg].seg_len);
|
|
}
|
|
|
|
ltophys(ccb->hashkey + offsetof(struct bha_ccb,
|
|
scat_gath), ccb->data_addr);
|
|
ltophys(ccb->dmamap_xfer->dm_nsegs *
|
|
sizeof(struct bha_scat_gath), ccb->data_length);
|
|
} else {
|
|
/*
|
|
* No data xfer, use non S/G values.
|
|
*/
|
|
ltophys(0, ccb->data_addr);
|
|
ltophys(0, ccb->data_length);
|
|
}
|
|
|
|
if (XS_CTL_TAGTYPE(xs) != 0) {
|
|
ccb->tag_enable = 1;
|
|
ccb->tag_type = xs->xs_tag_type & 0x03;
|
|
} else {
|
|
ccb->tag_enable = 0;
|
|
ccb->tag_type = 0;
|
|
}
|
|
|
|
ccb->data_out = 0;
|
|
ccb->data_in = 0;
|
|
ccb->target = periph->periph_target;
|
|
ccb->lun = periph->periph_lun;
|
|
ltophys(ccb->hashkey + offsetof(struct bha_ccb, scsi_sense),
|
|
ccb->sense_ptr);
|
|
ccb->req_sense_length = sizeof(ccb->scsi_sense);
|
|
ccb->host_stat = 0x00;
|
|
ccb->target_stat = 0x00;
|
|
ccb->link_id = 0;
|
|
ltophys(0, ccb->link_addr);
|
|
|
|
BHA_CCB_SYNC(sc, ccb, BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
|
|
|
|
s = splbio();
|
|
bha_queue_ccb(sc, ccb);
|
|
splx(s);
|
|
|
|
SC_DEBUG(periph, SCSIPI_DB3, ("cmd_sent\n"));
|
|
if ((flags & XS_CTL_POLL) == 0)
|
|
return;
|
|
|
|
/*
|
|
* If we can't use interrupts, poll on completion
|
|
*/
|
|
if (bha_poll(sc, xs, ccb->timeout)) {
|
|
bha_timeout(ccb);
|
|
if (bha_poll(sc, xs, ccb->timeout))
|
|
bha_timeout(ccb);
|
|
}
|
|
return;
|
|
|
|
case ADAPTER_REQ_GROW_RESOURCES:
|
|
if (sc->sc_cur_ccbs == sc->sc_max_ccbs) {
|
|
chan->chan_flags &= ~SCSIPI_CHAN_CANGROW;
|
|
return;
|
|
}
|
|
seg = sc->sc_cur_ccbs;
|
|
bha_create_ccbs(sc, bha_ccbs_per_group);
|
|
adapt->adapt_openings += sc->sc_cur_ccbs - seg;
|
|
return;
|
|
|
|
case ADAPTER_REQ_SET_XFER_MODE:
|
|
/*
|
|
* Can't really do this on the Buslogic. It has its
|
|
* own setup info. But we do know how to query what
|
|
* the settings are.
|
|
*/
|
|
bha_get_xfer_mode(sc, (struct scsipi_xfer_mode *)arg);
|
|
return;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* bha_minphys:
|
|
*
|
|
* Limit a transfer to our maximum transfer size.
|
|
*/
|
|
void
|
|
bha_minphys(struct buf *bp)
|
|
{
|
|
|
|
if (bp->b_bcount > BHA_MAXXFER)
|
|
bp->b_bcount = BHA_MAXXFER;
|
|
minphys(bp);
|
|
}
|
|
|
|
/*****************************************************************************
|
|
* SCSI job execution helper routines
|
|
*****************************************************************************/
|
|
|
|
/*
|
|
* bha_get_xfer_mode;
|
|
*
|
|
* Negotiate the xfer mode for the specified periph, and report
|
|
* back the mode to the midlayer.
|
|
*
|
|
* NOTE: we must be called at splbio().
|
|
*/
|
|
static void
|
|
bha_get_xfer_mode(struct bha_softc *sc, struct scsipi_xfer_mode *xm)
|
|
{
|
|
struct bha_setup hwsetup;
|
|
struct bha_period hwperiod;
|
|
struct bha_sync *bs;
|
|
int toff = xm->xm_target & 7, tmask = (1 << toff);
|
|
int wide, period, offset, rlen;
|
|
|
|
/*
|
|
* Issue an Inquire Setup Information. We can extract
|
|
* sync and wide information from here.
|
|
*/
|
|
rlen = sizeof(hwsetup.reply) +
|
|
((sc->sc_flags & BHAF_WIDE) ? sizeof(hwsetup.reply_w) : 0);
|
|
hwsetup.cmd.opcode = BHA_INQUIRE_SETUP;
|
|
hwsetup.cmd.len = rlen;
|
|
bha_cmd(sc->sc_iot, sc->sc_ioh, device_xname(sc->sc_dev),
|
|
sizeof(hwsetup.cmd), (u_char *)&hwsetup.cmd,
|
|
rlen, (u_char *)&hwsetup.reply);
|
|
|
|
xm->xm_mode = 0;
|
|
xm->xm_period = 0;
|
|
xm->xm_offset = 0;
|
|
|
|
/*
|
|
* First check for wide. On later boards, we can check
|
|
* directly in the setup info if wide is currently active.
|
|
*
|
|
* On earlier boards, we have to make an educated guess.
|
|
*/
|
|
if (sc->sc_flags & BHAF_WIDE) {
|
|
if (strcmp(sc->sc_firmware, "5.06L") >= 0) {
|
|
if (xm->xm_target > 7) {
|
|
wide =
|
|
hwsetup.reply_w.high_wide_active & tmask;
|
|
} else {
|
|
wide =
|
|
hwsetup.reply_w.low_wide_active & tmask;
|
|
}
|
|
if (wide)
|
|
xm->xm_mode |= PERIPH_CAP_WIDE16;
|
|
} else {
|
|
/* XXX Check `wide permitted' in the config info. */
|
|
xm->xm_mode |= PERIPH_CAP_WIDE16;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Now get basic sync info.
|
|
*/
|
|
bs = (xm->xm_target > 7) ?
|
|
&hwsetup.reply_w.sync_high[toff] :
|
|
&hwsetup.reply.sync_low[toff];
|
|
|
|
if (bs->valid) {
|
|
xm->xm_mode |= PERIPH_CAP_SYNC;
|
|
period = (bs->period * 50) + 20;
|
|
offset = bs->offset;
|
|
|
|
/*
|
|
* On boards that can do Fast and Ultra, use the Inquire Period
|
|
* command to get the period.
|
|
*/
|
|
if (sc->sc_firmware[0] >= '3') {
|
|
rlen = sizeof(hwperiod.reply) +
|
|
((sc->sc_flags & BHAF_WIDE) ?
|
|
sizeof(hwperiod.reply_w) : 0);
|
|
hwperiod.cmd.opcode = BHA_INQUIRE_PERIOD;
|
|
hwperiod.cmd.len = rlen;
|
|
bha_cmd(sc->sc_iot, sc->sc_ioh, device_xname(sc->sc_dev),
|
|
sizeof(hwperiod.cmd), (u_char *)&hwperiod.cmd,
|
|
rlen, (u_char *)&hwperiod.reply);
|
|
|
|
if (xm->xm_target > 7)
|
|
period = hwperiod.reply_w.period[toff];
|
|
else
|
|
period = hwperiod.reply.period[toff];
|
|
|
|
period *= 10;
|
|
}
|
|
|
|
xm->xm_period =
|
|
scsipi_sync_period_to_factor(period * 100);
|
|
xm->xm_offset = offset;
|
|
}
|
|
|
|
/*
|
|
* Now check for tagged queueing support.
|
|
*
|
|
* XXX Check `tags permitted' in the config info.
|
|
*/
|
|
if (sc->sc_flags & BHAF_TAGGED_QUEUEING)
|
|
xm->xm_mode |= PERIPH_CAP_TQING;
|
|
|
|
scsipi_async_event(&sc->sc_channel, ASYNC_EVENT_XFER_MODE, xm);
|
|
}
|
|
|
|
/*
|
|
* bha_done:
|
|
*
|
|
* A CCB has completed execution. Pass the status back to the
|
|
* upper layer.
|
|
*/
|
|
static void
|
|
bha_done(struct bha_softc *sc, struct bha_ccb *ccb)
|
|
{
|
|
bus_dma_tag_t dmat = sc->sc_dmat;
|
|
struct scsipi_xfer *xs = ccb->xs;
|
|
|
|
SC_DEBUG(xs->xs_periph, SCSIPI_DB2, ("bha_done\n"));
|
|
|
|
#ifdef BHADIAG
|
|
if (ccb->flags & CCB_SENDING) {
|
|
printf("%s: exiting ccb still in transit!\n",
|
|
device_xname(sc->sc_dev));
|
|
Debugger();
|
|
return;
|
|
}
|
|
#endif
|
|
if ((ccb->flags & CCB_ALLOC) == 0) {
|
|
aprint_error_dev(sc->sc_dev, "exiting ccb not allocated!\n");
|
|
Debugger();
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* If we were a data transfer, unload the map that described
|
|
* the data buffer.
|
|
*/
|
|
if (xs->datalen) {
|
|
bus_dmamap_sync(dmat, ccb->dmamap_xfer, 0,
|
|
ccb->dmamap_xfer->dm_mapsize,
|
|
(xs->xs_control & XS_CTL_DATA_IN) ? BUS_DMASYNC_POSTREAD :
|
|
BUS_DMASYNC_POSTWRITE);
|
|
bus_dmamap_unload(dmat, ccb->dmamap_xfer);
|
|
}
|
|
|
|
if (xs->error == XS_NOERROR) {
|
|
if (ccb->host_stat != BHA_OK) {
|
|
switch (ccb->host_stat) {
|
|
case BHA_SEL_TIMEOUT: /* No response */
|
|
xs->error = XS_SELTIMEOUT;
|
|
break;
|
|
default: /* Other scsi protocol messes */
|
|
printf("%s: host_stat %x\n",
|
|
device_xname(sc->sc_dev), ccb->host_stat);
|
|
xs->error = XS_DRIVER_STUFFUP;
|
|
break;
|
|
}
|
|
} else if (ccb->target_stat != SCSI_OK) {
|
|
switch (ccb->target_stat) {
|
|
case SCSI_CHECK:
|
|
memcpy(&xs->sense.scsi_sense,
|
|
&ccb->scsi_sense,
|
|
sizeof(xs->sense.scsi_sense));
|
|
xs->error = XS_SENSE;
|
|
break;
|
|
case SCSI_BUSY:
|
|
xs->error = XS_BUSY;
|
|
break;
|
|
default:
|
|
printf("%s: target_stat %x\n",
|
|
device_xname(sc->sc_dev), ccb->target_stat);
|
|
xs->error = XS_DRIVER_STUFFUP;
|
|
break;
|
|
}
|
|
} else
|
|
xs->resid = 0;
|
|
}
|
|
|
|
bha_free_ccb(sc, ccb);
|
|
scsipi_done(xs);
|
|
}
|
|
|
|
/*
|
|
* bha_poll:
|
|
*
|
|
* Poll for completion of the specified job.
|
|
*/
|
|
static int
|
|
bha_poll(struct bha_softc *sc, struct scsipi_xfer *xs, int count)
|
|
{
|
|
bus_space_tag_t iot = sc->sc_iot;
|
|
bus_space_handle_t ioh = sc->sc_ioh;
|
|
|
|
/* timeouts are in msec, so we loop in 1000 usec cycles */
|
|
while (count) {
|
|
/*
|
|
* If we had interrupts enabled, would we
|
|
* have got an interrupt?
|
|
*/
|
|
if (bus_space_read_1(iot, ioh, BHA_INTR_PORT) &
|
|
BHA_INTR_ANYINTR)
|
|
bha_intr(sc);
|
|
if (xs->xs_status & XS_STS_DONE)
|
|
return (0);
|
|
delay(1000); /* only happens in boot so ok */
|
|
count--;
|
|
}
|
|
return (1);
|
|
}
|
|
|
|
/*
|
|
* bha_timeout:
|
|
*
|
|
* CCB timeout handler.
|
|
*/
|
|
static void
|
|
bha_timeout(void *arg)
|
|
{
|
|
struct bha_ccb *ccb = arg;
|
|
struct scsipi_xfer *xs = ccb->xs;
|
|
struct scsipi_periph *periph = xs->xs_periph;
|
|
struct bha_softc *sc =
|
|
device_private(periph->periph_channel->chan_adapter->adapt_dev);
|
|
int s;
|
|
|
|
scsipi_printaddr(periph);
|
|
printf("timed out");
|
|
|
|
s = splbio();
|
|
|
|
#ifdef BHADIAG
|
|
/*
|
|
* If the ccb's mbx is not free, then the board has gone Far East?
|
|
*/
|
|
bha_collect_mbo(sc);
|
|
if (ccb->flags & CCB_SENDING) {
|
|
aprint_error_dev(sc->sc_dev, "not taking commands!\n");
|
|
Debugger();
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* If it has been through before, then
|
|
* a previous abort has failed, don't
|
|
* try abort again
|
|
*/
|
|
if (ccb->flags & CCB_ABORT) {
|
|
/* abort timed out */
|
|
printf(" AGAIN\n");
|
|
/* XXX Must reset! */
|
|
} else {
|
|
/* abort the operation that has timed out */
|
|
printf("\n");
|
|
ccb->xs->error = XS_TIMEOUT;
|
|
ccb->timeout = BHA_ABORT_TIMEOUT;
|
|
ccb->flags |= CCB_ABORT;
|
|
bha_queue_ccb(sc, ccb);
|
|
}
|
|
|
|
splx(s);
|
|
}
|
|
|
|
/*****************************************************************************
|
|
* Misc. subroutines.
|
|
*****************************************************************************/
|
|
|
|
/*
|
|
* bha_cmd:
|
|
*
|
|
* Send a command to the Buglogic controller.
|
|
*/
|
|
static int
|
|
bha_cmd(bus_space_tag_t iot, bus_space_handle_t ioh, const char *name, int icnt,
|
|
u_char *ibuf, int ocnt, u_char *obuf)
|
|
{
|
|
int i;
|
|
int wait;
|
|
u_char sts;
|
|
u_char opcode = ibuf[0];
|
|
|
|
/*
|
|
* Calculate a reasonable timeout for the command.
|
|
*/
|
|
switch (opcode) {
|
|
case BHA_INQUIRE_DEVICES:
|
|
case BHA_INQUIRE_DEVICES_2:
|
|
wait = 90 * 20000;
|
|
break;
|
|
default:
|
|
wait = 1 * 20000;
|
|
break;
|
|
}
|
|
|
|
/*
|
|
* Wait for the adapter to go idle, unless it's one of
|
|
* the commands which don't need this
|
|
*/
|
|
if (opcode != BHA_MBO_INTR_EN) {
|
|
for (i = 20000; i; i--) { /* 1 sec? */
|
|
sts = bus_space_read_1(iot, ioh, BHA_STAT_PORT);
|
|
if (sts & BHA_STAT_IDLE)
|
|
break;
|
|
delay(50);
|
|
}
|
|
if (!i) {
|
|
printf("%s: bha_cmd, host not idle(0x%x)\n",
|
|
name, sts);
|
|
return (1);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Now that it is idle, if we expect output, preflush the
|
|
* queue feeding to us.
|
|
*/
|
|
if (ocnt) {
|
|
while ((bus_space_read_1(iot, ioh, BHA_STAT_PORT)) &
|
|
BHA_STAT_DF)
|
|
(void)bus_space_read_1(iot, ioh, BHA_DATA_PORT);
|
|
}
|
|
|
|
/*
|
|
* Output the command and the number of arguments given
|
|
* for each byte, first check the port is empty.
|
|
*/
|
|
while (icnt--) {
|
|
for (i = wait; i; i--) {
|
|
sts = bus_space_read_1(iot, ioh, BHA_STAT_PORT);
|
|
if (!(sts & BHA_STAT_CDF))
|
|
break;
|
|
delay(50);
|
|
}
|
|
if (!i) {
|
|
if (opcode != BHA_INQUIRE_REVISION)
|
|
printf("%s: bha_cmd, cmd/data port full\n",
|
|
name);
|
|
goto bad;
|
|
}
|
|
bus_space_write_1(iot, ioh, BHA_CMD_PORT, *ibuf++);
|
|
}
|
|
|
|
/*
|
|
* If we expect input, loop that many times, each time,
|
|
* looking for the data register to have valid data
|
|
*/
|
|
while (ocnt--) {
|
|
for (i = wait; i; i--) {
|
|
sts = bus_space_read_1(iot, ioh, BHA_STAT_PORT);
|
|
if (sts & BHA_STAT_DF)
|
|
break;
|
|
delay(50);
|
|
}
|
|
if (!i) {
|
|
#ifdef BHADEBUG
|
|
if (opcode != BHA_INQUIRE_REVISION)
|
|
printf("%s: bha_cmd, cmd/data port empty %d\n",
|
|
name, ocnt);
|
|
#endif /* BHADEBUG */
|
|
goto bad;
|
|
}
|
|
*obuf++ = bus_space_read_1(iot, ioh, BHA_DATA_PORT);
|
|
}
|
|
|
|
/*
|
|
* Wait for the board to report a finished instruction.
|
|
* We may get an extra interrupt for the HACC signal, but this is
|
|
* unimportant.
|
|
*/
|
|
if (opcode != BHA_MBO_INTR_EN && opcode != BHA_MODIFY_IOPORT) {
|
|
for (i = 20000; i; i--) { /* 1 sec? */
|
|
sts = bus_space_read_1(iot, ioh, BHA_INTR_PORT);
|
|
/* XXX Need to save this in the interrupt handler? */
|
|
if (sts & BHA_INTR_HACC)
|
|
break;
|
|
delay(50);
|
|
}
|
|
if (!i) {
|
|
printf("%s: bha_cmd, host not finished(0x%x)\n",
|
|
name, sts);
|
|
return (1);
|
|
}
|
|
}
|
|
bus_space_write_1(iot, ioh, BHA_CTRL_PORT, BHA_CTRL_IRST);
|
|
return (0);
|
|
|
|
bad:
|
|
bus_space_write_1(iot, ioh, BHA_CTRL_PORT, BHA_CTRL_SRST);
|
|
return (1);
|
|
}
|
|
|
|
/*
|
|
* bha_find:
|
|
*
|
|
* Find the board.
|
|
*/
|
|
int
|
|
bha_find(bus_space_tag_t iot, bus_space_handle_t ioh)
|
|
{
|
|
int i;
|
|
u_char sts;
|
|
struct bha_extended_inquire inquire;
|
|
|
|
/* Check something is at the ports we need to access */
|
|
sts = bus_space_read_1(iot, ioh, BHA_STAT_PORT);
|
|
if (sts == 0xFF)
|
|
return (0);
|
|
|
|
/*
|
|
* Reset board, If it doesn't respond, assume
|
|
* that it's not there.. good for the probe
|
|
*/
|
|
|
|
bus_space_write_1(iot, ioh, BHA_CTRL_PORT,
|
|
BHA_CTRL_HRST | BHA_CTRL_SRST);
|
|
|
|
delay(100);
|
|
for (i = BHA_RESET_TIMEOUT; i; i--) {
|
|
sts = bus_space_read_1(iot, ioh, BHA_STAT_PORT);
|
|
if (sts == (BHA_STAT_IDLE | BHA_STAT_INIT))
|
|
break;
|
|
delay(1000);
|
|
}
|
|
if (!i) {
|
|
#ifdef BHADEBUG
|
|
if (bha_debug)
|
|
printf("bha_find: No answer from buslogic board\n");
|
|
#endif /* BHADEBUG */
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* The BusLogic cards implement an Adaptec 1542 (aha)-compatible
|
|
* interface. The native bha interface is not compatible with
|
|
* an aha. 1542. We need to ensure that we never match an
|
|
* Adaptec 1542. We must also avoid sending Adaptec-compatible
|
|
* commands to a real bha, lest it go into 1542 emulation mode.
|
|
* (On an indirect bus like ISA, we should always probe for BusLogic
|
|
* interfaces before Adaptec interfaces).
|
|
*/
|
|
|
|
/*
|
|
* Make sure we don't match an AHA-1542A or AHA-1542B, by checking
|
|
* for an extended-geometry register. The 1542[AB] don't have one.
|
|
*/
|
|
sts = bus_space_read_1(iot, ioh, BHA_EXTGEOM_PORT);
|
|
if (sts == 0xFF)
|
|
return (0);
|
|
|
|
/*
|
|
* Check that we actually know how to use this board.
|
|
*/
|
|
delay(1000);
|
|
inquire.cmd.opcode = BHA_INQUIRE_EXTENDED;
|
|
inquire.cmd.len = sizeof(inquire.reply);
|
|
i = bha_cmd(iot, ioh, "(bha_find)",
|
|
sizeof(inquire.cmd), (u_char *)&inquire.cmd,
|
|
sizeof(inquire.reply), (u_char *)&inquire.reply);
|
|
|
|
/*
|
|
* Some 1542Cs (CP, perhaps not CF, may depend on firmware rev)
|
|
* have the extended-geometry register and also respond to
|
|
* BHA_INQUIRE_EXTENDED. Make sure we never match such cards,
|
|
* by checking the size of the reply is what a BusLogic card returns.
|
|
*/
|
|
if (i) {
|
|
#ifdef BHADEBUG
|
|
printf("bha_find: board returned %d instead of %zu to %s\n",
|
|
i, sizeof(inquire.reply), "INQUIRE_EXTENDED");
|
|
#endif
|
|
return (0);
|
|
}
|
|
|
|
/* OK, we know we've found a buslogic adaptor. */
|
|
|
|
switch (inquire.reply.bus_type) {
|
|
case BHA_BUS_TYPE_24BIT:
|
|
case BHA_BUS_TYPE_32BIT:
|
|
break;
|
|
case BHA_BUS_TYPE_MCA:
|
|
/* We don't grok MicroChannel (yet). */
|
|
return (0);
|
|
default:
|
|
printf("bha_find: illegal bus type %c\n",
|
|
inquire.reply.bus_type);
|
|
return (0);
|
|
}
|
|
|
|
return (1);
|
|
}
|
|
|
|
|
|
/*
|
|
* bha_inquire_config:
|
|
*
|
|
* Determine irq/drq.
|
|
*/
|
|
int
|
|
bha_inquire_config(bus_space_tag_t iot, bus_space_handle_t ioh,
|
|
struct bha_probe_data *sc)
|
|
{
|
|
int irq, drq;
|
|
struct bha_config config;
|
|
|
|
/*
|
|
* Assume we have a board at this stage setup DMA channel from
|
|
* jumpers and save int level
|
|
*/
|
|
delay(1000);
|
|
config.cmd.opcode = BHA_INQUIRE_CONFIG;
|
|
bha_cmd(iot, ioh, "(bha_inquire_config)",
|
|
sizeof(config.cmd), (u_char *)&config.cmd,
|
|
sizeof(config.reply), (u_char *)&config.reply);
|
|
switch (config.reply.chan) {
|
|
case EISADMA:
|
|
drq = -1;
|
|
break;
|
|
case CHAN0:
|
|
drq = 0;
|
|
break;
|
|
case CHAN5:
|
|
drq = 5;
|
|
break;
|
|
case CHAN6:
|
|
drq = 6;
|
|
break;
|
|
case CHAN7:
|
|
drq = 7;
|
|
break;
|
|
default:
|
|
printf("bha: illegal drq setting %x\n",
|
|
config.reply.chan);
|
|
return (0);
|
|
}
|
|
|
|
switch (config.reply.intr) {
|
|
case INT9:
|
|
irq = 9;
|
|
break;
|
|
case INT10:
|
|
irq = 10;
|
|
break;
|
|
case INT11:
|
|
irq = 11;
|
|
break;
|
|
case INT12:
|
|
irq = 12;
|
|
break;
|
|
case INT14:
|
|
irq = 14;
|
|
break;
|
|
case INT15:
|
|
irq = 15;
|
|
break;
|
|
default:
|
|
printf("bha: illegal irq setting %x\n",
|
|
config.reply.intr);
|
|
return (0);
|
|
}
|
|
|
|
/* if we want to fill in softc, do so now */
|
|
if (sc != NULL) {
|
|
sc->sc_irq = irq;
|
|
sc->sc_drq = drq;
|
|
}
|
|
|
|
return (1);
|
|
}
|
|
|
|
int
|
|
bha_probe_inquiry(bus_space_tag_t iot, bus_space_handle_t ioh,
|
|
struct bha_probe_data *bpd)
|
|
{
|
|
return bha_find(iot, ioh) && bha_inquire_config(iot, ioh, bpd);
|
|
}
|
|
|
|
/*
|
|
* bha_disable_isacompat:
|
|
*
|
|
* Disable the ISA-compatibility ioports on PCI bha devices,
|
|
* to ensure they're not autoconfigured a second time as an ISA bha.
|
|
*/
|
|
int
|
|
bha_disable_isacompat(struct bha_softc *sc)
|
|
{
|
|
struct bha_isadisable isa_disable;
|
|
|
|
isa_disable.cmd.opcode = BHA_MODIFY_IOPORT;
|
|
isa_disable.cmd.modifier = BHA_IOMODIFY_DISABLE1;
|
|
bha_cmd(sc->sc_iot, sc->sc_ioh, device_xname(sc->sc_dev),
|
|
sizeof(isa_disable.cmd), (u_char*)&isa_disable.cmd,
|
|
0, (u_char *)0);
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* bha_info:
|
|
*
|
|
* Get information about the board, and report it. We
|
|
* return the initial number of CCBs, 0 if we failed.
|
|
*/
|
|
int
|
|
bha_info(struct bha_softc *sc)
|
|
{
|
|
bus_space_tag_t iot = sc->sc_iot;
|
|
bus_space_handle_t ioh = sc->sc_ioh;
|
|
struct bha_extended_inquire inquire;
|
|
struct bha_config config;
|
|
struct bha_devices devices;
|
|
struct bha_setup setup;
|
|
struct bha_model model;
|
|
struct bha_revision revision;
|
|
struct bha_digit digit;
|
|
int i, j, initial_ccbs, rlen;
|
|
const char *name = device_xname(sc->sc_dev);
|
|
char *p;
|
|
|
|
/*
|
|
* Fetch the extended inquire information.
|
|
*/
|
|
inquire.cmd.opcode = BHA_INQUIRE_EXTENDED;
|
|
inquire.cmd.len = sizeof(inquire.reply);
|
|
bha_cmd(iot, ioh, name,
|
|
sizeof(inquire.cmd), (u_char *)&inquire.cmd,
|
|
sizeof(inquire.reply), (u_char *)&inquire.reply);
|
|
|
|
/*
|
|
* Fetch the configuration information.
|
|
*/
|
|
config.cmd.opcode = BHA_INQUIRE_CONFIG;
|
|
bha_cmd(iot, ioh, name,
|
|
sizeof(config.cmd), (u_char *)&config.cmd,
|
|
sizeof(config.reply), (u_char *)&config.reply);
|
|
|
|
sc->sc_scsi_id = config.reply.scsi_dev;
|
|
|
|
/*
|
|
* Get the firmware revision.
|
|
*/
|
|
p = sc->sc_firmware;
|
|
revision.cmd.opcode = BHA_INQUIRE_REVISION;
|
|
bha_cmd(iot, ioh, name,
|
|
sizeof(revision.cmd), (u_char *)&revision.cmd,
|
|
sizeof(revision.reply), (u_char *)&revision.reply);
|
|
*p++ = revision.reply.firm_revision;
|
|
*p++ = '.';
|
|
*p++ = revision.reply.firm_version;
|
|
digit.cmd.opcode = BHA_INQUIRE_REVISION_3;
|
|
bha_cmd(iot, ioh, name,
|
|
sizeof(digit.cmd), (u_char *)&digit.cmd,
|
|
sizeof(digit.reply), (u_char *)&digit.reply);
|
|
*p++ = digit.reply.digit;
|
|
if (revision.reply.firm_revision >= '3' ||
|
|
(revision.reply.firm_revision == '3' &&
|
|
revision.reply.firm_version >= '3')) {
|
|
digit.cmd.opcode = BHA_INQUIRE_REVISION_4;
|
|
bha_cmd(iot, ioh, name,
|
|
sizeof(digit.cmd), (u_char *)&digit.cmd,
|
|
sizeof(digit.reply), (u_char *)&digit.reply);
|
|
*p++ = digit.reply.digit;
|
|
}
|
|
while (p > sc->sc_firmware && (p[-1] == ' ' || p[-1] == '\0'))
|
|
p--;
|
|
*p = '\0';
|
|
|
|
/*
|
|
* Get the model number.
|
|
*
|
|
* Some boards do not handle the Inquire Board Model Number
|
|
* command correctly, or don't give correct information.
|
|
*
|
|
* So, we use the Firmware Revision and Extended Setup
|
|
* information to fixup the model number in these cases.
|
|
*
|
|
* The firmware version indicates:
|
|
*
|
|
* 5.xx BusLogic "W" Series Host Adapters
|
|
* BT-948/958/958D
|
|
*
|
|
* 4.xx BusLogic "C" Series Host Adapters
|
|
* BT-946C/956C/956CD/747C/757C/757CD/445C/545C/540CF
|
|
*
|
|
* 3.xx BusLogic "S" Series Host Adapters
|
|
* BT-747S/747D/757S/757D/445S/545S/542D
|
|
* BT-542B/742A (revision H)
|
|
*
|
|
* 2.xx BusLogic "A" Series Host Adapters
|
|
* BT-542B/742A (revision G and below)
|
|
*
|
|
* 0.xx AMI FastDisk VLB/EISA BusLogic Clone Host Adapter
|
|
*/
|
|
if (inquire.reply.bus_type == BHA_BUS_TYPE_24BIT &&
|
|
sc->sc_firmware[0] < '3')
|
|
snprintf(sc->sc_model, sizeof(sc->sc_model), "542B");
|
|
else if (inquire.reply.bus_type == BHA_BUS_TYPE_32BIT &&
|
|
sc->sc_firmware[0] == '2' &&
|
|
(sc->sc_firmware[2] == '1' ||
|
|
(sc->sc_firmware[2] == '2' && sc->sc_firmware[3] == '0')))
|
|
snprintf(sc->sc_model, sizeof(sc->sc_model), "742A");
|
|
else if (inquire.reply.bus_type == BHA_BUS_TYPE_32BIT &&
|
|
sc->sc_firmware[0] == '0')
|
|
snprintf(sc->sc_model, sizeof(sc->sc_model), "747A");
|
|
else {
|
|
p = sc->sc_model;
|
|
model.cmd.opcode = BHA_INQUIRE_MODEL;
|
|
model.cmd.len = sizeof(model.reply);
|
|
bha_cmd(iot, ioh, name,
|
|
sizeof(model.cmd), (u_char *)&model.cmd,
|
|
sizeof(model.reply), (u_char *)&model.reply);
|
|
*p++ = model.reply.id[0];
|
|
*p++ = model.reply.id[1];
|
|
*p++ = model.reply.id[2];
|
|
*p++ = model.reply.id[3];
|
|
while (p > sc->sc_model && (p[-1] == ' ' || p[-1] == '\0'))
|
|
p--;
|
|
*p++ = model.reply.version[0];
|
|
*p++ = model.reply.version[1];
|
|
while (p > sc->sc_model && (p[-1] == ' ' || p[-1] == '\0'))
|
|
p--;
|
|
*p = '\0';
|
|
}
|
|
|
|
/* Enable round-robin scheme - appeared at firmware rev. 3.31. */
|
|
if (strcmp(sc->sc_firmware, "3.31") >= 0)
|
|
sc->sc_flags |= BHAF_STRICT_ROUND_ROBIN;
|
|
|
|
/*
|
|
* Determine some characteristics about our bus.
|
|
*/
|
|
if (inquire.reply.scsi_flags & BHA_SCSI_WIDE)
|
|
sc->sc_flags |= BHAF_WIDE;
|
|
if (inquire.reply.scsi_flags & BHA_SCSI_DIFFERENTIAL)
|
|
sc->sc_flags |= BHAF_DIFFERENTIAL;
|
|
if (inquire.reply.scsi_flags & BHA_SCSI_ULTRA)
|
|
sc->sc_flags |= BHAF_ULTRA;
|
|
|
|
/*
|
|
* Determine some characterists of the board.
|
|
*/
|
|
sc->sc_max_dmaseg = inquire.reply.sg_limit;
|
|
|
|
/*
|
|
* Determine the maximum CCB count and whether or not
|
|
* tagged queueing is available on this host adapter.
|
|
*
|
|
* Tagged queueing works on:
|
|
*
|
|
* "W" Series adapters
|
|
* "C" Series adapters with firmware >= 4.22
|
|
* "S" Series adapters with firmware >= 3.35
|
|
*
|
|
* The internal CCB counts are:
|
|
*
|
|
* 192 BT-948/958/958D
|
|
* 100 BT-946C/956C/956CD/747C/757C/757CD/445C
|
|
* 50 BT-545C/540CF
|
|
* 30 BT-747S/747D/757S/757D/445S/545S/542D/542B/742A
|
|
*/
|
|
switch (sc->sc_firmware[0]) {
|
|
case '5':
|
|
sc->sc_max_ccbs = 192;
|
|
sc->sc_flags |= BHAF_TAGGED_QUEUEING;
|
|
break;
|
|
|
|
case '4':
|
|
if (sc->sc_model[0] == '5')
|
|
sc->sc_max_ccbs = 50;
|
|
else
|
|
sc->sc_max_ccbs = 100;
|
|
if (strcmp(sc->sc_firmware, "4.22") >= 0)
|
|
sc->sc_flags |= BHAF_TAGGED_QUEUEING;
|
|
break;
|
|
|
|
case '3':
|
|
if (strcmp(sc->sc_firmware, "3.35") >= 0)
|
|
sc->sc_flags |= BHAF_TAGGED_QUEUEING;
|
|
/* FALLTHROUGH */
|
|
|
|
default:
|
|
sc->sc_max_ccbs = 30;
|
|
}
|
|
|
|
/*
|
|
* Set the mailbox count to precisely the number of HW CCBs
|
|
* available. A mailbox isn't required while a CCB is executing,
|
|
* but this allows us to actually enqueue up to our resource
|
|
* limit.
|
|
*
|
|
* This will keep the mailbox count small on boards which don't
|
|
* have strict round-robin (they have to scan the entire set of
|
|
* mailboxes each time they run a command).
|
|
*/
|
|
sc->sc_mbox_count = sc->sc_max_ccbs;
|
|
|
|
/*
|
|
* Obtain setup information.
|
|
*/
|
|
rlen = sizeof(setup.reply) +
|
|
((sc->sc_flags & BHAF_WIDE) ? sizeof(setup.reply_w) : 0);
|
|
setup.cmd.opcode = BHA_INQUIRE_SETUP;
|
|
setup.cmd.len = rlen;
|
|
bha_cmd(iot, ioh, name,
|
|
sizeof(setup.cmd), (u_char *)&setup.cmd,
|
|
rlen, (u_char *)&setup.reply);
|
|
|
|
aprint_normal_dev(sc->sc_dev, "model BT-%s, firmware %s\n",
|
|
sc->sc_model, sc->sc_firmware);
|
|
|
|
aprint_normal_dev(sc->sc_dev, "%d H/W CCBs", sc->sc_max_ccbs);
|
|
if (setup.reply.sync_neg)
|
|
aprint_normal(", sync");
|
|
if (setup.reply.parity)
|
|
aprint_normal(", parity");
|
|
if (sc->sc_flags & BHAF_TAGGED_QUEUEING)
|
|
aprint_normal(", tagged queueing");
|
|
if (sc->sc_flags & BHAF_WIDE_LUN)
|
|
aprint_normal(", wide LUN support");
|
|
aprint_normal("\n");
|
|
|
|
/*
|
|
* Poll targets 0 - 7.
|
|
*/
|
|
devices.cmd.opcode = BHA_INQUIRE_DEVICES;
|
|
bha_cmd(iot, ioh, name,
|
|
sizeof(devices.cmd), (u_char *)&devices.cmd,
|
|
sizeof(devices.reply), (u_char *)&devices.reply);
|
|
|
|
/* Count installed units. */
|
|
initial_ccbs = 0;
|
|
for (i = 0; i < 8; i++) {
|
|
for (j = 0; j < 8; j++) {
|
|
if (((devices.reply.lun_map[i] >> j) & 1) == 1)
|
|
initial_ccbs++;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Poll targets 8 - 15 if we have a wide bus.
|
|
*/
|
|
if (sc->sc_flags & BHAF_WIDE) {
|
|
devices.cmd.opcode = BHA_INQUIRE_DEVICES_2;
|
|
bha_cmd(iot, ioh, name,
|
|
sizeof(devices.cmd), (u_char *)&devices.cmd,
|
|
sizeof(devices.reply), (u_char *)&devices.reply);
|
|
|
|
for (i = 0; i < 8; i++) {
|
|
for (j = 0; j < 8; j++) {
|
|
if (((devices.reply.lun_map[i] >> j) & 1) == 1)
|
|
initial_ccbs++;
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Double the initial CCB count, for good measure.
|
|
*/
|
|
initial_ccbs *= 2;
|
|
|
|
/*
|
|
* Sanity check the initial CCB count; don't create more than
|
|
* we can enqueue (sc_max_ccbs), and make sure there are some
|
|
* at all.
|
|
*/
|
|
if (initial_ccbs > sc->sc_max_ccbs)
|
|
initial_ccbs = sc->sc_max_ccbs;
|
|
if (initial_ccbs == 0)
|
|
initial_ccbs = 2;
|
|
|
|
return (initial_ccbs);
|
|
}
|
|
|
|
/*
|
|
* bha_init:
|
|
*
|
|
* Initialize the board.
|
|
*/
|
|
static int
|
|
bha_init(struct bha_softc *sc)
|
|
{
|
|
const char *name = device_xname(sc->sc_dev);
|
|
struct bha_toggle toggle;
|
|
struct bha_mailbox mailbox;
|
|
struct bha_mbx_out *mbo;
|
|
struct bha_mbx_in *mbi;
|
|
int i;
|
|
|
|
/*
|
|
* Set up the mailbox. We always run the mailbox in round-robin.
|
|
*/
|
|
for (i = 0; i < sc->sc_mbox_count; i++) {
|
|
mbo = &sc->sc_mbo[i];
|
|
mbi = &sc->sc_mbi[i];
|
|
|
|
mbo->cmd = BHA_MBO_FREE;
|
|
BHA_MBO_SYNC(sc, mbo, BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
|
|
|
|
mbi->comp_stat = BHA_MBI_FREE;
|
|
BHA_MBI_SYNC(sc, mbi, BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
|
|
}
|
|
|
|
sc->sc_cmbo = sc->sc_tmbo = &sc->sc_mbo[0];
|
|
sc->sc_tmbi = &sc->sc_mbi[0];
|
|
|
|
sc->sc_mbofull = 0;
|
|
|
|
/*
|
|
* If the board supports strict round-robin, enable that.
|
|
*/
|
|
if (sc->sc_flags & BHAF_STRICT_ROUND_ROBIN) {
|
|
toggle.cmd.opcode = BHA_ROUND_ROBIN;
|
|
toggle.cmd.enable = 1;
|
|
bha_cmd(sc->sc_iot, sc->sc_ioh, name,
|
|
sizeof(toggle.cmd), (u_char *)&toggle.cmd,
|
|
0, NULL);
|
|
}
|
|
|
|
/*
|
|
* Give the mailbox to the board.
|
|
*/
|
|
mailbox.cmd.opcode = BHA_MBX_INIT_EXTENDED;
|
|
mailbox.cmd.nmbx = sc->sc_mbox_count;
|
|
ltophys(sc->sc_dmamap_mbox->dm_segs[0].ds_addr, mailbox.cmd.addr);
|
|
bha_cmd(sc->sc_iot, sc->sc_ioh, name,
|
|
sizeof(mailbox.cmd), (u_char *)&mailbox.cmd,
|
|
0, (u_char *)0);
|
|
|
|
return (0);
|
|
}
|
|
|
|
/*****************************************************************************
|
|
* CCB execution engine
|
|
*****************************************************************************/
|
|
|
|
/*
|
|
* bha_queue_ccb:
|
|
*
|
|
* Queue a CCB to be sent to the controller, and send it if possible.
|
|
*/
|
|
static void
|
|
bha_queue_ccb(struct bha_softc *sc, struct bha_ccb *ccb)
|
|
{
|
|
|
|
TAILQ_INSERT_TAIL(&sc->sc_waiting_ccb, ccb, chain);
|
|
bha_start_ccbs(sc);
|
|
}
|
|
|
|
/*
|
|
* bha_start_ccbs:
|
|
*
|
|
* Send as many CCBs as we have empty mailboxes for.
|
|
*/
|
|
static void
|
|
bha_start_ccbs(struct bha_softc *sc)
|
|
{
|
|
bus_space_tag_t iot = sc->sc_iot;
|
|
bus_space_handle_t ioh = sc->sc_ioh;
|
|
struct bha_ccb_group *bcg;
|
|
struct bha_mbx_out *mbo;
|
|
struct bha_ccb *ccb;
|
|
|
|
mbo = sc->sc_tmbo;
|
|
|
|
while ((ccb = TAILQ_FIRST(&sc->sc_waiting_ccb)) != NULL) {
|
|
if (sc->sc_mbofull >= sc->sc_mbox_count) {
|
|
#ifdef DIAGNOSTIC
|
|
if (sc->sc_mbofull > sc->sc_mbox_count)
|
|
panic("bha_start_ccbs: mbofull > mbox_count");
|
|
#endif
|
|
/*
|
|
* No mailboxes available; attempt to collect ones
|
|
* that have already been used.
|
|
*/
|
|
bha_collect_mbo(sc);
|
|
if (sc->sc_mbofull == sc->sc_mbox_count) {
|
|
/*
|
|
* Still no more available; have the
|
|
* controller interrupt us when it
|
|
* frees one.
|
|
*/
|
|
struct bha_toggle toggle;
|
|
|
|
toggle.cmd.opcode = BHA_MBO_INTR_EN;
|
|
toggle.cmd.enable = 1;
|
|
bha_cmd(iot, ioh, device_xname(sc->sc_dev),
|
|
sizeof(toggle.cmd), (u_char *)&toggle.cmd,
|
|
0, (u_char *)0);
|
|
break;
|
|
}
|
|
}
|
|
|
|
TAILQ_REMOVE(&sc->sc_waiting_ccb, ccb, chain);
|
|
#ifdef BHADIAG
|
|
ccb->flags |= CCB_SENDING;
|
|
#endif
|
|
|
|
/*
|
|
* Put the CCB in the mailbox.
|
|
*/
|
|
bcg = BHA_CCB_GROUP(ccb);
|
|
ltophys(bcg->bcg_dmamap->dm_segs[0].ds_addr +
|
|
BHA_CCB_OFFSET(ccb), mbo->ccb_addr);
|
|
if (ccb->flags & CCB_ABORT)
|
|
mbo->cmd = BHA_MBO_ABORT;
|
|
else
|
|
mbo->cmd = BHA_MBO_START;
|
|
|
|
BHA_MBO_SYNC(sc, mbo,
|
|
BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
|
|
|
|
/* Tell the card to poll immediately. */
|
|
bus_space_write_1(iot, ioh, BHA_CMD_PORT, BHA_START_SCSI);
|
|
|
|
if ((ccb->xs->xs_control & XS_CTL_POLL) == 0)
|
|
callout_reset(&ccb->xs->xs_callout,
|
|
mstohz(ccb->timeout), bha_timeout, ccb);
|
|
|
|
++sc->sc_mbofull;
|
|
mbo = bha_nextmbo(sc, mbo);
|
|
}
|
|
|
|
sc->sc_tmbo = mbo;
|
|
}
|
|
|
|
/*
|
|
* bha_finish_ccbs:
|
|
*
|
|
* Finalize the execution of CCBs in our incoming mailbox.
|
|
*/
|
|
static void
|
|
bha_finish_ccbs(struct bha_softc *sc)
|
|
{
|
|
struct bha_mbx_in *mbi;
|
|
struct bha_ccb *ccb;
|
|
int i;
|
|
|
|
mbi = sc->sc_tmbi;
|
|
|
|
BHA_MBI_SYNC(sc, mbi, BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
|
|
|
|
if (mbi->comp_stat == BHA_MBI_FREE) {
|
|
for (i = 0; i < sc->sc_mbox_count; i++) {
|
|
if (mbi->comp_stat != BHA_MBI_FREE) {
|
|
#ifdef BHADIAG
|
|
/*
|
|
* This can happen in normal operation if
|
|
* we use all mailbox slots.
|
|
*/
|
|
printf("%s: mbi not in round-robin order\n",
|
|
device_xname(sc->sc_dev));
|
|
#endif
|
|
goto again;
|
|
}
|
|
mbi = bha_nextmbi(sc, mbi);
|
|
BHA_MBI_SYNC(sc, mbi,
|
|
BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
|
|
}
|
|
#ifdef BHADIAGnot
|
|
printf("%s: mbi interrupt with no full mailboxes\n",
|
|
device_xname(sc->sc_dev));
|
|
#endif
|
|
return;
|
|
}
|
|
|
|
again:
|
|
do {
|
|
ccb = bha_ccb_phys_kv(sc, phystol(mbi->ccb_addr));
|
|
if (ccb == NULL) {
|
|
aprint_error_dev(sc->sc_dev, "bad mbi ccb pointer 0x%08x; skipping\n",
|
|
phystol(mbi->ccb_addr));
|
|
goto next;
|
|
}
|
|
|
|
BHA_CCB_SYNC(sc, ccb,
|
|
BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
|
|
|
|
#ifdef BHADEBUG
|
|
if (bha_debug) {
|
|
u_char *cp = ccb->scsi_cmd;
|
|
printf("op=%x %x %x %x %x %x\n",
|
|
cp[0], cp[1], cp[2], cp[3], cp[4], cp[5]);
|
|
printf("comp_stat %x for mbi addr = %p, ",
|
|
mbi->comp_stat, mbi);
|
|
printf("ccb addr = %p\n", ccb);
|
|
}
|
|
#endif /* BHADEBUG */
|
|
|
|
switch (mbi->comp_stat) {
|
|
case BHA_MBI_OK:
|
|
case BHA_MBI_ERROR:
|
|
if ((ccb->flags & CCB_ABORT) != 0) {
|
|
/*
|
|
* If we already started an abort, wait for it
|
|
* to complete before clearing the CCB. We
|
|
* could instead just clear CCB_SENDING, but
|
|
* what if the mailbox was already received?
|
|
* The worst that happens here is that we clear
|
|
* the CCB a bit later than we need to. BFD.
|
|
*/
|
|
goto next;
|
|
}
|
|
break;
|
|
|
|
case BHA_MBI_ABORT:
|
|
case BHA_MBI_UNKNOWN:
|
|
/*
|
|
* Even if the CCB wasn't found, we clear it anyway.
|
|
* See preceding comment.
|
|
*/
|
|
break;
|
|
|
|
default:
|
|
aprint_error_dev(sc->sc_dev, "bad mbi comp_stat %02x; skipping\n",
|
|
mbi->comp_stat);
|
|
goto next;
|
|
}
|
|
|
|
callout_stop(&ccb->xs->xs_callout);
|
|
bha_done(sc, ccb);
|
|
|
|
next:
|
|
mbi->comp_stat = BHA_MBI_FREE;
|
|
BHA_CCB_SYNC(sc, ccb,
|
|
BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
|
|
|
|
mbi = bha_nextmbi(sc, mbi);
|
|
BHA_MBI_SYNC(sc, mbi,
|
|
BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
|
|
} while (mbi->comp_stat != BHA_MBI_FREE);
|
|
|
|
sc->sc_tmbi = mbi;
|
|
}
|
|
|
|
/*****************************************************************************
|
|
* Mailbox management functions.
|
|
*****************************************************************************/
|
|
|
|
/*
|
|
* bha_create_mailbox:
|
|
*
|
|
* Create the mailbox structures. Helper function for bha_attach().
|
|
*
|
|
* NOTE: The Buslogic hardware only gets one DMA address for the
|
|
* mailbox! It expects:
|
|
*
|
|
* mailbox_out[mailbox_size]
|
|
* mailbox_in[mailbox_size]
|
|
*/
|
|
static int
|
|
bha_create_mailbox(struct bha_softc *sc)
|
|
{
|
|
bus_dma_segment_t seg;
|
|
size_t size;
|
|
int error, rseg;
|
|
|
|
size = (sizeof(struct bha_mbx_out) * sc->sc_mbox_count) +
|
|
(sizeof(struct bha_mbx_in) * sc->sc_mbox_count);
|
|
|
|
error = bus_dmamem_alloc(sc->sc_dmat, size, PAGE_SIZE, 0, &seg,
|
|
1, &rseg, sc->sc_dmaflags);
|
|
if (error) {
|
|
aprint_error_dev(sc->sc_dev, "unable to allocate mailboxes, error = %d\n",
|
|
error);
|
|
goto bad_0;
|
|
}
|
|
|
|
error = bus_dmamem_map(sc->sc_dmat, &seg, rseg, size,
|
|
(void **)&sc->sc_mbo, sc->sc_dmaflags | BUS_DMA_COHERENT);
|
|
if (error) {
|
|
aprint_error_dev(sc->sc_dev, "unable to map mailboxes, error = %d\n",
|
|
error);
|
|
goto bad_1;
|
|
}
|
|
|
|
memset(sc->sc_mbo, 0, size);
|
|
|
|
error = bus_dmamap_create(sc->sc_dmat, size, 1, size, 0,
|
|
sc->sc_dmaflags, &sc->sc_dmamap_mbox);
|
|
if (error) {
|
|
aprint_error_dev(sc->sc_dev,
|
|
"unable to create mailbox DMA map, error = %d\n",
|
|
error);
|
|
goto bad_2;
|
|
}
|
|
|
|
error = bus_dmamap_load(sc->sc_dmat, sc->sc_dmamap_mbox,
|
|
sc->sc_mbo, size, NULL, 0);
|
|
if (error) {
|
|
aprint_error_dev(sc->sc_dev, "unable to load mailbox DMA map, error = %d\n",
|
|
error);
|
|
goto bad_3;
|
|
}
|
|
|
|
sc->sc_mbi = (struct bha_mbx_in *)(sc->sc_mbo + sc->sc_mbox_count);
|
|
|
|
return (0);
|
|
|
|
bad_3:
|
|
bus_dmamap_destroy(sc->sc_dmat, sc->sc_dmamap_mbox);
|
|
bad_2:
|
|
bus_dmamem_unmap(sc->sc_dmat, (void *)sc->sc_mbo, size);
|
|
bad_1:
|
|
bus_dmamem_free(sc->sc_dmat, &seg, rseg);
|
|
bad_0:
|
|
return (error);
|
|
}
|
|
|
|
/*
|
|
* bha_collect_mbo:
|
|
*
|
|
* Garbage collect mailboxes that are no longer in use.
|
|
*/
|
|
static void
|
|
bha_collect_mbo(struct bha_softc *sc)
|
|
{
|
|
struct bha_mbx_out *mbo;
|
|
#ifdef BHADIAG
|
|
struct bha_ccb *ccb;
|
|
#endif
|
|
|
|
mbo = sc->sc_cmbo;
|
|
|
|
while (sc->sc_mbofull > 0) {
|
|
BHA_MBO_SYNC(sc, mbo,
|
|
BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
|
|
if (mbo->cmd != BHA_MBO_FREE)
|
|
break;
|
|
|
|
#ifdef BHADIAG
|
|
ccb = bha_ccb_phys_kv(sc, phystol(mbo->ccb_addr));
|
|
ccb->flags &= ~CCB_SENDING;
|
|
#endif
|
|
|
|
--sc->sc_mbofull;
|
|
mbo = bha_nextmbo(sc, mbo);
|
|
}
|
|
|
|
sc->sc_cmbo = mbo;
|
|
}
|
|
|
|
/*****************************************************************************
|
|
* CCB management functions
|
|
*****************************************************************************/
|
|
|
|
static inline void
|
|
bha_reset_ccb(struct bha_ccb *ccb)
|
|
{
|
|
|
|
ccb->flags = 0;
|
|
}
|
|
|
|
/*
|
|
* bha_create_ccbs:
|
|
*
|
|
* Create a set of CCBs.
|
|
*
|
|
* We determine the target CCB count, and then keep creating them
|
|
* until we reach the target, or fail. CCBs that are allocated
|
|
* but not "created" are left on the allocating list.
|
|
*
|
|
* XXX AB_QUIET/AB_SILENT lossage here; this is called during
|
|
* boot as well as at run-time.
|
|
*/
|
|
static void
|
|
bha_create_ccbs(struct bha_softc *sc, int count)
|
|
{
|
|
struct bha_ccb_group *bcg;
|
|
struct bha_ccb *ccb;
|
|
bus_dma_segment_t seg;
|
|
bus_dmamap_t ccbmap;
|
|
int target, i, error, rseg;
|
|
|
|
/*
|
|
* If the current CCB count is already the max number we're
|
|
* allowed to have, bail out now.
|
|
*/
|
|
if (sc->sc_cur_ccbs == sc->sc_max_ccbs)
|
|
return;
|
|
|
|
/*
|
|
* Compute our target count, and clamp it down to the max
|
|
* number we're allowed to have.
|
|
*/
|
|
target = sc->sc_cur_ccbs + count;
|
|
if (target > sc->sc_max_ccbs)
|
|
target = sc->sc_max_ccbs;
|
|
|
|
/*
|
|
* If there are CCBs on the allocating list, don't allocate a
|
|
* CCB group yet.
|
|
*/
|
|
if (TAILQ_FIRST(&sc->sc_allocating_ccbs) != NULL)
|
|
goto have_allocating_ccbs;
|
|
|
|
allocate_group:
|
|
error = bus_dmamem_alloc(sc->sc_dmat, PAGE_SIZE,
|
|
PAGE_SIZE, 0, &seg, 1, &rseg, sc->sc_dmaflags | BUS_DMA_NOWAIT);
|
|
if (error) {
|
|
aprint_error_dev(sc->sc_dev, "unable to allocate CCB group, error = %d\n",
|
|
error);
|
|
goto bad_0;
|
|
}
|
|
|
|
error = bus_dmamem_map(sc->sc_dmat, &seg, rseg, PAGE_SIZE,
|
|
(void *)&bcg,
|
|
sc->sc_dmaflags | BUS_DMA_NOWAIT | BUS_DMA_COHERENT);
|
|
if (error) {
|
|
aprint_error_dev(sc->sc_dev, "unable to map CCB group, error = %d\n",
|
|
error);
|
|
goto bad_1;
|
|
}
|
|
|
|
memset(bcg, 0, PAGE_SIZE);
|
|
|
|
error = bus_dmamap_create(sc->sc_dmat, PAGE_SIZE,
|
|
1, PAGE_SIZE, 0, sc->sc_dmaflags | BUS_DMA_NOWAIT, &ccbmap);
|
|
if (error) {
|
|
aprint_error_dev(sc->sc_dev, "unable to create CCB group DMA map, error = %d\n",
|
|
error);
|
|
goto bad_2;
|
|
}
|
|
|
|
error = bus_dmamap_load(sc->sc_dmat, ccbmap, bcg, PAGE_SIZE, NULL,
|
|
sc->sc_dmaflags | BUS_DMA_NOWAIT);
|
|
if (error) {
|
|
aprint_error_dev(sc->sc_dev, "unable to load CCB group DMA map, error = %d\n",
|
|
error);
|
|
goto bad_3;
|
|
}
|
|
|
|
bcg->bcg_dmamap = ccbmap;
|
|
|
|
#ifdef DIAGNOSTIC
|
|
if (BHA_CCB_GROUP(&bcg->bcg_ccbs[0]) !=
|
|
BHA_CCB_GROUP(&bcg->bcg_ccbs[bha_ccbs_per_group - 1]))
|
|
panic("bha_create_ccbs: CCB group size botch");
|
|
#endif
|
|
|
|
/*
|
|
* Add all of the CCBs in this group to the allocating list.
|
|
*/
|
|
for (i = 0; i < bha_ccbs_per_group; i++) {
|
|
ccb = &bcg->bcg_ccbs[i];
|
|
TAILQ_INSERT_TAIL(&sc->sc_allocating_ccbs, ccb, chain);
|
|
}
|
|
|
|
have_allocating_ccbs:
|
|
/*
|
|
* Loop over the allocating list until we reach our CCB target.
|
|
* If we run out on the list, we'll allocate another group's
|
|
* worth.
|
|
*/
|
|
while (sc->sc_cur_ccbs < target) {
|
|
ccb = TAILQ_FIRST(&sc->sc_allocating_ccbs);
|
|
if (ccb == NULL)
|
|
goto allocate_group;
|
|
if (bha_init_ccb(sc, ccb) != 0) {
|
|
/*
|
|
* We were unable to initialize the CCB.
|
|
* This is likely due to a resource shortage,
|
|
* so bail out now.
|
|
*/
|
|
return;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* If we got here, we've reached our target!
|
|
*/
|
|
return;
|
|
|
|
bad_3:
|
|
bus_dmamap_destroy(sc->sc_dmat, ccbmap);
|
|
bad_2:
|
|
bus_dmamem_unmap(sc->sc_dmat, (void *)bcg, PAGE_SIZE);
|
|
bad_1:
|
|
bus_dmamem_free(sc->sc_dmat, &seg, rseg);
|
|
bad_0:
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* bha_init_ccb:
|
|
*
|
|
* Initialize a CCB; helper function for bha_create_ccbs().
|
|
*/
|
|
static int
|
|
bha_init_ccb(struct bha_softc *sc, struct bha_ccb *ccb)
|
|
{
|
|
struct bha_ccb_group *bcg = BHA_CCB_GROUP(ccb);
|
|
int hashnum, error;
|
|
|
|
/*
|
|
* Create the DMA map for this CCB.
|
|
*
|
|
* XXX ALLOCNOW is a hack to prevent bounce buffer shortages
|
|
* XXX in the ISA case. A better solution is needed.
|
|
*/
|
|
error = bus_dmamap_create(sc->sc_dmat, BHA_MAXXFER, BHA_NSEG,
|
|
BHA_MAXXFER, 0, BUS_DMA_NOWAIT | BUS_DMA_ALLOCNOW | sc->sc_dmaflags,
|
|
&ccb->dmamap_xfer);
|
|
if (error) {
|
|
aprint_error_dev(sc->sc_dev, "unable to create CCB DMA map, error = %d\n",
|
|
error);
|
|
return (error);
|
|
}
|
|
|
|
TAILQ_REMOVE(&sc->sc_allocating_ccbs, ccb, chain);
|
|
|
|
/*
|
|
* Put the CCB into the phystokv hash table.
|
|
*/
|
|
ccb->hashkey = bcg->bcg_dmamap->dm_segs[0].ds_addr +
|
|
BHA_CCB_OFFSET(ccb);
|
|
hashnum = CCB_HASH(ccb->hashkey);
|
|
ccb->nexthash = sc->sc_ccbhash[hashnum];
|
|
sc->sc_ccbhash[hashnum] = ccb;
|
|
bha_reset_ccb(ccb);
|
|
|
|
TAILQ_INSERT_HEAD(&sc->sc_free_ccb, ccb, chain);
|
|
sc->sc_cur_ccbs++;
|
|
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* bha_get_ccb:
|
|
*
|
|
* Get a CCB for the SCSI operation. If there are none left,
|
|
* wait until one becomes available, if we can.
|
|
*/
|
|
static struct bha_ccb *
|
|
bha_get_ccb(struct bha_softc *sc)
|
|
{
|
|
struct bha_ccb *ccb;
|
|
int s;
|
|
|
|
s = splbio();
|
|
ccb = TAILQ_FIRST(&sc->sc_free_ccb);
|
|
if (ccb != NULL) {
|
|
TAILQ_REMOVE(&sc->sc_free_ccb, ccb, chain);
|
|
ccb->flags |= CCB_ALLOC;
|
|
}
|
|
splx(s);
|
|
return (ccb);
|
|
}
|
|
|
|
/*
|
|
* bha_free_ccb:
|
|
*
|
|
* Put a CCB back onto the free list.
|
|
*/
|
|
static void
|
|
bha_free_ccb(struct bha_softc *sc, struct bha_ccb *ccb)
|
|
{
|
|
int s;
|
|
|
|
s = splbio();
|
|
bha_reset_ccb(ccb);
|
|
TAILQ_INSERT_HEAD(&sc->sc_free_ccb, ccb, chain);
|
|
splx(s);
|
|
}
|
|
|
|
/*
|
|
* bha_ccb_phys_kv:
|
|
*
|
|
* Given a CCB DMA address, locate the CCB in kernel virtual space.
|
|
*/
|
|
static struct bha_ccb *
|
|
bha_ccb_phys_kv(struct bha_softc *sc, bus_addr_t ccb_phys)
|
|
{
|
|
int hashnum = CCB_HASH(ccb_phys);
|
|
struct bha_ccb *ccb = sc->sc_ccbhash[hashnum];
|
|
|
|
while (ccb) {
|
|
if (ccb->hashkey == ccb_phys)
|
|
break;
|
|
ccb = ccb->nexthash;
|
|
}
|
|
return (ccb);
|
|
}
|