2004-02-28 09:28:47 +03:00
|
|
|
/* $NetBSD: subr_disk.c,v 1.59 2004/02/28 06:28:48 yamt Exp $ */
|
1997-10-05 22:37:01 +04:00
|
|
|
|
|
|
|
/*-
|
2000-01-22 02:20:51 +03:00
|
|
|
* Copyright (c) 1996, 1997, 1999, 2000 The NetBSD Foundation, Inc.
|
1997-10-05 22:37:01 +04:00
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* This code is derived from software contributed to The NetBSD Foundation
|
|
|
|
* by Jason R. Thorpe of the Numerical Aerospace Simulation Facility,
|
|
|
|
* NASA Ames Research Center.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
|
|
* must display the following acknowledgement:
|
|
|
|
* This product includes software developed by the NetBSD
|
|
|
|
* Foundation, Inc. and its contributors.
|
|
|
|
* 4. Neither the name of The NetBSD Foundation nor the names of its
|
|
|
|
* contributors may be used to endorse or promote products derived
|
|
|
|
* from this software without specific prior written permission.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
|
|
|
|
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
|
|
|
|
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
|
|
|
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
|
|
|
|
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
*/
|
1994-06-29 10:29:24 +04:00
|
|
|
|
1994-05-19 07:43:13 +04:00
|
|
|
/*
|
|
|
|
* Copyright (c) 1982, 1986, 1988, 1993
|
|
|
|
* The Regents of the University of California. All rights reserved.
|
|
|
|
* (c) UNIX System Laboratories, Inc.
|
|
|
|
* All or some portions of this file are derived from material licensed
|
|
|
|
* to the University of California by American Telephone and Telegraph
|
|
|
|
* Co. or Unix System Laboratories, Inc. and are reproduced herein with
|
|
|
|
* the permission of UNIX System Laboratories, Inc.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
2003-08-07 20:26:28 +04:00
|
|
|
* 3. Neither the name of the University nor the names of its contributors
|
1994-05-19 07:43:13 +04:00
|
|
|
* may be used to endorse or promote products derived from this software
|
|
|
|
* without specific prior written permission.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
|
|
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
|
|
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
|
|
* SUCH DAMAGE.
|
|
|
|
*
|
1994-06-29 10:29:24 +04:00
|
|
|
* @(#)ufs_disksubr.c 8.5 (Berkeley) 1/21/94
|
1994-05-19 07:43:13 +04:00
|
|
|
*/
|
|
|
|
|
2001-11-13 16:33:44 +03:00
|
|
|
#include <sys/cdefs.h>
|
2004-02-28 09:28:47 +03:00
|
|
|
__KERNEL_RCSID(0, "$NetBSD: subr_disk.c,v 1.59 2004/02/28 06:28:48 yamt Exp $");
|
2002-11-05 16:22:32 +03:00
|
|
|
|
|
|
|
#include "opt_compat_netbsd.h"
|
2004-02-28 09:28:47 +03:00
|
|
|
#include "opt_bufq.h"
|
2001-11-13 16:33:44 +03:00
|
|
|
|
1994-05-19 07:43:13 +04:00
|
|
|
#include <sys/param.h>
|
1996-01-08 01:01:38 +03:00
|
|
|
#include <sys/kernel.h>
|
|
|
|
#include <sys/malloc.h>
|
1994-05-19 07:43:13 +04:00
|
|
|
#include <sys/buf.h>
|
|
|
|
#include <sys/syslog.h>
|
1996-01-08 01:01:38 +03:00
|
|
|
#include <sys/disklabel.h>
|
|
|
|
#include <sys/disk.h>
|
2002-01-27 15:41:07 +03:00
|
|
|
#include <sys/sysctl.h>
|
2002-11-04 06:50:07 +03:00
|
|
|
#include <lib/libkern/libkern.h>
|
1995-12-28 22:16:31 +03:00
|
|
|
|
1996-01-08 01:01:38 +03:00
|
|
|
/*
|
|
|
|
* A global list of all disks attached to the system. May grow or
|
|
|
|
* shrink over time.
|
|
|
|
*/
|
|
|
|
struct disklist_head disklist; /* TAILQ_HEAD */
|
|
|
|
int disk_count; /* number of drives in global disklist */
|
2002-01-27 15:41:07 +03:00
|
|
|
struct simplelock disklist_slock = SIMPLELOCK_INITIALIZER;
|
1996-01-08 01:01:38 +03:00
|
|
|
|
2004-02-28 09:28:47 +03:00
|
|
|
#ifdef NEW_BUFQ_STRATEGY
|
|
|
|
int bufq_disk_default_strat = BUFQ_READ_PRIO;
|
|
|
|
#else /* NEW_BUFQ_STRATEGY */
|
|
|
|
int bufq_disk_default_strat = BUFQ_DISKSORT;
|
|
|
|
#endif /* NEW_BUFQ_STRATEGY */
|
|
|
|
|
1994-05-19 07:43:13 +04:00
|
|
|
/*
|
|
|
|
* Compute checksum for disk label.
|
|
|
|
*/
|
|
|
|
u_int
|
2001-07-09 14:54:12 +04:00
|
|
|
dkcksum(struct disklabel *lp)
|
1994-05-19 07:43:13 +04:00
|
|
|
{
|
2000-03-30 13:27:11 +04:00
|
|
|
u_short *start, *end;
|
|
|
|
u_short sum = 0;
|
1994-05-19 07:43:13 +04:00
|
|
|
|
|
|
|
start = (u_short *)lp;
|
|
|
|
end = (u_short *)&lp->d_partitions[lp->d_npartitions];
|
|
|
|
while (start < end)
|
|
|
|
sum ^= *start++;
|
|
|
|
return (sum);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Disk error is the preface to plaintive error messages
|
|
|
|
* about failing disk transfers. It prints messages of the form
|
|
|
|
|
|
|
|
hp0g: hard error reading fsbn 12345 of 12344-12347 (hp0 bn %d cn %d tn %d sn %d)
|
|
|
|
|
|
|
|
* if the offset of the error in the transfer and a disk label
|
|
|
|
* are both available. blkdone should be -1 if the position of the error
|
|
|
|
* is unknown; the disklabel pointer may be null from drivers that have not
|
1996-10-13 06:32:29 +04:00
|
|
|
* been converted to use them. The message is printed with printf
|
1994-05-19 07:43:13 +04:00
|
|
|
* if pri is LOG_PRINTF, otherwise it uses log at the specified priority.
|
1996-10-13 06:32:29 +04:00
|
|
|
* The message should be completed (with at least a newline) with printf
|
1994-05-19 07:43:13 +04:00
|
|
|
* or addlog, respectively. There is no trailing space.
|
|
|
|
*/
|
2003-04-13 11:51:30 +04:00
|
|
|
#ifndef PRIdaddr
|
|
|
|
#define PRIdaddr PRId64
|
|
|
|
#endif
|
1994-05-19 07:43:13 +04:00
|
|
|
void
|
2002-06-28 20:37:20 +04:00
|
|
|
diskerr(const struct buf *bp, const char *dname, const char *what, int pri,
|
|
|
|
int blkdone, const struct disklabel *lp)
|
1994-05-19 07:43:13 +04:00
|
|
|
{
|
1999-02-22 19:00:01 +03:00
|
|
|
int unit = DISKUNIT(bp->b_dev), part = DISKPART(bp->b_dev);
|
2001-07-09 14:54:12 +04:00
|
|
|
void (*pr)(const char *, ...);
|
1994-05-19 07:43:13 +04:00
|
|
|
char partname = 'a' + part;
|
2003-04-13 11:51:30 +04:00
|
|
|
daddr_t sn;
|
|
|
|
|
2003-04-13 13:08:04 +04:00
|
|
|
if (/*CONSTCOND*/0)
|
2003-04-13 11:51:30 +04:00
|
|
|
/* Compiler will error this is the format is wrong... */
|
|
|
|
printf("%" PRIdaddr, bp->b_blkno);
|
1994-05-19 07:43:13 +04:00
|
|
|
|
|
|
|
if (pri != LOG_PRINTF) {
|
1996-03-17 02:17:04 +03:00
|
|
|
static const char fmt[] = "";
|
|
|
|
log(pri, fmt);
|
1994-05-19 07:43:13 +04:00
|
|
|
pr = addlog;
|
|
|
|
} else
|
1996-10-13 06:32:29 +04:00
|
|
|
pr = printf;
|
1994-05-19 07:43:13 +04:00
|
|
|
(*pr)("%s%d%c: %s %sing fsbn ", dname, unit, partname, what,
|
|
|
|
bp->b_flags & B_READ ? "read" : "writ");
|
|
|
|
sn = bp->b_blkno;
|
|
|
|
if (bp->b_bcount <= DEV_BSIZE)
|
2003-04-13 11:51:30 +04:00
|
|
|
(*pr)("%" PRIdaddr, sn);
|
1994-05-19 07:43:13 +04:00
|
|
|
else {
|
|
|
|
if (blkdone >= 0) {
|
|
|
|
sn += blkdone;
|
2003-04-13 11:51:30 +04:00
|
|
|
(*pr)("%" PRIdaddr " of ", sn);
|
1994-05-19 07:43:13 +04:00
|
|
|
}
|
2003-04-13 11:51:30 +04:00
|
|
|
(*pr)("%" PRIdaddr "-%" PRIdaddr "", bp->b_blkno,
|
1994-05-19 07:43:13 +04:00
|
|
|
bp->b_blkno + (bp->b_bcount - 1) / DEV_BSIZE);
|
|
|
|
}
|
|
|
|
if (lp && (blkdone >= 0 || bp->b_bcount <= lp->d_secsize)) {
|
|
|
|
sn += lp->d_partitions[part].p_offset;
|
2003-04-13 11:51:30 +04:00
|
|
|
(*pr)(" (%s%d bn %" PRIdaddr "; cn %" PRIdaddr "",
|
|
|
|
dname, unit, sn, sn / lp->d_secpercyl);
|
1994-05-19 07:43:13 +04:00
|
|
|
sn %= lp->d_secpercyl;
|
2003-04-13 11:51:30 +04:00
|
|
|
(*pr)(" tn %" PRIdaddr " sn %" PRIdaddr ")",
|
|
|
|
sn / lp->d_nsectors, sn % lp->d_nsectors);
|
1994-05-19 07:43:13 +04:00
|
|
|
}
|
|
|
|
}
|
1996-01-08 01:01:38 +03:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Initialize the disklist. Called by main() before autoconfiguration.
|
|
|
|
*/
|
|
|
|
void
|
2001-07-09 14:54:12 +04:00
|
|
|
disk_init(void)
|
1996-01-08 01:01:38 +03:00
|
|
|
{
|
|
|
|
|
|
|
|
TAILQ_INIT(&disklist);
|
|
|
|
disk_count = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Searches the disklist for the disk corresponding to the
|
|
|
|
* name provided.
|
|
|
|
*/
|
|
|
|
struct disk *
|
2001-07-09 14:54:12 +04:00
|
|
|
disk_find(char *name)
|
1996-01-08 01:01:38 +03:00
|
|
|
{
|
|
|
|
struct disk *diskp;
|
|
|
|
|
|
|
|
if ((name == NULL) || (disk_count <= 0))
|
|
|
|
return (NULL);
|
|
|
|
|
2002-01-27 15:41:07 +03:00
|
|
|
simple_lock(&disklist_slock);
|
|
|
|
for (diskp = TAILQ_FIRST(&disklist); diskp != NULL;
|
|
|
|
diskp = TAILQ_NEXT(diskp, dk_link))
|
|
|
|
if (strcmp(diskp->dk_name, name) == 0) {
|
|
|
|
simple_unlock(&disklist_slock);
|
1996-01-08 01:01:38 +03:00
|
|
|
return (diskp);
|
2002-01-27 15:41:07 +03:00
|
|
|
}
|
|
|
|
simple_unlock(&disklist_slock);
|
1996-01-08 01:01:38 +03:00
|
|
|
|
|
|
|
return (NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Attach a disk.
|
|
|
|
*/
|
|
|
|
void
|
2001-07-09 14:54:12 +04:00
|
|
|
disk_attach(struct disk *diskp)
|
1996-01-08 01:01:38 +03:00
|
|
|
{
|
|
|
|
int s;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate and initialize the disklabel structures. Note that
|
|
|
|
* it's not safe to sleep here, since we're probably going to be
|
|
|
|
* called during autoconfiguration.
|
|
|
|
*/
|
|
|
|
diskp->dk_label = malloc(sizeof(struct disklabel), M_DEVBUF, M_NOWAIT);
|
|
|
|
diskp->dk_cpulabel = malloc(sizeof(struct cpu_disklabel), M_DEVBUF,
|
|
|
|
M_NOWAIT);
|
|
|
|
if ((diskp->dk_label == NULL) || (diskp->dk_cpulabel == NULL))
|
|
|
|
panic("disk_attach: can't allocate storage for disklabel");
|
|
|
|
|
Abolition of bcopy, ovbcopy, bcmp, and bzero, phase one.
bcopy(x, y, z) -> memcpy(y, x, z)
ovbcopy(x, y, z) -> memmove(y, x, z)
bcmp(x, y, z) -> memcmp(x, y, z)
bzero(x, y) -> memset(x, 0, y)
1998-08-04 08:03:10 +04:00
|
|
|
memset(diskp->dk_label, 0, sizeof(struct disklabel));
|
|
|
|
memset(diskp->dk_cpulabel, 0, sizeof(struct cpu_disklabel));
|
1996-01-08 01:01:38 +03:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Set the attached timestamp.
|
|
|
|
*/
|
|
|
|
s = splclock();
|
|
|
|
diskp->dk_attachtime = mono_time;
|
|
|
|
splx(s);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Link into the disklist.
|
|
|
|
*/
|
2002-01-27 15:41:07 +03:00
|
|
|
simple_lock(&disklist_slock);
|
1996-01-08 01:01:38 +03:00
|
|
|
TAILQ_INSERT_TAIL(&disklist, diskp, dk_link);
|
2002-01-27 15:41:07 +03:00
|
|
|
simple_unlock(&disklist_slock);
|
1996-01-08 01:01:38 +03:00
|
|
|
++disk_count;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
1996-02-09 21:59:18 +03:00
|
|
|
* Detach a disk.
|
1996-01-08 01:01:38 +03:00
|
|
|
*/
|
|
|
|
void
|
2001-07-09 14:54:12 +04:00
|
|
|
disk_detach(struct disk *diskp)
|
1996-01-08 01:01:38 +03:00
|
|
|
{
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Remove from the disklist.
|
|
|
|
*/
|
|
|
|
if (--disk_count < 0)
|
1996-02-09 21:59:18 +03:00
|
|
|
panic("disk_detach: disk_count < 0");
|
2002-01-27 15:41:07 +03:00
|
|
|
simple_lock(&disklist_slock);
|
1997-12-30 12:51:24 +03:00
|
|
|
TAILQ_REMOVE(&disklist, diskp, dk_link);
|
2002-01-27 15:41:07 +03:00
|
|
|
simple_unlock(&disklist_slock);
|
1997-12-30 12:51:24 +03:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Free the space used by the disklabel structures.
|
|
|
|
*/
|
|
|
|
free(diskp->dk_label, M_DEVBUF);
|
|
|
|
free(diskp->dk_cpulabel, M_DEVBUF);
|
1996-01-08 01:01:38 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Increment a disk's busy counter. If the counter is going from
|
|
|
|
* 0 to 1, set the timestamp.
|
|
|
|
*/
|
|
|
|
void
|
2001-07-09 14:54:12 +04:00
|
|
|
disk_busy(struct disk *diskp)
|
1996-01-08 01:01:38 +03:00
|
|
|
{
|
|
|
|
int s;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* XXX We'd like to use something as accurate as microtime(),
|
|
|
|
* but that doesn't depend on the system TOD clock.
|
|
|
|
*/
|
|
|
|
if (diskp->dk_busy++ == 0) {
|
|
|
|
s = splclock();
|
|
|
|
diskp->dk_timestamp = mono_time;
|
|
|
|
splx(s);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Decrement a disk's busy counter, increment the byte count, total busy
|
|
|
|
* time, and reset the timestamp.
|
|
|
|
*/
|
|
|
|
void
|
2002-11-01 14:31:50 +03:00
|
|
|
disk_unbusy(struct disk *diskp, long bcount, int read)
|
1996-01-08 01:01:38 +03:00
|
|
|
{
|
|
|
|
int s;
|
|
|
|
struct timeval dv_time, diff_time;
|
|
|
|
|
1997-12-30 12:51:24 +03:00
|
|
|
if (diskp->dk_busy-- == 0) {
|
|
|
|
printf("%s: dk_busy < 0\n", diskp->dk_name);
|
|
|
|
panic("disk_unbusy");
|
|
|
|
}
|
1996-01-08 01:01:38 +03:00
|
|
|
|
|
|
|
s = splclock();
|
|
|
|
dv_time = mono_time;
|
|
|
|
splx(s);
|
|
|
|
|
|
|
|
timersub(&dv_time, &diskp->dk_timestamp, &diff_time);
|
|
|
|
timeradd(&diskp->dk_time, &diff_time, &diskp->dk_time);
|
|
|
|
|
|
|
|
diskp->dk_timestamp = dv_time;
|
|
|
|
if (bcount > 0) {
|
2002-11-01 14:31:50 +03:00
|
|
|
if (read) {
|
|
|
|
diskp->dk_rbytes += bcount;
|
|
|
|
diskp->dk_rxfer++;
|
|
|
|
} else {
|
|
|
|
diskp->dk_wbytes += bcount;
|
|
|
|
diskp->dk_wxfer++;
|
|
|
|
}
|
1996-01-08 01:01:38 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Reset the metrics counters on the given disk. Note that we cannot
|
|
|
|
* reset the busy counter, as it may case a panic in disk_unbusy().
|
|
|
|
* We also must avoid playing with the timestamp information, as it
|
|
|
|
* may skew any pending transfer results.
|
|
|
|
*/
|
|
|
|
void
|
2001-07-09 14:54:12 +04:00
|
|
|
disk_resetstat(struct disk *diskp)
|
1996-01-08 01:01:38 +03:00
|
|
|
{
|
|
|
|
int s = splbio(), t;
|
|
|
|
|
2002-11-01 14:31:50 +03:00
|
|
|
diskp->dk_rxfer = 0;
|
|
|
|
diskp->dk_rbytes = 0;
|
|
|
|
diskp->dk_wxfer = 0;
|
|
|
|
diskp->dk_wbytes = 0;
|
1996-01-08 01:01:38 +03:00
|
|
|
|
|
|
|
t = splclock();
|
|
|
|
diskp->dk_attachtime = mono_time;
|
|
|
|
splx(t);
|
|
|
|
|
|
|
|
timerclear(&diskp->dk_time);
|
|
|
|
|
|
|
|
splx(s);
|
|
|
|
}
|
2002-01-27 15:41:07 +03:00
|
|
|
|
|
|
|
int
|
Dynamic sysctl.
Gone are the old kern_sysctl(), cpu_sysctl(), hw_sysctl(),
vfs_sysctl(), etc, routines, along with sysctl_int() et al. Now all
nodes are registered with the tree, and nodes can be added (or
removed) easily, and I/O to and from the tree is handled generically.
Since the nodes are registered with the tree, the mapping from name to
number (and back again) can now be discovered, instead of having to be
hard coded. Adding new nodes to the tree is likewise much simpler --
the new infrastructure handles almost all the work for simple types,
and just about anything else can be done with a small helper function.
All existing nodes are where they were before (numerically speaking),
so all existing consumers of sysctl information should notice no
difference.
PS - I'm sorry, but there's a distinct lack of documentation at the
moment. I'm working on sysctl(3/8/9) right now, and I promise to
watch out for buses.
2003-12-04 22:38:21 +03:00
|
|
|
sysctl_hw_disknames(SYSCTLFN_ARGS)
|
2002-01-27 15:41:07 +03:00
|
|
|
{
|
|
|
|
char buf[DK_DISKNAMELEN + 1];
|
Dynamic sysctl.
Gone are the old kern_sysctl(), cpu_sysctl(), hw_sysctl(),
vfs_sysctl(), etc, routines, along with sysctl_int() et al. Now all
nodes are registered with the tree, and nodes can be added (or
removed) easily, and I/O to and from the tree is handled generically.
Since the nodes are registered with the tree, the mapping from name to
number (and back again) can now be discovered, instead of having to be
hard coded. Adding new nodes to the tree is likewise much simpler --
the new infrastructure handles almost all the work for simple types,
and just about anything else can be done with a small helper function.
All existing nodes are where they were before (numerically speaking),
so all existing consumers of sysctl information should notice no
difference.
PS - I'm sorry, but there's a distinct lack of documentation at the
moment. I'm working on sysctl(3/8/9) right now, and I promise to
watch out for buses.
2003-12-04 22:38:21 +03:00
|
|
|
char *where = oldp;
|
2002-01-27 15:41:07 +03:00
|
|
|
struct disk *diskp;
|
|
|
|
size_t needed, left, slen;
|
|
|
|
int error, first;
|
|
|
|
|
Dynamic sysctl.
Gone are the old kern_sysctl(), cpu_sysctl(), hw_sysctl(),
vfs_sysctl(), etc, routines, along with sysctl_int() et al. Now all
nodes are registered with the tree, and nodes can be added (or
removed) easily, and I/O to and from the tree is handled generically.
Since the nodes are registered with the tree, the mapping from name to
number (and back again) can now be discovered, instead of having to be
hard coded. Adding new nodes to the tree is likewise much simpler --
the new infrastructure handles almost all the work for simple types,
and just about anything else can be done with a small helper function.
All existing nodes are where they were before (numerically speaking),
so all existing consumers of sysctl information should notice no
difference.
PS - I'm sorry, but there's a distinct lack of documentation at the
moment. I'm working on sysctl(3/8/9) right now, and I promise to
watch out for buses.
2003-12-04 22:38:21 +03:00
|
|
|
if (newp != NULL)
|
|
|
|
return (EPERM);
|
|
|
|
if (namelen != 0)
|
|
|
|
return (EINVAL);
|
|
|
|
|
2002-01-27 15:41:07 +03:00
|
|
|
first = 1;
|
|
|
|
error = 0;
|
|
|
|
needed = 0;
|
Dynamic sysctl.
Gone are the old kern_sysctl(), cpu_sysctl(), hw_sysctl(),
vfs_sysctl(), etc, routines, along with sysctl_int() et al. Now all
nodes are registered with the tree, and nodes can be added (or
removed) easily, and I/O to and from the tree is handled generically.
Since the nodes are registered with the tree, the mapping from name to
number (and back again) can now be discovered, instead of having to be
hard coded. Adding new nodes to the tree is likewise much simpler --
the new infrastructure handles almost all the work for simple types,
and just about anything else can be done with a small helper function.
All existing nodes are where they were before (numerically speaking),
so all existing consumers of sysctl information should notice no
difference.
PS - I'm sorry, but there's a distinct lack of documentation at the
moment. I'm working on sysctl(3/8/9) right now, and I promise to
watch out for buses.
2003-12-04 22:38:21 +03:00
|
|
|
left = *oldlenp;
|
2002-01-28 06:33:55 +03:00
|
|
|
|
2002-01-27 15:41:07 +03:00
|
|
|
simple_lock(&disklist_slock);
|
|
|
|
for (diskp = TAILQ_FIRST(&disklist); diskp != NULL;
|
|
|
|
diskp = TAILQ_NEXT(diskp, dk_link)) {
|
|
|
|
if (where == NULL)
|
|
|
|
needed += strlen(diskp->dk_name) + 1;
|
|
|
|
else {
|
|
|
|
memset(buf, 0, sizeof(buf));
|
|
|
|
if (first) {
|
|
|
|
strncpy(buf, diskp->dk_name, sizeof(buf));
|
|
|
|
first = 0;
|
|
|
|
} else {
|
|
|
|
buf[0] = ' ';
|
2002-02-16 05:07:56 +03:00
|
|
|
strncpy(buf + 1, diskp->dk_name,
|
2002-02-16 05:11:43 +03:00
|
|
|
sizeof(buf) - 1);
|
2002-01-27 15:41:07 +03:00
|
|
|
}
|
|
|
|
buf[DK_DISKNAMELEN] = '\0';
|
|
|
|
slen = strlen(buf);
|
|
|
|
if (left < slen + 1)
|
|
|
|
break;
|
|
|
|
/* +1 to copy out the trailing NUL byte */
|
|
|
|
error = copyout(buf, where, slen + 1);
|
|
|
|
if (error)
|
|
|
|
break;
|
|
|
|
where += slen;
|
|
|
|
needed += slen;
|
|
|
|
left -= slen;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
simple_unlock(&disklist_slock);
|
Dynamic sysctl.
Gone are the old kern_sysctl(), cpu_sysctl(), hw_sysctl(),
vfs_sysctl(), etc, routines, along with sysctl_int() et al. Now all
nodes are registered with the tree, and nodes can be added (or
removed) easily, and I/O to and from the tree is handled generically.
Since the nodes are registered with the tree, the mapping from name to
number (and back again) can now be discovered, instead of having to be
hard coded. Adding new nodes to the tree is likewise much simpler --
the new infrastructure handles almost all the work for simple types,
and just about anything else can be done with a small helper function.
All existing nodes are where they were before (numerically speaking),
so all existing consumers of sysctl information should notice no
difference.
PS - I'm sorry, but there's a distinct lack of documentation at the
moment. I'm working on sysctl(3/8/9) right now, and I promise to
watch out for buses.
2003-12-04 22:38:21 +03:00
|
|
|
*oldlenp = needed;
|
2002-01-27 15:41:07 +03:00
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
Dynamic sysctl.
Gone are the old kern_sysctl(), cpu_sysctl(), hw_sysctl(),
vfs_sysctl(), etc, routines, along with sysctl_int() et al. Now all
nodes are registered with the tree, and nodes can be added (or
removed) easily, and I/O to and from the tree is handled generically.
Since the nodes are registered with the tree, the mapping from name to
number (and back again) can now be discovered, instead of having to be
hard coded. Adding new nodes to the tree is likewise much simpler --
the new infrastructure handles almost all the work for simple types,
and just about anything else can be done with a small helper function.
All existing nodes are where they were before (numerically speaking),
so all existing consumers of sysctl information should notice no
difference.
PS - I'm sorry, but there's a distinct lack of documentation at the
moment. I'm working on sysctl(3/8/9) right now, and I promise to
watch out for buses.
2003-12-04 22:38:21 +03:00
|
|
|
sysctl_hw_diskstats(SYSCTLFN_ARGS)
|
2002-01-27 15:41:07 +03:00
|
|
|
{
|
|
|
|
struct disk_sysctl sdisk;
|
|
|
|
struct disk *diskp;
|
Dynamic sysctl.
Gone are the old kern_sysctl(), cpu_sysctl(), hw_sysctl(),
vfs_sysctl(), etc, routines, along with sysctl_int() et al. Now all
nodes are registered with the tree, and nodes can be added (or
removed) easily, and I/O to and from the tree is handled generically.
Since the nodes are registered with the tree, the mapping from name to
number (and back again) can now be discovered, instead of having to be
hard coded. Adding new nodes to the tree is likewise much simpler --
the new infrastructure handles almost all the work for simple types,
and just about anything else can be done with a small helper function.
All existing nodes are where they were before (numerically speaking),
so all existing consumers of sysctl information should notice no
difference.
PS - I'm sorry, but there's a distinct lack of documentation at the
moment. I'm working on sysctl(3/8/9) right now, and I promise to
watch out for buses.
2003-12-04 22:38:21 +03:00
|
|
|
char *where = oldp;
|
2002-01-27 15:41:07 +03:00
|
|
|
size_t tocopy, left;
|
|
|
|
int error;
|
|
|
|
|
Dynamic sysctl.
Gone are the old kern_sysctl(), cpu_sysctl(), hw_sysctl(),
vfs_sysctl(), etc, routines, along with sysctl_int() et al. Now all
nodes are registered with the tree, and nodes can be added (or
removed) easily, and I/O to and from the tree is handled generically.
Since the nodes are registered with the tree, the mapping from name to
number (and back again) can now be discovered, instead of having to be
hard coded. Adding new nodes to the tree is likewise much simpler --
the new infrastructure handles almost all the work for simple types,
and just about anything else can be done with a small helper function.
All existing nodes are where they were before (numerically speaking),
so all existing consumers of sysctl information should notice no
difference.
PS - I'm sorry, but there's a distinct lack of documentation at the
moment. I'm working on sysctl(3/8/9) right now, and I promise to
watch out for buses.
2003-12-04 22:38:21 +03:00
|
|
|
if (newp != NULL)
|
|
|
|
return (EPERM);
|
|
|
|
|
2002-11-05 16:22:32 +03:00
|
|
|
/*
|
|
|
|
* The original hw.diskstats call was broken and did not require
|
|
|
|
* the userland to pass in it's size of struct disk_sysctl. This
|
|
|
|
* was fixed after NetBSD 1.6 was released, and any applications
|
|
|
|
* that do not pass in the size are given an error only, unless
|
|
|
|
* we care about 1.6 compatibility.
|
|
|
|
*/
|
2002-01-27 15:41:07 +03:00
|
|
|
if (namelen == 0)
|
2002-11-04 06:50:07 +03:00
|
|
|
#ifdef COMPAT_16
|
2002-11-06 05:31:34 +03:00
|
|
|
tocopy = offsetof(struct disk_sysctl, dk_rxfer);
|
2002-11-04 06:50:07 +03:00
|
|
|
#else
|
|
|
|
return (EINVAL);
|
|
|
|
#endif
|
2002-01-27 15:41:07 +03:00
|
|
|
else
|
|
|
|
tocopy = name[0];
|
|
|
|
|
2002-11-06 05:31:34 +03:00
|
|
|
if (where == NULL) {
|
Dynamic sysctl.
Gone are the old kern_sysctl(), cpu_sysctl(), hw_sysctl(),
vfs_sysctl(), etc, routines, along with sysctl_int() et al. Now all
nodes are registered with the tree, and nodes can be added (or
removed) easily, and I/O to and from the tree is handled generically.
Since the nodes are registered with the tree, the mapping from name to
number (and back again) can now be discovered, instead of having to be
hard coded. Adding new nodes to the tree is likewise much simpler --
the new infrastructure handles almost all the work for simple types,
and just about anything else can be done with a small helper function.
All existing nodes are where they were before (numerically speaking),
so all existing consumers of sysctl information should notice no
difference.
PS - I'm sorry, but there's a distinct lack of documentation at the
moment. I'm working on sysctl(3/8/9) right now, and I promise to
watch out for buses.
2003-12-04 22:38:21 +03:00
|
|
|
*oldlenp = disk_count * tocopy;
|
2002-11-06 05:31:34 +03:00
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
|
2002-01-27 15:41:07 +03:00
|
|
|
error = 0;
|
Dynamic sysctl.
Gone are the old kern_sysctl(), cpu_sysctl(), hw_sysctl(),
vfs_sysctl(), etc, routines, along with sysctl_int() et al. Now all
nodes are registered with the tree, and nodes can be added (or
removed) easily, and I/O to and from the tree is handled generically.
Since the nodes are registered with the tree, the mapping from name to
number (and back again) can now be discovered, instead of having to be
hard coded. Adding new nodes to the tree is likewise much simpler --
the new infrastructure handles almost all the work for simple types,
and just about anything else can be done with a small helper function.
All existing nodes are where they were before (numerically speaking),
so all existing consumers of sysctl information should notice no
difference.
PS - I'm sorry, but there's a distinct lack of documentation at the
moment. I'm working on sysctl(3/8/9) right now, and I promise to
watch out for buses.
2003-12-04 22:38:21 +03:00
|
|
|
left = *oldlenp;
|
2002-01-27 15:41:07 +03:00
|
|
|
memset(&sdisk, 0, sizeof(sdisk));
|
Dynamic sysctl.
Gone are the old kern_sysctl(), cpu_sysctl(), hw_sysctl(),
vfs_sysctl(), etc, routines, along with sysctl_int() et al. Now all
nodes are registered with the tree, and nodes can be added (or
removed) easily, and I/O to and from the tree is handled generically.
Since the nodes are registered with the tree, the mapping from name to
number (and back again) can now be discovered, instead of having to be
hard coded. Adding new nodes to the tree is likewise much simpler --
the new infrastructure handles almost all the work for simple types,
and just about anything else can be done with a small helper function.
All existing nodes are where they were before (numerically speaking),
so all existing consumers of sysctl information should notice no
difference.
PS - I'm sorry, but there's a distinct lack of documentation at the
moment. I'm working on sysctl(3/8/9) right now, and I promise to
watch out for buses.
2003-12-04 22:38:21 +03:00
|
|
|
*oldlenp = 0;
|
2002-01-27 15:41:07 +03:00
|
|
|
|
|
|
|
simple_lock(&disklist_slock);
|
2002-01-28 06:12:13 +03:00
|
|
|
TAILQ_FOREACH(diskp, &disklist, dk_link) {
|
2002-11-01 18:20:03 +03:00
|
|
|
if (left < tocopy)
|
2002-01-27 15:41:07 +03:00
|
|
|
break;
|
2002-02-16 05:07:56 +03:00
|
|
|
strncpy(sdisk.dk_name, diskp->dk_name, sizeof(sdisk.dk_name));
|
2002-11-01 14:31:50 +03:00
|
|
|
sdisk.dk_xfer = diskp->dk_rxfer + diskp->dk_wxfer;
|
|
|
|
sdisk.dk_rxfer = diskp->dk_rxfer;
|
|
|
|
sdisk.dk_wxfer = diskp->dk_wxfer;
|
2002-01-27 15:41:07 +03:00
|
|
|
sdisk.dk_seek = diskp->dk_seek;
|
2002-11-01 14:31:50 +03:00
|
|
|
sdisk.dk_bytes = diskp->dk_rbytes + diskp->dk_wbytes;
|
|
|
|
sdisk.dk_rbytes = diskp->dk_rbytes;
|
|
|
|
sdisk.dk_wbytes = diskp->dk_wbytes;
|
2002-01-27 15:41:07 +03:00
|
|
|
sdisk.dk_attachtime_sec = diskp->dk_attachtime.tv_sec;
|
|
|
|
sdisk.dk_attachtime_usec = diskp->dk_attachtime.tv_usec;
|
|
|
|
sdisk.dk_timestamp_sec = diskp->dk_timestamp.tv_sec;
|
|
|
|
sdisk.dk_timestamp_usec = diskp->dk_timestamp.tv_usec;
|
|
|
|
sdisk.dk_time_sec = diskp->dk_time.tv_sec;
|
|
|
|
sdisk.dk_time_usec = diskp->dk_time.tv_usec;
|
|
|
|
sdisk.dk_busy = diskp->dk_busy;
|
2002-01-28 06:33:55 +03:00
|
|
|
|
2002-01-27 15:41:07 +03:00
|
|
|
error = copyout(&sdisk, where, min(tocopy, sizeof(sdisk)));
|
|
|
|
if (error)
|
|
|
|
break;
|
|
|
|
where += tocopy;
|
Dynamic sysctl.
Gone are the old kern_sysctl(), cpu_sysctl(), hw_sysctl(),
vfs_sysctl(), etc, routines, along with sysctl_int() et al. Now all
nodes are registered with the tree, and nodes can be added (or
removed) easily, and I/O to and from the tree is handled generically.
Since the nodes are registered with the tree, the mapping from name to
number (and back again) can now be discovered, instead of having to be
hard coded. Adding new nodes to the tree is likewise much simpler --
the new infrastructure handles almost all the work for simple types,
and just about anything else can be done with a small helper function.
All existing nodes are where they were before (numerically speaking),
so all existing consumers of sysctl information should notice no
difference.
PS - I'm sorry, but there's a distinct lack of documentation at the
moment. I'm working on sysctl(3/8/9) right now, and I promise to
watch out for buses.
2003-12-04 22:38:21 +03:00
|
|
|
*oldlenp += tocopy;
|
2002-01-27 15:41:07 +03:00
|
|
|
left -= tocopy;
|
|
|
|
}
|
|
|
|
simple_unlock(&disklist_slock);
|
|
|
|
return (error);
|
|
|
|
}
|
2002-07-16 22:03:17 +04:00
|
|
|
|
|
|
|
struct bufq_fcfs {
|
|
|
|
TAILQ_HEAD(, buf) bq_head; /* actual list of buffers */
|
|
|
|
};
|
|
|
|
|
|
|
|
struct bufq_disksort {
|
|
|
|
TAILQ_HEAD(, buf) bq_head; /* actual list of buffers */
|
|
|
|
};
|
|
|
|
|
|
|
|
#define PRIO_READ_BURST 48
|
|
|
|
#define PRIO_WRITE_REQ 16
|
|
|
|
|
|
|
|
struct bufq_prio {
|
|
|
|
TAILQ_HEAD(, buf) bq_read, bq_write; /* actual list of buffers */
|
|
|
|
struct buf *bq_write_next; /* next request in bq_write */
|
2002-07-23 18:00:16 +04:00
|
|
|
struct buf *bq_next; /* current request */
|
2002-07-16 22:03:17 +04:00
|
|
|
int bq_read_burst; /* # of consecutive reads */
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check if two buf's are in ascending order.
|
|
|
|
*/
|
|
|
|
static __inline int
|
|
|
|
buf_inorder(struct buf *bp, struct buf *bq, int sortby)
|
|
|
|
{
|
|
|
|
|
|
|
|
if (bp == NULL || bq == NULL)
|
2002-11-01 06:32:21 +03:00
|
|
|
return (bq == NULL);
|
2002-07-16 22:03:17 +04:00
|
|
|
|
2003-12-06 20:23:22 +03:00
|
|
|
if (sortby == BUFQ_SORT_CYLINDER) {
|
|
|
|
if (bp->b_cylinder != bq->b_cylinder)
|
|
|
|
return bp->b_cylinder < bq->b_cylinder;
|
|
|
|
else
|
|
|
|
return bp->b_rawblkno < bq->b_rawblkno;
|
|
|
|
} else
|
|
|
|
return bp->b_rawblkno < bq->b_rawblkno;
|
2002-07-16 22:03:17 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
* First-come first-served sort for disks.
|
|
|
|
*
|
|
|
|
* Requests are appended to the queue without any reordering.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
bufq_fcfs_put(struct bufq_state *bufq, struct buf *bp)
|
|
|
|
{
|
|
|
|
struct bufq_fcfs *fcfs = bufq->bq_private;
|
|
|
|
|
|
|
|
TAILQ_INSERT_TAIL(&fcfs->bq_head, bp, b_actq);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct buf *
|
|
|
|
bufq_fcfs_get(struct bufq_state *bufq, int remove)
|
|
|
|
{
|
|
|
|
struct bufq_fcfs *fcfs = bufq->bq_private;
|
|
|
|
struct buf *bp;
|
|
|
|
|
|
|
|
bp = TAILQ_FIRST(&fcfs->bq_head);
|
|
|
|
|
|
|
|
if (bp != NULL && remove)
|
|
|
|
TAILQ_REMOVE(&fcfs->bq_head, bp, b_actq);
|
|
|
|
|
2002-11-01 06:32:21 +03:00
|
|
|
return (bp);
|
2002-07-16 22:03:17 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Seek sort for disks.
|
|
|
|
*
|
|
|
|
* There are actually two queues, sorted in ascendening order. The first
|
|
|
|
* queue holds those requests which are positioned after the current block;
|
|
|
|
* the second holds requests which came in after their position was passed.
|
|
|
|
* Thus we implement a one-way scan, retracting after reaching the end of
|
|
|
|
* the drive to the first request on the second queue, at which time it
|
|
|
|
* becomes the first queue.
|
|
|
|
*
|
|
|
|
* A one-way scan is natural because of the way UNIX read-ahead blocks are
|
|
|
|
* allocated.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
bufq_disksort_put(struct bufq_state *bufq, struct buf *bp)
|
|
|
|
{
|
|
|
|
struct bufq_disksort *disksort = bufq->bq_private;
|
|
|
|
struct buf *bq, *nbq;
|
|
|
|
int sortby;
|
|
|
|
|
|
|
|
sortby = bufq->bq_flags & BUFQ_SORT_MASK;
|
|
|
|
|
|
|
|
bq = TAILQ_FIRST(&disksort->bq_head);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the queue is empty it's easy; we just go on the end.
|
|
|
|
*/
|
|
|
|
if (bq == NULL) {
|
|
|
|
TAILQ_INSERT_TAIL(&disksort->bq_head, bp, b_actq);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we lie before the currently active request, then we
|
|
|
|
* must locate the second request list and add ourselves to it.
|
|
|
|
*/
|
|
|
|
if (buf_inorder(bp, bq, sortby)) {
|
|
|
|
while ((nbq = TAILQ_NEXT(bq, b_actq)) != NULL) {
|
|
|
|
/*
|
|
|
|
* Check for an ``inversion'' in the normally ascending
|
|
|
|
* block numbers, indicating the start of the second
|
|
|
|
* request list.
|
|
|
|
*/
|
|
|
|
if (buf_inorder(nbq, bq, sortby)) {
|
|
|
|
/*
|
|
|
|
* Search the second request list for the first
|
|
|
|
* request at a larger block number. We go
|
|
|
|
* after that; if there is no such request, we
|
|
|
|
* go at the end.
|
|
|
|
*/
|
|
|
|
do {
|
|
|
|
if (buf_inorder(bp, nbq, sortby))
|
|
|
|
goto insert;
|
|
|
|
bq = nbq;
|
2002-11-01 06:32:21 +03:00
|
|
|
} while ((nbq =
|
|
|
|
TAILQ_NEXT(bq, b_actq)) != NULL);
|
2002-07-16 22:03:17 +04:00
|
|
|
goto insert; /* after last */
|
|
|
|
}
|
|
|
|
bq = nbq;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* No inversions... we will go after the last, and
|
|
|
|
* be the first request in the second request list.
|
|
|
|
*/
|
|
|
|
goto insert;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Request is at/after the current request...
|
|
|
|
* sort in the first request list.
|
|
|
|
*/
|
|
|
|
while ((nbq = TAILQ_NEXT(bq, b_actq)) != NULL) {
|
|
|
|
/*
|
|
|
|
* We want to go after the current request if there is an
|
|
|
|
* inversion after it (i.e. it is the end of the first
|
|
|
|
* request list), or if the next request is a larger cylinder
|
|
|
|
* than our request.
|
|
|
|
*/
|
|
|
|
if (buf_inorder(nbq, bq, sortby) ||
|
|
|
|
buf_inorder(bp, nbq, sortby))
|
|
|
|
goto insert;
|
|
|
|
bq = nbq;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Neither a second list nor a larger request... we go at the end of
|
|
|
|
* the first list, which is the same as the end of the whole schebang.
|
|
|
|
*/
|
|
|
|
insert: TAILQ_INSERT_AFTER(&disksort->bq_head, bq, bp, b_actq);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct buf *
|
|
|
|
bufq_disksort_get(struct bufq_state *bufq, int remove)
|
|
|
|
{
|
|
|
|
struct bufq_disksort *disksort = bufq->bq_private;
|
|
|
|
struct buf *bp;
|
|
|
|
|
|
|
|
bp = TAILQ_FIRST(&disksort->bq_head);
|
|
|
|
|
|
|
|
if (bp != NULL && remove)
|
|
|
|
TAILQ_REMOVE(&disksort->bq_head, bp, b_actq);
|
|
|
|
|
2002-11-01 06:32:21 +03:00
|
|
|
return (bp);
|
2002-07-16 22:03:17 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Seek sort for disks.
|
|
|
|
*
|
|
|
|
* There are two queues. The first queue holds read requests; the second
|
|
|
|
* holds write requests. The read queue is first-come first-served; the
|
|
|
|
* write queue is sorted in ascendening block order.
|
|
|
|
* The read queue is processed first. After PRIO_READ_BURST consecutive
|
|
|
|
* read requests with non-empty write queue PRIO_WRITE_REQ requests from
|
|
|
|
* the write queue will be processed.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
bufq_prio_put(struct bufq_state *bufq, struct buf *bp)
|
|
|
|
{
|
|
|
|
struct bufq_prio *prio = bufq->bq_private;
|
|
|
|
struct buf *bq;
|
|
|
|
int sortby;
|
|
|
|
|
|
|
|
sortby = bufq->bq_flags & BUFQ_SORT_MASK;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If it's a read request append it to the list.
|
|
|
|
*/
|
|
|
|
if ((bp->b_flags & B_READ) == B_READ) {
|
|
|
|
TAILQ_INSERT_TAIL(&prio->bq_read, bp, b_actq);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
bq = TAILQ_FIRST(&prio->bq_write);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the write list is empty, simply append it to the list.
|
|
|
|
*/
|
|
|
|
if (bq == NULL) {
|
|
|
|
TAILQ_INSERT_TAIL(&prio->bq_write, bp, b_actq);
|
|
|
|
prio->bq_write_next = bp;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we lie after the next request, insert after this request.
|
|
|
|
*/
|
|
|
|
if (buf_inorder(prio->bq_write_next, bp, sortby))
|
|
|
|
bq = prio->bq_write_next;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Search for the first request at a larger block number.
|
|
|
|
* We go before this request if it exists.
|
|
|
|
*/
|
|
|
|
while (bq != NULL && buf_inorder(bq, bp, sortby))
|
|
|
|
bq = TAILQ_NEXT(bq, b_actq);
|
|
|
|
|
|
|
|
if (bq != NULL)
|
|
|
|
TAILQ_INSERT_BEFORE(bq, bp, b_actq);
|
|
|
|
else
|
|
|
|
TAILQ_INSERT_TAIL(&prio->bq_write, bp, b_actq);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct buf *
|
|
|
|
bufq_prio_get(struct bufq_state *bufq, int remove)
|
|
|
|
{
|
|
|
|
struct bufq_prio *prio = bufq->bq_private;
|
|
|
|
struct buf *bp;
|
|
|
|
|
|
|
|
/*
|
2002-07-23 18:00:16 +04:00
|
|
|
* If no current request, get next from the lists.
|
2002-07-16 22:03:17 +04:00
|
|
|
*/
|
2002-07-23 18:00:16 +04:00
|
|
|
if (prio->bq_next == NULL) {
|
2002-07-16 22:03:17 +04:00
|
|
|
/*
|
2002-07-23 18:00:16 +04:00
|
|
|
* If at least one list is empty, select the other.
|
2002-07-16 22:03:17 +04:00
|
|
|
*/
|
2002-07-23 18:00:16 +04:00
|
|
|
if (TAILQ_FIRST(&prio->bq_read) == NULL) {
|
|
|
|
prio->bq_next = prio->bq_write_next;
|
2002-07-16 22:03:17 +04:00
|
|
|
prio->bq_read_burst = 0;
|
2002-07-23 18:00:16 +04:00
|
|
|
} else if (prio->bq_write_next == NULL) {
|
|
|
|
prio->bq_next = TAILQ_FIRST(&prio->bq_read);
|
|
|
|
prio->bq_read_burst = 0;
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* Both list have requests. Select the read list up
|
|
|
|
* to PRIO_READ_BURST times, then select the write
|
|
|
|
* list PRIO_WRITE_REQ times.
|
|
|
|
*/
|
|
|
|
if (prio->bq_read_burst++ < PRIO_READ_BURST)
|
|
|
|
prio->bq_next = TAILQ_FIRST(&prio->bq_read);
|
|
|
|
else if (prio->bq_read_burst <
|
2002-11-01 06:32:21 +03:00
|
|
|
PRIO_READ_BURST + PRIO_WRITE_REQ)
|
2002-07-23 18:00:16 +04:00
|
|
|
prio->bq_next = prio->bq_write_next;
|
|
|
|
else {
|
|
|
|
prio->bq_next = TAILQ_FIRST(&prio->bq_read);
|
|
|
|
prio->bq_read_burst = 0;
|
|
|
|
}
|
2002-07-16 22:03:17 +04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2002-07-23 18:00:16 +04:00
|
|
|
bp = prio->bq_next;
|
|
|
|
|
2002-11-01 06:34:07 +03:00
|
|
|
if (bp != NULL && remove) {
|
|
|
|
if ((bp->b_flags & B_READ) == B_READ)
|
|
|
|
TAILQ_REMOVE(&prio->bq_read, bp, b_actq);
|
2002-07-16 22:03:17 +04:00
|
|
|
else {
|
|
|
|
/*
|
2002-11-01 06:34:07 +03:00
|
|
|
* Advance the write pointer before removing
|
|
|
|
* bp since it is actually prio->bq_write_next.
|
2002-07-16 22:03:17 +04:00
|
|
|
*/
|
|
|
|
prio->bq_write_next =
|
|
|
|
TAILQ_NEXT(prio->bq_write_next, b_actq);
|
2002-11-01 06:34:07 +03:00
|
|
|
TAILQ_REMOVE(&prio->bq_write, bp, b_actq);
|
2002-07-16 22:03:17 +04:00
|
|
|
if (prio->bq_write_next == NULL)
|
|
|
|
prio->bq_write_next =
|
|
|
|
TAILQ_FIRST(&prio->bq_write);
|
|
|
|
}
|
2002-07-23 18:00:16 +04:00
|
|
|
|
|
|
|
prio->bq_next = NULL;
|
2002-07-16 22:03:17 +04:00
|
|
|
}
|
|
|
|
|
2002-11-01 06:32:21 +03:00
|
|
|
return (bp);
|
2002-07-16 22:03:17 +04:00
|
|
|
}
|
|
|
|
|
2004-01-10 17:49:44 +03:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Cyclical scan (CSCAN)
|
|
|
|
*/
|
|
|
|
TAILQ_HEAD(bqhead, buf);
|
|
|
|
struct cscan_queue {
|
|
|
|
struct bqhead cq_head[2]; /* actual lists of buffers */
|
|
|
|
int cq_idx; /* current list index */
|
|
|
|
int cq_lastcylinder; /* b_cylinder of the last request */
|
|
|
|
daddr_t cq_lastrawblkno; /* b_rawblkno of the last request */
|
|
|
|
};
|
|
|
|
|
|
|
|
static int __inline cscan_empty(const struct cscan_queue *);
|
|
|
|
static void cscan_put(struct cscan_queue *, struct buf *, int);
|
|
|
|
static struct buf *cscan_get(struct cscan_queue *, int);
|
|
|
|
static void cscan_init(struct cscan_queue *);
|
|
|
|
|
|
|
|
static __inline int
|
|
|
|
cscan_empty(const struct cscan_queue *q)
|
|
|
|
{
|
|
|
|
|
|
|
|
return TAILQ_EMPTY(&q->cq_head[0]) && TAILQ_EMPTY(&q->cq_head[1]);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
cscan_put(struct cscan_queue *q, struct buf *bp, int sortby)
|
|
|
|
{
|
|
|
|
struct buf tmp;
|
|
|
|
struct buf *it;
|
|
|
|
struct bqhead *bqh;
|
|
|
|
int idx;
|
|
|
|
|
|
|
|
tmp.b_cylinder = q->cq_lastcylinder;
|
|
|
|
tmp.b_rawblkno = q->cq_lastrawblkno;
|
|
|
|
|
|
|
|
if (buf_inorder(bp, &tmp, sortby))
|
|
|
|
idx = 1 - q->cq_idx;
|
|
|
|
else
|
|
|
|
idx = q->cq_idx;
|
|
|
|
|
|
|
|
bqh = &q->cq_head[idx];
|
|
|
|
|
|
|
|
TAILQ_FOREACH(it, bqh, b_actq)
|
|
|
|
if (buf_inorder(bp, it, sortby))
|
|
|
|
break;
|
|
|
|
|
|
|
|
if (it != NULL)
|
|
|
|
TAILQ_INSERT_BEFORE(it, bp, b_actq);
|
|
|
|
else
|
|
|
|
TAILQ_INSERT_TAIL(bqh, bp, b_actq);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct buf *
|
|
|
|
cscan_get(struct cscan_queue *q, int remove)
|
|
|
|
{
|
|
|
|
int idx = q->cq_idx;
|
|
|
|
struct bqhead *bqh;
|
|
|
|
struct buf *bp;
|
|
|
|
|
|
|
|
bqh = &q->cq_head[idx];
|
|
|
|
bp = TAILQ_FIRST(bqh);
|
|
|
|
|
|
|
|
if (bp == NULL) {
|
|
|
|
/* switch queue */
|
|
|
|
idx = 1 - idx;
|
|
|
|
bqh = &q->cq_head[idx];
|
|
|
|
bp = TAILQ_FIRST(bqh);
|
|
|
|
}
|
|
|
|
|
|
|
|
KDASSERT((bp != NULL && !cscan_empty(q)) ||
|
|
|
|
(bp == NULL && cscan_empty(q)));
|
|
|
|
|
|
|
|
if (bp != NULL && remove) {
|
|
|
|
q->cq_idx = idx;
|
|
|
|
TAILQ_REMOVE(bqh, bp, b_actq);
|
|
|
|
|
|
|
|
q->cq_lastcylinder = bp->b_cylinder;
|
|
|
|
q->cq_lastrawblkno =
|
|
|
|
bp->b_rawblkno + (bp->b_bcount >> DEV_BSHIFT);
|
|
|
|
}
|
|
|
|
|
|
|
|
return (bp);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
cscan_init(struct cscan_queue *q)
|
|
|
|
{
|
|
|
|
|
|
|
|
TAILQ_INIT(&q->cq_head[0]);
|
|
|
|
TAILQ_INIT(&q->cq_head[1]);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Per-prioritiy CSCAN.
|
|
|
|
*
|
|
|
|
* XXX probably we should have a way to raise
|
|
|
|
* priority of the on-queue requests.
|
|
|
|
*/
|
|
|
|
#define PRIOCSCAN_NQUEUE 3
|
|
|
|
|
|
|
|
struct priocscan_queue {
|
|
|
|
struct cscan_queue q_queue;
|
|
|
|
int q_burst;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct bufq_priocscan {
|
|
|
|
struct priocscan_queue bq_queue[PRIOCSCAN_NQUEUE];
|
|
|
|
|
|
|
|
#if 0
|
|
|
|
/*
|
|
|
|
* XXX using "global" head position can reduce positioning time
|
|
|
|
* when switching between queues.
|
|
|
|
* although it might affect against fairness.
|
|
|
|
*/
|
|
|
|
daddr_t bq_lastrawblkno;
|
|
|
|
int bq_lastcylinder;
|
|
|
|
#endif
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* how many requests to serve when having pending requests on other queues.
|
|
|
|
*
|
|
|
|
* XXX tune
|
|
|
|
*/
|
|
|
|
const int priocscan_burst[] = {
|
|
|
|
64, 16, 4
|
|
|
|
};
|
|
|
|
|
|
|
|
static void bufq_priocscan_put(struct bufq_state *, struct buf *);
|
|
|
|
static struct buf *bufq_priocscan_get(struct bufq_state *, int);
|
|
|
|
static void bufq_priocscan_init(struct bufq_state *);
|
|
|
|
static __inline struct cscan_queue *bufq_priocscan_selectqueue(
|
|
|
|
struct bufq_priocscan *, const struct buf *);
|
|
|
|
|
|
|
|
static __inline struct cscan_queue *
|
|
|
|
bufq_priocscan_selectqueue(struct bufq_priocscan *q, const struct buf *bp)
|
|
|
|
{
|
|
|
|
static const int priocscan_priomap[] = {
|
|
|
|
[BPRIO_TIMENONCRITICAL] = 2,
|
|
|
|
[BPRIO_TIMELIMITED] = 1,
|
|
|
|
[BPRIO_TIMECRITICAL] = 0
|
|
|
|
};
|
|
|
|
|
|
|
|
return &q->bq_queue[priocscan_priomap[BIO_GETPRIO(bp)]].q_queue;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
bufq_priocscan_put(struct bufq_state *bufq, struct buf *bp)
|
|
|
|
{
|
|
|
|
struct bufq_priocscan *q = bufq->bq_private;
|
|
|
|
struct cscan_queue *cq;
|
|
|
|
const int sortby = bufq->bq_flags & BUFQ_SORT_MASK;
|
|
|
|
|
|
|
|
cq = bufq_priocscan_selectqueue(q, bp);
|
|
|
|
cscan_put(cq, bp, sortby);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct buf *
|
|
|
|
bufq_priocscan_get(struct bufq_state *bufq, int remove)
|
|
|
|
{
|
|
|
|
struct bufq_priocscan *q = bufq->bq_private;
|
|
|
|
struct priocscan_queue *pq, *npq;
|
|
|
|
struct priocscan_queue *first; /* first non-empty queue */
|
|
|
|
const struct priocscan_queue *epq;
|
|
|
|
const struct cscan_queue *cq;
|
|
|
|
struct buf *bp;
|
|
|
|
boolean_t single; /* true if there's only one non-empty queue */
|
|
|
|
|
|
|
|
pq = &q->bq_queue[0];
|
|
|
|
epq = pq + PRIOCSCAN_NQUEUE;
|
|
|
|
for (; pq < epq; pq++) {
|
|
|
|
cq = &pq->q_queue;
|
|
|
|
if (!cscan_empty(cq))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (pq == epq) {
|
|
|
|
/* there's no requests */
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
first = pq;
|
|
|
|
single = TRUE;
|
|
|
|
for (npq = first + 1; npq < epq; npq++) {
|
|
|
|
cq = &npq->q_queue;
|
|
|
|
if (!cscan_empty(cq)) {
|
|
|
|
single = FALSE;
|
|
|
|
if (pq->q_burst > 0)
|
|
|
|
break;
|
|
|
|
pq = npq;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (single) {
|
|
|
|
/*
|
|
|
|
* there's only a non-empty queue. just serve it.
|
|
|
|
*/
|
|
|
|
pq = first;
|
|
|
|
} else if (pq->q_burst > 0) {
|
|
|
|
/*
|
|
|
|
* XXX account only by number of requests. is it good enough?
|
|
|
|
*/
|
|
|
|
pq->q_burst--;
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* no queue was selected due to burst counts
|
|
|
|
*/
|
|
|
|
int i;
|
|
|
|
#ifdef DEBUG
|
|
|
|
for (i = 0; i < PRIOCSCAN_NQUEUE; i++) {
|
|
|
|
pq = &q->bq_queue[i];
|
|
|
|
cq = &pq->q_queue;
|
|
|
|
if (!cscan_empty(cq) && pq->q_burst)
|
|
|
|
panic("%s: inconsist", __func__);
|
|
|
|
}
|
|
|
|
#endif /* DEBUG */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* reset burst counts
|
|
|
|
*/
|
|
|
|
for (i = 0; i < PRIOCSCAN_NQUEUE; i++) {
|
|
|
|
pq = &q->bq_queue[i];
|
|
|
|
pq->q_burst = priocscan_burst[i];
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* serve first non-empty queue.
|
|
|
|
*/
|
|
|
|
pq = first;
|
|
|
|
}
|
|
|
|
|
|
|
|
KDASSERT(!cscan_empty(&pq->q_queue));
|
|
|
|
bp = cscan_get(&pq->q_queue, remove);
|
|
|
|
KDASSERT(bp != NULL);
|
|
|
|
KDASSERT(&pq->q_queue == bufq_priocscan_selectqueue(q, bp));
|
|
|
|
|
|
|
|
return bp;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
bufq_priocscan_init(struct bufq_state *bufq)
|
|
|
|
{
|
|
|
|
struct bufq_priocscan *q;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
bufq->bq_get = bufq_priocscan_get;
|
|
|
|
bufq->bq_put = bufq_priocscan_put;
|
|
|
|
bufq->bq_private = malloc(sizeof(struct bufq_priocscan),
|
|
|
|
M_DEVBUF, M_ZERO);
|
|
|
|
|
|
|
|
q = bufq->bq_private;
|
|
|
|
for (i = 0; i < PRIOCSCAN_NQUEUE; i++) {
|
|
|
|
struct cscan_queue *cq = &q->bq_queue[i].q_queue;
|
|
|
|
|
|
|
|
cscan_init(cq);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2002-07-21 19:32:17 +04:00
|
|
|
/*
|
|
|
|
* Create a device buffer queue.
|
|
|
|
*/
|
2002-07-16 22:03:17 +04:00
|
|
|
void
|
2002-07-21 19:32:17 +04:00
|
|
|
bufq_alloc(struct bufq_state *bufq, int flags)
|
2002-07-16 22:03:17 +04:00
|
|
|
{
|
|
|
|
struct bufq_fcfs *fcfs;
|
|
|
|
struct bufq_disksort *disksort;
|
|
|
|
struct bufq_prio *prio;
|
|
|
|
|
|
|
|
bufq->bq_flags = flags;
|
|
|
|
|
|
|
|
switch (flags & BUFQ_SORT_MASK) {
|
|
|
|
case BUFQ_SORT_RAWBLOCK:
|
|
|
|
case BUFQ_SORT_CYLINDER:
|
|
|
|
break;
|
|
|
|
case 0:
|
|
|
|
if ((flags & BUFQ_METHOD_MASK) == BUFQ_FCFS)
|
|
|
|
break;
|
|
|
|
/* FALLTHROUGH */
|
|
|
|
default:
|
2002-07-21 19:32:17 +04:00
|
|
|
panic("bufq_alloc: sort out of range");
|
2002-07-16 22:03:17 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
switch (flags & BUFQ_METHOD_MASK) {
|
|
|
|
case BUFQ_FCFS:
|
|
|
|
bufq->bq_get = bufq_fcfs_get;
|
|
|
|
bufq->bq_put = bufq_fcfs_put;
|
2002-07-21 19:32:17 +04:00
|
|
|
MALLOC(bufq->bq_private, struct bufq_fcfs *,
|
|
|
|
sizeof(struct bufq_fcfs), M_DEVBUF, M_ZERO);
|
2002-07-16 22:03:17 +04:00
|
|
|
fcfs = (struct bufq_fcfs *)bufq->bq_private;
|
|
|
|
TAILQ_INIT(&fcfs->bq_head);
|
|
|
|
break;
|
|
|
|
case BUFQ_DISKSORT:
|
|
|
|
bufq->bq_get = bufq_disksort_get;
|
|
|
|
bufq->bq_put = bufq_disksort_put;
|
2002-07-21 19:32:17 +04:00
|
|
|
MALLOC(bufq->bq_private, struct bufq_disksort *,
|
|
|
|
sizeof(struct bufq_disksort), M_DEVBUF, M_ZERO);
|
2002-07-16 22:03:17 +04:00
|
|
|
disksort = (struct bufq_disksort *)bufq->bq_private;
|
|
|
|
TAILQ_INIT(&disksort->bq_head);
|
|
|
|
break;
|
|
|
|
case BUFQ_READ_PRIO:
|
|
|
|
bufq->bq_get = bufq_prio_get;
|
|
|
|
bufq->bq_put = bufq_prio_put;
|
2002-07-21 19:32:17 +04:00
|
|
|
MALLOC(bufq->bq_private, struct bufq_prio *,
|
|
|
|
sizeof(struct bufq_prio), M_DEVBUF, M_ZERO);
|
2002-07-16 22:03:17 +04:00
|
|
|
prio = (struct bufq_prio *)bufq->bq_private;
|
|
|
|
TAILQ_INIT(&prio->bq_read);
|
|
|
|
TAILQ_INIT(&prio->bq_write);
|
|
|
|
break;
|
2004-01-10 17:49:44 +03:00
|
|
|
case BUFQ_PRIOCSCAN:
|
|
|
|
bufq_priocscan_init(bufq);
|
|
|
|
break;
|
2002-07-16 22:03:17 +04:00
|
|
|
default:
|
2002-07-21 19:32:17 +04:00
|
|
|
panic("bufq_alloc: method out of range");
|
2002-07-16 22:03:17 +04:00
|
|
|
}
|
|
|
|
}
|
2002-07-21 19:32:17 +04:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Destroy a device buffer queue.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
bufq_free(struct bufq_state *bufq)
|
|
|
|
{
|
2002-11-01 06:32:21 +03:00
|
|
|
|
2002-07-21 19:32:17 +04:00
|
|
|
KASSERT(bufq->bq_private != NULL);
|
|
|
|
KASSERT(BUFQ_PEEK(bufq) == NULL);
|
|
|
|
|
|
|
|
FREE(bufq->bq_private, M_DEVBUF);
|
|
|
|
bufq->bq_get = NULL;
|
|
|
|
bufq->bq_put = NULL;
|
|
|
|
}
|
2003-04-04 02:20:24 +04:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Bounds checking against the media size, used for the raw partition.
|
|
|
|
* The sector size passed in should currently always be DEV_BSIZE,
|
|
|
|
* and the media size the size of the device in DEV_BSIZE sectors.
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
bounds_check_with_mediasize(struct buf *bp, int secsize, u_int64_t mediasize)
|
|
|
|
{
|
|
|
|
int sz;
|
|
|
|
|
|
|
|
sz = howmany(bp->b_bcount, secsize);
|
|
|
|
|
|
|
|
if (bp->b_blkno + sz > mediasize) {
|
|
|
|
sz = mediasize - bp->b_blkno;
|
|
|
|
if (sz == 0) {
|
|
|
|
/* If exactly at end of disk, return EOF. */
|
|
|
|
bp->b_resid = bp->b_bcount;
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
if (sz < 0) {
|
|
|
|
/* If past end of disk, return EINVAL. */
|
|
|
|
bp->b_error = EINVAL;
|
|
|
|
goto bad;
|
|
|
|
}
|
|
|
|
/* Otherwise, truncate request. */
|
|
|
|
bp->b_bcount = sz << DEV_BSHIFT;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
bad:
|
|
|
|
bp->b_flags |= B_ERROR;
|
|
|
|
done:
|
|
|
|
return 0;
|
|
|
|
}
|