NetBSD/sys/altq/altq_rmclass.c

1877 lines
44 KiB
C

/* $NetBSD: altq_rmclass.c,v 1.9 2005/02/26 23:04:16 perry Exp $ */
/* $KAME: altq_rmclass.c,v 1.9 2000/12/14 08:12:46 thorpej Exp $ */
/*
* Copyright (c) 1991-1997 Regents of the University of California.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. All advertising materials mentioning features or use of this software
* must display the following acknowledgement:
* This product includes software developed by the Network Research
* Group at Lawrence Berkeley Laboratory.
* 4. Neither the name of the University nor of the Laboratory may be used
* to endorse or promote products derived from this software without
* specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*
* LBL code modified by speer@eng.sun.com, May 1977.
* For questions and/or comments, please send mail to cbq@ee.lbl.gov
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: altq_rmclass.c,v 1.9 2005/02/26 23:04:16 perry Exp $");
#ident "@(#)rm_class.c 1.48 97/12/05 SMI"
#if defined(__FreeBSD__) || defined(__NetBSD__)
#include "opt_altq.h"
#if (__FreeBSD__ != 2)
#include "opt_inet.h"
#ifdef __FreeBSD__
#include "opt_inet6.h"
#endif
#endif
#endif /* __FreeBSD__ || __NetBSD__ */
#ifdef ALTQ_CBQ /* cbq is enabled by ALTQ_CBQ option in opt_altq.h */
#include <sys/param.h>
#include <sys/malloc.h>
#include <sys/mbuf.h>
#include <sys/socket.h>
#include <sys/systm.h>
#include <sys/errno.h>
#include <sys/time.h>
#include <sys/kernel.h>
#include <net/if.h>
#include <netinet/in.h>
#include <netinet/in_systm.h>
#include <netinet/ip.h>
#include <altq/altq.h>
#include <altq/altq_rmclass.h>
#include <altq/altq_rmclass_debug.h>
#include <altq/altq_red.h>
#include <altq/altq_rio.h>
#ifdef CBQ_TRACE
struct cbqtrace cbqtrace_buffer[NCBQTRACE+1];
struct cbqtrace *cbqtrace_ptr = NULL;
int cbqtrace_count;
#endif
/*
* Local Macros
*/
#define reset_cutoff(ifd) { ifd->cutoff_ = RM_MAXDEPTH; }
/*
* Local routines.
*/
static int rmc_satisfied __P((struct rm_class *, struct timeval *));
static void rmc_wrr_set_weights __P((struct rm_ifdat *));
static void rmc_depth_compute __P((struct rm_class *));
static void rmc_depth_recompute __P((rm_class_t *));
static mbuf_t *_rmc_wrr_dequeue_next __P((struct rm_ifdat *, int));
static mbuf_t *_rmc_prr_dequeue_next __P((struct rm_ifdat *, int));
static int _rmc_addq __P((rm_class_t *, mbuf_t *));
static void _rmc_dropq __P((rm_class_t *));
static mbuf_t *_rmc_getq __P((rm_class_t *));
static mbuf_t *_rmc_pollq __P((rm_class_t *));
static int rmc_under_limit __P((struct rm_class *, struct timeval *));
static void rmc_tl_satisfied __P((struct rm_ifdat *, struct timeval *));
static void rmc_drop_action __P((struct rm_class *));
static void rmc_restart __P((struct rm_class *));
static void rmc_root_overlimit __P((struct rm_class *, struct rm_class *));
/*
* BORROW_OFFTIME (experimental):
* borrow the offtime of the class borrowing from.
* the reason is that when its own offtime is set, the class is unable
* to borrow much, especially when cutoff is taking effect.
* but when the borrowed class is overloaded (advidle is close to minidle),
* use the borrowing class's offtime to avoid overload.
*/
/*
* ADJUST_CUTOFF (experimental):
* if no underlimit class is found due to cutoff, increase cutoff and
* retry the scheduling loop.
* also, don't invoke delay_actions while cutoff is taking effect,
* since a sleeping class won't have a chance to be scheduled in the
* next loop.
*
* now heuristics for setting the top-level variable (cutoff_) becomes:
* 1. if a packet arrives for a not-overlimit class, set cutoff
* to the depth of the class.
* 2. if cutoff is i, and a packet arrives for an overlimit class
* with an underlimit ancestor at a lower level than i (say j),
* then set cutoff to j.
* 3. at scheduling a packet, if there is no underlimit class
* due to the current cutoff level, increase cutoff by 1 and
* then try to schedule again.
*/
/*
* rm_class_t *
* rmc_newclass(...) - Create a new resource management class at priority
* 'pri' on the interface given by 'ifd'.
*
* nsecPerByte is the data rate of the interface in nanoseconds/byte.
* E.g., 800 for a 10Mb/s ethernet. If the class gets less
* than 100% of the bandwidth, this number should be the
* 'effective' rate for the class. Let f be the
* bandwidth fraction allocated to this class, and let
* nsPerByte be the data rate of the output link in
* nanoseconds/byte. Then nsecPerByte is set to
* nsPerByte / f. E.g., 1600 (= 800 / .5)
* for a class that gets 50% of an ethernet's bandwidth.
*
* action the routine to call when the class is over limit.
*
* maxq max allowable queue size for class (in packets).
*
* parent parent class pointer.
*
* borrow class to borrow from (should be either 'parent' or null).
*
* maxidle max value allowed for class 'idle' time estimate (this
* parameter determines how large an initial burst of packets
* can be before overlimit action is invoked.
*
* offtime how long 'delay' action will delay when class goes over
* limit (this parameter determines the steady-state burst
* size when a class is running over its limit).
*
* Maxidle and offtime have to be computed from the following: If the
* average packet size is s, the bandwidth fraction allocated to this
* class is f, we want to allow b packet bursts, and the gain of the
* averaging filter is g (= 1 - 2^(-RM_FILTER_GAIN)), then:
*
* ptime = s * nsPerByte * (1 - f) / f
* maxidle = ptime * (1 - g^b) / g^b
* minidle = -ptime * (1 / (f - 1))
* offtime = ptime * (1 + 1/(1 - g) * (1 - g^(b - 1)) / g^(b - 1)
*
* Operationally, it's convenient to specify maxidle & offtime in units
* independent of the link bandwidth so the maxidle & offtime passed to
* this routine are the above values multiplied by 8*f/(1000*nsPerByte).
* (The constant factor is a scale factor needed to make the parameters
* integers. This scaling also means that the 'unscaled' values of
* maxidle*nsecPerByte/8 and offtime*nsecPerByte/8 will be in microseconds,
* not nanoseconds.) Also note that the 'idle' filter computation keeps
* an estimate scaled upward by 2^RM_FILTER_GAIN so the passed value of
* maxidle also must be scaled upward by this value. Thus, the passed
* values for maxidle and offtime can be computed as follows:
*
* maxidle = maxidle * 2^RM_FILTER_GAIN * 8 / (1000 * nsecPerByte)
* offtime = offtime * 8 / (1000 * nsecPerByte)
*
* When USE_HRTIME is employed, then maxidle and offtime become:
* maxidle = maxilde * (8.0 / nsecPerByte);
* offtime = offtime * (8.0 / nsecPerByte);
*/
struct rm_class *
rmc_newclass(pri, ifd, nsecPerByte, action, maxq, parent, borrow,
maxidle, minidle, offtime, pktsize, flags)
int pri;
struct rm_ifdat *ifd;
u_int nsecPerByte;
void (*action)(rm_class_t *, rm_class_t *);
int maxq;
struct rm_class *parent;
struct rm_class *borrow;
u_int maxidle;
int minidle;
u_int offtime;
int pktsize;
int flags;
{
struct rm_class *cl;
struct rm_class *peer;
int s;
if (pri >= RM_MAXPRIO)
return (NULL);
#ifndef ALTQ_RED
if (flags & RMCF_RED) {
printf("rmc_newclass: RED not configured for CBQ!\n");
return (NULL);
}
#endif
#ifndef ALTQ_RIO
if (flags & RMCF_RIO) {
printf("rmc_newclass: RIO not configured for CBQ!\n");
return (NULL);
}
#endif
MALLOC(cl, struct rm_class *, sizeof(struct rm_class),
M_DEVBUF, M_WAITOK);
if (cl == NULL)
return (NULL);
(void)memset(cl, 0, sizeof(struct rm_class));
CALLOUT_INIT(&cl->callout_);
MALLOC(cl->q_, class_queue_t *, sizeof(class_queue_t),
M_DEVBUF, M_WAITOK);
if (cl->q_ == NULL) {
FREE(cl, M_DEVBUF);
return (NULL);
}
(void)memset(cl->q_, 0, sizeof(class_queue_t));
/*
* Class initialization.
*/
cl->children_ = NULL;
cl->parent_ = parent;
cl->borrow_ = borrow;
cl->leaf_ = 1;
cl->ifdat_ = ifd;
cl->pri_ = pri;
cl->allotment_ = RM_NS_PER_SEC / nsecPerByte; /* Bytes per sec */
cl->depth_ = 0;
cl->qthresh_ = 0;
cl->ns_per_byte_ = nsecPerByte;
qlimit(cl->q_) = maxq;
qtype(cl->q_) = Q_DROPHEAD;
qlen(cl->q_) = 0;
cl->flags_ = flags;
#if 1 /* minidle is also scaled in ALTQ */
cl->minidle_ = (minidle * (int)nsecPerByte) / 8;
if (cl->minidle_ > 0)
cl->minidle_ = 0;
#else
cl->minidle_ = minidle;
#endif
cl->maxidle_ = (maxidle * nsecPerByte) / 8;
if (cl->maxidle_ == 0)
cl->maxidle_ = 1;
#if 1 /* offtime is also scaled in ALTQ */
cl->avgidle_ = cl->maxidle_;
cl->offtime_ = ((offtime * nsecPerByte) / 8) >> RM_FILTER_GAIN;
if (cl->offtime_ == 0)
cl->offtime_ = 1;
#else
cl->avgidle_ = 0;
cl->offtime_ = (offtime * nsecPerByte) / 8;
#endif
cl->overlimit = action;
#ifdef ALTQ_RED
if (flags & (RMCF_RED|RMCF_RIO)) {
int red_flags, red_pkttime;
red_flags = 0;
if (flags & RMCF_ECN)
red_flags |= REDF_ECN;
if (flags & RMCF_FLOWVALVE)
red_flags |= REDF_FLOWVALVE;
#ifdef ALTQ_RIO
if (flags & RMCF_CLEARDSCP)
red_flags |= RIOF_CLEARDSCP;
#endif
red_pkttime = nsecPerByte * pktsize / 1000;
if (flags & RMCF_RED) {
cl->red_ = red_alloc(0, 0, 0, 0,
red_flags, red_pkttime);
if (cl->red_ != NULL)
qtype(cl->q_) = Q_RED;
}
#ifdef ALTQ_RIO
else {
cl->red_ = (red_t *)rio_alloc(0, NULL,
red_flags, red_pkttime);
if (cl->red_ != NULL)
qtype(cl->q_) = Q_RIO;
}
#endif
}
#endif /* ALTQ_RED */
/*
* put the class into the class tree
*/
s = splnet();
if ((peer = ifd->active_[pri]) != NULL) {
/* find the last class at this pri */
cl->peer_ = peer;
while (peer->peer_ != ifd->active_[pri])
peer = peer->peer_;
peer->peer_ = cl;
} else {
ifd->active_[pri] = cl;
cl->peer_ = cl;
}
if (cl->parent_) {
cl->next_ = parent->children_;
parent->children_ = cl;
parent->leaf_ = 0;
}
/*
* Compute the depth of this class and it's ancestors in the class
* hierarchy.
*/
rmc_depth_compute(cl);
/*
* If CBQ's WRR is enabled, then initailize the class WRR state.
*/
if (ifd->wrr_) {
ifd->num_[pri]++;
ifd->alloc_[pri] += cl->allotment_;
rmc_wrr_set_weights(ifd);
}
splx(s);
return (cl);
}
int
rmc_modclass(cl, nsecPerByte, maxq, maxidle, minidle, offtime, pktsize)
struct rm_class *cl;
u_int nsecPerByte;
int maxq;
u_int maxidle;
int minidle;
u_int offtime;
int pktsize;
{
struct rm_ifdat *ifd;
u_int old_allotment;
int s;
ifd = cl->ifdat_;
old_allotment = cl->allotment_;
s = splnet();
cl->allotment_ = RM_NS_PER_SEC / nsecPerByte; /* Bytes per sec */
cl->qthresh_ = 0;
cl->ns_per_byte_ = nsecPerByte;
qlimit(cl->q_) = maxq;
#if 1 /* minidle is also scaled in ALTQ */
cl->minidle_ = (minidle * nsecPerByte) / 8;
if (cl->minidle_ > 0)
cl->minidle_ = 0;
#else
cl->minidle_ = minidle;
#endif
cl->maxidle_ = (maxidle * nsecPerByte) / 8;
if (cl->maxidle_ == 0)
cl->maxidle_ = 1;
#if 1 /* offtime is also scaled in ALTQ */
cl->avgidle_ = cl->maxidle_;
cl->offtime_ = ((offtime * nsecPerByte) / 8) >> RM_FILTER_GAIN;
if (cl->offtime_ == 0)
cl->offtime_ = 1;
#else
cl->avgidle_ = 0;
cl->offtime_ = (offtime * nsecPerByte) / 8;
#endif
/*
* If CBQ's WRR is enabled, then initailize the class WRR state.
*/
if (ifd->wrr_) {
ifd->alloc_[cl->pri_] += cl->allotment_ - old_allotment;
rmc_wrr_set_weights(ifd);
}
splx(s);
return (0);
}
/*
* static void
* rmc_wrr_set_weights(struct rm_ifdat *ifdat) - This function computes
* the appropriate run robin weights for the CBQ weighted round robin
* algorithm.
*
* Returns: NONE
*/
static void
rmc_wrr_set_weights(ifd)
struct rm_ifdat *ifd;
{
int i;
struct rm_class *cl, *clh;
for (i = 0; i < RM_MAXPRIO; i++) {
/*
* This is inverted from that of the simulator to
* maintain precision.
*/
if (ifd->num_[i] == 0)
ifd->M_[i] = 0;
else
ifd->M_[i] = ifd->alloc_[i] /
(ifd->num_[i] * ifd->maxpkt_);
/*
* Compute the weigthed allotment for each class.
* This takes the expensive div instruction out
* of the main loop for the wrr scheduling path.
* These only get recomputed when a class comes or
* goes.
*/
if (ifd->active_[i] != NULL) {
clh = cl = ifd->active_[i];
do {
/* safe-guard for slow link or alloc_ == 0 */
if (ifd->M_[i] == 0)
cl->w_allotment_ = 0;
else
cl->w_allotment_ = cl->allotment_ /
ifd->M_[i];
cl = cl->peer_;
} while ((cl != NULL) && (cl != clh));
}
}
}
int
rmc_get_weight(ifd, pri)
struct rm_ifdat *ifd;
int pri;
{
if ((pri >= 0) && (pri < RM_MAXPRIO))
return (ifd->M_[pri]);
else
return (0);
}
/*
* static void
* rmc_depth_compute(struct rm_class *cl) - This function computes the
* appropriate depth of class 'cl' and its ancestors.
*
* Returns: NONE
*/
static void
rmc_depth_compute(cl)
struct rm_class *cl;
{
rm_class_t *t = cl, *p;
/*
* Recompute the depth for the branch of the tree.
*/
while (t != NULL) {
p = t->parent_;
if (p && (t->depth_ >= p->depth_)) {
p->depth_ = t->depth_ + 1;
t = p;
} else
t = NULL;
}
}
/*
* static void
* rmc_depth_recompute(struct rm_class *cl) - This function re-computes
* the depth of the tree after a class has been deleted.
*
* Returns: NONE
*/
static void
rmc_depth_recompute(rm_class_t *cl)
{
#if 1 /* ALTQ */
rm_class_t *p, *t;
p = cl;
while (p != NULL) {
if ((t = p->children_) == NULL) {
p->depth_ = 0;
} else {
int cdepth = 0;
while (t != NULL) {
if (t->depth_ > cdepth)
cdepth = t->depth_;
t = t->next_;
}
if (p->depth_ == cdepth + 1)
/* no change to this parent */
return;
p->depth_ = cdepth + 1;
}
p = p->parent_;
}
#else
rm_class_t *t;
if (cl->depth_ >= 1) {
if (cl->children_ == NULL) {
cl->depth_ = 0;
} else if ((t = cl->children_) != NULL) {
while (t != NULL) {
if (t->children_ != NULL)
rmc_depth_recompute(t);
t = t->next_;
}
} else
rmc_depth_compute(cl);
}
#endif
}
/*
* void
* rmc_delete_class(struct rm_ifdat *ifdat, struct rm_class *cl) - This
* function deletes a class from the link-sharing stucture and frees
* all resources associated with the class.
*
* Returns: NONE
*/
void
rmc_delete_class(ifd, cl)
struct rm_ifdat *ifd;
struct rm_class *cl;
{
struct rm_class *p, *head, *previous;
int s;
ASSERT(cl->children_ == NULL);
if (cl->sleeping_)
CALLOUT_STOP(&cl->callout_);
s = splnet();
/*
* Free packets in the packet queue.
* XXX - this may not be a desired behavior. Packets should be
* re-queued.
*/
rmc_dropall(cl);
/*
* If the class has a parent, then remove the class from the
* class from the parent's children chain.
*/
if (cl->parent_ != NULL) {
head = cl->parent_->children_;
p = previous = head;
if (head->next_ == NULL) {
ASSERT(head == cl);
cl->parent_->children_ = NULL;
cl->parent_->leaf_ = 1;
} else while (p != NULL) {
if (p == cl) {
if (cl == head)
cl->parent_->children_ = cl->next_;
else
previous->next_ = cl->next_;
cl->next_ = NULL;
p = NULL;
} else {
previous = p;
p = p->next_;
}
}
}
/*
* Delete class from class priority peer list.
*/
if ((p = ifd->active_[cl->pri_]) != NULL) {
/*
* If there is more than one member of this priority
* level, then look for class(cl) in the priority level.
*/
if (p != p->peer_) {
while (p->peer_ != cl)
p = p->peer_;
p->peer_ = cl->peer_;
if (ifd->active_[cl->pri_] == cl)
ifd->active_[cl->pri_] = cl->peer_;
} else {
ASSERT(p == cl);
ifd->active_[cl->pri_] = NULL;
}
}
/*
* Recompute the WRR weights.
*/
if (ifd->wrr_) {
ifd->alloc_[cl->pri_] -= cl->allotment_;
ifd->num_[cl->pri_]--;
rmc_wrr_set_weights(ifd);
}
/*
* Re-compute the depth of the tree.
*/
#if 1 /* ALTQ */
rmc_depth_recompute(cl->parent_);
#else
rmc_depth_recompute(ifd->root_);
#endif
splx(s);
/*
* Free the class structure.
*/
if (cl->red_ != NULL) {
#ifdef ALTQ_RIO
if (q_is_rio(cl->q_))
rio_destroy((rio_t *)cl->red_);
#endif
#ifdef ALTQ_RED
if (q_is_red(cl->q_))
red_destroy(cl->red_);
#endif
}
FREE(cl->q_, M_DEVBUF);
FREE(cl, M_DEVBUF);
}
/*
* void
* rmc_init(...) - Initialize the resource management data structures
* associated with the output portion of interface 'ifp'. 'ifd' is
* where the structures will be built (for backwards compatibility, the
* structures aren't kept in the ifnet struct). 'nsecPerByte'
* gives the link speed (inverse of bandwidth) in nanoseconds/byte.
* 'restart' is the driver-specific routine that the generic 'delay
* until under limit' action will call to restart output. `maxq'
* is the queue size of the 'link' & 'default' classes. 'maxqueued'
* is the maximum number of packets that the resource management
* code will allow to be queued 'downstream' (this is typically 1).
*
* Returns: NONE
*/
void
rmc_init(ifq, ifd, nsecPerByte, restart, maxq, maxqueued, maxidle,
minidle, offtime, flags)
struct ifaltq *ifq;
struct rm_ifdat *ifd;
u_int nsecPerByte;
void (*restart)(struct ifaltq *);
int maxq, maxqueued;
u_int maxidle;
int minidle;
u_int offtime;
int flags;
{
int i, mtu;
/*
* Initialize the CBQ traciing/debug facility.
*/
CBQTRACEINIT();
(void)memset(ifd, 0, sizeof (*ifd));
mtu = ifq->altq_ifp->if_mtu;
ifd->ifq_ = ifq;
ifd->restart = restart;
ifd->maxqueued_ = maxqueued;
ifd->ns_per_byte_ = nsecPerByte;
ifd->maxpkt_ = mtu;
ifd->wrr_ = (flags & RMCF_WRR) ? 1 : 0;
ifd->efficient_ = (flags & RMCF_EFFICIENT) ? 1 : 0;
#if 1
ifd->maxiftime_ = mtu * nsecPerByte / 1000 * 16;
if (mtu * nsecPerByte > 10 * 1000000)
ifd->maxiftime_ /= 4;
#endif
reset_cutoff(ifd);
CBQTRACE(rmc_init, "INIT", ifd->cutoff_);
/*
* Initialize the CBQ's WRR state.
*/
for (i = 0; i < RM_MAXPRIO; i++) {
ifd->alloc_[i] = 0;
ifd->M_[i] = 0;
ifd->num_[i] = 0;
ifd->na_[i] = 0;
ifd->active_[i] = NULL;
}
/*
* Initialize current packet state.
*/
ifd->qi_ = 0;
ifd->qo_ = 0;
for (i = 0; i < RM_MAXQUEUED; i++) {
ifd->class_[i] = NULL;
ifd->curlen_[i] = 0;
ifd->borrowed_[i] = NULL;
}
/*
* Create the root class of the link-sharing structure.
*/
if ((ifd->root_ = rmc_newclass(0, ifd,
nsecPerByte,
rmc_root_overlimit, maxq, 0, 0,
maxidle, minidle, offtime,
0, 0)) == NULL) {
printf("rmc_init: root class not allocated\n");
return ;
}
ifd->root_->depth_ = 0;
}
/*
* void
* rmc_queue_packet(struct rm_class *cl, mbuf_t *m) - Add packet given by
* mbuf 'm' to queue for resource class 'cl'. This routine is called
* by a driver's if_output routine. This routine must be called with
* output packet completion interrupts locked out (to avoid racing with
* rmc_dequeue_next).
*
* Returns: 0 on successful queueing
* -1 when packet drop occurs
*/
int
rmc_queue_packet(cl, m)
struct rm_class *cl;
mbuf_t *m;
{
struct timeval now;
struct rm_ifdat *ifd = cl->ifdat_;
int cpri = cl->pri_;
int is_empty = qempty(cl->q_);
RM_GETTIME(now);
if (ifd->cutoff_ > 0) {
if (TV_LT(&cl->undertime_, &now)) {
if (ifd->cutoff_ > cl->depth_)
ifd->cutoff_ = cl->depth_;
CBQTRACE(rmc_queue_packet, "ffoc", cl->depth_);
}
#if 1 /* ALTQ */
else {
/*
* the class is overlimit. if the class has
* underlimit ancestors, set cutoff to the lowest
* depth among them.
*/
struct rm_class *borrow = cl->borrow_;
while (borrow != NULL &&
borrow->depth_ < ifd->cutoff_) {
if (TV_LT(&borrow->undertime_, &now)) {
ifd->cutoff_ = borrow->depth_;
CBQTRACE(rmc_queue_packet, "ffob", ifd->cutoff_);
break;
}
borrow = borrow->borrow_;
}
}
#else /* !ALTQ */
else if ((ifd->cutoff_ > 1) && cl->borrow_) {
if (TV_LT(&cl->borrow_->undertime_, &now)) {
ifd->cutoff_ = cl->borrow_->depth_;
CBQTRACE(rmc_queue_packet, "ffob",
cl->borrow_->depth_);
}
}
#endif /* !ALTQ */
}
if (_rmc_addq(cl, m) < 0)
/* failed */
return (-1);
if (is_empty) {
CBQTRACE(rmc_queue_packet, "ytpe", cl->stats_.handle);
ifd->na_[cpri]++;
}
if (qlen(cl->q_) > qlimit(cl->q_)) {
/* note: qlimit can be set to 0 or 1 */
rmc_drop_action(cl);
return (-1);
}
return (0);
}
/*
* void
* rmc_tl_satisfied(struct rm_ifdat *ifd, struct timeval *now) - Check all
* classes to see if there are satified.
*/
static void
rmc_tl_satisfied(ifd, now)
struct rm_ifdat *ifd;
struct timeval *now;
{
int i;
rm_class_t *p, *bp;
for (i = RM_MAXPRIO - 1; i >= 0; i--) {
if ((bp = ifd->active_[i]) != NULL) {
p = bp;
do {
if (!rmc_satisfied(p, now)) {
ifd->cutoff_ = p->depth_;
return;
}
p = p->peer_;
} while (p != bp);
}
}
reset_cutoff(ifd);
}
/*
* rmc_satisfied - Return 1 of the class is satisfied. O, otherwise.
*/
static int
rmc_satisfied(cl, now)
struct rm_class *cl;
struct timeval *now;
{
rm_class_t *p;
if (cl == NULL)
return (1);
if (TV_LT(now, &cl->undertime_))
return (1);
if (cl->depth_ == 0) {
if (!cl->sleeping_ && (qlen(cl->q_) > cl->qthresh_))
return (0);
else
return (1);
}
if (cl->children_ != NULL) {
p = cl->children_;
while (p != NULL) {
if (!rmc_satisfied(p, now))
return (0);
p = p->next_;
}
}
return (1);
}
/*
* Return 1 if class 'cl' is under limit or can borrow from a parent,
* 0 if overlimit. As a side-effect, this routine will invoke the
* class overlimit action if the class if overlimit.
*/
static int
rmc_under_limit(cl, now)
struct rm_class *cl;
struct timeval *now;
{
rm_class_t *p = cl;
rm_class_t *top;
struct rm_ifdat *ifd = cl->ifdat_;
ifd->borrowed_[ifd->qi_] = NULL;
/*
* If cl is the root class, then always return that it is
* underlimit. Otherwise, check to see if the class is underlimit.
*/
if (cl->parent_ == NULL)
return (1);
if (cl->sleeping_) {
if (TV_LT(now, &cl->undertime_))
return (0);
CALLOUT_STOP(&cl->callout_);
cl->sleeping_ = 0;
cl->undertime_.tv_sec = 0;
return (1);
}
top = NULL;
while (cl->undertime_.tv_sec && TV_LT(now, &cl->undertime_)) {
if (((cl = cl->borrow_) == NULL) ||
(cl->depth_ > ifd->cutoff_)) {
#ifdef ADJUST_CUTOFF
if (cl != NULL)
/* cutoff is taking effect, just
return false without calling
the delay action. */
return (0);
#endif
#ifdef BORROW_OFFTIME
/*
* check if the class can borrow offtime too.
* borrow offtime from the top of the borrow
* chain if the top class is not overloaded.
*/
if (cl != NULL) {
/* cutoff is taking effect, use this class as top. */
top = cl;
CBQTRACE(rmc_under_limit, "ffou", ifd->cutoff_);
}
if (top != NULL && top->avgidle_ == top->minidle_)
top = NULL;
p->overtime_ = *now;
(p->overlimit)(p, top);
#else
p->overtime_ = *now;
(p->overlimit)(p, NULL);
#endif
return (0);
}
top = cl;
}
if (cl != p)
ifd->borrowed_[ifd->qi_] = cl;
return (1);
}
/*
* _rmc_wrr_dequeue_next() - This is scheduler for WRR as opposed to
* Packet-by-packet round robin.
*
* The heart of the weigthed round-robin scheduler, which decides which
* class next gets to send a packet. Highest priority first, then
* weighted round-robin within priorites.
*
* Each able-to-send class gets to send until its byte allocation is
* exhausted. Thus, the active pointer is only changed after a class has
* exhausted its allocation.
*
* If the scheduler finds no class that is underlimit or able to borrow,
* then the first class found that had a nonzero queue and is allowed to
* borrow gets to send.
*/
static mbuf_t *
_rmc_wrr_dequeue_next(ifd, op)
struct rm_ifdat *ifd;
int op;
{
struct rm_class *cl = NULL, *first = NULL;
u_int deficit;
int cpri;
mbuf_t *m;
struct timeval now;
RM_GETTIME(now);
/*
* if the driver polls the top of the queue and then removes
* the polled packet, we must return the same packet.
*/
if (op == ALTDQ_REMOVE && ifd->pollcache_) {
cl = ifd->pollcache_;
cpri = cl->pri_;
if (ifd->efficient_) {
/* check if this class is overlimit */
if (cl->undertime_.tv_sec != 0 &&
rmc_under_limit(cl, &now) == 0)
first = cl;
}
ifd->pollcache_ = NULL;
goto _wrr_out;
}
else {
/* mode == ALTDQ_POLL || pollcache == NULL */
ifd->pollcache_ = NULL;
ifd->borrowed_[ifd->qi_] = NULL;
}
#ifdef ADJUST_CUTOFF
_again:
#endif
for (cpri = RM_MAXPRIO - 1; cpri >= 0; cpri--) {
if (ifd->na_[cpri] == 0)
continue;
deficit = 0;
/*
* Loop through twice for a priority level, if some class
* was unable to send a packet the first round because
* of the weighted round-robin mechanism.
* During the second loop at this level, deficit==2.
* (This second loop is not needed if for every class,
* "M[cl->pri_])" times "cl->allotment" is greater than
* the byte size for the largest packet in the class.)
*/
_wrr_loop:
cl = ifd->active_[cpri];
ASSERT(cl != NULL);
do {
if ((deficit < 2) && (cl->bytes_alloc_ <= 0))
cl->bytes_alloc_ += cl->w_allotment_;
if (!qempty(cl->q_)) {
if ((cl->undertime_.tv_sec == 0) ||
rmc_under_limit(cl, &now)) {
if (cl->bytes_alloc_ > 0 || deficit > 1)
goto _wrr_out;
/* underlimit but no alloc */
deficit = 1;
#if 1
ifd->borrowed_[ifd->qi_] = NULL;
#endif
}
else if (first == NULL && cl->borrow_ != NULL)
first = cl; /* borrowing candidate */
}
cl->bytes_alloc_ = 0;
cl = cl->peer_;
} while (cl != ifd->active_[cpri]);
if (deficit == 1) {
/* first loop found an underlimit class with deficit */
/* Loop on same priority level, with new deficit. */
deficit = 2;
goto _wrr_loop;
}
}
#ifdef ADJUST_CUTOFF
/*
* no underlimit class found. if cutoff is taking effect,
* increase cutoff and try again.
*/
if (first != NULL && ifd->cutoff_ < ifd->root_->depth_) {
ifd->cutoff_++;
CBQTRACE(_rmc_wrr_dequeue_next, "ojda", ifd->cutoff_);
goto _again;
}
#endif /* ADJUST_CUTOFF */
/*
* If LINK_EFFICIENCY is turned on, then the first overlimit
* class we encounter will send a packet if all the classes
* of the link-sharing structure are overlimit.
*/
reset_cutoff(ifd);
CBQTRACE(_rmc_wrr_dequeue_next, "otsr", ifd->cutoff_);
if (!ifd->efficient_ || first == NULL)
return (NULL);
cl = first;
cpri = cl->pri_;
#if 0 /* too time-consuming for nothing */
if (cl->sleeping_)
CALLOUT_STOP(&cl->callout_);
cl->sleeping_ = 0;
cl->undertime_.tv_sec = 0;
#endif
ifd->borrowed_[ifd->qi_] = cl->borrow_;
ifd->cutoff_ = cl->borrow_->depth_;
/*
* Deque the packet and do the book keeping...
*/
_wrr_out:
if (op == ALTDQ_REMOVE) {
m = _rmc_getq(cl);
if (m == NULL)
panic("_rmc_wrr_dequeue_next");
if (qempty(cl->q_))
ifd->na_[cpri]--;
/*
* Update class statistics and link data.
*/
if (cl->bytes_alloc_ > 0)
cl->bytes_alloc_ -= m_pktlen(m);
if ((cl->bytes_alloc_ <= 0) || first == cl)
ifd->active_[cl->pri_] = cl->peer_;
else
ifd->active_[cl->pri_] = cl;
ifd->class_[ifd->qi_] = cl;
ifd->curlen_[ifd->qi_] = m_pktlen(m);
ifd->now_[ifd->qi_] = now;
ifd->qi_ = (ifd->qi_ + 1) % ifd->maxqueued_;
ifd->queued_++;
} else {
/* mode == ALTDQ_PPOLL */
m = _rmc_pollq(cl);
ifd->pollcache_ = cl;
}
return (m);
}
/*
* Dequeue & return next packet from the highest priority class that
* has a packet to send & has enough allocation to send it. This
* routine is called by a driver whenever it needs a new packet to
* output.
*/
static mbuf_t *
_rmc_prr_dequeue_next(ifd, op)
struct rm_ifdat *ifd;
int op;
{
mbuf_t *m;
int cpri;
struct rm_class *cl, *first = NULL;
struct timeval now;
RM_GETTIME(now);
/*
* if the driver polls the top of the queue and then removes
* the polled packet, we must return the same packet.
*/
if (op == ALTDQ_REMOVE && ifd->pollcache_) {
cl = ifd->pollcache_;
cpri = cl->pri_;
ifd->pollcache_ = NULL;
goto _prr_out;
} else {
/* mode == ALTDQ_POLL || pollcache == NULL */
ifd->pollcache_ = NULL;
ifd->borrowed_[ifd->qi_] = NULL;
}
#ifdef ADJUST_CUTOFF
_again:
#endif
for (cpri = RM_MAXPRIO - 1; cpri >= 0; cpri--) {
if (ifd->na_[cpri] == 0)
continue;
cl = ifd->active_[cpri];
ASSERT(cl != NULL);
do {
if (!qempty(cl->q_)) {
if ((cl->undertime_.tv_sec == 0) ||
rmc_under_limit(cl, &now))
goto _prr_out;
if (first == NULL && cl->borrow_ != NULL)
first = cl;
}
cl = cl->peer_;
} while (cl != ifd->active_[cpri]);
}
#ifdef ADJUST_CUTOFF
/*
* no underlimit class found. if cutoff is taking effect, increase
* cutoff and try again.
*/
if (first != NULL && ifd->cutoff_ < ifd->root_->depth_) {
ifd->cutoff_++;
goto _again;
}
#endif /* ADJUST_CUTOFF */
/*
* If LINK_EFFICIENCY is turned on, then the first overlimit
* class we encounter will send a packet if all the classes
* of the link-sharing structure are overlimit.
*/
reset_cutoff(ifd);
if (!ifd->efficient_ || first == NULL)
return (NULL);
cl = first;
cpri = cl->pri_;
#if 0 /* too time-consuming for nothing */
if (cl->sleeping_)
CALLOUT_STOP(&cl->callout_);
cl->sleeping_ = 0;
cl->undertime_.tv_sec = 0;
#endif
ifd->borrowed_[ifd->qi_] = cl->borrow_;
ifd->cutoff_ = cl->borrow_->depth_;
/*
* Deque the packet and do the book keeping...
*/
_prr_out:
if (op == ALTDQ_REMOVE) {
m = _rmc_getq(cl);
if (m == NULL)
panic("_rmc_prr_dequeue_next");
if (qempty(cl->q_))
ifd->na_[cpri]--;
ifd->active_[cpri] = cl->peer_;
ifd->class_[ifd->qi_] = cl;
ifd->curlen_[ifd->qi_] = m_pktlen(m);
ifd->now_[ifd->qi_] = now;
ifd->qi_ = (ifd->qi_ + 1) % ifd->maxqueued_;
ifd->queued_++;
} else {
/* mode == ALTDQ_POLL */
m = _rmc_pollq(cl);
ifd->pollcache_ = cl;
}
return (m);
}
/*
* mbuf_t *
* rmc_dequeue_next(struct rm_ifdat *ifd, struct timeval *now) - this function
* is invoked by the packet driver to get the next packet to be
* dequeued and output on the link. If WRR is enabled, then the
* WRR dequeue next routine will determine the next packet to sent.
* Otherwise, packet-by-packet round robin is invoked.
*
* Returns: NULL, if a packet is not available or if all
* classes are overlimit.
*
* Otherwise, Pointer to the next packet.
*/
mbuf_t *
rmc_dequeue_next(ifd, mode)
struct rm_ifdat *ifd;
int mode;
{
if (ifd->queued_ >= ifd->maxqueued_)
return (NULL);
else if (ifd->wrr_)
return (_rmc_wrr_dequeue_next(ifd, mode));
else
return (_rmc_prr_dequeue_next(ifd, mode));
}
/*
* Update the utilization estimate for the packet that just completed.
* The packet's class & the parent(s) of that class all get their
* estimators updated. This routine is called by the driver's output-
* packet-completion interrupt service routine.
*/
/*
* a macro to approximate "divide by 1000" that gives 0.000999,
* if a value has enough effective digits.
* (on pentium, mul takes 9 cycles but div takes 46!)
*/
#define NSEC_TO_USEC(t) (((t) >> 10) + ((t) >> 16) + ((t) >> 17))
void
rmc_update_class_util(ifd)
struct rm_ifdat *ifd;
{
int idle, avgidle, pktlen;
int pkt_time, tidle;
rm_class_t *cl, *borrowed;
rm_class_t *borrows;
struct timeval *nowp;
/*
* Get the most recent completed class.
*/
if ((cl = ifd->class_[ifd->qo_]) == NULL)
return;
pktlen = ifd->curlen_[ifd->qo_];
borrowed = ifd->borrowed_[ifd->qo_];
borrows = borrowed;
PKTCNTR_ADD(&cl->stats_.xmit_cnt, pktlen);
/*
* Run estimator on class and it's ancesstors.
*/
/*
* rm_update_class_util is designed to be called when the
* transfer is completed from a xmit complete interrupt,
* but most drivers don't implement an upcall for that.
* so, just use estimated completion time.
* as a result, ifd->qi_ and ifd->qo_ are always synced.
*/
nowp = &ifd->now_[ifd->qo_];
/* get pkt_time (for link) in usec */
#if 1 /* use approximation */
pkt_time = ifd->curlen_[ifd->qo_] * ifd->ns_per_byte_;
pkt_time = NSEC_TO_USEC(pkt_time);
#else
pkt_time = ifd->curlen_[ifd->qo_] * ifd->ns_per_byte_ / 1000;
#endif
#if 1 /* ALTQ4PPP */
if (TV_LT(nowp, &ifd->ifnow_)) {
int iftime;
/*
* make sure the estimated completion time does not go
* too far. it can happen when the link layer supports
* data compression or the interface speed is set to
* a much lower value.
*/
TV_DELTA(&ifd->ifnow_, nowp, iftime);
if (iftime+pkt_time < ifd->maxiftime_) {
TV_ADD_DELTA(&ifd->ifnow_, pkt_time, &ifd->ifnow_);
} else {
TV_ADD_DELTA(nowp, ifd->maxiftime_, &ifd->ifnow_);
}
} else {
TV_ADD_DELTA(nowp, pkt_time, &ifd->ifnow_);
}
#else
if (TV_LT(nowp, &ifd->ifnow_)) {
TV_ADD_DELTA(&ifd->ifnow_, pkt_time, &ifd->ifnow_);
} else {
TV_ADD_DELTA(nowp, pkt_time, &ifd->ifnow_);
}
#endif
while (cl != NULL) {
TV_DELTA(&ifd->ifnow_, &cl->last_, idle);
if (idle >= 2000000)
/*
* this class is idle enough, reset avgidle.
* (TV_DELTA returns 2000000 us when delta is large.)
*/
cl->avgidle_ = cl->maxidle_;
/* get pkt_time (for class) in usec */
#if 1 /* use approximation */
pkt_time = pktlen * cl->ns_per_byte_;
pkt_time = NSEC_TO_USEC(pkt_time);
#else
pkt_time = pktlen * cl->ns_per_byte_ / 1000;
#endif
idle -= pkt_time;
avgidle = cl->avgidle_;
avgidle += idle - (avgidle >> RM_FILTER_GAIN);
cl->avgidle_ = avgidle;
/* Are we overlimit ? */
if (avgidle <= 0) {
CBQTRACE(rmc_update_class_util, "milo", cl->stats_.handle);
#if 1 /* ALTQ */
/*
* need some lower bound for avgidle, otherwise
* a borrowing class gets unbounded penalty.
*/
if (avgidle < cl->minidle_)
avgidle = cl->avgidle_ = cl->minidle_;
#endif
/* set next idle to make avgidle 0 */
tidle = pkt_time +
(((1 - RM_POWER) * avgidle) >> RM_FILTER_GAIN);
TV_ADD_DELTA(nowp, tidle, &cl->undertime_);
++cl->stats_.over;
} else {
cl->avgidle_ =
(avgidle > cl->maxidle_) ? cl->maxidle_ : avgidle;
cl->undertime_.tv_sec = 0;
if (cl->sleeping_) {
CALLOUT_STOP(&cl->callout_);
cl->sleeping_ = 0;
}
}
if (borrows != NULL) {
if (borrows != cl)
++cl->stats_.borrows;
else
borrows = NULL;
}
cl->last_ = ifd->ifnow_;
cl->last_pkttime_ = pkt_time;
#if 1
if (cl->parent_ == NULL) {
/* take stats of root class */
PKTCNTR_ADD(&cl->stats_.xmit_cnt, pktlen);
}
#endif
cl = cl->parent_;
}
/*
* Check to see if cutoff needs to set to a new level.
*/
cl = ifd->class_[ifd->qo_];
if (borrowed && (ifd->cutoff_ >= borrowed->depth_)) {
#if 1 /* ALTQ */
if ((qlen(cl->q_) <= 0) || TV_LT(nowp, &borrowed->undertime_)) {
rmc_tl_satisfied(ifd, nowp);
CBQTRACE(rmc_update_class_util, "broe", ifd->cutoff_);
} else {
ifd->cutoff_ = borrowed->depth_;
CBQTRACE(rmc_update_class_util, "ffob", borrowed->depth_);
}
#else /* !ALTQ */
if ((qlen(cl->q_) <= 1) || TV_LT(&now, &borrowed->undertime_)) {
reset_cutoff(ifd);
#ifdef notdef
rmc_tl_satisfied(ifd, &now);
#endif
CBQTRACE(rmc_update_class_util, "broe", ifd->cutoff_);
} else {
ifd->cutoff_ = borrowed->depth_;
CBQTRACE(rmc_update_class_util, "ffob", borrowed->depth_);
}
#endif /* !ALTQ */
}
/*
* Release class slot
*/
ifd->borrowed_[ifd->qo_] = NULL;
ifd->class_[ifd->qo_] = NULL;
ifd->qo_ = (ifd->qo_ + 1) % ifd->maxqueued_;
ifd->queued_--;
}
/*
* void
* rmc_drop_action(struct rm_class *cl) - Generic (not protocol-specific)
* over-limit action routines. These get invoked by rmc_under_limit()
* if a class with packets to send if over its bandwidth limit & can't
* borrow from a parent class.
*
* Returns: NONE
*/
static void
rmc_drop_action(cl)
struct rm_class *cl;
{
struct rm_ifdat *ifd = cl->ifdat_;
ASSERT(qlen(cl->q_) > 0);
_rmc_dropq(cl);
if (qempty(cl->q_))
ifd->na_[cl->pri_]--;
}
void rmc_dropall(cl)
struct rm_class *cl;
{
struct rm_ifdat *ifd = cl->ifdat_;
if (!qempty(cl->q_)) {
_flushq(cl->q_);
ifd->na_[cl->pri_]--;
}
}
#if (__FreeBSD_version > 300000)
/* hzto() is removed from FreeBSD-3.0 */
static int hzto __P((struct timeval *));
static int
hzto(tv)
struct timeval *tv;
{
struct timeval t2;
getmicrotime(&t2);
t2.tv_sec = tv->tv_sec - t2.tv_sec;
t2.tv_usec = tv->tv_usec - t2.tv_usec;
return (tvtohz(&t2));
}
#endif /* __FreeBSD_version > 300000 */
/*
* void
* rmc_delay_action(struct rm_class *cl) - This function is the generic CBQ
* delay action routine. It is invoked via rmc_under_limit when the
* packet is discoverd to be overlimit.
*
* If the delay action is result of borrow class being overlimit, then
* delay for the offtime of the borrowing class that is overlimit.
*
* Returns: NONE
*/
void
rmc_delay_action(cl, borrow)
struct rm_class *cl, *borrow;
{
int delay, t, extradelay;
cl->stats_.overactions++;
TV_DELTA(&cl->undertime_, &cl->overtime_, delay);
#ifndef BORROW_OFFTIME
delay += cl->offtime_;
#endif
if (!cl->sleeping_) {
CBQTRACE(rmc_delay_action, "yled", cl->stats_.handle);
#ifdef BORROW_OFFTIME
if (borrow != NULL)
extradelay = borrow->offtime_;
else
#endif
extradelay = cl->offtime_;
#ifdef ALTQ
/*
* XXX recalculate suspend time:
* current undertime is (tidle + pkt_time) calculated
* from the last transmission.
* tidle: time required to bring avgidle back to 0
* pkt_time: target waiting time for this class
* we need to replace pkt_time by offtime
*/
extradelay -= cl->last_pkttime_;
#endif
if (extradelay > 0) {
TV_ADD_DELTA(&cl->undertime_, extradelay, &cl->undertime_);
delay += extradelay;
}
cl->sleeping_ = 1;
cl->stats_.delays++;
/*
* Since packets are phased randomly with respect to the
* clock, 1 tick (the next clock tick) can be an arbitrarily
* short time so we have to wait for at least two ticks.
* NOTE: If there's no other traffic, we need the timer as
* a 'backstop' to restart this class.
*/
if (delay > tick * 2) {
#ifdef __FreeBSD__
/* FreeBSD rounds up the tick */
t = hzto(&cl->undertime_);
#else
/* other BSDs round down the tick */
t = hzto(&cl->undertime_) + 1;
#endif
} else
t = 2;
CALLOUT_RESET(&cl->callout_, t,
(timeout_t *)rmc_restart, (caddr_t)cl);
}
}
/*
* void
* rmc_restart() - is just a helper routine for rmc_delay_action -- it is
* called by the system timer code & is responsible checking if the
* class is still sleeping (it might have been restarted as a side
* effect of the queue scan on a packet arrival) and, if so, restarting
* output for the class. Inspecting the class state & restarting output
* require locking the class structure. In general the driver is
* responsible for locking but this is the only routine that is not
* called directly or indirectly from the interface driver so it has
* know about system locking conventions. Under bsd, locking is done
* by raising IPL to splnet so that's what's implemented here. On a
* different system this would probably need to be changed.
*
* Returns: NONE
*/
static void
rmc_restart(cl)
struct rm_class *cl;
{
struct rm_ifdat *ifd = cl->ifdat_;
int s;
s = splnet();
if (cl->sleeping_) {
cl->sleeping_ = 0;
cl->undertime_.tv_sec = 0;
if (ifd->queued_ < ifd->maxqueued_ && ifd->restart != NULL) {
CBQTRACE(rmc_restart, "trts", cl->stats_.handle);
(ifd->restart)(ifd->ifq_);
}
}
splx(s);
}
/*
* void
* rmc_root_overlimit(struct rm_class *cl) - This the generic overlimit
* handling routine for the root class of the link sharing structure.
*
* Returns: NONE
*/
static void
rmc_root_overlimit(cl, borrow)
struct rm_class *cl, *borrow;
{
panic("rmc_root_overlimit");
}
/*
* Packet Queue handling routines. Eventually, this is to localize the
* effects on the code whether queues are red queues or droptail
* queues.
*/
static int
_rmc_addq(cl, m)
rm_class_t *cl;
mbuf_t *m;
{
#ifdef ALTQ_RIO
if (q_is_rio(cl->q_))
return rio_addq((rio_t *)cl->red_, cl->q_, m, cl->pktattr_);
#endif
#ifdef ALTQ_RED
if (q_is_red(cl->q_))
return red_addq(cl->red_, cl->q_, m, cl->pktattr_);
#endif /* ALTQ_RED */
if (cl->flags_ & RMCF_CLEARDSCP)
write_dsfield(m, cl->pktattr_, 0);
_addq(cl->q_, m);
return (0);
}
/* note: _rmc_dropq is not called for red */
static void
_rmc_dropq(cl)
rm_class_t *cl;
{
mbuf_t *m;
if ((m = _getq(cl->q_)) != NULL)
m_freem(m);
}
static mbuf_t *
_rmc_getq(cl)
rm_class_t *cl;
{
#ifdef ALTQ_RIO
if (q_is_rio(cl->q_))
return rio_getq((rio_t *)cl->red_, cl->q_);
#endif
#ifdef ALTQ_RED
if (q_is_red(cl->q_))
return red_getq(cl->red_, cl->q_);
#endif
return _getq(cl->q_);
}
static mbuf_t *
_rmc_pollq(cl)
rm_class_t *cl;
{
return qhead(cl->q_);
}
#ifdef CBQ_TRACE
/*
* DDB hook to trace cbq events:
* the last 1024 events are held in a circular buffer.
* use "call cbqtrace_dump(N)" to display 20 events from Nth event.
*/
void cbqtrace_dump(int);
static char *rmc_funcname(void *);
static struct rmc_funcs {
void *func;
char *name;
} rmc_funcs[] = {
{ rmc_init, "rmc_init" },
{ rmc_queue_packet, "rmc_queue_packet" },
{ rmc_under_limit, "rmc_under_limit" },
{ rmc_update_class_util, "rmc_update_class_util" },
{ rmc_delay_action, "rmc_delay_action" },
{ rmc_restart, "rmc_restart" },
{ _rmc_wrr_dequeue_next, "_rmc_wrr_dequeue_next" },
{ NULL, NULL }
};
static char *rmc_funcname(func)
void *func;
{
struct rmc_funcs *fp;
for (fp = rmc_funcs; fp->func != NULL; fp++)
if (fp->func == func)
return (fp->name);
return ("unknown");
}
void cbqtrace_dump(counter)
int counter;
{
int i, *p;
char *cp;
counter = counter % NCBQTRACE;
p = (int *)&cbqtrace_buffer[counter];
for (i=0; i<20; i++) {
printf("[0x%x] ", *p++);
printf("%s: ", rmc_funcname((void *)*p++));
cp = (char *)p++;
printf("%c%c%c%c: ", cp[0], cp[1], cp[2], cp[3]);
printf("%d\n",*p++);
if (p >= (int *)&cbqtrace_buffer[NCBQTRACE])
p = (int *)cbqtrace_buffer;
}
}
#endif /* CBQ_TRACE */
#endif /* ALTQ_CBQ */
#if defined(ALTQ_CBQ) || defined(ALTQ_RED) || defined(ALTQ_RIO) || defined(ALTQ_HFSC) || defined(ALTQ_PRIQ)
#if !defined(__GNUC__) || defined(ALTQ_DEBUG)
void
_addq(q, m)
class_queue_t *q;
mbuf_t *m;
{
mbuf_t *m0;
if ((m0 = qtail(q)) != NULL)
m->m_nextpkt = m0->m_nextpkt;
else
m0 = m;
m0->m_nextpkt = m;
qtail(q) = m;
qlen(q)++;
}
mbuf_t *
_getq(q)
class_queue_t *q;
{
mbuf_t *m, *m0;
if ((m = qtail(q)) == NULL)
return (NULL);
if ((m0 = m->m_nextpkt) != m)
m->m_nextpkt = m0->m_nextpkt;
else {
ASSERT(qlen(q) == 1);
qtail(q) = NULL;
}
qlen(q)--;
m0->m_nextpkt = NULL;
return (m0);
}
/* drop a packet at the tail of the queue */
mbuf_t *
_getq_tail(q)
class_queue_t *q;
{
mbuf_t *m, *m0, *prev;
if ((m = m0 = qtail(q)) == NULL)
return NULL;
do {
prev = m0;
m0 = m0->m_nextpkt;
} while (m0 != m);
prev->m_nextpkt = m->m_nextpkt;
if (prev == m) {
ASSERT(qlen(q) == 1);
qtail(q) = NULL;
} else
qtail(q) = prev;
qlen(q)--;
m->m_nextpkt = NULL;
return (m);
}
/* randomly select a packet in the queue */
mbuf_t *
_getq_random(q)
class_queue_t *q;
{
struct mbuf *m;
int i, n;
if ((m = qtail(q)) == NULL)
return NULL;
if (m->m_nextpkt == m) {
ASSERT(qlen(q) == 1);
qtail(q) = NULL;
} else {
struct mbuf *prev = NULL;
n = random() % qlen(q) + 1;
for (i = 0; i < n; i++) {
prev = m;
m = m->m_nextpkt;
}
prev->m_nextpkt = m->m_nextpkt;
if (m == qtail(q))
qtail(q) = prev;
}
qlen(q)--;
m->m_nextpkt = NULL;
return (m);
}
void
_removeq(q, m)
class_queue_t *q;
mbuf_t *m;
{
mbuf_t *m0, *prev;
m0 = qtail(q);
do {
prev = m0;
m0 = m0->m_nextpkt;
} while (m0 != m);
prev->m_nextpkt = m->m_nextpkt;
if (prev == m)
qtail(q) = NULL;
else if (qtail(q) == m)
qtail(q) = prev;
qlen(q)--;
}
void
_flushq(q)
class_queue_t *q;
{
mbuf_t *m;
while ((m = _getq(q)) != NULL)
m_freem(m);
ASSERT(qlen(q) == 0);
}
#endif /* !__GNUC__ || ALTQ_DEBUG */
#endif /* ALTQ_CBQ || ALTQ_RED || ALTQ_RIO || ALTQ_HFSC || ALTQ_PRIQ */