NetBSD/sys/external/bsd/ipf/netinet/ip_fil_netbsd.c

2289 lines
49 KiB
C

/* $NetBSD: ip_fil_netbsd.c,v 1.37 2022/03/28 12:33:21 riastradh Exp $ */
/*
* Copyright (C) 2012 by Darren Reed.
*
* See the IPFILTER.LICENCE file for details on licencing.
*/
#if !defined(lint)
#if defined(__NetBSD__)
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: ip_fil_netbsd.c,v 1.37 2022/03/28 12:33:21 riastradh Exp $");
#else
static const char sccsid[] = "@(#)ip_fil.c 2.41 6/5/96 (C) 1993-2000 Darren Reed";
static const char rcsid[] = "@(#)Id: ip_fil_netbsd.c,v 1.1.1.2 2012/07/22 13:45:17 darrenr Exp";
#endif
#endif
#if defined(KERNEL) || defined(_KERNEL)
# undef KERNEL
# undef _KERNEL
# define KERNEL 1
# define _KERNEL 1
#endif
#include <sys/param.h>
#if (NetBSD >= 199905) && !defined(IPFILTER_LKM)
# if (__NetBSD_Version__ >= 799003000)
# ifdef _KERNEL_OPT
# include "opt_ipsec.h"
# endif
# else
# include "opt_ipsec.h"
# endif
#endif
#include <sys/errno.h>
#include <sys/types.h>
#include <sys/file.h>
#include <sys/ioctl.h>
#include <sys/time.h>
#include <sys/systm.h>
#include <sys/select.h>
#if (NetBSD > 199609)
# include <sys/dirent.h>
#else
# include <sys/dir.h>
#endif
#if (__NetBSD_Version__ >= 599005900)
# include <sys/cprng.h>
#endif
#include <sys/mbuf.h>
#include <sys/protosw.h>
#include <sys/socket.h>
#include <sys/poll.h>
#if (__NetBSD_Version__ >= 399002000)
# include <sys/kauth.h>
#endif
#if (__NetBSD_Version__ >= 799003000)
#include <sys/module.h>
#include <sys/mutex.h>
#endif
#if defined(__NetBSD__)
#include <netinet/in_offload.h>
#endif
#include <net/if.h>
#include <net/route.h>
#include <netinet/in.h>
#include <netinet/in_var.h>
#include <netinet/in_systm.h>
#include <netinet/ip.h>
#include <netinet/ip_var.h>
#include <netinet/tcp.h>
#if __NetBSD_Version__ >= 105190000 /* 1.5T */
# include <netinet/tcp_timer.h>
# include <netinet/tcp_var.h>
#endif
#include <netinet/udp.h>
#include <netinet/ip_icmp.h>
#include "netinet/ip_compat.h"
#ifdef USE_INET6
# include <netinet/icmp6.h>
# if (__NetBSD_Version__ >= 106000000)
# include <netinet6/nd6.h>
# endif
# if __NetBSD_Version__ >= 499001100
# include <netinet6/scope6_var.h>
# include <netinet6/in6_offload.h>
# endif
#endif
#include "netinet/ip_fil.h"
#include "netinet/ip_nat.h"
#include "netinet/ip_frag.h"
#include "netinet/ip_state.h"
#include "netinet/ip_proxy.h"
#include "netinet/ip_auth.h"
#include "netinet/ip_sync.h"
#include "netinet/ip_lookup.h"
#include "netinet/ip_dstlist.h"
#ifdef IPFILTER_SCAN
#include "netinet/ip_scan.h"
#endif
#include <sys/md5.h>
#include <sys/kernel.h>
#include <sys/conf.h>
#ifdef INET
extern int ip_optcopy (struct ip *, struct ip *);
#endif
#ifdef IPFILTER_M_IPFILTER
MALLOC_DEFINE(M_IPFILTER, "IP Filter", "IP Filter packet filter data structures");
#endif
#if __NetBSD_Version__ >= 105009999
# define csuminfo csum_flags
#endif
#if __NetBSD_Version__ < 200000000
extern struct protosw inetsw[];
#endif
#if (__NetBSD_Version__ >= 599002000)
static kauth_listener_t ipf_listener;
#endif
#if (__NetBSD_Version__ < 399001400)
extern int ip6_getpmtu (struct route_in6 *, struct route_in6 *,
struct ifnet *, struct in6_addr *, u_long *,
int *);
#endif
#if (NetBSD >= 199511)
static int ipfopen(dev_t dev, int flags, int devtype, PROC_T *p);
static int ipfclose(dev_t dev, int flags, int devtype, PROC_T *p);
#else
# if (__NetBSD_Version__ >= 399001400)
static int ipfopen(dev_t dev, int flags, struct lwp *);
static int ipfclose(dev_t dev, int flags, struct lwp *);
# else
static int ipfopen(dev_t dev, int flags);
static int ipfclose(dev_t dev, int flags);
# endif /* __NetBSD_Version__ >= 399001400 */
#endif
static int ipfread(dev_t, struct uio *, int ioflag);
static int ipfwrite(dev_t, struct uio *, int ioflag);
static int ipfpoll(dev_t, int events, PROC_T *);
static void ipf_timer_func(void *ptr);
const struct cdevsw ipl_cdevsw = {
.d_open = ipfopen,
.d_close = ipfclose,
.d_read = ipfread,
.d_write = ipfwrite,
.d_ioctl = ipfioctl,
.d_stop = nostop,
.d_tty = notty,
.d_poll = ipfpoll,
.d_mmap = nommap,
#if (__NetBSD_Version__ >= 200000000)
.d_kqfilter = nokqfilter,
#endif
.d_discard = nodiscard,
#ifdef D_OTHER
.d_flag = D_OTHER
#else
.d_flag = 0
#endif
};
#if (__NetBSD_Version__ >= 799003000)
kmutex_t ipf_ref_mutex;
int ipf_active;
#endif
ipf_main_softc_t ipfmain;
static u_short ipid = 0;
static int (*ipf_savep)(void *, ip_t *, int, void *, int, struct mbuf **);
static int ipf_send_ip(fr_info_t *, mb_t *);
#ifdef USE_INET6
static int ipf_fastroute6(struct mbuf *, struct mbuf **,
fr_info_t *, frdest_t *);
#endif
#if defined(NETBSD_PF)
# include <net/pfil.h>
/*
* We provide the ipf_checkp name just to minimize changes later.
*/
int (*ipf_checkp)(void *, ip_t *ip, int hlen, void *ifp, int out, mb_t **mp);
#endif /* NETBSD_PF */
#if defined(__NetBSD_Version__) && (__NetBSD_Version__ >= 105110000)
# include <net/pfil.h>
static int ipf_check_wrapper(void *, struct mbuf **, struct ifnet *, int );
static int
ipf_check_wrapper(void *arg, struct mbuf **mp, struct ifnet *ifp, int dir)
{
struct ip *ip;
int rv, hlen;
#if __NetBSD_Version__ >= 200080000
/*
* ensure that mbufs are writable beforehand
* as it's assumed by ipf code.
* XXX inefficient
*/
int error = m_makewritable(mp, 0, M_COPYALL, M_DONTWAIT);
if (error) {
m_freem(*mp);
*mp = NULL;
return error;
}
#endif
ip = mtod(*mp, struct ip *);
hlen = ip->ip_hl << 2;
#ifdef INET
#if defined(M_CSUM_TCPv4)
/*
* If the packet is out-bound, we can't delay checksums
* here. For in-bound, the checksum has already been
* validated.
*/
if (dir == PFIL_OUT) {
if ((*mp)->m_pkthdr.csum_flags & (M_CSUM_TCPv4|M_CSUM_UDPv4)) {
in_undefer_cksum_tcpudp(*mp);
(*mp)->m_pkthdr.csum_flags &=
~(M_CSUM_TCPv4|M_CSUM_UDPv4);
}
}
#endif /* M_CSUM_TCPv4 */
#endif /* INET */
/*
* Note, we don't need to update the checksum, because
* it has already been verified.
*/
rv = ipf_check(&ipfmain, ip, hlen, ifp, (dir == PFIL_OUT), mp);
return (rv);
}
# ifdef USE_INET6
# include <netinet/ip6.h>
static int ipf_check_wrapper6(void *, struct mbuf **, struct ifnet *, int );
static int
ipf_check_wrapper6(void *arg, struct mbuf **mp, struct ifnet *ifp, int dir)
{
#if defined(INET6)
# if defined(M_CSUM_TCPv6) && (__NetBSD_Version__ > 200000000)
/*
* If the packet is out-bound, we can't delay checksums
* here. For in-bound, the checksum has already been
* validated.
*/
if (dir == PFIL_OUT) {
if ((*mp)->m_pkthdr.csum_flags & (M_CSUM_TCPv6|M_CSUM_UDPv6)) {
# if (__NetBSD_Version__ > 399000600)
in6_undefer_cksum_tcpudp(*mp);
# endif
(*mp)->m_pkthdr.csum_flags &= ~(M_CSUM_TCPv6|
M_CSUM_UDPv6);
}
}
# endif
#endif /* INET6 */
return (ipf_check(&ipfmain, mtod(*mp, struct ip *), sizeof(struct ip6_hdr),
ifp, (dir == PFIL_OUT), mp));
}
# endif
# if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET)
# if (__NetBSD_Version__ >= 799000400)
static void ipf_pfilsync(void *, unsigned long, void *);
static void
ipf_pfilsync(void *hdr, unsigned long cmd, void *arg2)
{
/*
* The interface pointer is useless for create (we have nothing to
* compare it to) and at detach, the interface name is still in the
* list of active NICs (albeit, down, but that's not any real
* indicator) and doing ifunit() on the name will still return the
* pointer, so it's not much use then, either.
*/
ipf_sync(&ipfmain, NULL);
}
# else
static int ipf_pfilsync(void *, struct mbuf **, struct ifnet *, int);
static int
ipf_pfilsync(void *hdr, struct mbuf **mp, struct ifnet *ifp, int dir)
{
ipf_sync(&ipfmain, NULL);
return 0;
}
# endif
# endif
#endif /* __NetBSD_Version__ >= 105110000 */
#if defined(IPFILTER_LKM)
int
ipf_identify(s)
char *s;
{
if (strcmp(s, "ipl") == 0)
return 1;
return 0;
}
#endif /* IPFILTER_LKM */
#if (__NetBSD_Version__ >= 599002000)
static int
ipf_listener_cb(kauth_cred_t cred, kauth_action_t action, void *cookie,
void *arg0, void *arg1, void *arg2, void *arg3)
{
int result;
enum kauth_network_req req;
result = KAUTH_RESULT_DEFER;
req = (enum kauth_network_req)(uintptr_t)arg0;
if (action != KAUTH_NETWORK_FIREWALL)
return result;
/* These must have came from device context. */
if ((req == KAUTH_REQ_NETWORK_FIREWALL_FW) ||
(req == KAUTH_REQ_NETWORK_FIREWALL_NAT))
result = KAUTH_RESULT_ALLOW;
return result;
}
#endif
/*
* Try to detect the case when compiling for NetBSD with pseudo-device
*/
void
ipfilterattach(int count)
{
#if (__NetBSD_Version__ >= 799003000)
return;
#else
#if (__NetBSD_Version__ >= 599002000)
ipf_listener = kauth_listen_scope(KAUTH_SCOPE_NETWORK,
ipf_listener_cb, NULL);
#endif
if (ipf_load_all() == 0)
(void) ipf_create_all(&ipfmain);
#endif
}
int
ipfattach(ipf_main_softc_t *softc)
{
SPL_INT(s);
#if (__NetBSD_Version__ >= 499005500)
int i;
#endif
#if defined(NETBSD_PF) && (__NetBSD_Version__ >= 104200000)
int error = 0;
# if defined(__NetBSD_Version__) && (__NetBSD_Version__ >= 105110000)
pfil_head_t *ph_inet;
# ifdef USE_INET6
pfil_head_t *ph_inet6;
# endif
# if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET)
pfil_head_t *ph_ifsync;
# endif
# endif
#endif
SPL_NET(s);
if ((softc->ipf_running > 0) || (ipf_checkp == ipf_check)) {
printf("IP Filter: already initialized\n");
SPL_X(s);
IPFERROR(130017);
return EBUSY;
}
if (ipf_init_all(softc) < 0) {
SPL_X(s);
IPFERROR(130015);
return EIO;
}
#ifdef NETBSD_PF
# if (__NetBSD_Version__ >= 104200000)
# if __NetBSD_Version__ >= 105110000
ph_inet = pfil_head_get(PFIL_TYPE_AF, (void *)AF_INET);
# ifdef USE_INET6
ph_inet6 = pfil_head_get(PFIL_TYPE_AF, (void *)AF_INET6);
# endif
# if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET)
ph_ifsync = pfil_head_get(PFIL_TYPE_IFNET, 0);
# endif
if (ph_inet == NULL
# ifdef USE_INET6
&& ph_inet6 == NULL
# endif
# if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET)
&& ph_ifsync == NULL
# endif
) {
SPL_X(s);
IPFERROR(130016);
return ENODEV;
}
if (ph_inet != NULL)
error = pfil_add_hook((void *)ipf_check_wrapper, NULL,
PFIL_IN|PFIL_OUT, ph_inet);
else
error = 0;
# else
error = pfil_add_hook((void *)ipf_check, PFIL_IN|PFIL_OUT,
&inetsw[ip_protox[IPPROTO_IP]].pr_pfh);
# endif
if (error) {
IPFERROR(130013);
goto pfil_error;
}
# else
pfil_add_hook((void *)ipf_check, PFIL_IN|PFIL_OUT);
# endif
# ifdef USE_INET6
# if __NetBSD_Version__ >= 105110000
if (ph_inet6 != NULL)
error = pfil_add_hook((void *)ipf_check_wrapper6, NULL,
PFIL_IN|PFIL_OUT, ph_inet6);
else
error = 0;
if (error) {
pfil_remove_hook((void *)ipf_check_wrapper6, NULL,
PFIL_IN|PFIL_OUT, ph_inet6);
ipfmain.ipf_interror = 130014;
goto pfil_error;
}
# else
error = pfil_add_hook((void *)ipf_check, PFIL_IN|PFIL_OUT,
&inetsw[ip_protox[IPPROTO_IPV6]].pr_pfh);
if (error) {
pfil_remove_hook((void *)ipf_check, PFIL_IN|PFIL_OUT,
&inetsw[ip_protox[IPPROTO_IP]].pr_pfh);
IPFERROR(130014);
goto pfil_error;
}
# endif
# endif
# if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET)
if (ph_ifsync != NULL)
#if (__NetBSD_Version__ >= 799000400)
(void) pfil_add_ihook((void *)ipf_pfilsync, NULL,
PFIL_IFNET, ph_ifsync);
#else
(void) pfil_add_hook((void *)ipf_pfilsync, NULL,
PFIL_IFNET, ph_ifsync);
#endif
# endif
#endif
#if (__NetBSD_Version__ >= 499005500)
for (i = 0; i < IPL_LOGSIZE; i++)
selinit(&ipfmain.ipf_selwait[i]);
#else
bzero((char *)ipfmain.ipf_selwait, sizeof(ipfmain.ipf_selwait));
#endif
ipf_savep = ipf_checkp;
ipf_checkp = ipf_check;
#ifdef INET
if (softc->ipf_control_forwarding & 1)
ipforwarding = 1;
#endif
ipid = 0;
SPL_X(s);
#if (__NetBSD_Version__ >= 104010000)
# if (__NetBSD_Version__ >= 499002000)
callout_init(&softc->ipf_slow_ch, 0);
# else
callout_init(&softc->ipf_slow_ch);
# endif
callout_reset(&softc->ipf_slow_ch, (hz / IPF_HZ_DIVIDE) * IPF_HZ_MULT,
ipf_timer_func, softc);
#else
timeout(ipf_timer_func, softc, (hz / IPF_HZ_DIVIDE) * IPF_HZ_MULT);
#endif
return 0;
#if __NetBSD_Version__ >= 105110000
pfil_error:
SPL_X(s);
ipf_fini_all(softc);
return error;
#endif
}
static void
ipf_timer_func(void *ptr)
{
ipf_main_softc_t *softc = ptr;
SPL_INT(s);
SPL_NET(s);
READ_ENTER(&softc->ipf_global);
if (softc->ipf_running > 0)
ipf_slowtimer(softc);
if (softc->ipf_running == -1 || softc->ipf_running == 1) {
#if NETBSD_GE_REV(104240000)
callout_reset(&softc->ipf_slow_ch, hz / 2,
ipf_timer_func, softc);
#else
timeout(ipf_timer_func, softc,
(hz / IPF_HZ_DIVIDE) * IPF_HZ_MULT);
#endif
}
RWLOCK_EXIT(&softc->ipf_global);
SPL_X(s);
}
/*
* Disable the filter by removing the hooks from the IP input/output
* stream.
*/
int
ipfdetach(ipf_main_softc_t *softc)
{
SPL_INT(s);
#if (__NetBSD_Version__ >= 499005500)
int i;
#endif
#if defined(NETBSD_PF) && (__NetBSD_Version__ >= 104200000)
int error = 0;
# if __NetBSD_Version__ >= 105150000
pfil_head_t *ph_inet = pfil_head_get(PFIL_TYPE_AF, (void *)AF_INET);
# ifdef USE_INET6
pfil_head_t *ph_inet6 = pfil_head_get(PFIL_TYPE_AF, (void *)AF_INET6);
# endif
# if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET)
struct pfil_head *ph_ifsync = pfil_head_get(PFIL_TYPE_IFNET, 0);
# endif
# endif
#endif
SPL_NET(s);
#if (__NetBSD_Version__ >= 104010000)
if (softc->ipf_running > 0)
callout_stop(&softc->ipf_slow_ch);
#else
untimeout(ipf_slowtimer, NULL);
#endif /* NetBSD */
ipf_checkp = ipf_savep;
(void) ipf_flush(softc, IPL_LOGIPF, FR_INQUE|FR_OUTQUE|FR_INACTIVE);
(void) ipf_flush(softc, IPL_LOGIPF, FR_INQUE|FR_OUTQUE);
#ifdef INET
if (softc->ipf_control_forwarding & 2)
ipforwarding = 0;
#endif
#ifdef NETBSD_PF
# if (__NetBSD_Version__ >= 104200000)
# if __NetBSD_Version__ >= 105110000
# if defined(PFIL_TYPE_IFNET) && defined(PFIL_IFNET)
# if __NetBSD_Version__ >= 799000400
(void) pfil_remove_ihook((void *)ipf_pfilsync, NULL,
PFIL_IFNET, ph_ifsync);
# else
(void) pfil_remove_hook((void *)ipf_pfilsync, NULL,
PFIL_IFNET, ph_ifsync);
# endif
# endif
if (ph_inet != NULL)
error = pfil_remove_hook((void *)ipf_check_wrapper, NULL,
PFIL_IN|PFIL_OUT, ph_inet);
else
error = 0;
# else
error = pfil_remove_hook((void *)ipf_check, PFIL_IN|PFIL_OUT,
&inetsw[ip_protox[IPPROTO_IP]].pr_pfh);
# endif
if (error) {
SPL_X(s);
IPFERROR(130011);
return error;
}
# else
pfil_remove_hook((void *)ipf_check, PFIL_IN|PFIL_OUT);
# endif
# ifdef USE_INET6
# if __NetBSD_Version__ >= 105110000
if (ph_inet6 != NULL)
error = pfil_remove_hook((void *)ipf_check_wrapper6, NULL,
PFIL_IN|PFIL_OUT, ph_inet6);
else
error = 0;
# else
error = pfil_remove_hook((void *)ipf_check, PFIL_IN|PFIL_OUT,
&inetsw[ip_protox[IPPROTO_IPV6]].pr_pfh);
# endif
if (error) {
SPL_X(s);
IPFERROR(130012);
return error;
}
# endif
#endif
SPL_X(s);
#if (__NetBSD_Version__ >= 499005500)
for (i = 0; i < IPL_LOGSIZE; i++)
seldestroy(&ipfmain.ipf_selwait[i]);
#endif
ipf_fini_all(softc);
return 0;
}
/*
* Filter ioctl interface.
*/
int
ipfioctl(dev_t dev, u_long cmd,
#if (__NetBSD_Version__ >= 499001000)
void *data,
#else
caddr_t data,
#endif
int mode
#if (NetBSD >= 199511)
# if (__NetBSD_Version__ >= 399001400)
, struct lwp *p
# if (__NetBSD_Version__ >= 399002000)
# define UID(l) kauth_cred_getuid((l)->l_cred)
# else
# define UID(l) ((l)->l_proc->p_cred->p_ruid)
# endif
# else
, struct proc *p
# define UID(p) ((p)->p_cred->p_ruid)
# endif
#endif
)
{
int error = 0, unit = 0;
SPL_INT(s);
#if (__NetBSD_Version__ >= 399002000)
if ((mode & FWRITE) &&
kauth_authorize_network(p->l_cred, KAUTH_NETWORK_FIREWALL,
KAUTH_REQ_NETWORK_FIREWALL_FW, NULL,
NULL, NULL)) {
ipfmain.ipf_interror = 130005;
return EPERM;
}
#else
if ((securelevel >= 2) && (mode & FWRITE)) {
ipfmain.ipf_interror = 130001;
return EPERM;
}
#endif
unit = GET_MINOR(dev);
if ((IPL_LOGMAX < unit) || (unit < 0)) {
ipfmain.ipf_interror = 130002;
return ENXIO;
}
if (ipfmain.ipf_running <= 0) {
if (unit != IPL_LOGIPF && cmd != SIOCIPFINTERROR) {
ipfmain.ipf_interror = 130003;
return EIO;
}
if (cmd != SIOCIPFGETNEXT && cmd != SIOCIPFGET &&
cmd != SIOCIPFSET && cmd != SIOCFRENB &&
cmd != SIOCGETFS && cmd != SIOCGETFF &&
cmd != SIOCIPFINTERROR) {
ipfmain.ipf_interror = 130004;
return EIO;
}
}
SPL_NET(s);
error = ipf_ioctlswitch(&ipfmain, unit, data, cmd, mode, UID(p), p);
if (error != -1) {
SPL_X(s);
return error;
}
SPL_X(s);
return error;
}
/*
* ipf_send_reset - this could conceivably be a call to tcp_respond(), but that
* requires a large amount of setting up and isn't any more efficient.
*/
int
ipf_send_reset(fr_info_t *fin)
{
struct tcphdr *tcp, *tcp2;
int tlen = 0, hlen;
struct mbuf *m;
#ifdef USE_INET6
ip6_t *ip6;
#endif
ip_t *ip;
tcp = fin->fin_dp;
if (tcp->th_flags & TH_RST)
return -1; /* feedback loop */
if (ipf_checkl4sum(fin) == -1)
return -1;
tlen = fin->fin_dlen - (TCP_OFF(tcp) << 2) +
((tcp->th_flags & TH_SYN) ? 1 : 0) +
((tcp->th_flags & TH_FIN) ? 1 : 0);
#ifdef USE_INET6
hlen = (fin->fin_v == 6) ? sizeof(ip6_t) : sizeof(ip_t);
#else
hlen = sizeof(ip_t);
#endif
#ifdef MGETHDR
MGETHDR(m, M_DONTWAIT, MT_HEADER);
#else
MGET(m, M_DONTWAIT, MT_HEADER);
#endif
if (m == NULL)
return -1;
if (sizeof(*tcp2) + hlen > MHLEN) {
MCLGET(m, M_DONTWAIT);
if (m == NULL)
return -1;
if ((m->m_flags & M_EXT) == 0) {
FREE_MB_T(m);
return -1;
}
}
m->m_len = sizeof(*tcp2) + hlen;
m->m_data += max_linkhdr;
m->m_pkthdr.len = m->m_len;
m_reset_rcvif(m);
ip = mtod(m, struct ip *);
bzero((char *)ip, hlen);
#ifdef USE_INET6
ip6 = (ip6_t *)ip;
#endif
bzero((char *)ip, sizeof(*tcp2) + hlen);
tcp2 = (struct tcphdr *)((char *)ip + hlen);
tcp2->th_sport = tcp->th_dport;
tcp2->th_dport = tcp->th_sport;
if (tcp->th_flags & TH_ACK) {
tcp2->th_seq = tcp->th_ack;
tcp2->th_flags = TH_RST;
tcp2->th_ack = 0;
} else {
tcp2->th_seq = 0;
tcp2->th_ack = ntohl(tcp->th_seq);
tcp2->th_ack += tlen;
tcp2->th_ack = htonl(tcp2->th_ack);
tcp2->th_flags = TH_RST|TH_ACK;
}
tcp2->th_x2 = 0;
TCP_OFF_A(tcp2, sizeof(*tcp2) >> 2);
tcp2->th_win = tcp->th_win;
tcp2->th_sum = 0;
tcp2->th_urp = 0;
#ifdef USE_INET6
if (fin->fin_v == 6) {
ip6->ip6_flow = ((ip6_t *)fin->fin_ip)->ip6_flow;
ip6->ip6_plen = htons(sizeof(struct tcphdr));
ip6->ip6_nxt = IPPROTO_TCP;
ip6->ip6_hlim = 0;
ip6->ip6_src = fin->fin_dst6.in6;
ip6->ip6_dst = fin->fin_src6.in6;
tcp2->th_sum = in6_cksum(m, IPPROTO_TCP,
sizeof(*ip6), sizeof(*tcp2));
return ipf_send_ip(fin, m);
}
#endif
#ifdef INET
ip->ip_p = IPPROTO_TCP;
ip->ip_len = htons(sizeof(struct tcphdr));
ip->ip_src.s_addr = fin->fin_daddr;
ip->ip_dst.s_addr = fin->fin_saddr;
tcp2->th_sum = in_cksum(m, hlen + sizeof(*tcp2));
ip->ip_len = hlen + sizeof(*tcp2);
return ipf_send_ip(fin, m);
#else
return 0;
#endif
}
/*
* Expects ip_len to be in host byte order when called.
*/
static int
ipf_send_ip(fr_info_t *fin, mb_t *m)
{
fr_info_t fnew;
#ifdef INET
ip_t *oip;
#endif
ip_t *ip;
int hlen;
ip = mtod(m, ip_t *);
bzero((char *)&fnew, sizeof(fnew));
fnew.fin_main_soft = fin->fin_main_soft;
IP_V_A(ip, fin->fin_v);
switch (fin->fin_v)
{
#ifdef INET
case 4 :
oip = fin->fin_ip;
hlen = sizeof(*oip);
fnew.fin_v = 4;
fnew.fin_p = ip->ip_p;
fnew.fin_plen = ntohs(ip->ip_len);
HTONS(ip->ip_len);
IP_HL_A(ip, sizeof(*oip) >> 2);
ip->ip_tos = oip->ip_tos;
ip->ip_id = ipf_nextipid(fin);
ip->ip_off = htons(ip_mtudisc ? IP_DF : 0);
ip->ip_ttl = ip_defttl;
ip->ip_sum = 0;
break;
#endif
#ifdef USE_INET6
case 6 :
{
ip6_t *ip6 = (ip6_t *)ip;
ip6->ip6_vfc = 0x60;
ip6->ip6_hlim = IPDEFTTL;
hlen = sizeof(*ip6);
fnew.fin_p = ip6->ip6_nxt;
fnew.fin_v = 6;
fnew.fin_plen = ntohs(ip6->ip6_plen) + hlen;
break;
}
#endif
default :
return EINVAL;
}
#ifdef KAME_IPSEC
m_reset_rcvif(m);
#endif
fnew.fin_ifp = fin->fin_ifp;
fnew.fin_flx = FI_NOCKSUM;
fnew.fin_m = m;
fnew.fin_ip = ip;
fnew.fin_mp = &m;
fnew.fin_hlen = hlen;
fnew.fin_dp = (char *)ip + hlen;
(void) ipf_makefrip(hlen, ip, &fnew);
return ipf_fastroute(m, &m, &fnew, NULL);
}
int
ipf_send_icmp_err(int type, fr_info_t *fin, int dst)
{
int err, hlen, xtra, iclen, ohlen, avail;
struct in_addr dst4;
struct icmp *icmp;
struct mbuf *m;
i6addr_t dst6;
void *ifp;
#ifdef USE_INET6
int code;
ip6_t *ip6;
#endif
ip_t *ip, *ip2;
if ((type < 0) || (type > ICMP_MAXTYPE))
return -1;
#ifdef USE_INET6
code = fin->fin_icode;
if ((code < 0) || (code >= sizeof(icmptoicmp6unreach)/sizeof(int)))
return -1;
#endif
if (ipf_checkl4sum(fin) == -1)
return -1;
#ifdef MGETHDR
MGETHDR(m, M_DONTWAIT, MT_HEADER);
#else
MGET(m, M_DONTWAIT, MT_HEADER);
#endif
if (m == NULL)
return -1;
avail = MHLEN;
xtra = 0;
hlen = 0;
ohlen = 0;
dst4.s_addr = 0;
ifp = fin->fin_ifp;
if (fin->fin_v == 4) {
if ((fin->fin_p == IPPROTO_ICMP) && !(fin->fin_flx & FI_SHORT))
switch (ntohs(fin->fin_data[0]) >> 8)
{
case ICMP_ECHO :
case ICMP_TSTAMP :
case ICMP_IREQ :
case ICMP_MASKREQ :
break;
default :
FREE_MB_T(m);
return 0;
}
if (dst == 0) {
if (ipf_ifpaddr(&ipfmain, 4, FRI_NORMAL, ifp,
&dst6, NULL) == -1) {
FREE_MB_T(m);
return -1;
}
dst4 = dst6.in4;
} else
dst4.s_addr = fin->fin_daddr;
hlen = sizeof(ip_t);
ohlen = fin->fin_hlen;
iclen = hlen + offsetof(struct icmp, icmp_ip) + ohlen;
if (fin->fin_hlen < fin->fin_plen)
xtra = MIN(fin->fin_dlen, 8);
else
xtra = 0;
}
#ifdef USE_INET6
else if (fin->fin_v == 6) {
hlen = sizeof(ip6_t);
ohlen = sizeof(ip6_t);
iclen = hlen + offsetof(struct icmp, icmp_ip) + ohlen;
type = icmptoicmp6types[type];
if (type == ICMP6_DST_UNREACH)
code = icmptoicmp6unreach[code];
if (iclen + max_linkhdr + fin->fin_plen > avail) {
MCLGET(m, M_DONTWAIT);
if (m == NULL)
return -1;
if ((m->m_flags & M_EXT) == 0) {
FREE_MB_T(m);
return -1;
}
avail = MCLBYTES;
}
xtra = MIN(fin->fin_plen, avail - iclen - max_linkhdr);
xtra = MIN(xtra, IPV6_MMTU - iclen);
if (dst == 0 && !IN6_IS_ADDR_LINKLOCAL(&fin->fin_dst6.in6)) {
if (ipf_ifpaddr(&ipfmain, 6, FRI_NORMAL, ifp,
&dst6, NULL) == -1) {
FREE_MB_T(m);
return -1;
}
} else
dst6 = fin->fin_dst6;
}
#endif
else {
FREE_MB_T(m);
return -1;
}
avail -= (max_linkhdr + iclen);
if (avail < 0) {
FREE_MB_T(m);
return -1;
}
if (xtra > avail)
xtra = avail;
iclen += xtra;
m->m_data += max_linkhdr;
m_reset_rcvif(m);
m->m_pkthdr.len = iclen;
m->m_len = iclen;
ip = mtod(m, ip_t *);
icmp = (struct icmp *)((char *)ip + hlen);
ip2 = (ip_t *)&icmp->icmp_ip;
icmp->icmp_type = type;
icmp->icmp_code = fin->fin_icode;
icmp->icmp_cksum = 0;
#ifdef icmp_nextmtu
if (type == ICMP_UNREACH && fin->fin_icode == ICMP_UNREACH_NEEDFRAG) {
if (fin->fin_mtu != 0) {
icmp->icmp_nextmtu = htons(fin->fin_mtu);
} else if (ifp != NULL) {
icmp->icmp_nextmtu = htons(GETIFMTU_4(ifp));
} else { /* make up a number... */
icmp->icmp_nextmtu = htons(fin->fin_plen - 20);
}
}
#endif
bcopy((char *)fin->fin_ip, (char *)ip2, ohlen);
#if defined(M_CSUM_IPv4)
/*
* Clear any in-bound checksum flags for this packet.
*/
m->m_pkthdr.csuminfo = 0;
#endif /* __NetBSD__ && M_CSUM_IPv4 */
#ifdef USE_INET6
ip6 = (ip6_t *)ip;
if (fin->fin_v == 6) {
ip6->ip6_flow = ((ip6_t *)fin->fin_ip)->ip6_flow;
ip6->ip6_plen = htons(iclen - hlen);
ip6->ip6_nxt = IPPROTO_ICMPV6;
ip6->ip6_hlim = 0;
ip6->ip6_src = dst6.in6;
ip6->ip6_dst = fin->fin_src6.in6;
if (xtra > 0)
bcopy((char *)fin->fin_ip + ohlen,
(char *)&icmp->icmp_ip + ohlen, xtra);
icmp->icmp_cksum = in6_cksum(m, IPPROTO_ICMPV6,
sizeof(*ip6), iclen - hlen);
} else
#endif
{
ip->ip_p = IPPROTO_ICMP;
ip->ip_src.s_addr = dst4.s_addr;
ip->ip_dst.s_addr = fin->fin_saddr;
if (xtra > 0)
bcopy((char *)fin->fin_ip + ohlen,
(char *)&icmp->icmp_ip + ohlen, xtra);
icmp->icmp_cksum = ipf_cksum((u_short *)icmp,
sizeof(*icmp) + 8);
ip->ip_len = iclen;
ip->ip_p = IPPROTO_ICMP;
}
err = ipf_send_ip(fin, m);
return err;
}
/*
* m0 - pointer to mbuf where the IP packet starts
* mpp - pointer to the mbuf pointer that is the start of the mbuf chain
*/
int
ipf_fastroute(mb_t *m0, mb_t **mpp, fr_info_t *fin, frdest_t *fdp)
{
register struct ip *ip, *mhip;
register struct mbuf *m = *mpp;
register struct route *ro;
int len, off, error = 0, hlen, code;
struct ifnet *ifp, *sifp;
ipf_main_softc_t *softc;
#if __NetBSD_Version__ >= 499001100
union {
struct sockaddr dst;
struct sockaddr_in dst4;
} u;
#else
struct sockaddr_in *dst4;
#endif
struct sockaddr *dst;
u_short ip_off, ip_len;
struct route iproute;
struct rtentry *rt;
frdest_t node;
frentry_t *fr;
if (fin->fin_v == 6) {
#ifdef USE_INET6
error = ipf_fastroute6(m0, mpp, fin, fdp);
#else
error = EPROTONOSUPPORT;
#endif
if ((error != 0) && (*mpp != NULL))
FREE_MB_T(*mpp);
return error;
}
#ifndef INET
FREE_MB_T(*mpp);
return EPROTONOSUPPORT;
#else
hlen = fin->fin_hlen;
ip = mtod(m0, struct ip *);
softc = fin->fin_main_soft;
rt = NULL;
ifp = NULL;
# if defined(M_CSUM_IPv4)
/*
* Clear any in-bound checksum flags for this packet.
*/
m0->m_pkthdr.csuminfo = 0;
# endif /* __NetBSD__ && M_CSUM_IPv4 */
/*
* Route packet.
*/
ro = &iproute;
memset(ro, 0, sizeof(*ro));
fr = fin->fin_fr;
if ((fr != NULL) && !(fr->fr_flags & FR_KEEPSTATE) && (fdp != NULL) &&
(fdp->fd_type == FRD_DSTLIST)) {
if (ipf_dstlist_select_node(fin, fdp->fd_ptr, NULL, &node) == 0)
fdp = &node;
}
if (fdp != NULL)
ifp = fdp->fd_ptr;
else
ifp = fin->fin_ifp;
if ((ifp == NULL) && ((fr == NULL) || !(fr->fr_flags & FR_FASTROUTE))) {
error = -2;
goto bad;
}
# if __NetBSD_Version__ >= 499001100
if ((fdp != NULL) && (fdp->fd_ip.s_addr != 0))
sockaddr_in_init(&u.dst4, &fdp->fd_ip, 0);
else
sockaddr_in_init(&u.dst4, &ip->ip_dst, 0);
dst = &u.dst;
rtcache_setdst(ro, dst);
rt = rtcache_init(ro);
# else
dst4 = (struct sockaddr_in *)&ro->ro_dst;
dst = (struct sockaddr *)dst4;
dst4->sin_family = AF_INET;
dst4->sin_addr = ip->ip_dst;
if ((fdp != NULL) && (fdp->fd_ip.s_addr != 0))
dst4->sin_addr = fdp->fd_ip;
dst4->sin_len = sizeof(*dst);
rtalloc(ro);
rt = ro->ro_rt;
# endif
if ((ifp == NULL) && (rt != NULL))
ifp = rt->rt_ifp;
if ((rt == NULL) || (ifp == NULL)) {
#ifdef INET
if (in_localaddr(ip->ip_dst))
error = EHOSTUNREACH;
else
#endif
error = ENETUNREACH;
goto bad;
}
if (rt->rt_flags & RTF_GATEWAY)
dst = rt->rt_gateway;
rt->rt_use++;
/*
* For input packets which are being "fastrouted", they won't
* go back through output filtering and miss their chance to get
* NAT'd and counted. Duplicated packets aren't considered to be
* part of the normal packet stream, so do not NAT them or pass
* them through stateful checking, etc.
*/
if ((fdp != &fr->fr_dif) && (fin->fin_out == 0)) {
sifp = fin->fin_ifp;
fin->fin_ifp = ifp;
fin->fin_out = 1;
(void) ipf_acctpkt(fin, NULL);
fin->fin_fr = NULL;
if (!fr || !(fr->fr_flags & FR_RETMASK)) {
u_32_t pass;
(void) ipf_state_check(fin, &pass);
}
switch (ipf_nat_checkout(fin, NULL))
{
case 0 :
break;
case 1 :
ip->ip_sum = 0;
break;
case -1 :
error = -1;
goto bad;
break;
}
fin->fin_ifp = sifp;
fin->fin_out = 0;
} else
ip->ip_sum = 0;
/*
* If small enough for interface, can just send directly.
*/
m_set_rcvif(m, ifp);
ip_len = ntohs(ip->ip_len);
if (ip_len <= ifp->if_mtu) {
# if defined(M_CSUM_IPv4)
# if (__NetBSD_Version__ >= 105009999)
if (ifp->if_csum_flags_tx & M_CSUM_IPv4)
m->m_pkthdr.csuminfo |= M_CSUM_IPv4;
# else
if (ifp->if_capabilities & IFCAP_CSUM_IPv4)
m->m_pkthdr.csuminfo |= M_CSUM_IPv4;
# endif /* (__NetBSD_Version__ >= 105009999) */
else if (ip->ip_sum == 0)
ip->ip_sum = in_cksum(m, hlen);
# else
if (!ip->ip_sum)
ip->ip_sum = in_cksum(m, hlen);
# endif /* M_CSUM_IPv4 */
error = if_output_lock(ifp, ifp, m, dst, rt);
goto done;
}
/*
* Too large for interface; fragment if possible.
* Must be able to put at least 8 bytes per fragment.
*/
ip_off = ntohs(ip->ip_off);
if (ip_off & IP_DF) {
error = EMSGSIZE;
goto bad;
}
len = (ifp->if_mtu - hlen) &~ 7;
if (len < 8) {
error = EMSGSIZE;
goto bad;
}
{
int mhlen, firstlen = len;
struct mbuf **mnext = &m->m_act;
/*
* Loop through length of segment after first fragment,
* make new header and copy data of each part and link onto chain.
*/
m0 = m;
mhlen = sizeof (struct ip);
for (off = hlen + len; off < ip_len; off += len) {
# ifdef MGETHDR
MGETHDR(m, M_DONTWAIT, MT_HEADER);
# else
MGET(m, M_DONTWAIT, MT_HEADER);
# endif
if (m == 0) {
m = m0;
error = ENOBUFS;
goto bad;
}
m->m_data += max_linkhdr;
mhip = mtod(m, struct ip *);
bcopy((char *)ip, (char *)mhip, sizeof(*ip));
#ifdef INET
if (hlen > sizeof (struct ip)) {
mhlen = ip_optcopy(ip, mhip) + sizeof (struct ip);
IP_HL_A(mhip, mhlen >> 2);
}
#endif
m->m_len = mhlen;
mhip->ip_off = ((off - hlen) >> 3) + ip_off;
if (off + len >= ip_len)
len = ip_len - off;
else
mhip->ip_off |= IP_MF;
mhip->ip_len = htons((u_short)(len + mhlen));
m->m_next = m_copym(m0, off, len, M_DONTWAIT);
if (m->m_next == 0) {
error = ENOBUFS; /* ??? */
goto sendorfree;
}
m->m_pkthdr.len = mhlen + len;
m_reset_rcvif(m);
mhip->ip_off = htons((u_short)mhip->ip_off);
mhip->ip_sum = 0;
#ifdef INET
mhip->ip_sum = in_cksum(m, mhlen);
#endif
*mnext = m;
mnext = &m->m_act;
}
/*
* Update first fragment by trimming what's been copied out
* and updating header, then send each fragment (in order).
*/
m_adj(m0, hlen + firstlen - ip_len);
ip->ip_len = htons((u_short)(hlen + firstlen));
ip->ip_off = htons((u_short)IP_MF);
ip->ip_sum = 0;
#ifdef INET
ip->ip_sum = in_cksum(m0, hlen);
#endif
sendorfree:
for (m = m0; m; m = m0) {
m0 = m->m_act;
m->m_act = 0;
if (error == 0) {
KERNEL_LOCK(1, NULL);
error = (*ifp->if_output)(ifp, m, dst, rt);
KERNEL_UNLOCK_ONE(NULL);
} else {
FREE_MB_T(m);
}
}
}
done:
if (!error)
softc->ipf_frouteok[0]++;
else
softc->ipf_frouteok[1]++;
# if __NetBSD_Version__ >= 499001100
rtcache_unref(rt, ro);
rtcache_free(ro);
# else
if (rt) {
RTFREE(rt);
}
# endif
return error;
bad:
if (error == EMSGSIZE) {
sifp = fin->fin_ifp;
code = fin->fin_icode;
fin->fin_icode = ICMP_UNREACH_NEEDFRAG;
fin->fin_ifp = ifp;
(void) ipf_send_icmp_err(ICMP_UNREACH, fin, 1);
fin->fin_ifp = sifp;
fin->fin_icode = code;
}
FREE_MB_T(m);
goto done;
#endif /* INET */
}
#if defined(USE_INET6)
/*
* This is the IPv6 specific fastroute code. It doesn't clean up the mbuf's
* or ensure that it is an IPv6 packet that is being forwarded, those are
* expected to be done by the called (ipf_fastroute).
*/
static int
ipf_fastroute6(struct mbuf *m0, struct mbuf **mpp, fr_info_t *fin,
frdest_t *fdp)
{
# if __NetBSD_Version__ >= 499001100
struct route ip6route;
const struct sockaddr *dst;
union {
struct sockaddr dst;
struct sockaddr_in6 dst6;
} u;
struct route *ro;
# else
struct route_in6 ip6route;
struct sockaddr_in6 *dst6;
struct route_in6 *ro;
# endif
struct rtentry *rt;
struct ifnet *ifp;
u_long mtu;
int error;
error = 0;
ro = &ip6route;
if (fdp != NULL)
ifp = fdp->fd_ptr;
else
ifp = fin->fin_ifp;
memset(ro, 0, sizeof(*ro));
# if __NetBSD_Version__ >= 499001100
if (fdp != NULL && IP6_NOTZERO(&fdp->fd_ip6))
sockaddr_in6_init(&u.dst6, &fdp->fd_ip6.in6, 0, 0, 0);
else
sockaddr_in6_init(&u.dst6, &fin->fin_fi.fi_dst.in6, 0, 0, 0);
if ((error = in6_setscope(&u.dst6.sin6_addr, ifp,
&u.dst6.sin6_scope_id)) != 0)
return error;
if ((error = sa6_embedscope(&u.dst6, 0)) != 0)
return error;
dst = &u.dst;
rtcache_setdst(ro, dst);
rt = rtcache_init(ro);
if ((ifp == NULL) && (rt != NULL))
ifp = rt->rt_ifp;
# else
dst6 = (struct sockaddr_in6 *)&ro->ro_dst;
dst6->sin6_family = AF_INET6;
dst6->sin6_len = sizeof(struct sockaddr_in6);
dst6->sin6_addr = fin->fin_fi.fi_dst.in6;
/* KAME */
if (IN6_IS_ADDR_LINKLOCAL(&dst6->sin6_addr))
dst6->sin6_addr.s6_addr16[1] = htons(ifp->if_index);
if (fdp != NULL) {
if (IP6_NOTZERO(&fdp->fd_ip6))
dst6->sin6_addr = fdp->fd_ip6.in6;
}
rtalloc((struct route *)ro);
if ((ifp == NULL) && (ro->ro_rt != NULL))
ifp = ro->ro_rt->rt_ifp;
rt = ro->ro_rt;
# endif
if ((rt == NULL) || (ifp == NULL)) {
error = EHOSTUNREACH;
goto bad;
}
{
# if (__NetBSD_Version__ >= 106010000) && !defined(IN6_LINKMTU) \
&& defined(IPV6CTL_ACCEPT_RTADV)
struct in6_ifextra *ife;
# endif
if (rt->rt_flags & RTF_GATEWAY)
# if __NetBSD_Version__ >= 499001100
dst = rt->rt_gateway;
# else
dst6 = (struct sockaddr_in6 *)rt->rt_gateway;
# endif
rt->rt_use++;
/* Determine path MTU. */
# if (__NetBSD_Version__ <= 106009999)
mtu = nd_ifinfo[ifp->if_index].linkmtu;
# elif defined(IPV6CTL_ACCEPT_RTADV)
# ifdef IN6_LINKMTU
mtu = IN6_LINKMTU(ifp);
# else
ife = (struct in6_ifextra *)(ifp)->if_afdata[AF_INET6];
mtu = ife->nd_ifinfo[ifp->if_index].linkmtu;
# endif
# else
mtu = ifp->if_mtu;
# endif
if ((error == 0) && (m0->m_pkthdr.len <= mtu)) {
# if __NetBSD_Version__ >= 499001100
error = ip6_if_output(ifp, ifp, m0, satocsin6(dst), rt);
# else
error = nd6_output(ifp, ifp, m0, dst6, rt);
# endif
if (error)
*mpp = NULL; /* m0 has been freed */
} else {
error = EMSGSIZE;
}
}
bad:
# if __NetBSD_Version__ >= 499001100
rtcache_unref(rt, ro);
rtcache_free(ro);
# else
if (ro->ro_rt != NULL) {
RTFREE(((struct route *)ro)->ro_rt);
}
# endif
return error;
}
#endif /* INET6 */
int
ipf_verifysrc(fr_info_t *fin)
{
#if __NetBSD_Version__ >= 499001100
union {
struct sockaddr dst;
struct sockaddr_in dst4;
} u;
struct rtentry *rt;
#else
struct sockaddr_in *dst;
#endif
struct route iproute;
int rc;
#if __NetBSD_Version__ >= 499001100
sockaddr_in_init(&u.dst4, &fin->fin_src, 0);
rtcache_setdst(&iproute, &u.dst);
rt = rtcache_init(&iproute);
if (rt == NULL)
rc = 0;
else
rc = (fin->fin_ifp == rt->rt_ifp);
rtcache_unref(rt, &iproute);
rtcache_free(&iproute);
#else
dst = (struct sockaddr_in *)&iproute.ro_dst;
dst->sin_len = sizeof(*dst);
dst->sin_family = AF_INET;
dst->sin_addr = fin->fin_src;
rtalloc(&iproute);
if (iproute.ro_rt == NULL)
return 0;
rc = (fin->fin_ifp == iproute.ro_rt->rt_ifp);
RTFREE(iproute.ro_rt);
#endif
return rc;
}
/*
* return the first IP Address associated with an interface
*/
int
ipf_ifpaddr(ipf_main_softc_t *softc, int v, int atype, void *ifptr,
i6addr_t *inp, i6addr_t *inpmask)
{
#ifdef USE_INET6
struct in6_addr *inp6 = NULL;
#endif
struct sockaddr *sock, *mask;
struct sockaddr_in *sin;
struct ifaddr *ifa;
struct ifnet *ifp;
if ((ifptr == NULL) || (ifptr == (void *)-1))
return -1;
ifp = ifptr;
mask = NULL;
if (v == 4)
inp->in4.s_addr = 0;
#ifdef USE_INET6
else if (v == 6)
bzero((char *)inp, sizeof(*inp));
#endif
ifa = IFADDR_READER_FIRST(ifp);
sock = ifa ? ifa->ifa_addr : NULL;
while (sock != NULL && ifa != NULL) {
sin = (struct sockaddr_in *)sock;
if ((v == 4) && (sin->sin_family == AF_INET))
break;
#ifdef USE_INET6
if ((v == 6) && (sin->sin_family == AF_INET6)) {
inp6 = &((struct sockaddr_in6 *)sin)->sin6_addr;
if (!IN6_IS_ADDR_LINKLOCAL(inp6) &&
!IN6_IS_ADDR_LOOPBACK(inp6))
break;
}
#endif
ifa = IFADDR_READER_NEXT(ifa);
if (ifa != NULL)
sock = ifa->ifa_addr;
}
if (ifa == NULL || sock == NULL)
return -1;
mask = ifa->ifa_netmask;
if (atype == FRI_BROADCAST)
sock = ifa->ifa_broadaddr;
else if (atype == FRI_PEERADDR)
sock = ifa->ifa_dstaddr;
#ifdef USE_INET6
if (v == 6)
return ipf_ifpfillv6addr(atype, (struct sockaddr_in6 *)sock,
(struct sockaddr_in6 *)mask,
inp, inpmask);
#endif
return ipf_ifpfillv4addr(atype, (struct sockaddr_in *)sock,
(struct sockaddr_in *)mask,
&inp->in4, &inpmask->in4);
}
u_32_t
ipf_newisn(fr_info_t *fin)
{
#if __NetBSD_Version__ >= 105190000 /* 1.5T */
size_t asz;
if (fin->fin_v == 4)
asz = sizeof(struct in_addr);
else if (fin->fin_v == 6)
asz = sizeof(fin->fin_src);
else /* XXX: no way to return error */
return 0;
#ifdef INET
return tcp_new_iss1((void *)&fin->fin_src, (void *)&fin->fin_dst,
fin->fin_sport, fin->fin_dport, asz);
#else
return ENOSYS;
#endif
#else
static int iss_seq_off = 0;
u_char hash[16];
u_32_t newiss;
MD5_CTX ctx;
/*
* Compute the base value of the ISS. It is a hash
* of (saddr, sport, daddr, dport, secret).
*/
MD5Init(&ctx);
MD5Update(&ctx, (u_char *) &fin->fin_fi.fi_src,
sizeof(fin->fin_fi.fi_src));
MD5Update(&ctx, (u_char *) &fin->fin_fi.fi_dst,
sizeof(fin->fin_fi.fi_dst));
MD5Update(&ctx, (u_char *) &fin->fin_dat, sizeof(fin->fin_dat));
MD5Update(&ctx, ipf_iss_secret, sizeof(ipf_iss_secret));
MD5Final(hash, &ctx);
memcpy(&newiss, hash, sizeof(newiss));
/*
* Now increment our "timer", and add it in to
* the computed value.
*
* XXX Use `addin'?
* XXX TCP_ISSINCR too large to use?
*/
iss_seq_off += 0x00010000;
newiss += iss_seq_off;
return newiss;
#endif
}
/* ------------------------------------------------------------------------ */
/* Function: ipf_nextipid */
/* Returns: int - 0 == success, -1 == error (packet should be droppped) */
/* Parameters: fin(I) - pointer to packet information */
/* */
/* Returns the next IPv4 ID to use for this packet. */
/* ------------------------------------------------------------------------ */
u_short
ipf_nextipid(fr_info_t *fin)
{
#ifdef USE_MUTEXES
ipf_main_softc_t *softc = fin->fin_main_soft;
#endif
u_short id;
MUTEX_ENTER(&softc->ipf_rw);
id = ipid++;
MUTEX_EXIT(&softc->ipf_rw);
return id;
}
EXTERN_INLINE int
ipf_checkv4sum(fr_info_t *fin)
{
#ifdef M_CSUM_TCP_UDP_BAD
int manual, pflag, cflags, active;
mb_t *m;
if ((fin->fin_flx & FI_NOCKSUM) != 0)
return 0;
if ((fin->fin_flx & FI_SHORT) != 0)
return 1;
if (fin->fin_cksum != FI_CK_NEEDED)
return (fin->fin_cksum > FI_CK_NEEDED) ? 0 : -1;
manual = 0;
m = fin->fin_m;
if (m == NULL) {
manual = 1;
goto skipauto;
}
switch (fin->fin_p)
{
case IPPROTO_UDP :
pflag = M_CSUM_UDPv4;
break;
case IPPROTO_TCP :
pflag = M_CSUM_TCPv4;
break;
default :
pflag = 0;
manual = 1;
break;
}
active = ((struct ifnet *)fin->fin_ifp)->if_csum_flags_rx & pflag;
active |= M_CSUM_TCP_UDP_BAD | M_CSUM_DATA;
cflags = m->m_pkthdr.csum_flags & active;
if (pflag != 0) {
if (cflags == (pflag | M_CSUM_TCP_UDP_BAD)) {
fin->fin_flx |= FI_BAD;
fin->fin_cksum = FI_CK_BAD;
} else if (cflags == (pflag | M_CSUM_DATA)) {
if ((m->m_pkthdr.csum_data ^ 0xffff) != 0) {
fin->fin_flx |= FI_BAD;
fin->fin_cksum = FI_CK_BAD;
} else {
fin->fin_cksum = FI_CK_SUMOK;
}
} else if (cflags == pflag) {
fin->fin_cksum = FI_CK_SUMOK;
} else {
manual = 1;
}
}
skipauto:
if (manual != 0) {
if (ipf_checkl4sum(fin) == -1) {
fin->fin_flx |= FI_BAD;
return -1;
}
}
#else
if (ipf_checkl4sum(fin) == -1) {
fin->fin_flx |= FI_BAD;
return -1;
}
#endif
return 0;
}
#ifdef USE_INET6
EXTERN_INLINE int
ipf_checkv6sum(fr_info_t *fin)
{
# ifdef M_CSUM_TCP_UDP_BAD
int manual, pflag, cflags, active;
mb_t *m;
if ((fin->fin_flx & FI_NOCKSUM) != 0)
return 0;
if ((fin->fin_flx & FI_SHORT) != 0)
return 1;
if (fin->fin_cksum != FI_CK_SUMOK)
return (fin->fin_cksum > FI_CK_NEEDED) ? 0 : -1;
manual = 0;
m = fin->fin_m;
switch (fin->fin_p)
{
case IPPROTO_UDP :
pflag = M_CSUM_UDPv6;
break;
case IPPROTO_TCP :
pflag = M_CSUM_TCPv6;
break;
default :
pflag = 0;
manual = 1;
break;
}
active = ((struct ifnet *)fin->fin_ifp)->if_csum_flags_rx & pflag;
active |= M_CSUM_TCP_UDP_BAD | M_CSUM_DATA;
cflags = m->m_pkthdr.csum_flags & active;
if (pflag != 0) {
if (cflags == (pflag | M_CSUM_TCP_UDP_BAD)) {
fin->fin_flx |= FI_BAD;
} else if (cflags == (pflag | M_CSUM_DATA)) {
if ((m->m_pkthdr.csum_data ^ 0xffff) != 0)
fin->fin_flx |= FI_BAD;
} else if (cflags == pflag) {
;
} else {
manual = 1;
}
}
if (manual != 0) {
if (ipf_checkl4sum(fin) == -1) {
fin->fin_flx |= FI_BAD;
return -1;
}
}
# else
if (ipf_checkl4sum(fin) == -1) {
fin->fin_flx |= FI_BAD;
return -1;
}
# endif
return 0;
}
#endif /* USE_INET6 */
size_t
mbufchainlen(struct mbuf *m0)
{
size_t len;
if ((m0->m_flags & M_PKTHDR) != 0) {
len = m0->m_pkthdr.len;
} else {
struct mbuf *m;
for (m = m0, len = 0; m != NULL; m = m->m_next)
len += m->m_len;
}
return len;
}
/* ------------------------------------------------------------------------ */
/* Function: ipf_pullup */
/* Returns: NULL == pullup failed, else pointer to protocol header */
/* Parameters: xmin(I)- pointer to buffer where data packet starts */
/* fin(I) - pointer to packet information */
/* len(I) - number of bytes to pullup */
/* */
/* Attempt to move at least len bytes (from the start of the buffer) into a */
/* single buffer for ease of access. Operating system native functions are */
/* used to manage buffers - if necessary. If the entire packet ends up in */
/* a single buffer, set the FI_COALESCE flag even though ipf_coalesce() has */
/* not been called. Both fin_ip and fin_dp are updated before exiting _IF_ */
/* and ONLY if the pullup succeeds. */
/* */
/* We assume that 'xmin' is a pointer to a buffer that is part of the chain */
/* of buffers that starts at *fin->fin_mp. */
/* ------------------------------------------------------------------------ */
void *
ipf_pullup(mb_t *xmin, fr_info_t *fin, int len)
{
int dpoff, ipoff;
mb_t *m = xmin;
char *ip;
if (m == NULL)
return NULL;
ip = (char *)fin->fin_ip;
if ((fin->fin_flx & FI_COALESCE) != 0)
return ip;
ipoff = fin->fin_ipoff;
if (fin->fin_dp != NULL)
dpoff = (char *)fin->fin_dp - (char *)ip;
else
dpoff = 0;
if (M_LEN(m) < len) {
mb_t *n = *fin->fin_mp;
/*
* Assume that M_PKTHDR is set and just work with what is left
* rather than check..
* Should not make any real difference, anyway.
*/
if (m != n) {
/*
* Record the mbuf that points to the mbuf that we're
* about to go to work on so that we can update the
* m_next appropriately later.
*/
for (; n->m_next != m; n = n->m_next)
;
} else {
n = NULL;
}
#ifdef MHLEN
if (len > MHLEN)
#else
if (len > MLEN)
#endif
{
#ifdef HAVE_M_PULLDOWN
if (m_pulldown(m, 0, len, NULL) == NULL)
m = NULL;
#else
FREE_MB_T(*fin->fin_mp);
m = NULL;
n = NULL;
#endif
} else
{
m = m_pullup(m, len);
}
if (n != NULL)
n->m_next = m;
if (m == NULL) {
/*
* When n is non-NULL, it indicates that m pointed to
* a sub-chain (tail) of the mbuf and that the head
* of this chain has not yet been free'd.
*/
if (n != NULL) {
FREE_MB_T(*fin->fin_mp);
}
*fin->fin_mp = NULL;
fin->fin_m = NULL;
return NULL;
}
if (n == NULL)
*fin->fin_mp = m;
while (M_LEN(m) == 0) {
m = m->m_next;
}
fin->fin_m = m;
ip = MTOD(m, char *) + ipoff;
fin->fin_ip = (ip_t *)ip;
if (fin->fin_dp != NULL)
fin->fin_dp = (char *)fin->fin_ip + dpoff;
if (fin->fin_fraghdr != NULL)
fin->fin_fraghdr = (char *)ip +
((char *)fin->fin_fraghdr -
(char *)fin->fin_ip);
}
if (len == fin->fin_plen)
fin->fin_flx |= FI_COALESCE;
return ip;
}
int
ipf_inject(fr_info_t *fin, mb_t *m)
{
int error;
if (fin->fin_out == 0) {
if (__predict_false(!pktq_enqueue(ip_pktq, m, 0))) {
FREE_MB_T(m);
error = ENOBUFS;
} else {
error = 0;
}
} else {
error = ip_output(m, NULL, NULL, IP_FORWARDING, NULL, NULL);
}
return error;
}
u_32_t
ipf_random(void)
{
int number;
#ifdef _CPRNG_H
number = cprng_fast32();
#else
number = arc4random();
#endif
return number;
}
/*
* routines below for saving IP headers to buffer
*/
static int ipfopen(dev_t dev, int flags
#if (NetBSD >= 199511)
, int devtype, PROC_T *p
#endif
)
{
u_int unit = GET_MINOR(dev);
int error;
if (IPL_LOGMAX < unit) {
error = ENXIO;
} else {
switch (unit)
{
case IPL_LOGIPF :
case IPL_LOGNAT :
case IPL_LOGSTATE :
case IPL_LOGAUTH :
case IPL_LOGLOOKUP :
case IPL_LOGSYNC :
#ifdef IPFILTER_SCAN
case IPL_LOGSCAN :
#endif
error = 0;
break;
default :
error = ENXIO;
break;
}
}
#if (__NetBSD_Version__ >= 799003000)
if (error == 0) {
mutex_enter(&ipf_ref_mutex);
ipf_active = 1;
mutex_exit(&ipf_ref_mutex);
}
#endif
return error;
}
static int ipfclose(dev_t dev, int flags
#if (NetBSD >= 199511)
, int devtype, PROC_T *p
#endif
)
{
u_int unit = GET_MINOR(dev);
if (IPL_LOGMAX < unit)
return ENXIO;
else {
#if (__NetBSD_Version__ >= 799003000)
mutex_enter(&ipf_ref_mutex);
ipf_active = 0;
mutex_exit(&ipf_ref_mutex);
#endif
return 0;
}
}
/*
* ipfread/ipflog
* both of these must operate with at least splnet() lest they be
* called during packet processing and cause an inconsistancy to appear in
* the filter lists.
*/
static int ipfread(dev_t dev, struct uio *uio, int ioflag)
{
if (ipfmain.ipf_running < 1) {
ipfmain.ipf_interror = 130006;
return EIO;
}
if (GET_MINOR(dev) == IPL_LOGSYNC)
return ipf_sync_read(&ipfmain, uio);
#ifdef IPFILTER_LOG
return ipf_log_read(&ipfmain, GET_MINOR(dev), uio);
#else
ipfmain.ipf_interror = 130007;
return ENXIO;
#endif
}
/*
* ipfwrite
* both of these must operate with at least splnet() lest they be
* called during packet processing and cause an inconsistancy to appear in
* the filter lists.
*/
static int ipfwrite(dev_t dev, struct uio *uio, int ioflag)
{
if (ipfmain.ipf_running < 1) {
ipfmain.ipf_interror = 130008;
return EIO;
}
if (GET_MINOR(dev) == IPL_LOGSYNC)
return ipf_sync_write(&ipfmain, uio);
ipfmain.ipf_interror = 130009;
return ENXIO;
}
static int ipfpoll(dev_t dev, int events, PROC_T *p)
{
u_int unit = GET_MINOR(dev);
int revents = 0;
if (IPL_LOGMAX < unit) {
ipfmain.ipf_interror = 130010;
return ENXIO;
}
switch (unit)
{
case IPL_LOGIPF :
case IPL_LOGNAT :
case IPL_LOGSTATE :
#ifdef IPFILTER_LOG
if ((events & (POLLIN | POLLRDNORM)) &&
ipf_log_canread(&ipfmain, unit))
revents |= events & (POLLIN | POLLRDNORM);
#endif
break;
case IPL_LOGAUTH :
if ((events & (POLLIN | POLLRDNORM)) &&
ipf_auth_waiting(&ipfmain))
revents |= events & (POLLIN | POLLRDNORM);
break;
case IPL_LOGSYNC :
if ((events & (POLLIN | POLLRDNORM)) &&
ipf_sync_canread(&ipfmain))
revents |= events & (POLLIN | POLLRDNORM);
if ((events & (POLLOUT | POLLWRNORM)) &&
ipf_sync_canwrite(&ipfmain))
revents |= events & (POLLOUT | POLLWRNORM);
break;
case IPL_LOGSCAN :
case IPL_LOGLOOKUP :
default :
break;
}
if ((revents == 0) && (((events & (POLLIN|POLLRDNORM)) != 0)))
selrecord(p, &ipfmain.ipf_selwait[unit]);
return revents;
}
u_int
ipf_pcksum(fr_info_t *fin, int hlen, u_int sum)
{
struct mbuf *m;
u_int sum2;
int off;
m = fin->fin_m;
off = (char *)fin->fin_dp - (char *)fin->fin_ip;
m->m_data += hlen;
m->m_len -= hlen;
sum2 = in_cksum(fin->fin_m, fin->fin_plen - off);
m->m_len += hlen;
m->m_data -= hlen;
/*
* Both sum and sum2 are partial sums, so combine them together.
*/
sum += ~sum2 & 0xffff;
while (sum > 0xffff)
sum = (sum & 0xffff) + (sum >> 16);
sum2 = ~sum & 0xffff;
return sum2;
}
#if (__NetBSD_Version__ >= 799003000)
/* NetBSD module interface */
MODULE(MODULE_CLASS_DRIVER, ipl, "bpf_filter");
static int ipl_init(void *);
static int ipl_fini(void *);
static int ipl_modcmd(modcmd_t, void *);
#ifdef _MODULE
static devmajor_t ipl_cmaj = -1, ipl_bmaj = -1;
#endif
static int
ipl_modcmd(modcmd_t cmd, void *opaque)
{
switch (cmd) {
case MODULE_CMD_INIT:
return ipl_init(opaque);
case MODULE_CMD_FINI:
return ipl_fini(opaque);
default:
return ENOTTY;
}
}
static int
ipl_init(void *opaque)
{
int error;
ipf_listener = kauth_listen_scope(KAUTH_SCOPE_NETWORK,
ipf_listener_cb, NULL);
if ((error = ipf_load_all()) != 0)
return error;
if (ipf_create_all(&ipfmain) == NULL) {
ipf_unload_all();
return ENODEV;
}
/* Initialize our mutex and reference count */
mutex_init(&ipf_ref_mutex, MUTEX_DEFAULT, IPL_NONE);
ipf_active = 0;
#ifdef _MODULE
/*
* Insert ourself into the cdevsw list.
*/
error = devsw_attach("ipl", NULL, &ipl_bmaj, &ipl_cdevsw, &ipl_cmaj);
if (error)
ipl_fini(opaque);
#endif
return error;
}
static int
ipl_fini(void *opaque)
{
#ifdef _MODULE
devsw_detach(NULL, &ipl_cdevsw);
#endif
/*
* Grab the mutex, verify that there are no references
* and that there are no running filters. If either
* of these exists, reinsert our cdevsw entry and return
* an error.
*/
mutex_enter(&ipf_ref_mutex);
if (ipf_active != 0 || ipfmain.ipf_running > 0) {
#ifdef _MODULE
(void)devsw_attach("ipl", NULL, &ipl_bmaj,
&ipl_cdevsw, &ipl_cmaj);
#endif
mutex_exit(&ipf_ref_mutex);
return EBUSY;
}
/* Clean up the rest of our state before being unloaded */
mutex_exit(&ipf_ref_mutex);
mutex_destroy(&ipf_ref_mutex);
ipf_destroy_all(&ipfmain);
ipf_unload_all();
kauth_unlisten_scope(ipf_listener);
return 0;
}
#endif /* (__NetBSD_Version__ >= 799003000) */