7e62c90eef
This patch is part of a series that moves towards a consistent use of g_assert_not_reached() rather than an ad hoc mix of different assertion mechanisms. Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Signed-off-by: Pierrick Bouvier <pierrick.bouvier@linaro.org> Message-ID: <20240919044641.386068-19-pierrick.bouvier@linaro.org> Signed-off-by: Thomas Huth <thuth@redhat.com>
676 lines
18 KiB
C
676 lines
18 KiB
C
/*
|
|
* QEMU RX packets abstractions
|
|
*
|
|
* Copyright (c) 2012 Ravello Systems LTD (http://ravellosystems.com)
|
|
*
|
|
* Developed by Daynix Computing LTD (http://www.daynix.com)
|
|
*
|
|
* Authors:
|
|
* Dmitry Fleytman <dmitry@daynix.com>
|
|
* Tamir Shomer <tamirs@daynix.com>
|
|
* Yan Vugenfirer <yan@daynix.com>
|
|
*
|
|
* This work is licensed under the terms of the GNU GPL, version 2 or later.
|
|
* See the COPYING file in the top-level directory.
|
|
*
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "qemu/crc32c.h"
|
|
#include "trace.h"
|
|
#include "net_rx_pkt.h"
|
|
#include "net/checksum.h"
|
|
#include "net/tap.h"
|
|
|
|
struct NetRxPkt {
|
|
struct virtio_net_hdr virt_hdr;
|
|
struct {
|
|
struct eth_header eth;
|
|
struct vlan_header vlan;
|
|
} ehdr_buf;
|
|
struct iovec *vec;
|
|
uint16_t vec_len_total;
|
|
uint16_t vec_len;
|
|
uint32_t tot_len;
|
|
uint16_t tci;
|
|
size_t ehdr_buf_len;
|
|
eth_pkt_types_e packet_type;
|
|
|
|
/* Analysis results */
|
|
bool hasip4;
|
|
bool hasip6;
|
|
|
|
size_t l3hdr_off;
|
|
size_t l4hdr_off;
|
|
size_t l5hdr_off;
|
|
|
|
eth_ip6_hdr_info ip6hdr_info;
|
|
eth_ip4_hdr_info ip4hdr_info;
|
|
eth_l4_hdr_info l4hdr_info;
|
|
};
|
|
|
|
void net_rx_pkt_init(struct NetRxPkt **pkt)
|
|
{
|
|
struct NetRxPkt *p = g_malloc0(sizeof *p);
|
|
p->vec = NULL;
|
|
p->vec_len_total = 0;
|
|
*pkt = p;
|
|
}
|
|
|
|
void net_rx_pkt_uninit(struct NetRxPkt *pkt)
|
|
{
|
|
if (pkt->vec_len_total != 0) {
|
|
g_free(pkt->vec);
|
|
}
|
|
|
|
g_free(pkt);
|
|
}
|
|
|
|
struct virtio_net_hdr *net_rx_pkt_get_vhdr(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
return &pkt->virt_hdr;
|
|
}
|
|
|
|
static inline void
|
|
net_rx_pkt_iovec_realloc(struct NetRxPkt *pkt,
|
|
int new_iov_len)
|
|
{
|
|
if (pkt->vec_len_total < new_iov_len) {
|
|
g_free(pkt->vec);
|
|
pkt->vec = g_malloc(sizeof(*pkt->vec) * new_iov_len);
|
|
pkt->vec_len_total = new_iov_len;
|
|
}
|
|
}
|
|
|
|
static void
|
|
net_rx_pkt_pull_data(struct NetRxPkt *pkt,
|
|
const struct iovec *iov, int iovcnt,
|
|
size_t ploff)
|
|
{
|
|
uint32_t pllen = iov_size(iov, iovcnt) - ploff;
|
|
|
|
if (pkt->ehdr_buf_len) {
|
|
net_rx_pkt_iovec_realloc(pkt, iovcnt + 1);
|
|
|
|
pkt->vec[0].iov_base = &pkt->ehdr_buf;
|
|
pkt->vec[0].iov_len = pkt->ehdr_buf_len;
|
|
|
|
pkt->tot_len = pllen + pkt->ehdr_buf_len;
|
|
pkt->vec_len = iov_copy(pkt->vec + 1, pkt->vec_len_total - 1,
|
|
iov, iovcnt, ploff, pllen) + 1;
|
|
} else {
|
|
net_rx_pkt_iovec_realloc(pkt, iovcnt);
|
|
|
|
pkt->tot_len = pllen;
|
|
pkt->vec_len = iov_copy(pkt->vec, pkt->vec_len_total,
|
|
iov, iovcnt, ploff, pkt->tot_len);
|
|
}
|
|
|
|
eth_get_protocols(pkt->vec, pkt->vec_len, 0, &pkt->hasip4, &pkt->hasip6,
|
|
&pkt->l3hdr_off, &pkt->l4hdr_off, &pkt->l5hdr_off,
|
|
&pkt->ip6hdr_info, &pkt->ip4hdr_info, &pkt->l4hdr_info);
|
|
|
|
trace_net_rx_pkt_parsed(pkt->hasip4, pkt->hasip6, pkt->l4hdr_info.proto,
|
|
pkt->l3hdr_off, pkt->l4hdr_off, pkt->l5hdr_off);
|
|
}
|
|
|
|
void net_rx_pkt_attach_iovec(struct NetRxPkt *pkt,
|
|
const struct iovec *iov, int iovcnt,
|
|
size_t iovoff, bool strip_vlan)
|
|
{
|
|
uint16_t tci = 0;
|
|
uint16_t ploff = iovoff;
|
|
assert(pkt);
|
|
|
|
if (strip_vlan) {
|
|
pkt->ehdr_buf_len = eth_strip_vlan(iov, iovcnt, iovoff, &pkt->ehdr_buf,
|
|
&ploff, &tci);
|
|
} else {
|
|
pkt->ehdr_buf_len = 0;
|
|
}
|
|
|
|
pkt->tci = tci;
|
|
|
|
net_rx_pkt_pull_data(pkt, iov, iovcnt, ploff);
|
|
}
|
|
|
|
void net_rx_pkt_attach_iovec_ex(struct NetRxPkt *pkt,
|
|
const struct iovec *iov, int iovcnt,
|
|
size_t iovoff, int strip_vlan_index,
|
|
uint16_t vet, uint16_t vet_ext)
|
|
{
|
|
uint16_t tci = 0;
|
|
uint16_t ploff = iovoff;
|
|
assert(pkt);
|
|
|
|
pkt->ehdr_buf_len = eth_strip_vlan_ex(iov, iovcnt, iovoff,
|
|
strip_vlan_index, vet, vet_ext,
|
|
&pkt->ehdr_buf,
|
|
&ploff, &tci);
|
|
|
|
pkt->tci = tci;
|
|
|
|
net_rx_pkt_pull_data(pkt, iov, iovcnt, ploff);
|
|
}
|
|
|
|
void net_rx_pkt_dump(struct NetRxPkt *pkt)
|
|
{
|
|
#ifdef NET_RX_PKT_DEBUG
|
|
assert(pkt);
|
|
|
|
printf("RX PKT: tot_len: %d, ehdr_buf_len: %lu, vlan_tag: %d\n",
|
|
pkt->tot_len, pkt->ehdr_buf_len, pkt->tci);
|
|
#endif
|
|
}
|
|
|
|
void net_rx_pkt_set_packet_type(struct NetRxPkt *pkt,
|
|
eth_pkt_types_e packet_type)
|
|
{
|
|
assert(pkt);
|
|
|
|
pkt->packet_type = packet_type;
|
|
|
|
}
|
|
|
|
eth_pkt_types_e net_rx_pkt_get_packet_type(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
|
|
return pkt->packet_type;
|
|
}
|
|
|
|
size_t net_rx_pkt_get_total_len(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
|
|
return pkt->tot_len;
|
|
}
|
|
|
|
void net_rx_pkt_set_protocols(struct NetRxPkt *pkt,
|
|
const struct iovec *iov, size_t iovcnt,
|
|
size_t iovoff)
|
|
{
|
|
assert(pkt);
|
|
|
|
eth_get_protocols(iov, iovcnt, iovoff, &pkt->hasip4, &pkt->hasip6,
|
|
&pkt->l3hdr_off, &pkt->l4hdr_off, &pkt->l5hdr_off,
|
|
&pkt->ip6hdr_info, &pkt->ip4hdr_info, &pkt->l4hdr_info);
|
|
}
|
|
|
|
void net_rx_pkt_get_protocols(struct NetRxPkt *pkt,
|
|
bool *hasip4, bool *hasip6,
|
|
EthL4HdrProto *l4hdr_proto)
|
|
{
|
|
assert(pkt);
|
|
|
|
*hasip4 = pkt->hasip4;
|
|
*hasip6 = pkt->hasip6;
|
|
*l4hdr_proto = pkt->l4hdr_info.proto;
|
|
}
|
|
|
|
size_t net_rx_pkt_get_l3_hdr_offset(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
return pkt->l3hdr_off;
|
|
}
|
|
|
|
size_t net_rx_pkt_get_l4_hdr_offset(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
return pkt->l4hdr_off;
|
|
}
|
|
|
|
size_t net_rx_pkt_get_l5_hdr_offset(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
return pkt->l5hdr_off;
|
|
}
|
|
|
|
eth_ip6_hdr_info *net_rx_pkt_get_ip6_info(struct NetRxPkt *pkt)
|
|
{
|
|
return &pkt->ip6hdr_info;
|
|
}
|
|
|
|
eth_ip4_hdr_info *net_rx_pkt_get_ip4_info(struct NetRxPkt *pkt)
|
|
{
|
|
return &pkt->ip4hdr_info;
|
|
}
|
|
|
|
static inline void
|
|
_net_rx_rss_add_chunk(uint8_t *rss_input, size_t *bytes_written,
|
|
void *ptr, size_t size)
|
|
{
|
|
memcpy(&rss_input[*bytes_written], ptr, size);
|
|
trace_net_rx_pkt_rss_add_chunk(ptr, size, *bytes_written);
|
|
*bytes_written += size;
|
|
}
|
|
|
|
static inline void
|
|
_net_rx_rss_prepare_ip4(uint8_t *rss_input,
|
|
struct NetRxPkt *pkt,
|
|
size_t *bytes_written)
|
|
{
|
|
struct ip_header *ip4_hdr = &pkt->ip4hdr_info.ip4_hdr;
|
|
|
|
_net_rx_rss_add_chunk(rss_input, bytes_written,
|
|
&ip4_hdr->ip_src, sizeof(uint32_t));
|
|
|
|
_net_rx_rss_add_chunk(rss_input, bytes_written,
|
|
&ip4_hdr->ip_dst, sizeof(uint32_t));
|
|
}
|
|
|
|
static inline void
|
|
_net_rx_rss_prepare_ip6(uint8_t *rss_input,
|
|
struct NetRxPkt *pkt,
|
|
bool ipv6ex, size_t *bytes_written)
|
|
{
|
|
eth_ip6_hdr_info *ip6info = &pkt->ip6hdr_info;
|
|
|
|
_net_rx_rss_add_chunk(rss_input, bytes_written,
|
|
(ipv6ex && ip6info->rss_ex_src_valid) ? &ip6info->rss_ex_src
|
|
: &ip6info->ip6_hdr.ip6_src,
|
|
sizeof(struct in6_address));
|
|
|
|
_net_rx_rss_add_chunk(rss_input, bytes_written,
|
|
(ipv6ex && ip6info->rss_ex_dst_valid) ? &ip6info->rss_ex_dst
|
|
: &ip6info->ip6_hdr.ip6_dst,
|
|
sizeof(struct in6_address));
|
|
}
|
|
|
|
static inline void
|
|
_net_rx_rss_prepare_tcp(uint8_t *rss_input,
|
|
struct NetRxPkt *pkt,
|
|
size_t *bytes_written)
|
|
{
|
|
struct tcp_header *tcphdr = &pkt->l4hdr_info.hdr.tcp;
|
|
|
|
_net_rx_rss_add_chunk(rss_input, bytes_written,
|
|
&tcphdr->th_sport, sizeof(uint16_t));
|
|
|
|
_net_rx_rss_add_chunk(rss_input, bytes_written,
|
|
&tcphdr->th_dport, sizeof(uint16_t));
|
|
}
|
|
|
|
static inline void
|
|
_net_rx_rss_prepare_udp(uint8_t *rss_input,
|
|
struct NetRxPkt *pkt,
|
|
size_t *bytes_written)
|
|
{
|
|
struct udp_header *udphdr = &pkt->l4hdr_info.hdr.udp;
|
|
|
|
_net_rx_rss_add_chunk(rss_input, bytes_written,
|
|
&udphdr->uh_sport, sizeof(uint16_t));
|
|
|
|
_net_rx_rss_add_chunk(rss_input, bytes_written,
|
|
&udphdr->uh_dport, sizeof(uint16_t));
|
|
}
|
|
|
|
uint32_t
|
|
net_rx_pkt_calc_rss_hash(struct NetRxPkt *pkt,
|
|
NetRxPktRssType type,
|
|
uint8_t *key)
|
|
{
|
|
uint8_t rss_input[36];
|
|
size_t rss_length = 0;
|
|
uint32_t rss_hash = 0;
|
|
net_toeplitz_key key_data;
|
|
|
|
switch (type) {
|
|
case NetPktRssIpV4:
|
|
assert(pkt->hasip4);
|
|
trace_net_rx_pkt_rss_ip4();
|
|
_net_rx_rss_prepare_ip4(&rss_input[0], pkt, &rss_length);
|
|
break;
|
|
case NetPktRssIpV4Tcp:
|
|
assert(pkt->hasip4);
|
|
assert(pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_TCP);
|
|
trace_net_rx_pkt_rss_ip4_tcp();
|
|
_net_rx_rss_prepare_ip4(&rss_input[0], pkt, &rss_length);
|
|
_net_rx_rss_prepare_tcp(&rss_input[0], pkt, &rss_length);
|
|
break;
|
|
case NetPktRssIpV6Tcp:
|
|
assert(pkt->hasip6);
|
|
assert(pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_TCP);
|
|
trace_net_rx_pkt_rss_ip6_tcp();
|
|
_net_rx_rss_prepare_ip6(&rss_input[0], pkt, false, &rss_length);
|
|
_net_rx_rss_prepare_tcp(&rss_input[0], pkt, &rss_length);
|
|
break;
|
|
case NetPktRssIpV6:
|
|
assert(pkt->hasip6);
|
|
trace_net_rx_pkt_rss_ip6();
|
|
_net_rx_rss_prepare_ip6(&rss_input[0], pkt, false, &rss_length);
|
|
break;
|
|
case NetPktRssIpV6Ex:
|
|
assert(pkt->hasip6);
|
|
trace_net_rx_pkt_rss_ip6_ex();
|
|
_net_rx_rss_prepare_ip6(&rss_input[0], pkt, true, &rss_length);
|
|
break;
|
|
case NetPktRssIpV6TcpEx:
|
|
assert(pkt->hasip6);
|
|
assert(pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_TCP);
|
|
trace_net_rx_pkt_rss_ip6_ex_tcp();
|
|
_net_rx_rss_prepare_ip6(&rss_input[0], pkt, true, &rss_length);
|
|
_net_rx_rss_prepare_tcp(&rss_input[0], pkt, &rss_length);
|
|
break;
|
|
case NetPktRssIpV4Udp:
|
|
assert(pkt->hasip4);
|
|
assert(pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_UDP);
|
|
trace_net_rx_pkt_rss_ip4_udp();
|
|
_net_rx_rss_prepare_ip4(&rss_input[0], pkt, &rss_length);
|
|
_net_rx_rss_prepare_udp(&rss_input[0], pkt, &rss_length);
|
|
break;
|
|
case NetPktRssIpV6Udp:
|
|
assert(pkt->hasip6);
|
|
assert(pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_UDP);
|
|
trace_net_rx_pkt_rss_ip6_udp();
|
|
_net_rx_rss_prepare_ip6(&rss_input[0], pkt, false, &rss_length);
|
|
_net_rx_rss_prepare_udp(&rss_input[0], pkt, &rss_length);
|
|
break;
|
|
case NetPktRssIpV6UdpEx:
|
|
assert(pkt->hasip6);
|
|
assert(pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_UDP);
|
|
trace_net_rx_pkt_rss_ip6_ex_udp();
|
|
_net_rx_rss_prepare_ip6(&rss_input[0], pkt, true, &rss_length);
|
|
_net_rx_rss_prepare_udp(&rss_input[0], pkt, &rss_length);
|
|
break;
|
|
default:
|
|
g_assert_not_reached();
|
|
}
|
|
|
|
net_toeplitz_key_init(&key_data, key);
|
|
net_toeplitz_add(&rss_hash, rss_input, rss_length, &key_data);
|
|
|
|
trace_net_rx_pkt_rss_hash(rss_length, rss_hash);
|
|
|
|
return rss_hash;
|
|
}
|
|
|
|
uint16_t net_rx_pkt_get_ip_id(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
|
|
if (pkt->hasip4) {
|
|
return be16_to_cpu(pkt->ip4hdr_info.ip4_hdr.ip_id);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
bool net_rx_pkt_is_tcp_ack(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
|
|
if (pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_TCP) {
|
|
return TCP_HEADER_FLAGS(&pkt->l4hdr_info.hdr.tcp) & TCP_FLAG_ACK;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool net_rx_pkt_has_tcp_data(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
|
|
if (pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_TCP) {
|
|
return pkt->l4hdr_info.has_tcp_data;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
struct iovec *net_rx_pkt_get_iovec(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
|
|
return pkt->vec;
|
|
}
|
|
|
|
uint16_t net_rx_pkt_get_iovec_len(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
|
|
return pkt->vec_len;
|
|
}
|
|
|
|
void net_rx_pkt_set_vhdr(struct NetRxPkt *pkt,
|
|
struct virtio_net_hdr *vhdr)
|
|
{
|
|
assert(pkt);
|
|
|
|
memcpy(&pkt->virt_hdr, vhdr, sizeof pkt->virt_hdr);
|
|
}
|
|
|
|
void net_rx_pkt_set_vhdr_iovec(struct NetRxPkt *pkt,
|
|
const struct iovec *iov, int iovcnt)
|
|
{
|
|
assert(pkt);
|
|
|
|
iov_to_buf(iov, iovcnt, 0, &pkt->virt_hdr, sizeof pkt->virt_hdr);
|
|
}
|
|
|
|
void net_rx_pkt_unset_vhdr(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
|
|
memset(&pkt->virt_hdr, 0, sizeof(pkt->virt_hdr));
|
|
}
|
|
|
|
bool net_rx_pkt_is_vlan_stripped(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
|
|
return pkt->ehdr_buf_len ? true : false;
|
|
}
|
|
|
|
uint16_t net_rx_pkt_get_vlan_tag(struct NetRxPkt *pkt)
|
|
{
|
|
assert(pkt);
|
|
|
|
return pkt->tci;
|
|
}
|
|
|
|
bool net_rx_pkt_validate_l3_csum(struct NetRxPkt *pkt, bool *csum_valid)
|
|
{
|
|
uint32_t cntr;
|
|
uint16_t csum;
|
|
uint32_t csl;
|
|
|
|
trace_net_rx_pkt_l3_csum_validate_entry();
|
|
|
|
if (!pkt->hasip4) {
|
|
trace_net_rx_pkt_l3_csum_validate_not_ip4();
|
|
return false;
|
|
}
|
|
|
|
csl = pkt->l4hdr_off - pkt->l3hdr_off;
|
|
|
|
cntr = net_checksum_add_iov(pkt->vec, pkt->vec_len,
|
|
pkt->l3hdr_off,
|
|
csl, 0);
|
|
|
|
csum = net_checksum_finish(cntr);
|
|
|
|
*csum_valid = (csum == 0);
|
|
|
|
trace_net_rx_pkt_l3_csum_validate_csum(pkt->l3hdr_off, csl,
|
|
cntr, csum, *csum_valid);
|
|
|
|
return true;
|
|
}
|
|
|
|
static uint16_t
|
|
_net_rx_pkt_calc_l4_csum(struct NetRxPkt *pkt)
|
|
{
|
|
uint32_t cntr;
|
|
uint16_t csum;
|
|
uint16_t csl;
|
|
uint32_t cso;
|
|
|
|
trace_net_rx_pkt_l4_csum_calc_entry();
|
|
|
|
if (pkt->hasip4) {
|
|
if (pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_UDP) {
|
|
csl = be16_to_cpu(pkt->l4hdr_info.hdr.udp.uh_ulen);
|
|
trace_net_rx_pkt_l4_csum_calc_ip4_udp();
|
|
} else {
|
|
csl = be16_to_cpu(pkt->ip4hdr_info.ip4_hdr.ip_len) -
|
|
IP_HDR_GET_LEN(&pkt->ip4hdr_info.ip4_hdr);
|
|
trace_net_rx_pkt_l4_csum_calc_ip4_tcp();
|
|
}
|
|
|
|
cntr = eth_calc_ip4_pseudo_hdr_csum(&pkt->ip4hdr_info.ip4_hdr,
|
|
csl, &cso);
|
|
trace_net_rx_pkt_l4_csum_calc_ph_csum(cntr, csl);
|
|
} else {
|
|
if (pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_UDP) {
|
|
csl = be16_to_cpu(pkt->l4hdr_info.hdr.udp.uh_ulen);
|
|
trace_net_rx_pkt_l4_csum_calc_ip6_udp();
|
|
} else {
|
|
struct ip6_header *ip6hdr = &pkt->ip6hdr_info.ip6_hdr;
|
|
size_t full_ip6hdr_len = pkt->l4hdr_off - pkt->l3hdr_off;
|
|
size_t ip6opts_len = full_ip6hdr_len - sizeof(struct ip6_header);
|
|
|
|
csl = be16_to_cpu(ip6hdr->ip6_ctlun.ip6_un1.ip6_un1_plen) -
|
|
ip6opts_len;
|
|
trace_net_rx_pkt_l4_csum_calc_ip6_tcp();
|
|
}
|
|
|
|
cntr = eth_calc_ip6_pseudo_hdr_csum(&pkt->ip6hdr_info.ip6_hdr, csl,
|
|
pkt->ip6hdr_info.l4proto, &cso);
|
|
trace_net_rx_pkt_l4_csum_calc_ph_csum(cntr, csl);
|
|
}
|
|
|
|
cntr += net_checksum_add_iov(pkt->vec, pkt->vec_len,
|
|
pkt->l4hdr_off, csl, cso);
|
|
|
|
csum = net_checksum_finish_nozero(cntr);
|
|
|
|
trace_net_rx_pkt_l4_csum_calc_csum(pkt->l4hdr_off, csl, cntr, csum);
|
|
|
|
return csum;
|
|
}
|
|
|
|
static bool
|
|
_net_rx_pkt_validate_sctp_sum(struct NetRxPkt *pkt)
|
|
{
|
|
size_t csum_off;
|
|
size_t off = pkt->l4hdr_off;
|
|
size_t vec_len = pkt->vec_len;
|
|
struct iovec *vec;
|
|
uint32_t calculated = 0;
|
|
uint32_t original;
|
|
bool valid;
|
|
|
|
for (vec = pkt->vec; vec->iov_len < off; vec++) {
|
|
off -= vec->iov_len;
|
|
vec_len--;
|
|
}
|
|
|
|
csum_off = off + 8;
|
|
|
|
if (!iov_to_buf(vec, vec_len, csum_off, &original, sizeof(original))) {
|
|
return false;
|
|
}
|
|
|
|
if (!iov_from_buf(vec, vec_len, csum_off,
|
|
&calculated, sizeof(calculated))) {
|
|
return false;
|
|
}
|
|
|
|
calculated = crc32c(0xffffffff,
|
|
(uint8_t *)vec->iov_base + off, vec->iov_len - off);
|
|
calculated = iov_crc32c(calculated ^ 0xffffffff, vec + 1, vec_len - 1);
|
|
valid = calculated == le32_to_cpu(original);
|
|
iov_from_buf(vec, vec_len, csum_off, &original, sizeof(original));
|
|
|
|
return valid;
|
|
}
|
|
|
|
bool net_rx_pkt_validate_l4_csum(struct NetRxPkt *pkt, bool *csum_valid)
|
|
{
|
|
uint32_t csum;
|
|
|
|
trace_net_rx_pkt_l4_csum_validate_entry();
|
|
|
|
if (pkt->hasip4 && pkt->ip4hdr_info.fragment) {
|
|
trace_net_rx_pkt_l4_csum_validate_ip4_fragment();
|
|
return false;
|
|
}
|
|
|
|
switch (pkt->l4hdr_info.proto) {
|
|
case ETH_L4_HDR_PROTO_UDP:
|
|
if (pkt->l4hdr_info.hdr.udp.uh_sum == 0) {
|
|
trace_net_rx_pkt_l4_csum_validate_udp_with_no_checksum();
|
|
return false;
|
|
}
|
|
/* fall through */
|
|
case ETH_L4_HDR_PROTO_TCP:
|
|
csum = _net_rx_pkt_calc_l4_csum(pkt);
|
|
*csum_valid = ((csum == 0) || (csum == 0xFFFF));
|
|
break;
|
|
|
|
case ETH_L4_HDR_PROTO_SCTP:
|
|
*csum_valid = _net_rx_pkt_validate_sctp_sum(pkt);
|
|
break;
|
|
|
|
default:
|
|
trace_net_rx_pkt_l4_csum_validate_not_xxp();
|
|
return false;
|
|
}
|
|
|
|
trace_net_rx_pkt_l4_csum_validate_csum(*csum_valid);
|
|
|
|
return true;
|
|
}
|
|
|
|
bool net_rx_pkt_fix_l4_csum(struct NetRxPkt *pkt)
|
|
{
|
|
uint16_t csum = 0;
|
|
uint32_t l4_cso;
|
|
|
|
trace_net_rx_pkt_l4_csum_fix_entry();
|
|
|
|
switch (pkt->l4hdr_info.proto) {
|
|
case ETH_L4_HDR_PROTO_TCP:
|
|
l4_cso = offsetof(struct tcp_header, th_sum);
|
|
trace_net_rx_pkt_l4_csum_fix_tcp(l4_cso);
|
|
break;
|
|
|
|
case ETH_L4_HDR_PROTO_UDP:
|
|
if (pkt->l4hdr_info.hdr.udp.uh_sum == 0) {
|
|
trace_net_rx_pkt_l4_csum_fix_udp_with_no_checksum();
|
|
return false;
|
|
}
|
|
l4_cso = offsetof(struct udp_header, uh_sum);
|
|
trace_net_rx_pkt_l4_csum_fix_udp(l4_cso);
|
|
break;
|
|
|
|
default:
|
|
trace_net_rx_pkt_l4_csum_fix_not_xxp();
|
|
return false;
|
|
}
|
|
|
|
if (pkt->hasip4 && pkt->ip4hdr_info.fragment) {
|
|
trace_net_rx_pkt_l4_csum_fix_ip4_fragment();
|
|
return false;
|
|
}
|
|
|
|
/* Set zero to checksum word */
|
|
iov_from_buf(pkt->vec, pkt->vec_len,
|
|
pkt->l4hdr_off + l4_cso,
|
|
&csum, sizeof(csum));
|
|
|
|
/* Calculate L4 checksum */
|
|
csum = cpu_to_be16(_net_rx_pkt_calc_l4_csum(pkt));
|
|
|
|
/* Set calculated checksum to checksum word */
|
|
iov_from_buf(pkt->vec, pkt->vec_len,
|
|
pkt->l4hdr_off + l4_cso,
|
|
&csum, sizeof(csum));
|
|
|
|
trace_net_rx_pkt_l4_csum_fix_csum(pkt->l4hdr_off + l4_cso, csum);
|
|
|
|
return true;
|
|
}
|