mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
synced 2024-12-28 16:52:18 +00:00
6dd514f481
In PFCP receive path set metadata needed by flower code to do correct classification based on this metadata. Signed-off-by: Michal Swiatkowski <michal.swiatkowski@linux.intel.com> Signed-off-by: Marcin Szycik <marcin.szycik@linux.intel.com> Reviewed-by: Simon Horman <horms@kernel.org> Signed-off-by: Alexander Lobakin <aleksander.lobakin@intel.com> Signed-off-by: David S. Miller <davem@davemloft.net>
388 lines
9.9 KiB
C
388 lines
9.9 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
|
|
#include <kunit/test.h>
|
|
|
|
/* GSO */
|
|
|
|
#include <linux/skbuff.h>
|
|
|
|
static const char hdr[] = "abcdefgh";
|
|
#define GSO_TEST_SIZE 1000
|
|
|
|
static void __init_skb(struct sk_buff *skb)
|
|
{
|
|
skb_reset_mac_header(skb);
|
|
memcpy(skb_mac_header(skb), hdr, sizeof(hdr));
|
|
|
|
/* skb_segment expects skb->data at start of payload */
|
|
skb_pull(skb, sizeof(hdr));
|
|
skb_reset_network_header(skb);
|
|
skb_reset_transport_header(skb);
|
|
|
|
/* proto is arbitrary, as long as not ETH_P_TEB or vlan */
|
|
skb->protocol = htons(ETH_P_ATALK);
|
|
skb_shinfo(skb)->gso_size = GSO_TEST_SIZE;
|
|
}
|
|
|
|
enum gso_test_nr {
|
|
GSO_TEST_LINEAR,
|
|
GSO_TEST_NO_GSO,
|
|
GSO_TEST_FRAGS,
|
|
GSO_TEST_FRAGS_PURE,
|
|
GSO_TEST_GSO_PARTIAL,
|
|
GSO_TEST_FRAG_LIST,
|
|
GSO_TEST_FRAG_LIST_PURE,
|
|
GSO_TEST_FRAG_LIST_NON_UNIFORM,
|
|
GSO_TEST_GSO_BY_FRAGS,
|
|
};
|
|
|
|
struct gso_test_case {
|
|
enum gso_test_nr id;
|
|
const char *name;
|
|
|
|
/* input */
|
|
unsigned int linear_len;
|
|
unsigned int nr_frags;
|
|
const unsigned int *frags;
|
|
unsigned int nr_frag_skbs;
|
|
const unsigned int *frag_skbs;
|
|
|
|
/* output as expected */
|
|
unsigned int nr_segs;
|
|
const unsigned int *segs;
|
|
};
|
|
|
|
static struct gso_test_case cases[] = {
|
|
{
|
|
.id = GSO_TEST_NO_GSO,
|
|
.name = "no_gso",
|
|
.linear_len = GSO_TEST_SIZE,
|
|
.nr_segs = 1,
|
|
.segs = (const unsigned int[]) { GSO_TEST_SIZE },
|
|
},
|
|
{
|
|
.id = GSO_TEST_LINEAR,
|
|
.name = "linear",
|
|
.linear_len = GSO_TEST_SIZE + GSO_TEST_SIZE + 1,
|
|
.nr_segs = 3,
|
|
.segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 1 },
|
|
},
|
|
{
|
|
.id = GSO_TEST_FRAGS,
|
|
.name = "frags",
|
|
.linear_len = GSO_TEST_SIZE,
|
|
.nr_frags = 2,
|
|
.frags = (const unsigned int[]) { GSO_TEST_SIZE, 1 },
|
|
.nr_segs = 3,
|
|
.segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 1 },
|
|
},
|
|
{
|
|
.id = GSO_TEST_FRAGS_PURE,
|
|
.name = "frags_pure",
|
|
.nr_frags = 3,
|
|
.frags = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 2 },
|
|
.nr_segs = 3,
|
|
.segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 2 },
|
|
},
|
|
{
|
|
.id = GSO_TEST_GSO_PARTIAL,
|
|
.name = "gso_partial",
|
|
.linear_len = GSO_TEST_SIZE,
|
|
.nr_frags = 2,
|
|
.frags = (const unsigned int[]) { GSO_TEST_SIZE, 3 },
|
|
.nr_segs = 2,
|
|
.segs = (const unsigned int[]) { 2 * GSO_TEST_SIZE, 3 },
|
|
},
|
|
{
|
|
/* commit 89319d3801d1: frag_list on mss boundaries */
|
|
.id = GSO_TEST_FRAG_LIST,
|
|
.name = "frag_list",
|
|
.linear_len = GSO_TEST_SIZE,
|
|
.nr_frag_skbs = 2,
|
|
.frag_skbs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE },
|
|
.nr_segs = 3,
|
|
.segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, GSO_TEST_SIZE },
|
|
},
|
|
{
|
|
.id = GSO_TEST_FRAG_LIST_PURE,
|
|
.name = "frag_list_pure",
|
|
.nr_frag_skbs = 2,
|
|
.frag_skbs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE },
|
|
.nr_segs = 2,
|
|
.segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE },
|
|
},
|
|
{
|
|
/* commit 43170c4e0ba7: GRO of frag_list trains */
|
|
.id = GSO_TEST_FRAG_LIST_NON_UNIFORM,
|
|
.name = "frag_list_non_uniform",
|
|
.linear_len = GSO_TEST_SIZE,
|
|
.nr_frag_skbs = 4,
|
|
.frag_skbs = (const unsigned int[]) { GSO_TEST_SIZE, 1, GSO_TEST_SIZE, 2 },
|
|
.nr_segs = 4,
|
|
.segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, GSO_TEST_SIZE, 3 },
|
|
},
|
|
{
|
|
/* commit 3953c46c3ac7 ("sk_buff: allow segmenting based on frag sizes") and
|
|
* commit 90017accff61 ("sctp: Add GSO support")
|
|
*
|
|
* "there will be a cover skb with protocol headers and
|
|
* children ones containing the actual segments"
|
|
*/
|
|
.id = GSO_TEST_GSO_BY_FRAGS,
|
|
.name = "gso_by_frags",
|
|
.nr_frag_skbs = 4,
|
|
.frag_skbs = (const unsigned int[]) { 100, 200, 300, 400 },
|
|
.nr_segs = 4,
|
|
.segs = (const unsigned int[]) { 100, 200, 300, 400 },
|
|
},
|
|
};
|
|
|
|
static void gso_test_case_to_desc(struct gso_test_case *t, char *desc)
|
|
{
|
|
sprintf(desc, "%s", t->name);
|
|
}
|
|
|
|
KUNIT_ARRAY_PARAM(gso_test, cases, gso_test_case_to_desc);
|
|
|
|
static void gso_test_func(struct kunit *test)
|
|
{
|
|
const int shinfo_size = SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
|
|
struct sk_buff *skb, *segs, *cur, *next, *last;
|
|
const struct gso_test_case *tcase;
|
|
netdev_features_t features;
|
|
struct page *page;
|
|
int i;
|
|
|
|
tcase = test->param_value;
|
|
|
|
page = alloc_page(GFP_KERNEL);
|
|
KUNIT_ASSERT_NOT_NULL(test, page);
|
|
skb = build_skb(page_address(page), sizeof(hdr) + tcase->linear_len + shinfo_size);
|
|
KUNIT_ASSERT_NOT_NULL(test, skb);
|
|
__skb_put(skb, sizeof(hdr) + tcase->linear_len);
|
|
|
|
__init_skb(skb);
|
|
|
|
if (tcase->nr_frags) {
|
|
unsigned int pg_off = 0;
|
|
|
|
page = alloc_page(GFP_KERNEL);
|
|
KUNIT_ASSERT_NOT_NULL(test, page);
|
|
page_ref_add(page, tcase->nr_frags - 1);
|
|
|
|
for (i = 0; i < tcase->nr_frags; i++) {
|
|
skb_fill_page_desc(skb, i, page, pg_off, tcase->frags[i]);
|
|
pg_off += tcase->frags[i];
|
|
}
|
|
|
|
KUNIT_ASSERT_LE(test, pg_off, PAGE_SIZE);
|
|
|
|
skb->data_len = pg_off;
|
|
skb->len += skb->data_len;
|
|
skb->truesize += skb->data_len;
|
|
}
|
|
|
|
if (tcase->frag_skbs) {
|
|
unsigned int total_size = 0, total_true_size = 0;
|
|
struct sk_buff *frag_skb, *prev = NULL;
|
|
|
|
for (i = 0; i < tcase->nr_frag_skbs; i++) {
|
|
unsigned int frag_size;
|
|
|
|
page = alloc_page(GFP_KERNEL);
|
|
KUNIT_ASSERT_NOT_NULL(test, page);
|
|
|
|
frag_size = tcase->frag_skbs[i];
|
|
frag_skb = build_skb(page_address(page),
|
|
frag_size + shinfo_size);
|
|
KUNIT_ASSERT_NOT_NULL(test, frag_skb);
|
|
__skb_put(frag_skb, frag_size);
|
|
|
|
if (prev)
|
|
prev->next = frag_skb;
|
|
else
|
|
skb_shinfo(skb)->frag_list = frag_skb;
|
|
prev = frag_skb;
|
|
|
|
total_size += frag_size;
|
|
total_true_size += frag_skb->truesize;
|
|
}
|
|
|
|
skb->len += total_size;
|
|
skb->data_len += total_size;
|
|
skb->truesize += total_true_size;
|
|
|
|
if (tcase->id == GSO_TEST_GSO_BY_FRAGS)
|
|
skb_shinfo(skb)->gso_size = GSO_BY_FRAGS;
|
|
}
|
|
|
|
features = NETIF_F_SG | NETIF_F_HW_CSUM;
|
|
if (tcase->id == GSO_TEST_GSO_PARTIAL)
|
|
features |= NETIF_F_GSO_PARTIAL;
|
|
|
|
/* TODO: this should also work with SG,
|
|
* rather than hit BUG_ON(i >= nfrags)
|
|
*/
|
|
if (tcase->id == GSO_TEST_FRAG_LIST_NON_UNIFORM)
|
|
features &= ~NETIF_F_SG;
|
|
|
|
segs = skb_segment(skb, features);
|
|
if (IS_ERR(segs)) {
|
|
KUNIT_FAIL(test, "segs error %pe", segs);
|
|
goto free_gso_skb;
|
|
} else if (!segs) {
|
|
KUNIT_FAIL(test, "no segments");
|
|
goto free_gso_skb;
|
|
}
|
|
|
|
last = segs->prev;
|
|
for (cur = segs, i = 0; cur; cur = next, i++) {
|
|
next = cur->next;
|
|
|
|
KUNIT_ASSERT_EQ(test, cur->len, sizeof(hdr) + tcase->segs[i]);
|
|
|
|
/* segs have skb->data pointing to the mac header */
|
|
KUNIT_ASSERT_PTR_EQ(test, skb_mac_header(cur), cur->data);
|
|
KUNIT_ASSERT_PTR_EQ(test, skb_network_header(cur), cur->data + sizeof(hdr));
|
|
|
|
/* header was copied to all segs */
|
|
KUNIT_ASSERT_EQ(test, memcmp(skb_mac_header(cur), hdr, sizeof(hdr)), 0);
|
|
|
|
/* last seg can be found through segs->prev pointer */
|
|
if (!next)
|
|
KUNIT_ASSERT_PTR_EQ(test, cur, last);
|
|
|
|
consume_skb(cur);
|
|
}
|
|
|
|
KUNIT_ASSERT_EQ(test, i, tcase->nr_segs);
|
|
|
|
free_gso_skb:
|
|
consume_skb(skb);
|
|
}
|
|
|
|
/* IP tunnel flags */
|
|
|
|
#include <net/ip_tunnels.h>
|
|
|
|
struct ip_tunnel_flags_test {
|
|
const char *name;
|
|
|
|
const u16 *src_bits;
|
|
const u16 *exp_bits;
|
|
u8 src_num;
|
|
u8 exp_num;
|
|
|
|
__be16 exp_val;
|
|
bool exp_comp;
|
|
};
|
|
|
|
#define IP_TUNNEL_FLAGS_TEST(n, src, comp, eval, exp) { \
|
|
.name = (n), \
|
|
.src_bits = (src), \
|
|
.src_num = ARRAY_SIZE(src), \
|
|
.exp_comp = (comp), \
|
|
.exp_val = (eval), \
|
|
.exp_bits = (exp), \
|
|
.exp_num = ARRAY_SIZE(exp), \
|
|
}
|
|
|
|
/* These are __be16-compatible and can be compared as is */
|
|
static const u16 ip_tunnel_flags_1[] = {
|
|
IP_TUNNEL_KEY_BIT,
|
|
IP_TUNNEL_STRICT_BIT,
|
|
IP_TUNNEL_ERSPAN_OPT_BIT,
|
|
};
|
|
|
|
/* Due to the previous flags design limitation, setting either
|
|
* ``IP_TUNNEL_CSUM_BIT`` (on Big Endian) or ``IP_TUNNEL_DONT_FRAGMENT_BIT``
|
|
* (on Little) also sets VTI/ISATAP bit. In the bitmap implementation, they
|
|
* correspond to ``BIT(16)``, which is bigger than ``U16_MAX``, but still is
|
|
* backward-compatible.
|
|
*/
|
|
#ifdef __LITTLE_ENDIAN
|
|
#define IP_TUNNEL_CONFLICT_BIT IP_TUNNEL_DONT_FRAGMENT_BIT
|
|
#else
|
|
#define IP_TUNNEL_CONFLICT_BIT IP_TUNNEL_CSUM_BIT
|
|
#endif
|
|
|
|
static const u16 ip_tunnel_flags_2_src[] = {
|
|
IP_TUNNEL_CONFLICT_BIT,
|
|
};
|
|
|
|
static const u16 ip_tunnel_flags_2_exp[] = {
|
|
IP_TUNNEL_CONFLICT_BIT,
|
|
IP_TUNNEL_SIT_ISATAP_BIT,
|
|
};
|
|
|
|
/* Bits 17 and higher are not compatible with __be16 flags */
|
|
static const u16 ip_tunnel_flags_3_src[] = {
|
|
IP_TUNNEL_VXLAN_OPT_BIT,
|
|
17,
|
|
18,
|
|
20,
|
|
};
|
|
|
|
static const u16 ip_tunnel_flags_3_exp[] = {
|
|
IP_TUNNEL_VXLAN_OPT_BIT,
|
|
};
|
|
|
|
static const struct ip_tunnel_flags_test ip_tunnel_flags_test[] = {
|
|
IP_TUNNEL_FLAGS_TEST("compat", ip_tunnel_flags_1, true,
|
|
cpu_to_be16(BIT(IP_TUNNEL_KEY_BIT) |
|
|
BIT(IP_TUNNEL_STRICT_BIT) |
|
|
BIT(IP_TUNNEL_ERSPAN_OPT_BIT)),
|
|
ip_tunnel_flags_1),
|
|
IP_TUNNEL_FLAGS_TEST("conflict", ip_tunnel_flags_2_src, true,
|
|
VTI_ISVTI, ip_tunnel_flags_2_exp),
|
|
IP_TUNNEL_FLAGS_TEST("new", ip_tunnel_flags_3_src, false,
|
|
cpu_to_be16(BIT(IP_TUNNEL_VXLAN_OPT_BIT)),
|
|
ip_tunnel_flags_3_exp),
|
|
};
|
|
|
|
static void
|
|
ip_tunnel_flags_test_case_to_desc(const struct ip_tunnel_flags_test *t,
|
|
char *desc)
|
|
{
|
|
strscpy(desc, t->name, KUNIT_PARAM_DESC_SIZE);
|
|
}
|
|
KUNIT_ARRAY_PARAM(ip_tunnel_flags_test, ip_tunnel_flags_test,
|
|
ip_tunnel_flags_test_case_to_desc);
|
|
|
|
static void ip_tunnel_flags_test_run(struct kunit *test)
|
|
{
|
|
const struct ip_tunnel_flags_test *t = test->param_value;
|
|
IP_TUNNEL_DECLARE_FLAGS(src) = { };
|
|
IP_TUNNEL_DECLARE_FLAGS(exp) = { };
|
|
IP_TUNNEL_DECLARE_FLAGS(out);
|
|
|
|
for (u32 j = 0; j < t->src_num; j++)
|
|
__set_bit(t->src_bits[j], src);
|
|
for (u32 j = 0; j < t->exp_num; j++)
|
|
__set_bit(t->exp_bits[j], exp);
|
|
|
|
KUNIT_ASSERT_EQ(test, t->exp_comp,
|
|
ip_tunnel_flags_is_be16_compat(src));
|
|
KUNIT_ASSERT_EQ(test, (__force u16)t->exp_val,
|
|
(__force u16)ip_tunnel_flags_to_be16(src));
|
|
|
|
ip_tunnel_flags_from_be16(out, t->exp_val);
|
|
KUNIT_ASSERT_TRUE(test, __ipt_flag_op(bitmap_equal, exp, out));
|
|
}
|
|
|
|
static struct kunit_case net_test_cases[] = {
|
|
KUNIT_CASE_PARAM(gso_test_func, gso_test_gen_params),
|
|
KUNIT_CASE_PARAM(ip_tunnel_flags_test_run,
|
|
ip_tunnel_flags_test_gen_params),
|
|
{ },
|
|
};
|
|
|
|
static struct kunit_suite net_test_suite = {
|
|
.name = "net_core",
|
|
.test_cases = net_test_cases,
|
|
};
|
|
kunit_test_suite(net_test_suite);
|
|
|
|
MODULE_DESCRIPTION("KUnit tests for networking core");
|
|
MODULE_LICENSE("GPL");
|