| .. | .. |
|---|
| 1 | | -/* |
|---|
| 2 | | - * Copyright (C) 2017 Netronome Systems, Inc. |
|---|
| 3 | | - * |
|---|
| 4 | | - * This software is dual licensed under the GNU General License Version 2, |
|---|
| 5 | | - * June 1991 as shown in the file COPYING in the top-level directory of this |
|---|
| 6 | | - * source tree or the BSD 2-Clause License provided below. You have the |
|---|
| 7 | | - * option to license this software under the complete terms of either license. |
|---|
| 8 | | - * |
|---|
| 9 | | - * The BSD 2-Clause License: |
|---|
| 10 | | - * |
|---|
| 11 | | - * Redistribution and use in source and binary forms, with or |
|---|
| 12 | | - * without modification, are permitted provided that the following |
|---|
| 13 | | - * conditions are met: |
|---|
| 14 | | - * |
|---|
| 15 | | - * 1. Redistributions of source code must retain the above |
|---|
| 16 | | - * copyright notice, this list of conditions and the following |
|---|
| 17 | | - * disclaimer. |
|---|
| 18 | | - * |
|---|
| 19 | | - * 2. Redistributions in binary form must reproduce the above |
|---|
| 20 | | - * copyright notice, this list of conditions and the following |
|---|
| 21 | | - * disclaimer in the documentation and/or other materials |
|---|
| 22 | | - * provided with the distribution. |
|---|
| 23 | | - * |
|---|
| 24 | | - * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, |
|---|
| 25 | | - * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
|---|
| 26 | | - * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND |
|---|
| 27 | | - * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS |
|---|
| 28 | | - * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN |
|---|
| 29 | | - * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN |
|---|
| 30 | | - * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
|---|
| 31 | | - * SOFTWARE. |
|---|
| 32 | | - */ |
|---|
| 1 | +// SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause) |
|---|
| 2 | +/* Copyright (C) 2017-2018 Netronome Systems, Inc. */ |
|---|
| 33 | 3 | |
|---|
| 34 | 4 | #include <linux/bitfield.h> |
|---|
| 35 | 5 | #include <net/pkt_cls.h> |
|---|
| .. | .. |
|---|
| 38 | 8 | #include "main.h" |
|---|
| 39 | 9 | |
|---|
| 40 | 10 | static void |
|---|
| 41 | | -nfp_flower_compile_meta_tci(struct nfp_flower_meta_tci *frame, |
|---|
| 42 | | - struct tc_cls_flower_offload *flow, u8 key_type, |
|---|
| 43 | | - bool mask_version) |
|---|
| 11 | +nfp_flower_compile_meta_tci(struct nfp_flower_meta_tci *ext, |
|---|
| 12 | + struct nfp_flower_meta_tci *msk, |
|---|
| 13 | + struct flow_rule *rule, u8 key_type, bool qinq_sup) |
|---|
| 44 | 14 | { |
|---|
| 45 | | - struct fl_flow_key *target = mask_version ? flow->mask : flow->key; |
|---|
| 46 | | - struct flow_dissector_key_vlan *flow_vlan; |
|---|
| 47 | 15 | u16 tmp_tci; |
|---|
| 48 | 16 | |
|---|
| 49 | | - memset(frame, 0, sizeof(struct nfp_flower_meta_tci)); |
|---|
| 50 | | - /* Populate the metadata frame. */ |
|---|
| 51 | | - frame->nfp_flow_key_layer = key_type; |
|---|
| 52 | | - frame->mask_id = ~0; |
|---|
| 17 | + memset(ext, 0, sizeof(struct nfp_flower_meta_tci)); |
|---|
| 18 | + memset(msk, 0, sizeof(struct nfp_flower_meta_tci)); |
|---|
| 53 | 19 | |
|---|
| 54 | | - if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_VLAN)) { |
|---|
| 55 | | - flow_vlan = skb_flow_dissector_target(flow->dissector, |
|---|
| 56 | | - FLOW_DISSECTOR_KEY_VLAN, |
|---|
| 57 | | - target); |
|---|
| 20 | + /* Populate the metadata frame. */ |
|---|
| 21 | + ext->nfp_flow_key_layer = key_type; |
|---|
| 22 | + ext->mask_id = ~0; |
|---|
| 23 | + |
|---|
| 24 | + msk->nfp_flow_key_layer = key_type; |
|---|
| 25 | + msk->mask_id = ~0; |
|---|
| 26 | + |
|---|
| 27 | + if (!qinq_sup && flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_VLAN)) { |
|---|
| 28 | + struct flow_match_vlan match; |
|---|
| 29 | + |
|---|
| 30 | + flow_rule_match_vlan(rule, &match); |
|---|
| 58 | 31 | /* Populate the tci field. */ |
|---|
| 59 | 32 | tmp_tci = NFP_FLOWER_MASK_VLAN_PRESENT; |
|---|
| 60 | 33 | tmp_tci |= FIELD_PREP(NFP_FLOWER_MASK_VLAN_PRIO, |
|---|
| 61 | | - flow_vlan->vlan_priority) | |
|---|
| 34 | + match.key->vlan_priority) | |
|---|
| 62 | 35 | FIELD_PREP(NFP_FLOWER_MASK_VLAN_VID, |
|---|
| 63 | | - flow_vlan->vlan_id); |
|---|
| 64 | | - frame->tci = cpu_to_be16(tmp_tci); |
|---|
| 36 | + match.key->vlan_id); |
|---|
| 37 | + ext->tci = cpu_to_be16(tmp_tci); |
|---|
| 38 | + |
|---|
| 39 | + tmp_tci = NFP_FLOWER_MASK_VLAN_PRESENT; |
|---|
| 40 | + tmp_tci |= FIELD_PREP(NFP_FLOWER_MASK_VLAN_PRIO, |
|---|
| 41 | + match.mask->vlan_priority) | |
|---|
| 42 | + FIELD_PREP(NFP_FLOWER_MASK_VLAN_VID, |
|---|
| 43 | + match.mask->vlan_id); |
|---|
| 44 | + msk->tci = cpu_to_be16(tmp_tci); |
|---|
| 65 | 45 | } |
|---|
| 66 | 46 | } |
|---|
| 67 | 47 | |
|---|
| .. | .. |
|---|
| 73 | 53 | |
|---|
| 74 | 54 | static int |
|---|
| 75 | 55 | nfp_flower_compile_port(struct nfp_flower_in_port *frame, u32 cmsg_port, |
|---|
| 76 | | - bool mask_version, enum nfp_flower_tun_type tun_type) |
|---|
| 56 | + bool mask_version, enum nfp_flower_tun_type tun_type, |
|---|
| 57 | + struct netlink_ext_ack *extack) |
|---|
| 77 | 58 | { |
|---|
| 78 | 59 | if (mask_version) { |
|---|
| 79 | 60 | frame->in_port = cpu_to_be32(~0); |
|---|
| 80 | 61 | return 0; |
|---|
| 81 | 62 | } |
|---|
| 82 | 63 | |
|---|
| 83 | | - if (tun_type) |
|---|
| 64 | + if (tun_type) { |
|---|
| 84 | 65 | frame->in_port = cpu_to_be32(NFP_FL_PORT_TYPE_TUN | tun_type); |
|---|
| 85 | | - else |
|---|
| 66 | + } else { |
|---|
| 67 | + if (!cmsg_port) { |
|---|
| 68 | + NL_SET_ERR_MSG_MOD(extack, "unsupported offload: invalid ingress interface for match offload"); |
|---|
| 69 | + return -EOPNOTSUPP; |
|---|
| 70 | + } |
|---|
| 86 | 71 | frame->in_port = cpu_to_be32(cmsg_port); |
|---|
| 72 | + } |
|---|
| 87 | 73 | |
|---|
| 88 | 74 | return 0; |
|---|
| 89 | 75 | } |
|---|
| 90 | 76 | |
|---|
| 91 | | -static void |
|---|
| 92 | | -nfp_flower_compile_mac(struct nfp_flower_mac_mpls *frame, |
|---|
| 93 | | - struct tc_cls_flower_offload *flow, |
|---|
| 94 | | - bool mask_version) |
|---|
| 77 | +static int |
|---|
| 78 | +nfp_flower_compile_mac(struct nfp_flower_mac_mpls *ext, |
|---|
| 79 | + struct nfp_flower_mac_mpls *msk, struct flow_rule *rule, |
|---|
| 80 | + struct netlink_ext_ack *extack) |
|---|
| 95 | 81 | { |
|---|
| 96 | | - struct fl_flow_key *target = mask_version ? flow->mask : flow->key; |
|---|
| 97 | | - struct flow_dissector_key_eth_addrs *addr; |
|---|
| 82 | + memset(ext, 0, sizeof(struct nfp_flower_mac_mpls)); |
|---|
| 83 | + memset(msk, 0, sizeof(struct nfp_flower_mac_mpls)); |
|---|
| 98 | 84 | |
|---|
| 99 | | - memset(frame, 0, sizeof(struct nfp_flower_mac_mpls)); |
|---|
| 85 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_ETH_ADDRS)) { |
|---|
| 86 | + struct flow_match_eth_addrs match; |
|---|
| 100 | 87 | |
|---|
| 101 | | - if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_ETH_ADDRS)) { |
|---|
| 102 | | - addr = skb_flow_dissector_target(flow->dissector, |
|---|
| 103 | | - FLOW_DISSECTOR_KEY_ETH_ADDRS, |
|---|
| 104 | | - target); |
|---|
| 88 | + flow_rule_match_eth_addrs(rule, &match); |
|---|
| 105 | 89 | /* Populate mac frame. */ |
|---|
| 106 | | - ether_addr_copy(frame->mac_dst, &addr->dst[0]); |
|---|
| 107 | | - ether_addr_copy(frame->mac_src, &addr->src[0]); |
|---|
| 90 | + ether_addr_copy(ext->mac_dst, &match.key->dst[0]); |
|---|
| 91 | + ether_addr_copy(ext->mac_src, &match.key->src[0]); |
|---|
| 92 | + ether_addr_copy(msk->mac_dst, &match.mask->dst[0]); |
|---|
| 93 | + ether_addr_copy(msk->mac_src, &match.mask->src[0]); |
|---|
| 108 | 94 | } |
|---|
| 109 | 95 | |
|---|
| 110 | | - if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_MPLS)) { |
|---|
| 111 | | - struct flow_dissector_key_mpls *mpls; |
|---|
| 96 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_MPLS)) { |
|---|
| 97 | + struct flow_match_mpls match; |
|---|
| 112 | 98 | u32 t_mpls; |
|---|
| 113 | 99 | |
|---|
| 114 | | - mpls = skb_flow_dissector_target(flow->dissector, |
|---|
| 115 | | - FLOW_DISSECTOR_KEY_MPLS, |
|---|
| 116 | | - target); |
|---|
| 100 | + flow_rule_match_mpls(rule, &match); |
|---|
| 117 | 101 | |
|---|
| 118 | | - t_mpls = FIELD_PREP(NFP_FLOWER_MASK_MPLS_LB, mpls->mpls_label) | |
|---|
| 119 | | - FIELD_PREP(NFP_FLOWER_MASK_MPLS_TC, mpls->mpls_tc) | |
|---|
| 120 | | - FIELD_PREP(NFP_FLOWER_MASK_MPLS_BOS, mpls->mpls_bos) | |
|---|
| 102 | + /* Only support matching the first LSE */ |
|---|
| 103 | + if (match.mask->used_lses != 1) { |
|---|
| 104 | + NL_SET_ERR_MSG_MOD(extack, |
|---|
| 105 | + "unsupported offload: invalid LSE depth for MPLS match offload"); |
|---|
| 106 | + return -EOPNOTSUPP; |
|---|
| 107 | + } |
|---|
| 108 | + |
|---|
| 109 | + t_mpls = FIELD_PREP(NFP_FLOWER_MASK_MPLS_LB, |
|---|
| 110 | + match.key->ls[0].mpls_label) | |
|---|
| 111 | + FIELD_PREP(NFP_FLOWER_MASK_MPLS_TC, |
|---|
| 112 | + match.key->ls[0].mpls_tc) | |
|---|
| 113 | + FIELD_PREP(NFP_FLOWER_MASK_MPLS_BOS, |
|---|
| 114 | + match.key->ls[0].mpls_bos) | |
|---|
| 121 | 115 | NFP_FLOWER_MASK_MPLS_Q; |
|---|
| 122 | | - |
|---|
| 123 | | - frame->mpls_lse = cpu_to_be32(t_mpls); |
|---|
| 124 | | - } else if (dissector_uses_key(flow->dissector, |
|---|
| 125 | | - FLOW_DISSECTOR_KEY_BASIC)) { |
|---|
| 116 | + ext->mpls_lse = cpu_to_be32(t_mpls); |
|---|
| 117 | + t_mpls = FIELD_PREP(NFP_FLOWER_MASK_MPLS_LB, |
|---|
| 118 | + match.mask->ls[0].mpls_label) | |
|---|
| 119 | + FIELD_PREP(NFP_FLOWER_MASK_MPLS_TC, |
|---|
| 120 | + match.mask->ls[0].mpls_tc) | |
|---|
| 121 | + FIELD_PREP(NFP_FLOWER_MASK_MPLS_BOS, |
|---|
| 122 | + match.mask->ls[0].mpls_bos) | |
|---|
| 123 | + NFP_FLOWER_MASK_MPLS_Q; |
|---|
| 124 | + msk->mpls_lse = cpu_to_be32(t_mpls); |
|---|
| 125 | + } else if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_BASIC)) { |
|---|
| 126 | 126 | /* Check for mpls ether type and set NFP_FLOWER_MASK_MPLS_Q |
|---|
| 127 | 127 | * bit, which indicates an mpls ether type but without any |
|---|
| 128 | 128 | * mpls fields. |
|---|
| 129 | 129 | */ |
|---|
| 130 | | - struct flow_dissector_key_basic *key_basic; |
|---|
| 130 | + struct flow_match_basic match; |
|---|
| 131 | 131 | |
|---|
| 132 | | - key_basic = skb_flow_dissector_target(flow->dissector, |
|---|
| 133 | | - FLOW_DISSECTOR_KEY_BASIC, |
|---|
| 134 | | - flow->key); |
|---|
| 135 | | - if (key_basic->n_proto == cpu_to_be16(ETH_P_MPLS_UC) || |
|---|
| 136 | | - key_basic->n_proto == cpu_to_be16(ETH_P_MPLS_MC)) |
|---|
| 137 | | - frame->mpls_lse = cpu_to_be32(NFP_FLOWER_MASK_MPLS_Q); |
|---|
| 132 | + flow_rule_match_basic(rule, &match); |
|---|
| 133 | + if (match.key->n_proto == cpu_to_be16(ETH_P_MPLS_UC) || |
|---|
| 134 | + match.key->n_proto == cpu_to_be16(ETH_P_MPLS_MC)) { |
|---|
| 135 | + ext->mpls_lse = cpu_to_be32(NFP_FLOWER_MASK_MPLS_Q); |
|---|
| 136 | + msk->mpls_lse = cpu_to_be32(NFP_FLOWER_MASK_MPLS_Q); |
|---|
| 137 | + } |
|---|
| 138 | 138 | } |
|---|
| 139 | | -} |
|---|
| 140 | | - |
|---|
| 141 | | -static void |
|---|
| 142 | | -nfp_flower_compile_tport(struct nfp_flower_tp_ports *frame, |
|---|
| 143 | | - struct tc_cls_flower_offload *flow, |
|---|
| 144 | | - bool mask_version) |
|---|
| 145 | | -{ |
|---|
| 146 | | - struct fl_flow_key *target = mask_version ? flow->mask : flow->key; |
|---|
| 147 | | - struct flow_dissector_key_ports *tp; |
|---|
| 148 | | - |
|---|
| 149 | | - memset(frame, 0, sizeof(struct nfp_flower_tp_ports)); |
|---|
| 150 | | - |
|---|
| 151 | | - if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_PORTS)) { |
|---|
| 152 | | - tp = skb_flow_dissector_target(flow->dissector, |
|---|
| 153 | | - FLOW_DISSECTOR_KEY_PORTS, |
|---|
| 154 | | - target); |
|---|
| 155 | | - frame->port_src = tp->src; |
|---|
| 156 | | - frame->port_dst = tp->dst; |
|---|
| 157 | | - } |
|---|
| 158 | | -} |
|---|
| 159 | | - |
|---|
| 160 | | -static void |
|---|
| 161 | | -nfp_flower_compile_ip_ext(struct nfp_flower_ip_ext *frame, |
|---|
| 162 | | - struct tc_cls_flower_offload *flow, |
|---|
| 163 | | - bool mask_version) |
|---|
| 164 | | -{ |
|---|
| 165 | | - struct fl_flow_key *target = mask_version ? flow->mask : flow->key; |
|---|
| 166 | | - |
|---|
| 167 | | - if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_BASIC)) { |
|---|
| 168 | | - struct flow_dissector_key_basic *basic; |
|---|
| 169 | | - |
|---|
| 170 | | - basic = skb_flow_dissector_target(flow->dissector, |
|---|
| 171 | | - FLOW_DISSECTOR_KEY_BASIC, |
|---|
| 172 | | - target); |
|---|
| 173 | | - frame->proto = basic->ip_proto; |
|---|
| 174 | | - } |
|---|
| 175 | | - |
|---|
| 176 | | - if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_IP)) { |
|---|
| 177 | | - struct flow_dissector_key_ip *flow_ip; |
|---|
| 178 | | - |
|---|
| 179 | | - flow_ip = skb_flow_dissector_target(flow->dissector, |
|---|
| 180 | | - FLOW_DISSECTOR_KEY_IP, |
|---|
| 181 | | - target); |
|---|
| 182 | | - frame->tos = flow_ip->tos; |
|---|
| 183 | | - frame->ttl = flow_ip->ttl; |
|---|
| 184 | | - } |
|---|
| 185 | | - |
|---|
| 186 | | - if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_TCP)) { |
|---|
| 187 | | - struct flow_dissector_key_tcp *tcp; |
|---|
| 188 | | - u32 tcp_flags; |
|---|
| 189 | | - |
|---|
| 190 | | - tcp = skb_flow_dissector_target(flow->dissector, |
|---|
| 191 | | - FLOW_DISSECTOR_KEY_TCP, target); |
|---|
| 192 | | - tcp_flags = be16_to_cpu(tcp->flags); |
|---|
| 193 | | - |
|---|
| 194 | | - if (tcp_flags & TCPHDR_FIN) |
|---|
| 195 | | - frame->flags |= NFP_FL_TCP_FLAG_FIN; |
|---|
| 196 | | - if (tcp_flags & TCPHDR_SYN) |
|---|
| 197 | | - frame->flags |= NFP_FL_TCP_FLAG_SYN; |
|---|
| 198 | | - if (tcp_flags & TCPHDR_RST) |
|---|
| 199 | | - frame->flags |= NFP_FL_TCP_FLAG_RST; |
|---|
| 200 | | - if (tcp_flags & TCPHDR_PSH) |
|---|
| 201 | | - frame->flags |= NFP_FL_TCP_FLAG_PSH; |
|---|
| 202 | | - if (tcp_flags & TCPHDR_URG) |
|---|
| 203 | | - frame->flags |= NFP_FL_TCP_FLAG_URG; |
|---|
| 204 | | - } |
|---|
| 205 | | - |
|---|
| 206 | | - if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_CONTROL)) { |
|---|
| 207 | | - struct flow_dissector_key_control *key; |
|---|
| 208 | | - |
|---|
| 209 | | - key = skb_flow_dissector_target(flow->dissector, |
|---|
| 210 | | - FLOW_DISSECTOR_KEY_CONTROL, |
|---|
| 211 | | - target); |
|---|
| 212 | | - if (key->flags & FLOW_DIS_IS_FRAGMENT) |
|---|
| 213 | | - frame->flags |= NFP_FL_IP_FRAGMENTED; |
|---|
| 214 | | - if (key->flags & FLOW_DIS_FIRST_FRAG) |
|---|
| 215 | | - frame->flags |= NFP_FL_IP_FRAG_FIRST; |
|---|
| 216 | | - } |
|---|
| 217 | | -} |
|---|
| 218 | | - |
|---|
| 219 | | -static void |
|---|
| 220 | | -nfp_flower_compile_ipv4(struct nfp_flower_ipv4 *frame, |
|---|
| 221 | | - struct tc_cls_flower_offload *flow, |
|---|
| 222 | | - bool mask_version) |
|---|
| 223 | | -{ |
|---|
| 224 | | - struct fl_flow_key *target = mask_version ? flow->mask : flow->key; |
|---|
| 225 | | - struct flow_dissector_key_ipv4_addrs *addr; |
|---|
| 226 | | - |
|---|
| 227 | | - memset(frame, 0, sizeof(struct nfp_flower_ipv4)); |
|---|
| 228 | | - |
|---|
| 229 | | - if (dissector_uses_key(flow->dissector, |
|---|
| 230 | | - FLOW_DISSECTOR_KEY_IPV4_ADDRS)) { |
|---|
| 231 | | - addr = skb_flow_dissector_target(flow->dissector, |
|---|
| 232 | | - FLOW_DISSECTOR_KEY_IPV4_ADDRS, |
|---|
| 233 | | - target); |
|---|
| 234 | | - frame->ipv4_src = addr->src; |
|---|
| 235 | | - frame->ipv4_dst = addr->dst; |
|---|
| 236 | | - } |
|---|
| 237 | | - |
|---|
| 238 | | - nfp_flower_compile_ip_ext(&frame->ip_ext, flow, mask_version); |
|---|
| 239 | | -} |
|---|
| 240 | | - |
|---|
| 241 | | -static void |
|---|
| 242 | | -nfp_flower_compile_ipv6(struct nfp_flower_ipv6 *frame, |
|---|
| 243 | | - struct tc_cls_flower_offload *flow, |
|---|
| 244 | | - bool mask_version) |
|---|
| 245 | | -{ |
|---|
| 246 | | - struct fl_flow_key *target = mask_version ? flow->mask : flow->key; |
|---|
| 247 | | - struct flow_dissector_key_ipv6_addrs *addr; |
|---|
| 248 | | - |
|---|
| 249 | | - memset(frame, 0, sizeof(struct nfp_flower_ipv6)); |
|---|
| 250 | | - |
|---|
| 251 | | - if (dissector_uses_key(flow->dissector, |
|---|
| 252 | | - FLOW_DISSECTOR_KEY_IPV6_ADDRS)) { |
|---|
| 253 | | - addr = skb_flow_dissector_target(flow->dissector, |
|---|
| 254 | | - FLOW_DISSECTOR_KEY_IPV6_ADDRS, |
|---|
| 255 | | - target); |
|---|
| 256 | | - frame->ipv6_src = addr->src; |
|---|
| 257 | | - frame->ipv6_dst = addr->dst; |
|---|
| 258 | | - } |
|---|
| 259 | | - |
|---|
| 260 | | - nfp_flower_compile_ip_ext(&frame->ip_ext, flow, mask_version); |
|---|
| 261 | | -} |
|---|
| 262 | | - |
|---|
| 263 | | -static int |
|---|
| 264 | | -nfp_flower_compile_geneve_opt(void *key_buf, struct tc_cls_flower_offload *flow, |
|---|
| 265 | | - bool mask_version) |
|---|
| 266 | | -{ |
|---|
| 267 | | - struct fl_flow_key *target = mask_version ? flow->mask : flow->key; |
|---|
| 268 | | - struct flow_dissector_key_enc_opts *opts; |
|---|
| 269 | | - |
|---|
| 270 | | - opts = skb_flow_dissector_target(flow->dissector, |
|---|
| 271 | | - FLOW_DISSECTOR_KEY_ENC_OPTS, |
|---|
| 272 | | - target); |
|---|
| 273 | | - memcpy(key_buf, opts->data, opts->len); |
|---|
| 274 | 139 | |
|---|
| 275 | 140 | return 0; |
|---|
| 276 | 141 | } |
|---|
| 277 | 142 | |
|---|
| 278 | 143 | static void |
|---|
| 279 | | -nfp_flower_compile_ipv4_udp_tun(struct nfp_flower_ipv4_udp_tun *frame, |
|---|
| 280 | | - struct tc_cls_flower_offload *flow, |
|---|
| 281 | | - bool mask_version) |
|---|
| 144 | +nfp_flower_compile_tport(struct nfp_flower_tp_ports *ext, |
|---|
| 145 | + struct nfp_flower_tp_ports *msk, |
|---|
| 146 | + struct flow_rule *rule) |
|---|
| 282 | 147 | { |
|---|
| 283 | | - struct fl_flow_key *target = mask_version ? flow->mask : flow->key; |
|---|
| 284 | | - struct flow_dissector_key_ipv4_addrs *tun_ips; |
|---|
| 285 | | - struct flow_dissector_key_keyid *vni; |
|---|
| 286 | | - struct flow_dissector_key_ip *ip; |
|---|
| 148 | + memset(ext, 0, sizeof(struct nfp_flower_tp_ports)); |
|---|
| 149 | + memset(msk, 0, sizeof(struct nfp_flower_tp_ports)); |
|---|
| 287 | 150 | |
|---|
| 288 | | - memset(frame, 0, sizeof(struct nfp_flower_ipv4_udp_tun)); |
|---|
| 151 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_PORTS)) { |
|---|
| 152 | + struct flow_match_ports match; |
|---|
| 289 | 153 | |
|---|
| 290 | | - if (dissector_uses_key(flow->dissector, |
|---|
| 291 | | - FLOW_DISSECTOR_KEY_ENC_KEYID)) { |
|---|
| 292 | | - u32 temp_vni; |
|---|
| 293 | | - |
|---|
| 294 | | - vni = skb_flow_dissector_target(flow->dissector, |
|---|
| 295 | | - FLOW_DISSECTOR_KEY_ENC_KEYID, |
|---|
| 296 | | - target); |
|---|
| 297 | | - temp_vni = be32_to_cpu(vni->keyid) << NFP_FL_TUN_VNI_OFFSET; |
|---|
| 298 | | - frame->tun_id = cpu_to_be32(temp_vni); |
|---|
| 299 | | - } |
|---|
| 300 | | - |
|---|
| 301 | | - if (dissector_uses_key(flow->dissector, |
|---|
| 302 | | - FLOW_DISSECTOR_KEY_ENC_IPV4_ADDRS)) { |
|---|
| 303 | | - tun_ips = |
|---|
| 304 | | - skb_flow_dissector_target(flow->dissector, |
|---|
| 305 | | - FLOW_DISSECTOR_KEY_ENC_IPV4_ADDRS, |
|---|
| 306 | | - target); |
|---|
| 307 | | - frame->ip_src = tun_ips->src; |
|---|
| 308 | | - frame->ip_dst = tun_ips->dst; |
|---|
| 309 | | - } |
|---|
| 310 | | - |
|---|
| 311 | | - if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_ENC_IP)) { |
|---|
| 312 | | - ip = skb_flow_dissector_target(flow->dissector, |
|---|
| 313 | | - FLOW_DISSECTOR_KEY_ENC_IP, |
|---|
| 314 | | - target); |
|---|
| 315 | | - frame->tos = ip->tos; |
|---|
| 316 | | - frame->ttl = ip->ttl; |
|---|
| 154 | + flow_rule_match_ports(rule, &match); |
|---|
| 155 | + ext->port_src = match.key->src; |
|---|
| 156 | + ext->port_dst = match.key->dst; |
|---|
| 157 | + msk->port_src = match.mask->src; |
|---|
| 158 | + msk->port_dst = match.mask->dst; |
|---|
| 317 | 159 | } |
|---|
| 318 | 160 | } |
|---|
| 319 | 161 | |
|---|
| 320 | | -int nfp_flower_compile_flow_match(struct tc_cls_flower_offload *flow, |
|---|
| 162 | +static void |
|---|
| 163 | +nfp_flower_compile_ip_ext(struct nfp_flower_ip_ext *ext, |
|---|
| 164 | + struct nfp_flower_ip_ext *msk, struct flow_rule *rule) |
|---|
| 165 | +{ |
|---|
| 166 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_BASIC)) { |
|---|
| 167 | + struct flow_match_basic match; |
|---|
| 168 | + |
|---|
| 169 | + flow_rule_match_basic(rule, &match); |
|---|
| 170 | + ext->proto = match.key->ip_proto; |
|---|
| 171 | + msk->proto = match.mask->ip_proto; |
|---|
| 172 | + } |
|---|
| 173 | + |
|---|
| 174 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_IP)) { |
|---|
| 175 | + struct flow_match_ip match; |
|---|
| 176 | + |
|---|
| 177 | + flow_rule_match_ip(rule, &match); |
|---|
| 178 | + ext->tos = match.key->tos; |
|---|
| 179 | + ext->ttl = match.key->ttl; |
|---|
| 180 | + msk->tos = match.mask->tos; |
|---|
| 181 | + msk->ttl = match.mask->ttl; |
|---|
| 182 | + } |
|---|
| 183 | + |
|---|
| 184 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_TCP)) { |
|---|
| 185 | + u16 tcp_flags, tcp_flags_mask; |
|---|
| 186 | + struct flow_match_tcp match; |
|---|
| 187 | + |
|---|
| 188 | + flow_rule_match_tcp(rule, &match); |
|---|
| 189 | + tcp_flags = be16_to_cpu(match.key->flags); |
|---|
| 190 | + tcp_flags_mask = be16_to_cpu(match.mask->flags); |
|---|
| 191 | + |
|---|
| 192 | + if (tcp_flags & TCPHDR_FIN) |
|---|
| 193 | + ext->flags |= NFP_FL_TCP_FLAG_FIN; |
|---|
| 194 | + if (tcp_flags_mask & TCPHDR_FIN) |
|---|
| 195 | + msk->flags |= NFP_FL_TCP_FLAG_FIN; |
|---|
| 196 | + |
|---|
| 197 | + if (tcp_flags & TCPHDR_SYN) |
|---|
| 198 | + ext->flags |= NFP_FL_TCP_FLAG_SYN; |
|---|
| 199 | + if (tcp_flags_mask & TCPHDR_SYN) |
|---|
| 200 | + msk->flags |= NFP_FL_TCP_FLAG_SYN; |
|---|
| 201 | + |
|---|
| 202 | + if (tcp_flags & TCPHDR_RST) |
|---|
| 203 | + ext->flags |= NFP_FL_TCP_FLAG_RST; |
|---|
| 204 | + if (tcp_flags_mask & TCPHDR_RST) |
|---|
| 205 | + msk->flags |= NFP_FL_TCP_FLAG_RST; |
|---|
| 206 | + |
|---|
| 207 | + if (tcp_flags & TCPHDR_PSH) |
|---|
| 208 | + ext->flags |= NFP_FL_TCP_FLAG_PSH; |
|---|
| 209 | + if (tcp_flags_mask & TCPHDR_PSH) |
|---|
| 210 | + msk->flags |= NFP_FL_TCP_FLAG_PSH; |
|---|
| 211 | + |
|---|
| 212 | + if (tcp_flags & TCPHDR_URG) |
|---|
| 213 | + ext->flags |= NFP_FL_TCP_FLAG_URG; |
|---|
| 214 | + if (tcp_flags_mask & TCPHDR_URG) |
|---|
| 215 | + msk->flags |= NFP_FL_TCP_FLAG_URG; |
|---|
| 216 | + } |
|---|
| 217 | + |
|---|
| 218 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_CONTROL)) { |
|---|
| 219 | + struct flow_match_control match; |
|---|
| 220 | + |
|---|
| 221 | + flow_rule_match_control(rule, &match); |
|---|
| 222 | + if (match.key->flags & FLOW_DIS_IS_FRAGMENT) |
|---|
| 223 | + ext->flags |= NFP_FL_IP_FRAGMENTED; |
|---|
| 224 | + if (match.mask->flags & FLOW_DIS_IS_FRAGMENT) |
|---|
| 225 | + msk->flags |= NFP_FL_IP_FRAGMENTED; |
|---|
| 226 | + if (match.key->flags & FLOW_DIS_FIRST_FRAG) |
|---|
| 227 | + ext->flags |= NFP_FL_IP_FRAG_FIRST; |
|---|
| 228 | + if (match.mask->flags & FLOW_DIS_FIRST_FRAG) |
|---|
| 229 | + msk->flags |= NFP_FL_IP_FRAG_FIRST; |
|---|
| 230 | + } |
|---|
| 231 | +} |
|---|
| 232 | + |
|---|
| 233 | +static void |
|---|
| 234 | +nfp_flower_fill_vlan(struct flow_dissector_key_vlan *key, |
|---|
| 235 | + struct nfp_flower_vlan *frame, |
|---|
| 236 | + bool outer_vlan) |
|---|
| 237 | +{ |
|---|
| 238 | + u16 tci; |
|---|
| 239 | + |
|---|
| 240 | + tci = NFP_FLOWER_MASK_VLAN_PRESENT; |
|---|
| 241 | + tci |= FIELD_PREP(NFP_FLOWER_MASK_VLAN_PRIO, |
|---|
| 242 | + key->vlan_priority) | |
|---|
| 243 | + FIELD_PREP(NFP_FLOWER_MASK_VLAN_VID, |
|---|
| 244 | + key->vlan_id); |
|---|
| 245 | + |
|---|
| 246 | + if (outer_vlan) { |
|---|
| 247 | + frame->outer_tci = cpu_to_be16(tci); |
|---|
| 248 | + frame->outer_tpid = key->vlan_tpid; |
|---|
| 249 | + } else { |
|---|
| 250 | + frame->inner_tci = cpu_to_be16(tci); |
|---|
| 251 | + frame->inner_tpid = key->vlan_tpid; |
|---|
| 252 | + } |
|---|
| 253 | +} |
|---|
| 254 | + |
|---|
| 255 | +static void |
|---|
| 256 | +nfp_flower_compile_vlan(struct nfp_flower_vlan *ext, |
|---|
| 257 | + struct nfp_flower_vlan *msk, |
|---|
| 258 | + struct flow_rule *rule) |
|---|
| 259 | +{ |
|---|
| 260 | + struct flow_match_vlan match; |
|---|
| 261 | + |
|---|
| 262 | + memset(ext, 0, sizeof(struct nfp_flower_vlan)); |
|---|
| 263 | + memset(msk, 0, sizeof(struct nfp_flower_vlan)); |
|---|
| 264 | + |
|---|
| 265 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_VLAN)) { |
|---|
| 266 | + flow_rule_match_vlan(rule, &match); |
|---|
| 267 | + nfp_flower_fill_vlan(match.key, ext, true); |
|---|
| 268 | + nfp_flower_fill_vlan(match.mask, msk, true); |
|---|
| 269 | + } |
|---|
| 270 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_CVLAN)) { |
|---|
| 271 | + flow_rule_match_cvlan(rule, &match); |
|---|
| 272 | + nfp_flower_fill_vlan(match.key, ext, false); |
|---|
| 273 | + nfp_flower_fill_vlan(match.mask, msk, false); |
|---|
| 274 | + } |
|---|
| 275 | +} |
|---|
| 276 | + |
|---|
| 277 | +static void |
|---|
| 278 | +nfp_flower_compile_ipv4(struct nfp_flower_ipv4 *ext, |
|---|
| 279 | + struct nfp_flower_ipv4 *msk, struct flow_rule *rule) |
|---|
| 280 | +{ |
|---|
| 281 | + struct flow_match_ipv4_addrs match; |
|---|
| 282 | + |
|---|
| 283 | + memset(ext, 0, sizeof(struct nfp_flower_ipv4)); |
|---|
| 284 | + memset(msk, 0, sizeof(struct nfp_flower_ipv4)); |
|---|
| 285 | + |
|---|
| 286 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_IPV4_ADDRS)) { |
|---|
| 287 | + flow_rule_match_ipv4_addrs(rule, &match); |
|---|
| 288 | + ext->ipv4_src = match.key->src; |
|---|
| 289 | + ext->ipv4_dst = match.key->dst; |
|---|
| 290 | + msk->ipv4_src = match.mask->src; |
|---|
| 291 | + msk->ipv4_dst = match.mask->dst; |
|---|
| 292 | + } |
|---|
| 293 | + |
|---|
| 294 | + nfp_flower_compile_ip_ext(&ext->ip_ext, &msk->ip_ext, rule); |
|---|
| 295 | +} |
|---|
| 296 | + |
|---|
| 297 | +static void |
|---|
| 298 | +nfp_flower_compile_ipv6(struct nfp_flower_ipv6 *ext, |
|---|
| 299 | + struct nfp_flower_ipv6 *msk, struct flow_rule *rule) |
|---|
| 300 | +{ |
|---|
| 301 | + memset(ext, 0, sizeof(struct nfp_flower_ipv6)); |
|---|
| 302 | + memset(msk, 0, sizeof(struct nfp_flower_ipv6)); |
|---|
| 303 | + |
|---|
| 304 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_IPV6_ADDRS)) { |
|---|
| 305 | + struct flow_match_ipv6_addrs match; |
|---|
| 306 | + |
|---|
| 307 | + flow_rule_match_ipv6_addrs(rule, &match); |
|---|
| 308 | + ext->ipv6_src = match.key->src; |
|---|
| 309 | + ext->ipv6_dst = match.key->dst; |
|---|
| 310 | + msk->ipv6_src = match.mask->src; |
|---|
| 311 | + msk->ipv6_dst = match.mask->dst; |
|---|
| 312 | + } |
|---|
| 313 | + |
|---|
| 314 | + nfp_flower_compile_ip_ext(&ext->ip_ext, &msk->ip_ext, rule); |
|---|
| 315 | +} |
|---|
| 316 | + |
|---|
| 317 | +static int |
|---|
| 318 | +nfp_flower_compile_geneve_opt(void *ext, void *msk, struct flow_rule *rule) |
|---|
| 319 | +{ |
|---|
| 320 | + struct flow_match_enc_opts match; |
|---|
| 321 | + |
|---|
| 322 | + flow_rule_match_enc_opts(rule, &match); |
|---|
| 323 | + memcpy(ext, match.key->data, match.key->len); |
|---|
| 324 | + memcpy(msk, match.mask->data, match.mask->len); |
|---|
| 325 | + |
|---|
| 326 | + return 0; |
|---|
| 327 | +} |
|---|
| 328 | + |
|---|
| 329 | +static void |
|---|
| 330 | +nfp_flower_compile_tun_ipv4_addrs(struct nfp_flower_tun_ipv4 *ext, |
|---|
| 331 | + struct nfp_flower_tun_ipv4 *msk, |
|---|
| 332 | + struct flow_rule *rule) |
|---|
| 333 | +{ |
|---|
| 334 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_ENC_IPV4_ADDRS)) { |
|---|
| 335 | + struct flow_match_ipv4_addrs match; |
|---|
| 336 | + |
|---|
| 337 | + flow_rule_match_enc_ipv4_addrs(rule, &match); |
|---|
| 338 | + ext->src = match.key->src; |
|---|
| 339 | + ext->dst = match.key->dst; |
|---|
| 340 | + msk->src = match.mask->src; |
|---|
| 341 | + msk->dst = match.mask->dst; |
|---|
| 342 | + } |
|---|
| 343 | +} |
|---|
| 344 | + |
|---|
| 345 | +static void |
|---|
| 346 | +nfp_flower_compile_tun_ipv6_addrs(struct nfp_flower_tun_ipv6 *ext, |
|---|
| 347 | + struct nfp_flower_tun_ipv6 *msk, |
|---|
| 348 | + struct flow_rule *rule) |
|---|
| 349 | +{ |
|---|
| 350 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_ENC_IPV6_ADDRS)) { |
|---|
| 351 | + struct flow_match_ipv6_addrs match; |
|---|
| 352 | + |
|---|
| 353 | + flow_rule_match_enc_ipv6_addrs(rule, &match); |
|---|
| 354 | + ext->src = match.key->src; |
|---|
| 355 | + ext->dst = match.key->dst; |
|---|
| 356 | + msk->src = match.mask->src; |
|---|
| 357 | + msk->dst = match.mask->dst; |
|---|
| 358 | + } |
|---|
| 359 | +} |
|---|
| 360 | + |
|---|
| 361 | +static void |
|---|
| 362 | +nfp_flower_compile_tun_ip_ext(struct nfp_flower_tun_ip_ext *ext, |
|---|
| 363 | + struct nfp_flower_tun_ip_ext *msk, |
|---|
| 364 | + struct flow_rule *rule) |
|---|
| 365 | +{ |
|---|
| 366 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_ENC_IP)) { |
|---|
| 367 | + struct flow_match_ip match; |
|---|
| 368 | + |
|---|
| 369 | + flow_rule_match_enc_ip(rule, &match); |
|---|
| 370 | + ext->tos = match.key->tos; |
|---|
| 371 | + ext->ttl = match.key->ttl; |
|---|
| 372 | + msk->tos = match.mask->tos; |
|---|
| 373 | + msk->ttl = match.mask->ttl; |
|---|
| 374 | + } |
|---|
| 375 | +} |
|---|
| 376 | + |
|---|
| 377 | +static void |
|---|
| 378 | +nfp_flower_compile_tun_udp_key(__be32 *key, __be32 *key_msk, |
|---|
| 379 | + struct flow_rule *rule) |
|---|
| 380 | +{ |
|---|
| 381 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_ENC_KEYID)) { |
|---|
| 382 | + struct flow_match_enc_keyid match; |
|---|
| 383 | + u32 vni; |
|---|
| 384 | + |
|---|
| 385 | + flow_rule_match_enc_keyid(rule, &match); |
|---|
| 386 | + vni = be32_to_cpu(match.key->keyid) << NFP_FL_TUN_VNI_OFFSET; |
|---|
| 387 | + *key = cpu_to_be32(vni); |
|---|
| 388 | + vni = be32_to_cpu(match.mask->keyid) << NFP_FL_TUN_VNI_OFFSET; |
|---|
| 389 | + *key_msk = cpu_to_be32(vni); |
|---|
| 390 | + } |
|---|
| 391 | +} |
|---|
| 392 | + |
|---|
| 393 | +static void |
|---|
| 394 | +nfp_flower_compile_tun_gre_key(__be32 *key, __be32 *key_msk, __be16 *flags, |
|---|
| 395 | + __be16 *flags_msk, struct flow_rule *rule) |
|---|
| 396 | +{ |
|---|
| 397 | + if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_ENC_KEYID)) { |
|---|
| 398 | + struct flow_match_enc_keyid match; |
|---|
| 399 | + |
|---|
| 400 | + flow_rule_match_enc_keyid(rule, &match); |
|---|
| 401 | + *key = match.key->keyid; |
|---|
| 402 | + *key_msk = match.mask->keyid; |
|---|
| 403 | + |
|---|
| 404 | + *flags = cpu_to_be16(NFP_FL_GRE_FLAG_KEY); |
|---|
| 405 | + *flags_msk = cpu_to_be16(NFP_FL_GRE_FLAG_KEY); |
|---|
| 406 | + } |
|---|
| 407 | +} |
|---|
| 408 | + |
|---|
| 409 | +static void |
|---|
| 410 | +nfp_flower_compile_ipv4_gre_tun(struct nfp_flower_ipv4_gre_tun *ext, |
|---|
| 411 | + struct nfp_flower_ipv4_gre_tun *msk, |
|---|
| 412 | + struct flow_rule *rule) |
|---|
| 413 | +{ |
|---|
| 414 | + memset(ext, 0, sizeof(struct nfp_flower_ipv4_gre_tun)); |
|---|
| 415 | + memset(msk, 0, sizeof(struct nfp_flower_ipv4_gre_tun)); |
|---|
| 416 | + |
|---|
| 417 | + /* NVGRE is the only supported GRE tunnel type */ |
|---|
| 418 | + ext->ethertype = cpu_to_be16(ETH_P_TEB); |
|---|
| 419 | + msk->ethertype = cpu_to_be16(~0); |
|---|
| 420 | + |
|---|
| 421 | + nfp_flower_compile_tun_ipv4_addrs(&ext->ipv4, &msk->ipv4, rule); |
|---|
| 422 | + nfp_flower_compile_tun_ip_ext(&ext->ip_ext, &msk->ip_ext, rule); |
|---|
| 423 | + nfp_flower_compile_tun_gre_key(&ext->tun_key, &msk->tun_key, |
|---|
| 424 | + &ext->tun_flags, &msk->tun_flags, rule); |
|---|
| 425 | +} |
|---|
| 426 | + |
|---|
| 427 | +static void |
|---|
| 428 | +nfp_flower_compile_ipv4_udp_tun(struct nfp_flower_ipv4_udp_tun *ext, |
|---|
| 429 | + struct nfp_flower_ipv4_udp_tun *msk, |
|---|
| 430 | + struct flow_rule *rule) |
|---|
| 431 | +{ |
|---|
| 432 | + memset(ext, 0, sizeof(struct nfp_flower_ipv4_udp_tun)); |
|---|
| 433 | + memset(msk, 0, sizeof(struct nfp_flower_ipv4_udp_tun)); |
|---|
| 434 | + |
|---|
| 435 | + nfp_flower_compile_tun_ipv4_addrs(&ext->ipv4, &msk->ipv4, rule); |
|---|
| 436 | + nfp_flower_compile_tun_ip_ext(&ext->ip_ext, &msk->ip_ext, rule); |
|---|
| 437 | + nfp_flower_compile_tun_udp_key(&ext->tun_id, &msk->tun_id, rule); |
|---|
| 438 | +} |
|---|
| 439 | + |
|---|
| 440 | +static void |
|---|
| 441 | +nfp_flower_compile_ipv6_udp_tun(struct nfp_flower_ipv6_udp_tun *ext, |
|---|
| 442 | + struct nfp_flower_ipv6_udp_tun *msk, |
|---|
| 443 | + struct flow_rule *rule) |
|---|
| 444 | +{ |
|---|
| 445 | + memset(ext, 0, sizeof(struct nfp_flower_ipv6_udp_tun)); |
|---|
| 446 | + memset(msk, 0, sizeof(struct nfp_flower_ipv6_udp_tun)); |
|---|
| 447 | + |
|---|
| 448 | + nfp_flower_compile_tun_ipv6_addrs(&ext->ipv6, &msk->ipv6, rule); |
|---|
| 449 | + nfp_flower_compile_tun_ip_ext(&ext->ip_ext, &msk->ip_ext, rule); |
|---|
| 450 | + nfp_flower_compile_tun_udp_key(&ext->tun_id, &msk->tun_id, rule); |
|---|
| 451 | +} |
|---|
| 452 | + |
|---|
| 453 | +static void |
|---|
| 454 | +nfp_flower_compile_ipv6_gre_tun(struct nfp_flower_ipv6_gre_tun *ext, |
|---|
| 455 | + struct nfp_flower_ipv6_gre_tun *msk, |
|---|
| 456 | + struct flow_rule *rule) |
|---|
| 457 | +{ |
|---|
| 458 | + memset(ext, 0, sizeof(struct nfp_flower_ipv6_gre_tun)); |
|---|
| 459 | + memset(msk, 0, sizeof(struct nfp_flower_ipv6_gre_tun)); |
|---|
| 460 | + |
|---|
| 461 | + /* NVGRE is the only supported GRE tunnel type */ |
|---|
| 462 | + ext->ethertype = cpu_to_be16(ETH_P_TEB); |
|---|
| 463 | + msk->ethertype = cpu_to_be16(~0); |
|---|
| 464 | + |
|---|
| 465 | + nfp_flower_compile_tun_ipv6_addrs(&ext->ipv6, &msk->ipv6, rule); |
|---|
| 466 | + nfp_flower_compile_tun_ip_ext(&ext->ip_ext, &msk->ip_ext, rule); |
|---|
| 467 | + nfp_flower_compile_tun_gre_key(&ext->tun_key, &msk->tun_key, |
|---|
| 468 | + &ext->tun_flags, &msk->tun_flags, rule); |
|---|
| 469 | +} |
|---|
| 470 | + |
|---|
| 471 | +int nfp_flower_compile_flow_match(struct nfp_app *app, |
|---|
| 472 | + struct flow_cls_offload *flow, |
|---|
| 321 | 473 | struct nfp_fl_key_ls *key_ls, |
|---|
| 322 | 474 | struct net_device *netdev, |
|---|
| 323 | 475 | struct nfp_fl_payload *nfp_flow, |
|---|
| 324 | | - enum nfp_flower_tun_type tun_type) |
|---|
| 476 | + enum nfp_flower_tun_type tun_type, |
|---|
| 477 | + struct netlink_ext_ack *extack) |
|---|
| 325 | 478 | { |
|---|
| 326 | | - struct nfp_repr *netdev_repr; |
|---|
| 479 | + struct flow_rule *rule = flow_cls_offload_flow_rule(flow); |
|---|
| 480 | + struct nfp_flower_priv *priv = app->priv; |
|---|
| 481 | + bool qinq_sup; |
|---|
| 482 | + u32 port_id; |
|---|
| 483 | + int ext_len; |
|---|
| 327 | 484 | int err; |
|---|
| 328 | 485 | u8 *ext; |
|---|
| 329 | 486 | u8 *msk; |
|---|
| 487 | + |
|---|
| 488 | + port_id = nfp_flower_get_port_id_from_netdev(app, netdev); |
|---|
| 330 | 489 | |
|---|
| 331 | 490 | memset(nfp_flow->unmasked_data, 0, key_ls->key_size); |
|---|
| 332 | 491 | memset(nfp_flow->mask_data, 0, key_ls->key_size); |
|---|
| .. | .. |
|---|
| 334 | 493 | ext = nfp_flow->unmasked_data; |
|---|
| 335 | 494 | msk = nfp_flow->mask_data; |
|---|
| 336 | 495 | |
|---|
| 337 | | - /* Populate Exact Metadata. */ |
|---|
| 496 | + qinq_sup = !!(priv->flower_ext_feats & NFP_FL_FEATS_VLAN_QINQ); |
|---|
| 497 | + |
|---|
| 338 | 498 | nfp_flower_compile_meta_tci((struct nfp_flower_meta_tci *)ext, |
|---|
| 339 | | - flow, key_ls->key_layer, false); |
|---|
| 340 | | - /* Populate Mask Metadata. */ |
|---|
| 341 | | - nfp_flower_compile_meta_tci((struct nfp_flower_meta_tci *)msk, |
|---|
| 342 | | - flow, key_ls->key_layer, true); |
|---|
| 499 | + (struct nfp_flower_meta_tci *)msk, |
|---|
| 500 | + rule, key_ls->key_layer, qinq_sup); |
|---|
| 343 | 501 | ext += sizeof(struct nfp_flower_meta_tci); |
|---|
| 344 | 502 | msk += sizeof(struct nfp_flower_meta_tci); |
|---|
| 345 | 503 | |
|---|
| .. | .. |
|---|
| 355 | 513 | |
|---|
| 356 | 514 | /* Populate Exact Port data. */ |
|---|
| 357 | 515 | err = nfp_flower_compile_port((struct nfp_flower_in_port *)ext, |
|---|
| 358 | | - nfp_repr_get_port_id(netdev), |
|---|
| 359 | | - false, tun_type); |
|---|
| 516 | + port_id, false, tun_type, extack); |
|---|
| 360 | 517 | if (err) |
|---|
| 361 | 518 | return err; |
|---|
| 362 | 519 | |
|---|
| 363 | 520 | /* Populate Mask Port Data. */ |
|---|
| 364 | 521 | err = nfp_flower_compile_port((struct nfp_flower_in_port *)msk, |
|---|
| 365 | | - nfp_repr_get_port_id(netdev), |
|---|
| 366 | | - true, tun_type); |
|---|
| 522 | + port_id, true, tun_type, extack); |
|---|
| 367 | 523 | if (err) |
|---|
| 368 | 524 | return err; |
|---|
| 369 | 525 | |
|---|
| .. | .. |
|---|
| 371 | 527 | msk += sizeof(struct nfp_flower_in_port); |
|---|
| 372 | 528 | |
|---|
| 373 | 529 | if (NFP_FLOWER_LAYER_MAC & key_ls->key_layer) { |
|---|
| 374 | | - /* Populate Exact MAC Data. */ |
|---|
| 375 | | - nfp_flower_compile_mac((struct nfp_flower_mac_mpls *)ext, |
|---|
| 376 | | - flow, false); |
|---|
| 377 | | - /* Populate Mask MAC Data. */ |
|---|
| 378 | | - nfp_flower_compile_mac((struct nfp_flower_mac_mpls *)msk, |
|---|
| 379 | | - flow, true); |
|---|
| 530 | + err = nfp_flower_compile_mac((struct nfp_flower_mac_mpls *)ext, |
|---|
| 531 | + (struct nfp_flower_mac_mpls *)msk, |
|---|
| 532 | + rule, extack); |
|---|
| 533 | + if (err) |
|---|
| 534 | + return err; |
|---|
| 535 | + |
|---|
| 380 | 536 | ext += sizeof(struct nfp_flower_mac_mpls); |
|---|
| 381 | 537 | msk += sizeof(struct nfp_flower_mac_mpls); |
|---|
| 382 | 538 | } |
|---|
| 383 | 539 | |
|---|
| 384 | 540 | if (NFP_FLOWER_LAYER_TP & key_ls->key_layer) { |
|---|
| 385 | | - /* Populate Exact TP Data. */ |
|---|
| 386 | 541 | nfp_flower_compile_tport((struct nfp_flower_tp_ports *)ext, |
|---|
| 387 | | - flow, false); |
|---|
| 388 | | - /* Populate Mask TP Data. */ |
|---|
| 389 | | - nfp_flower_compile_tport((struct nfp_flower_tp_ports *)msk, |
|---|
| 390 | | - flow, true); |
|---|
| 542 | + (struct nfp_flower_tp_ports *)msk, |
|---|
| 543 | + rule); |
|---|
| 391 | 544 | ext += sizeof(struct nfp_flower_tp_ports); |
|---|
| 392 | 545 | msk += sizeof(struct nfp_flower_tp_ports); |
|---|
| 393 | 546 | } |
|---|
| 394 | 547 | |
|---|
| 395 | 548 | if (NFP_FLOWER_LAYER_IPV4 & key_ls->key_layer) { |
|---|
| 396 | | - /* Populate Exact IPv4 Data. */ |
|---|
| 397 | 549 | nfp_flower_compile_ipv4((struct nfp_flower_ipv4 *)ext, |
|---|
| 398 | | - flow, false); |
|---|
| 399 | | - /* Populate Mask IPv4 Data. */ |
|---|
| 400 | | - nfp_flower_compile_ipv4((struct nfp_flower_ipv4 *)msk, |
|---|
| 401 | | - flow, true); |
|---|
| 550 | + (struct nfp_flower_ipv4 *)msk, |
|---|
| 551 | + rule); |
|---|
| 402 | 552 | ext += sizeof(struct nfp_flower_ipv4); |
|---|
| 403 | 553 | msk += sizeof(struct nfp_flower_ipv4); |
|---|
| 404 | 554 | } |
|---|
| 405 | 555 | |
|---|
| 406 | 556 | if (NFP_FLOWER_LAYER_IPV6 & key_ls->key_layer) { |
|---|
| 407 | | - /* Populate Exact IPv4 Data. */ |
|---|
| 408 | 557 | nfp_flower_compile_ipv6((struct nfp_flower_ipv6 *)ext, |
|---|
| 409 | | - flow, false); |
|---|
| 410 | | - /* Populate Mask IPv4 Data. */ |
|---|
| 411 | | - nfp_flower_compile_ipv6((struct nfp_flower_ipv6 *)msk, |
|---|
| 412 | | - flow, true); |
|---|
| 558 | + (struct nfp_flower_ipv6 *)msk, |
|---|
| 559 | + rule); |
|---|
| 413 | 560 | ext += sizeof(struct nfp_flower_ipv6); |
|---|
| 414 | 561 | msk += sizeof(struct nfp_flower_ipv6); |
|---|
| 415 | 562 | } |
|---|
| 416 | 563 | |
|---|
| 417 | | - if (key_ls->key_layer & NFP_FLOWER_LAYER_VXLAN || |
|---|
| 418 | | - key_ls->key_layer_two & NFP_FLOWER_LAYER2_GENEVE) { |
|---|
| 419 | | - __be32 tun_dst; |
|---|
| 564 | + if (key_ls->key_layer_two & NFP_FLOWER_LAYER2_GRE) { |
|---|
| 565 | + if (key_ls->key_layer_two & NFP_FLOWER_LAYER2_TUN_IPV6) { |
|---|
| 566 | + struct nfp_flower_ipv6_gre_tun *gre_match; |
|---|
| 567 | + struct nfp_ipv6_addr_entry *entry; |
|---|
| 568 | + struct in6_addr *dst; |
|---|
| 420 | 569 | |
|---|
| 421 | | - /* Populate Exact VXLAN Data. */ |
|---|
| 422 | | - nfp_flower_compile_ipv4_udp_tun((void *)ext, flow, false); |
|---|
| 423 | | - /* Populate Mask VXLAN Data. */ |
|---|
| 424 | | - nfp_flower_compile_ipv4_udp_tun((void *)msk, flow, true); |
|---|
| 425 | | - tun_dst = ((struct nfp_flower_ipv4_udp_tun *)ext)->ip_dst; |
|---|
| 426 | | - ext += sizeof(struct nfp_flower_ipv4_udp_tun); |
|---|
| 427 | | - msk += sizeof(struct nfp_flower_ipv4_udp_tun); |
|---|
| 570 | + nfp_flower_compile_ipv6_gre_tun((void *)ext, |
|---|
| 571 | + (void *)msk, rule); |
|---|
| 572 | + gre_match = (struct nfp_flower_ipv6_gre_tun *)ext; |
|---|
| 573 | + dst = &gre_match->ipv6.dst; |
|---|
| 574 | + ext += sizeof(struct nfp_flower_ipv6_gre_tun); |
|---|
| 575 | + msk += sizeof(struct nfp_flower_ipv6_gre_tun); |
|---|
| 428 | 576 | |
|---|
| 429 | | - /* Configure tunnel end point MAC. */ |
|---|
| 430 | | - if (nfp_netdev_is_nfp_repr(netdev)) { |
|---|
| 431 | | - netdev_repr = netdev_priv(netdev); |
|---|
| 432 | | - nfp_tunnel_write_macs(netdev_repr->app); |
|---|
| 577 | + entry = nfp_tunnel_add_ipv6_off(app, dst); |
|---|
| 578 | + if (!entry) |
|---|
| 579 | + return -EOPNOTSUPP; |
|---|
| 580 | + |
|---|
| 581 | + nfp_flow->nfp_tun_ipv6 = entry; |
|---|
| 582 | + } else { |
|---|
| 583 | + __be32 dst; |
|---|
| 584 | + |
|---|
| 585 | + nfp_flower_compile_ipv4_gre_tun((void *)ext, |
|---|
| 586 | + (void *)msk, rule); |
|---|
| 587 | + dst = ((struct nfp_flower_ipv4_gre_tun *)ext)->ipv4.dst; |
|---|
| 588 | + ext += sizeof(struct nfp_flower_ipv4_gre_tun); |
|---|
| 589 | + msk += sizeof(struct nfp_flower_ipv4_gre_tun); |
|---|
| 433 | 590 | |
|---|
| 434 | 591 | /* Store the tunnel destination in the rule data. |
|---|
| 435 | 592 | * This must be present and be an exact match. |
|---|
| 436 | 593 | */ |
|---|
| 437 | | - nfp_flow->nfp_tun_ipv4_addr = tun_dst; |
|---|
| 438 | | - nfp_tunnel_add_ipv4_off(netdev_repr->app, tun_dst); |
|---|
| 594 | + nfp_flow->nfp_tun_ipv4_addr = dst; |
|---|
| 595 | + nfp_tunnel_add_ipv4_off(app, dst); |
|---|
| 596 | + } |
|---|
| 597 | + } |
|---|
| 598 | + |
|---|
| 599 | + if (NFP_FLOWER_LAYER2_QINQ & key_ls->key_layer_two) { |
|---|
| 600 | + nfp_flower_compile_vlan((struct nfp_flower_vlan *)ext, |
|---|
| 601 | + (struct nfp_flower_vlan *)msk, |
|---|
| 602 | + rule); |
|---|
| 603 | + ext += sizeof(struct nfp_flower_vlan); |
|---|
| 604 | + msk += sizeof(struct nfp_flower_vlan); |
|---|
| 605 | + } |
|---|
| 606 | + |
|---|
| 607 | + if (key_ls->key_layer & NFP_FLOWER_LAYER_VXLAN || |
|---|
| 608 | + key_ls->key_layer_two & NFP_FLOWER_LAYER2_GENEVE) { |
|---|
| 609 | + if (key_ls->key_layer_two & NFP_FLOWER_LAYER2_TUN_IPV6) { |
|---|
| 610 | + struct nfp_flower_ipv6_udp_tun *udp_match; |
|---|
| 611 | + struct nfp_ipv6_addr_entry *entry; |
|---|
| 612 | + struct in6_addr *dst; |
|---|
| 613 | + |
|---|
| 614 | + nfp_flower_compile_ipv6_udp_tun((void *)ext, |
|---|
| 615 | + (void *)msk, rule); |
|---|
| 616 | + udp_match = (struct nfp_flower_ipv6_udp_tun *)ext; |
|---|
| 617 | + dst = &udp_match->ipv6.dst; |
|---|
| 618 | + ext += sizeof(struct nfp_flower_ipv6_udp_tun); |
|---|
| 619 | + msk += sizeof(struct nfp_flower_ipv6_udp_tun); |
|---|
| 620 | + |
|---|
| 621 | + entry = nfp_tunnel_add_ipv6_off(app, dst); |
|---|
| 622 | + if (!entry) |
|---|
| 623 | + return -EOPNOTSUPP; |
|---|
| 624 | + |
|---|
| 625 | + nfp_flow->nfp_tun_ipv6 = entry; |
|---|
| 626 | + } else { |
|---|
| 627 | + __be32 dst; |
|---|
| 628 | + |
|---|
| 629 | + nfp_flower_compile_ipv4_udp_tun((void *)ext, |
|---|
| 630 | + (void *)msk, rule); |
|---|
| 631 | + dst = ((struct nfp_flower_ipv4_udp_tun *)ext)->ipv4.dst; |
|---|
| 632 | + ext += sizeof(struct nfp_flower_ipv4_udp_tun); |
|---|
| 633 | + msk += sizeof(struct nfp_flower_ipv4_udp_tun); |
|---|
| 634 | + |
|---|
| 635 | + /* Store the tunnel destination in the rule data. |
|---|
| 636 | + * This must be present and be an exact match. |
|---|
| 637 | + */ |
|---|
| 638 | + nfp_flow->nfp_tun_ipv4_addr = dst; |
|---|
| 639 | + nfp_tunnel_add_ipv4_off(app, dst); |
|---|
| 439 | 640 | } |
|---|
| 440 | 641 | |
|---|
| 441 | 642 | if (key_ls->key_layer_two & NFP_FLOWER_LAYER2_GENEVE_OP) { |
|---|
| 442 | | - err = nfp_flower_compile_geneve_opt(ext, flow, false); |
|---|
| 443 | | - if (err) |
|---|
| 444 | | - return err; |
|---|
| 445 | | - |
|---|
| 446 | | - err = nfp_flower_compile_geneve_opt(msk, flow, true); |
|---|
| 643 | + err = nfp_flower_compile_geneve_opt(ext, msk, rule); |
|---|
| 447 | 644 | if (err) |
|---|
| 448 | 645 | return err; |
|---|
| 449 | 646 | } |
|---|
| 450 | 647 | } |
|---|
| 451 | 648 | |
|---|
| 649 | + /* Check that the flow key does not exceed the maximum limit. |
|---|
| 650 | + * All structures in the key is multiples of 4 bytes, so use u32. |
|---|
| 651 | + */ |
|---|
| 652 | + ext_len = (u32 *)ext - (u32 *)nfp_flow->unmasked_data; |
|---|
| 653 | + if (ext_len > NFP_FLOWER_KEY_MAX_LW) { |
|---|
| 654 | + NL_SET_ERR_MSG_MOD(extack, |
|---|
| 655 | + "unsupported offload: flow key too long"); |
|---|
| 656 | + return -EOPNOTSUPP; |
|---|
| 657 | + } |
|---|
| 658 | + |
|---|
| 452 | 659 | return 0; |
|---|
| 453 | 660 | } |
|---|