Home
last modified time | relevance | path

Searched refs:eseg (Results 1 – 10 of 10) sorted by relevance

/linux-6.1.9/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
Dipsec_rxtx.h80 static inline bool mlx5e_ipsec_eseg_meta(struct mlx5_wqe_eth_seg *eseg) in mlx5e_ipsec_eseg_meta() argument
82 return eseg->flow_table_metadata & cpu_to_be32(MLX5_ETH_WQE_FT_META_IPSEC); in mlx5e_ipsec_eseg_meta()
86 struct mlx5_wqe_eth_seg *eseg);
124 struct mlx5_wqe_eth_seg *eseg) in mlx5e_ipsec_txwqe_build_eseg_csum() argument
128 if (!mlx5e_ipsec_eseg_meta(eseg)) in mlx5e_ipsec_txwqe_build_eseg_csum()
131 eseg->cs_flags = MLX5_ETH_WQE_L3_CSUM; in mlx5e_ipsec_txwqe_build_eseg_csum()
134 eseg->cs_flags |= MLX5_ETH_WQE_L3_INNER_CSUM; in mlx5e_ipsec_txwqe_build_eseg_csum()
136 eseg->cs_flags |= MLX5_ETH_WQE_L4_INNER_CSUM; in mlx5e_ipsec_txwqe_build_eseg_csum()
138 eseg->cs_flags |= MLX5_ETH_WQE_L4_CSUM; in mlx5e_ipsec_txwqe_build_eseg_csum()
151 static inline bool mlx5e_ipsec_eseg_meta(struct mlx5_wqe_eth_seg *eseg) in mlx5e_ipsec_eseg_meta() argument
[all …]
Dipsec_rxtx.c73 struct mlx5_wqe_eth_seg *eseg, u8 mode, in mlx5e_ipsec_set_swp() argument
90 eseg->swp_outer_l3_offset = skb_network_offset(skb) / 2; in mlx5e_ipsec_set_swp()
92 eseg->swp_flags |= MLX5_ETH_WQE_SWP_OUTER_L3_IPV6; in mlx5e_ipsec_set_swp()
96 eseg->swp_inner_l3_offset = skb_inner_network_offset(skb) / 2; in mlx5e_ipsec_set_swp()
98 eseg->swp_flags |= MLX5_ETH_WQE_SWP_INNER_L3_IPV6; in mlx5e_ipsec_set_swp()
102 eseg->swp_flags |= MLX5_ETH_WQE_SWP_INNER_L4_UDP; in mlx5e_ipsec_set_swp()
106 eseg->swp_inner_l4_offset = skb_inner_transport_offset(skb) / 2; in mlx5e_ipsec_set_swp()
121 eseg->swp_flags |= MLX5_ETH_WQE_SWP_OUTER_L4_UDP; in mlx5e_ipsec_set_swp()
125 eseg->swp_outer_l4_offset = skb_inner_transport_offset(skb) / 2; in mlx5e_ipsec_set_swp()
134 eseg->swp_flags |= MLX5_ETH_WQE_SWP_INNER_L4_UDP; in mlx5e_ipsec_set_swp()
[all …]
Den_accel.h55 mlx5e_tx_tunnel_accel(struct sk_buff *skb, struct mlx5_wqe_eth_seg *eseg, u16 ihs) in mlx5e_tx_tunnel_accel() argument
88 mlx5e_set_eseg_swp(skb, eseg, &swp_spec); in mlx5e_tx_tunnel_accel()
90 mlx5e_eseg_swp_offsets_add_vlan(eseg); in mlx5e_tx_tunnel_accel()
169 struct mlx5_wqe_eth_seg *eseg, u16 ihs) in mlx5e_accel_tx_eseg() argument
173 mlx5e_ipsec_tx_build_eseg(priv, skb, eseg); in mlx5e_accel_tx_eseg()
178 mlx5e_macsec_tx_build_eseg(priv->macsec, skb, eseg); in mlx5e_accel_tx_eseg()
183 mlx5e_tx_tunnel_accel(skb, eseg, ihs); in mlx5e_accel_tx_eseg()
Dmacsec.h39 struct mlx5_wqe_eth_seg *eseg);
Dmacsec.c1719 struct mlx5_wqe_eth_seg *eseg) in mlx5e_macsec_tx_build_eseg() argument
1728 eseg->flow_table_metadata = cpu_to_be32(MLX5_ETH_WQE_FT_META_MACSEC | fs_id << 2); in mlx5e_macsec_tx_build_eseg()
/linux-6.1.9/drivers/net/ethernet/mellanox/mlx5/core/
Den_tx.c121 struct mlx5_wqe_eth_seg *eseg) in mlx5e_txwqe_build_eseg_csum() argument
123 if (unlikely(mlx5e_ipsec_txwqe_build_eseg_csum(sq, skb, eseg))) in mlx5e_txwqe_build_eseg_csum()
127 eseg->cs_flags = MLX5_ETH_WQE_L3_CSUM; in mlx5e_txwqe_build_eseg_csum()
129 eseg->cs_flags |= MLX5_ETH_WQE_L3_INNER_CSUM | in mlx5e_txwqe_build_eseg_csum()
133 eseg->cs_flags |= MLX5_ETH_WQE_L4_CSUM; in mlx5e_txwqe_build_eseg_csum()
138 eseg->cs_flags = MLX5_ETH_WQE_L3_CSUM | MLX5_ETH_WQE_L4_CSUM; in mlx5e_txwqe_build_eseg_csum()
419 struct mlx5_wqe_eth_seg *eseg; in mlx5e_sq_xmit_wqe() local
432 eseg = &wqe->eth; in mlx5e_sq_xmit_wqe()
435 eseg->mss = attr->mss; in mlx5e_sq_xmit_wqe()
438 u8 *start = eseg->inline_hdr.start; in mlx5e_sq_xmit_wqe()
[all …]
Den_main.c1835 struct mlx5_wqe_eth_seg *eseg = &wqe->eth; in mlx5e_open_xdpsq() local
1844 eseg->inline_hdr.sz = cpu_to_be16(inline_hdr_sz); in mlx5e_open_xdpsq()
/linux-6.1.9/drivers/net/ethernet/mellanox/mlx5/core/en/
Dtxrx.h402 static inline void mlx5e_eseg_swp_offsets_add_vlan(struct mlx5_wqe_eth_seg *eseg) in mlx5e_eseg_swp_offsets_add_vlan() argument
405 eseg->swp_outer_l3_offset += VLAN_HLEN / 2; in mlx5e_eseg_swp_offsets_add_vlan()
406 eseg->swp_outer_l4_offset += VLAN_HLEN / 2; in mlx5e_eseg_swp_offsets_add_vlan()
407 eseg->swp_inner_l3_offset += VLAN_HLEN / 2; in mlx5e_eseg_swp_offsets_add_vlan()
408 eseg->swp_inner_l4_offset += VLAN_HLEN / 2; in mlx5e_eseg_swp_offsets_add_vlan()
412 mlx5e_set_eseg_swp(struct sk_buff *skb, struct mlx5_wqe_eth_seg *eseg, in mlx5e_set_eseg_swp() argument
416 eseg->swp_outer_l3_offset = skb_network_offset(skb) / 2; in mlx5e_set_eseg_swp()
418 eseg->swp_flags |= MLX5_ETH_WQE_SWP_OUTER_L3_IPV6; in mlx5e_set_eseg_swp()
420 eseg->swp_outer_l4_offset = skb_transport_offset(skb) / 2; in mlx5e_set_eseg_swp()
422 eseg->swp_flags |= MLX5_ETH_WQE_SWP_OUTER_L4_UDP; in mlx5e_set_eseg_swp()
[all …]
Dxdp.c367 struct mlx5_wqe_eth_seg *eseg; in mlx5e_xmit_xdp_frame() local
412 eseg = &wqe->eth; in mlx5e_xmit_xdp_frame()
419 memcpy(eseg->inline_hdr.start, xdptxd->data, sizeof(eseg->inline_hdr.start)); in mlx5e_xmit_xdp_frame()
420 memcpy(dseg, xdptxd->data + sizeof(eseg->inline_hdr.start), in mlx5e_xmit_xdp_frame()
421 MLX5E_XDP_MIN_INLINE - sizeof(eseg->inline_hdr.start)); in mlx5e_xmit_xdp_frame()
439 memset(eseg, 0, sizeof(*eseg) - sizeof(eseg->trailer)); in mlx5e_xmit_xdp_frame()
441 eseg->inline_hdr.sz = cpu_to_be16(inline_hdr_sz); in mlx5e_xmit_xdp_frame()
/linux-6.1.9/drivers/infiniband/hw/mlx5/
Dwr.c57 struct mlx5_wqe_eth_seg *eseg = *seg; in set_eth_seg() local
59 memset(eseg, 0, sizeof(struct mlx5_wqe_eth_seg)); in set_eth_seg()
62 eseg->cs_flags = MLX5_ETH_WQE_L3_CSUM | in set_eth_seg()
72 eseg->mss = cpu_to_be16(ud_wr->mss); in set_eth_seg()
73 eseg->inline_hdr.sz = cpu_to_be16(left); in set_eth_seg()
79 copysz = min_t(u64, *cur_edge - (void *)eseg->inline_hdr.start, in set_eth_seg()
81 memcpy(eseg->inline_hdr.start, pdata, copysz); in set_eth_seg()
83 sizeof(eseg->inline_hdr.start) + copysz, 16); in set_eth_seg()