/linux-6.1.9/tools/testing/selftests/bpf/progs/ |
D | test_tcp_check_syncookie_kern.c | 28 struct tcphdr *tcph) in gen_syncookie() argument 30 __u32 thlen = tcph->doff * 4; in gen_syncookie() 32 if (tcph->syn && !tcph->ack) { in gen_syncookie() 37 if ((void *)tcph + thlen > data_end) in gen_syncookie() 40 return bpf_tcp_gen_syncookie(sk, iph, ip_size, tcph, thlen); in gen_syncookie() 53 struct tcphdr *tcph; in check_syncookie() local 73 tcph = data + sizeof(struct ethhdr) + sizeof(struct iphdr); in check_syncookie() 74 if (tcph + 1 > data_end) in check_syncookie() 79 tup.ipv4.sport = tcph->source; in check_syncookie() 80 tup.ipv4.dport = tcph->dest; in check_syncookie() [all …]
|
D | cgroup_skb_sk_lookup_kern.c | 31 const struct tcphdr *tcph) in set_tuple() argument 35 tuple->ipv6.sport = tcph->dest; in set_tuple() 36 tuple->ipv6.dport = tcph->source; in set_tuple() 41 const struct tcphdr *tcph) in is_allowed_peer_cg() argument 48 set_tuple(&tuple, ip6h, tcph); in is_allowed_peer_cg() 71 struct tcphdr tcph; in ingress_lookup() local 86 if (bpf_skb_load_bytes(skb, sizeof(ip6h), &tcph, sizeof(tcph))) in ingress_lookup() 92 if (tcph.dest != g_serv_port) in ingress_lookup() 95 return is_allowed_peer_cg(skb, &ip6h, &tcph); in ingress_lookup()
|
D | test_tc_tunnel.c | 93 struct tcphdr tcph; in __encap_ipv4() local 142 &tcph, sizeof(tcph)) < 0) in __encap_ipv4() 145 if (tcph.dest != __bpf_constant_htons(cfg_port)) in __encap_ipv4() 261 struct tcphdr tcph; in __encap_ipv6() local 273 &tcph, sizeof(tcph)) < 0) in __encap_ipv6() 276 if (tcph.dest != __bpf_constant_htons(cfg_port)) in __encap_ipv6()
|
/linux-6.1.9/net/netfilter/ |
D | nf_conntrack_seqadj.c | 76 struct tcphdr *tcph, in nf_ct_sack_block_adjust() argument 106 inet_proto_csum_replace4(&tcph->check, skb, in nf_ct_sack_block_adjust() 108 inet_proto_csum_replace4(&tcph->check, skb, in nf_ct_sack_block_adjust() 122 struct tcphdr *tcph = (void *)skb->data + protoff; in nf_ct_sack_adjust() local 127 optend = protoff + tcph->doff * 4; in nf_ct_sack_adjust() 132 tcph = (void *)skb->data + protoff; in nf_ct_sack_adjust() 154 nf_ct_sack_block_adjust(skb, tcph, optoff + 2, in nf_ct_sack_adjust() 169 struct tcphdr *tcph; in nf_ct_seq_adjust() local 179 if (skb_ensure_writable(skb, protoff + sizeof(*tcph))) in nf_ct_seq_adjust() 182 tcph = (void *)skb->data + protoff; in nf_ct_seq_adjust() [all …]
|
D | xt_TCPMSS.c | 78 struct tcphdr *tcph; in tcpmss_mangle_packet() local 96 tcph = (struct tcphdr *)(skb_network_header(skb) + tcphoff); in tcpmss_mangle_packet() 97 tcp_hdrlen = tcph->doff * 4; in tcpmss_mangle_packet() 116 opt = (u_int8_t *)tcph; in tcpmss_mangle_packet() 133 inet_proto_csum_replace2(&tcph->check, skb, in tcpmss_mangle_packet() 159 tcph = (struct tcphdr *)(skb_network_header(skb) + tcphoff); in tcpmss_mangle_packet() 176 opt = (u_int8_t *)tcph + sizeof(struct tcphdr); in tcpmss_mangle_packet() 179 inet_proto_csum_replace2(&tcph->check, skb, in tcpmss_mangle_packet() 186 inet_proto_csum_replace4(&tcph->check, skb, 0, *((__be32 *)opt), false); in tcpmss_mangle_packet() 188 oldval = ((__be16 *)tcph)[6]; in tcpmss_mangle_packet() [all …]
|
D | nft_exthdr.c | 168 struct tcphdr *tcph; in nft_tcp_header_pointer() local 173 tcph = skb_header_pointer(pkt->skb, nft_thoff(pkt), sizeof(*tcph), buffer); in nft_tcp_header_pointer() 174 if (!tcph) in nft_tcp_header_pointer() 177 *tcphdr_len = __tcp_hdrlen(tcph); in nft_tcp_header_pointer() 178 if (*tcphdr_len < sizeof(*tcph) || *tcphdr_len > len) in nft_tcp_header_pointer() 192 struct tcphdr *tcph; in nft_exthdr_tcp_eval() local 195 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_eval() 196 if (!tcph) in nft_exthdr_tcp_eval() 199 opt = (u8 *)tcph; in nft_exthdr_tcp_eval() 200 for (i = sizeof(*tcph); i < tcphdr_len - 1; i += optl) { in nft_exthdr_tcp_eval() [all …]
|
D | xt_TCPOPTSTRIP.c | 34 struct tcphdr *tcph, _th; in tcpoptstrip_mangle_packet() local 44 tcph = skb_header_pointer(skb, tcphoff, sizeof(_th), &_th); in tcpoptstrip_mangle_packet() 45 if (!tcph) in tcpoptstrip_mangle_packet() 48 tcp_hdrlen = tcph->doff * 4; in tcpoptstrip_mangle_packet() 56 tcph = (struct tcphdr *)(skb_network_header(skb) + tcphoff); in tcpoptstrip_mangle_packet() 57 opt = (u8 *)tcph; in tcpoptstrip_mangle_packet() 79 inet_proto_csum_replace2(&tcph->check, skb, htons(o), in tcpoptstrip_mangle_packet()
|
D | nf_conntrack_proto_tcp.c | 275 static unsigned int get_conntrack_index(const struct tcphdr *tcph) in get_conntrack_index() argument 277 if (tcph->rst) return TCP_RST_SET; in get_conntrack_index() 278 else if (tcph->syn) return (tcph->ack ? TCP_SYNACK_SET : TCP_SYN_SET); in get_conntrack_index() 279 else if (tcph->fin) return TCP_FIN_SET; in get_conntrack_index() 280 else if (tcph->ack) return TCP_ACK_SET; in get_conntrack_index() 316 const struct tcphdr *tcph) in segment_seq_plus_len() argument 320 return (seq + len - dataoff - tcph->doff*4 in segment_seq_plus_len() 321 + (tcph->syn ? 1 : 0) + (tcph->fin ? 1 : 0)); in segment_seq_plus_len() 335 const struct tcphdr *tcph, in tcp_options() argument 340 int length = (tcph->doff*4) - sizeof(struct tcphdr); in tcp_options() [all …]
|
D | nf_nat_helper.c | 95 struct tcphdr *tcph; in __nf_nat_mangle_tcp_packet() local 106 tcph = (void *)skb->data + protoff; in __nf_nat_mangle_tcp_packet() 109 mangle_contents(skb, protoff + tcph->doff*4, in __nf_nat_mangle_tcp_packet() 115 tcph, &tcph->check, datalen, oldlen); in __nf_nat_mangle_tcp_packet() 118 nf_ct_seqadj_set(ct, ctinfo, tcph->seq, in __nf_nat_mangle_tcp_packet()
|
D | nf_flow_table_ip.c | 24 struct tcphdr *tcph; in nf_flow_state_check() local 29 tcph = (void *)(skb_network_header(skb) + thoff); in nf_flow_state_check() 30 if (unlikely(tcph->fin || tcph->rst)) { in nf_flow_state_check() 41 struct tcphdr *tcph; in nf_flow_nat_ip_tcp() local 43 tcph = (void *)(skb_network_header(skb) + thoff); in nf_flow_nat_ip_tcp() 44 inet_proto_csum_replace4(&tcph->check, skb, addr, new_addr, true); in nf_flow_nat_ip_tcp() 435 struct tcphdr *tcph; in nf_flow_nat_ipv6_tcp() local 437 tcph = (void *)(skb_network_header(skb) + thoff); in nf_flow_nat_ipv6_tcp() 438 inet_proto_csum_replace16(&tcph->check, skb, addr->s6_addr32, in nf_flow_nat_ipv6_tcp()
|
D | nft_flow_offload.c | 292 struct tcphdr _tcph, *tcph = NULL; in nft_flow_offload_eval() local 309 tcph = skb_header_pointer(pkt->skb, nft_thoff(pkt), in nft_flow_offload_eval() 311 if (unlikely(!tcph || tcph->fin || tcph->rst || in nft_flow_offload_eval() 355 if (tcph) { in nft_flow_offload_eval()
|
D | nf_conntrack_pptp.c | 497 const struct tcphdr *tcph; in conntrack_pptp_help() local 522 tcph = skb_header_pointer(skb, nexthdr_off, sizeof(_tcph), &_tcph); in conntrack_pptp_help() 523 if (!tcph) in conntrack_pptp_help() 526 nexthdr_off += tcph->doff * 4; in conntrack_pptp_help() 527 datalen = tcplen - tcph->doff * 4; in conntrack_pptp_help()
|
/linux-6.1.9/net/ipv4/netfilter/ |
D | ipt_ECN.c | 47 struct tcphdr _tcph, *tcph; in set_ect_tcp() local 51 tcph = skb_header_pointer(skb, ip_hdrlen(skb), sizeof(_tcph), &_tcph); in set_ect_tcp() 52 if (!tcph) in set_ect_tcp() 56 tcph->ece == einfo->proto.tcp.ece) && in set_ect_tcp() 58 tcph->cwr == einfo->proto.tcp.cwr)) in set_ect_tcp() 61 if (skb_ensure_writable(skb, ip_hdrlen(skb) + sizeof(*tcph))) in set_ect_tcp() 63 tcph = (void *)ip_hdr(skb) + ip_hdrlen(skb); in set_ect_tcp() 65 oldval = ((__be16 *)tcph)[6]; in set_ect_tcp() 67 tcph->ece = einfo->proto.tcp.ece; in set_ect_tcp() 69 tcph->cwr = einfo->proto.tcp.cwr; in set_ect_tcp() [all …]
|
D | nf_reject_ipv4.c | 197 struct tcphdr *tcph; in nf_reject_ip_tcphdr_put() local 200 tcph = skb_put_zero(nskb, sizeof(struct tcphdr)); in nf_reject_ip_tcphdr_put() 201 tcph->source = oth->dest; in nf_reject_ip_tcphdr_put() 202 tcph->dest = oth->source; in nf_reject_ip_tcphdr_put() 203 tcph->doff = sizeof(struct tcphdr) / 4; in nf_reject_ip_tcphdr_put() 206 tcph->seq = oth->ack_seq; in nf_reject_ip_tcphdr_put() 208 tcph->ack_seq = htonl(ntohl(oth->seq) + oth->syn + oth->fin + in nf_reject_ip_tcphdr_put() 211 tcph->ack = 1; in nf_reject_ip_tcphdr_put() 214 tcph->rst = 1; in nf_reject_ip_tcphdr_put() 215 tcph->check = ~tcp_v4_check(sizeof(struct tcphdr), niph->saddr, in nf_reject_ip_tcphdr_put() [all …]
|
/linux-6.1.9/net/netfilter/ipvs/ |
D | ip_vs_proto_tcp.c | 105 tcp_fast_csum_update(int af, struct tcphdr *tcph, in tcp_fast_csum_update() argument 112 tcph->check = in tcp_fast_csum_update() 115 ~csum_unfold(tcph->check)))); in tcp_fast_csum_update() 118 tcph->check = in tcp_fast_csum_update() 121 ~csum_unfold(tcph->check)))); in tcp_fast_csum_update() 126 tcp_partial_csum_update(int af, struct tcphdr *tcph, in tcp_partial_csum_update() argument 133 tcph->check = in tcp_partial_csum_update() 136 csum_unfold(tcph->check)))); in tcp_partial_csum_update() 139 tcph->check = in tcp_partial_csum_update() 142 csum_unfold(tcph->check)))); in tcp_partial_csum_update() [all …]
|
/linux-6.1.9/net/ipv6/netfilter/ |
D | nf_reject_ipv6.c | 225 struct tcphdr *tcph; in nf_reject_ip6_tcphdr_put() local 229 tcph = skb_put(nskb, sizeof(struct tcphdr)); in nf_reject_ip6_tcphdr_put() 231 tcph->doff = sizeof(struct tcphdr)/4; in nf_reject_ip6_tcphdr_put() 232 tcph->source = oth->dest; in nf_reject_ip6_tcphdr_put() 233 tcph->dest = oth->source; in nf_reject_ip6_tcphdr_put() 237 tcph->seq = oth->ack_seq; in nf_reject_ip6_tcphdr_put() 238 tcph->ack_seq = 0; in nf_reject_ip6_tcphdr_put() 241 tcph->ack_seq = htonl(ntohl(oth->seq) + oth->syn + oth->fin + in nf_reject_ip6_tcphdr_put() 243 tcph->seq = 0; in nf_reject_ip6_tcphdr_put() 247 ((u_int8_t *)tcph)[13] = 0; in nf_reject_ip6_tcphdr_put() [all …]
|
/linux-6.1.9/tools/testing/selftests/net/ |
D | gro.c | 258 struct tcphdr *tcph = buf; in fill_transportlayer() local 260 memset(tcph, 0, sizeof(*tcph)); in fill_transportlayer() 262 tcph->source = htons(SPORT); in fill_transportlayer() 263 tcph->dest = htons(DPORT); in fill_transportlayer() 264 tcph->seq = ntohl(START_SEQ + seq_offset); in fill_transportlayer() 265 tcph->ack_seq = ntohl(START_ACK + ack_offset); in fill_transportlayer() 266 tcph->ack = 1; in fill_transportlayer() 267 tcph->fin = fin; in fill_transportlayer() 268 tcph->doff = 5; in fill_transportlayer() 269 tcph->window = htons(TCP_MAXWIN); in fill_transportlayer() [all …]
|
/linux-6.1.9/drivers/infiniband/hw/irdma/ |
D | cm.c | 310 struct tcphdr *tcph; in irdma_form_ah_cm_frame() local 338 pktsize = sizeof(*tcph) + opts_len + hdr_len + pd_len; in irdma_form_ah_cm_frame() 343 sqbuf->tcphlen = sizeof(*tcph) + opts_len; in irdma_form_ah_cm_frame() 346 tcph = (struct tcphdr *)buf; in irdma_form_ah_cm_frame() 347 buf += sizeof(*tcph); in irdma_form_ah_cm_frame() 349 tcph->source = htons(cm_node->loc_port); in irdma_form_ah_cm_frame() 350 tcph->dest = htons(cm_node->rem_port); in irdma_form_ah_cm_frame() 351 tcph->seq = htonl(cm_node->tcp_cntxt.loc_seq_num); in irdma_form_ah_cm_frame() 355 tcph->ack_seq = htonl(cm_node->tcp_cntxt.loc_ack_num); in irdma_form_ah_cm_frame() 356 tcph->ack = 1; in irdma_form_ah_cm_frame() [all …]
|
D | utils.c | 1400 struct tcphdr *tcph = (struct tcphdr *)buf->tcph; in irdma_ieq_get_qp() local 1410 loc_port = ntohs(tcph->dest); in irdma_ieq_get_qp() 1411 rem_port = ntohs(tcph->source); in irdma_ieq_get_qp() 1431 struct tcphdr *tcph = (struct tcphdr *)buf->tcph; in irdma_send_ieq_ack() local 1434 cm_node->tcp_cntxt.loc_seq_num = ntohl(tcph->ack_seq); in irdma_send_ieq_ack() 1493 struct tcphdr *tcph; in irdma_gen1_ieq_update_tcpip_info() local 1501 tcph = (struct tcphdr *)(addr + buf->maclen + iphlen); in irdma_gen1_ieq_update_tcpip_info() 1504 tcph->seq = htonl(seqnum); in irdma_gen1_ieq_update_tcpip_info() 1516 struct tcphdr *tcph; in irdma_ieq_update_tcpip_info() local 1523 tcph = (struct tcphdr *)addr; in irdma_ieq_update_tcpip_info() [all …]
|
/linux-6.1.9/net/core/ |
D | tso.c | 36 struct tcphdr *tcph = (struct tcphdr *)hdr; in tso_build_hdr() local 38 put_unaligned_be32(tso->tcp_seq, &tcph->seq); in tso_build_hdr() 42 tcph->psh = 0; in tso_build_hdr() 43 tcph->fin = 0; in tso_build_hdr() 44 tcph->rst = 0; in tso_build_hdr()
|
/linux-6.1.9/net/sched/ |
D | act_csum.c | 207 struct tcphdr *tcph; in tcf_csum_ipv4_tcp() local 213 tcph = tcf_csum_skb_nextlayer(skb, ihl, ipl, sizeof(*tcph)); in tcf_csum_ipv4_tcp() 214 if (tcph == NULL) in tcf_csum_ipv4_tcp() 218 tcph->check = 0; in tcf_csum_ipv4_tcp() 219 skb->csum = csum_partial(tcph, ipl - ihl, 0); in tcf_csum_ipv4_tcp() 220 tcph->check = tcp_v4_check(ipl - ihl, in tcf_csum_ipv4_tcp() 231 struct tcphdr *tcph; in tcf_csum_ipv6_tcp() local 237 tcph = tcf_csum_skb_nextlayer(skb, ihl, ipl, sizeof(*tcph)); in tcf_csum_ipv6_tcp() 238 if (tcph == NULL) in tcf_csum_ipv6_tcp() 242 tcph->check = 0; in tcf_csum_ipv6_tcp() [all …]
|
D | act_nat.c | 168 struct tcphdr *tcph; in tcf_nat_act() local 170 if (!pskb_may_pull(skb, ihl + sizeof(*tcph) + noff) || in tcf_nat_act() 171 skb_try_make_writable(skb, ihl + sizeof(*tcph) + noff)) in tcf_nat_act() 174 tcph = (void *)(skb_network_header(skb) + ihl); in tcf_nat_act() 175 inet_proto_csum_replace4(&tcph->check, skb, addr, new_addr, in tcf_nat_act()
|
D | sch_cake.c | 906 const struct tcphdr *tcph; in cake_get_tcphdr() local 945 tcph = skb_header_pointer(skb, offset, sizeof(_tcph), &_tcph); in cake_get_tcphdr() 946 if (!tcph || tcph->doff < 5) in cake_get_tcphdr() 950 min(__tcp_hdrlen(tcph), bufsize), buf); in cake_get_tcphdr() 953 static const void *cake_get_tcpopt(const struct tcphdr *tcph, in cake_get_tcpopt() argument 957 int length = __tcp_hdrlen(tcph) - sizeof(struct tcphdr); in cake_get_tcpopt() 958 const u8 *ptr = (const u8 *)(tcph + 1); in cake_get_tcpopt() 1065 static void cake_tcph_get_tstamp(const struct tcphdr *tcph, in cake_tcph_get_tstamp() argument 1071 ptr = cake_get_tcpopt(tcph, TCPOPT_TIMESTAMP, &opsize); in cake_tcph_get_tstamp() 1079 static bool cake_tcph_may_drop(const struct tcphdr *tcph, in cake_tcph_may_drop() argument [all …]
|
D | act_ct.c | 451 struct tcphdr **tcph) in tcf_ct_flow_table_fill_tuple_ipv4() argument 494 *tcph = (void *)(skb_network_header(skb) + thoff); in tcf_ct_flow_table_fill_tuple_ipv4() 524 struct tcphdr **tcph) in tcf_ct_flow_table_fill_tuple_ipv6() argument 563 *tcph = (void *)(skb_network_header(skb) + thoff); in tcf_ct_flow_table_fill_tuple_ipv6() 598 struct tcphdr *tcph = NULL; in tcf_ct_flow_table_lookup() local 605 if (!tcf_ct_flow_table_fill_tuple_ipv4(skb, &tuple, &tcph)) in tcf_ct_flow_table_lookup() 609 if (!tcf_ct_flow_table_fill_tuple_ipv6(skb, &tuple, &tcph)) in tcf_ct_flow_table_lookup() 624 if (tcph && (unlikely(tcph->fin || tcph->rst))) { in tcf_ct_flow_table_lookup()
|
/linux-6.1.9/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
D | chtls_cm.c | 1012 const struct tcphdr *tcph; in chtls_pass_accept_rpl() local 1056 tcph = (struct tcphdr *)((u8 *)(req + 1) + in chtls_pass_accept_rpl() 1058 if (tcph->ece && tcph->cwr) in chtls_pass_accept_rpl() 1299 struct tcphdr *tcph; in chtls_pass_accept_request() local 1354 tcph = (struct tcphdr *)(iph + 1); in chtls_pass_accept_request() 1358 tcph = (struct tcphdr *)(ip6h + 1); in chtls_pass_accept_request() 1373 tcp_rsk(oreq)->rcv_isn = ntohl(tcph->seq); in chtls_pass_accept_request() 1374 chtls_set_req_port(oreq, tcph->source, tcph->dest); in chtls_pass_accept_request() 1391 th_ecn = tcph->ece && tcph->cwr; in chtls_pass_accept_request()
|