Home
last modified time | relevance | path

Searched refs:tcph (Results 1 – 25 of 39) sorted by relevance

12

/linux-6.1.9/tools/testing/selftests/bpf/progs/
Dtest_tcp_check_syncookie_kern.c28 struct tcphdr *tcph) in gen_syncookie() argument
30 __u32 thlen = tcph->doff * 4; in gen_syncookie()
32 if (tcph->syn && !tcph->ack) { in gen_syncookie()
37 if ((void *)tcph + thlen > data_end) in gen_syncookie()
40 return bpf_tcp_gen_syncookie(sk, iph, ip_size, tcph, thlen); in gen_syncookie()
53 struct tcphdr *tcph; in check_syncookie() local
73 tcph = data + sizeof(struct ethhdr) + sizeof(struct iphdr); in check_syncookie()
74 if (tcph + 1 > data_end) in check_syncookie()
79 tup.ipv4.sport = tcph->source; in check_syncookie()
80 tup.ipv4.dport = tcph->dest; in check_syncookie()
[all …]
Dcgroup_skb_sk_lookup_kern.c31 const struct tcphdr *tcph) in set_tuple() argument
35 tuple->ipv6.sport = tcph->dest; in set_tuple()
36 tuple->ipv6.dport = tcph->source; in set_tuple()
41 const struct tcphdr *tcph) in is_allowed_peer_cg() argument
48 set_tuple(&tuple, ip6h, tcph); in is_allowed_peer_cg()
71 struct tcphdr tcph; in ingress_lookup() local
86 if (bpf_skb_load_bytes(skb, sizeof(ip6h), &tcph, sizeof(tcph))) in ingress_lookup()
92 if (tcph.dest != g_serv_port) in ingress_lookup()
95 return is_allowed_peer_cg(skb, &ip6h, &tcph); in ingress_lookup()
Dtest_tc_tunnel.c93 struct tcphdr tcph; in __encap_ipv4() local
142 &tcph, sizeof(tcph)) < 0) in __encap_ipv4()
145 if (tcph.dest != __bpf_constant_htons(cfg_port)) in __encap_ipv4()
261 struct tcphdr tcph; in __encap_ipv6() local
273 &tcph, sizeof(tcph)) < 0) in __encap_ipv6()
276 if (tcph.dest != __bpf_constant_htons(cfg_port)) in __encap_ipv6()
/linux-6.1.9/net/netfilter/
Dnf_conntrack_seqadj.c76 struct tcphdr *tcph, in nf_ct_sack_block_adjust() argument
106 inet_proto_csum_replace4(&tcph->check, skb, in nf_ct_sack_block_adjust()
108 inet_proto_csum_replace4(&tcph->check, skb, in nf_ct_sack_block_adjust()
122 struct tcphdr *tcph = (void *)skb->data + protoff; in nf_ct_sack_adjust() local
127 optend = protoff + tcph->doff * 4; in nf_ct_sack_adjust()
132 tcph = (void *)skb->data + protoff; in nf_ct_sack_adjust()
154 nf_ct_sack_block_adjust(skb, tcph, optoff + 2, in nf_ct_sack_adjust()
169 struct tcphdr *tcph; in nf_ct_seq_adjust() local
179 if (skb_ensure_writable(skb, protoff + sizeof(*tcph))) in nf_ct_seq_adjust()
182 tcph = (void *)skb->data + protoff; in nf_ct_seq_adjust()
[all …]
Dxt_TCPMSS.c78 struct tcphdr *tcph; in tcpmss_mangle_packet() local
96 tcph = (struct tcphdr *)(skb_network_header(skb) + tcphoff); in tcpmss_mangle_packet()
97 tcp_hdrlen = tcph->doff * 4; in tcpmss_mangle_packet()
116 opt = (u_int8_t *)tcph; in tcpmss_mangle_packet()
133 inet_proto_csum_replace2(&tcph->check, skb, in tcpmss_mangle_packet()
159 tcph = (struct tcphdr *)(skb_network_header(skb) + tcphoff); in tcpmss_mangle_packet()
176 opt = (u_int8_t *)tcph + sizeof(struct tcphdr); in tcpmss_mangle_packet()
179 inet_proto_csum_replace2(&tcph->check, skb, in tcpmss_mangle_packet()
186 inet_proto_csum_replace4(&tcph->check, skb, 0, *((__be32 *)opt), false); in tcpmss_mangle_packet()
188 oldval = ((__be16 *)tcph)[6]; in tcpmss_mangle_packet()
[all …]
Dnft_exthdr.c168 struct tcphdr *tcph; in nft_tcp_header_pointer() local
173 tcph = skb_header_pointer(pkt->skb, nft_thoff(pkt), sizeof(*tcph), buffer); in nft_tcp_header_pointer()
174 if (!tcph) in nft_tcp_header_pointer()
177 *tcphdr_len = __tcp_hdrlen(tcph); in nft_tcp_header_pointer()
178 if (*tcphdr_len < sizeof(*tcph) || *tcphdr_len > len) in nft_tcp_header_pointer()
192 struct tcphdr *tcph; in nft_exthdr_tcp_eval() local
195 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_eval()
196 if (!tcph) in nft_exthdr_tcp_eval()
199 opt = (u8 *)tcph; in nft_exthdr_tcp_eval()
200 for (i = sizeof(*tcph); i < tcphdr_len - 1; i += optl) { in nft_exthdr_tcp_eval()
[all …]
Dxt_TCPOPTSTRIP.c34 struct tcphdr *tcph, _th; in tcpoptstrip_mangle_packet() local
44 tcph = skb_header_pointer(skb, tcphoff, sizeof(_th), &_th); in tcpoptstrip_mangle_packet()
45 if (!tcph) in tcpoptstrip_mangle_packet()
48 tcp_hdrlen = tcph->doff * 4; in tcpoptstrip_mangle_packet()
56 tcph = (struct tcphdr *)(skb_network_header(skb) + tcphoff); in tcpoptstrip_mangle_packet()
57 opt = (u8 *)tcph; in tcpoptstrip_mangle_packet()
79 inet_proto_csum_replace2(&tcph->check, skb, htons(o), in tcpoptstrip_mangle_packet()
Dnf_conntrack_proto_tcp.c275 static unsigned int get_conntrack_index(const struct tcphdr *tcph) in get_conntrack_index() argument
277 if (tcph->rst) return TCP_RST_SET; in get_conntrack_index()
278 else if (tcph->syn) return (tcph->ack ? TCP_SYNACK_SET : TCP_SYN_SET); in get_conntrack_index()
279 else if (tcph->fin) return TCP_FIN_SET; in get_conntrack_index()
280 else if (tcph->ack) return TCP_ACK_SET; in get_conntrack_index()
316 const struct tcphdr *tcph) in segment_seq_plus_len() argument
320 return (seq + len - dataoff - tcph->doff*4 in segment_seq_plus_len()
321 + (tcph->syn ? 1 : 0) + (tcph->fin ? 1 : 0)); in segment_seq_plus_len()
335 const struct tcphdr *tcph, in tcp_options() argument
340 int length = (tcph->doff*4) - sizeof(struct tcphdr); in tcp_options()
[all …]
Dnf_nat_helper.c95 struct tcphdr *tcph; in __nf_nat_mangle_tcp_packet() local
106 tcph = (void *)skb->data + protoff; in __nf_nat_mangle_tcp_packet()
109 mangle_contents(skb, protoff + tcph->doff*4, in __nf_nat_mangle_tcp_packet()
115 tcph, &tcph->check, datalen, oldlen); in __nf_nat_mangle_tcp_packet()
118 nf_ct_seqadj_set(ct, ctinfo, tcph->seq, in __nf_nat_mangle_tcp_packet()
Dnf_flow_table_ip.c24 struct tcphdr *tcph; in nf_flow_state_check() local
29 tcph = (void *)(skb_network_header(skb) + thoff); in nf_flow_state_check()
30 if (unlikely(tcph->fin || tcph->rst)) { in nf_flow_state_check()
41 struct tcphdr *tcph; in nf_flow_nat_ip_tcp() local
43 tcph = (void *)(skb_network_header(skb) + thoff); in nf_flow_nat_ip_tcp()
44 inet_proto_csum_replace4(&tcph->check, skb, addr, new_addr, true); in nf_flow_nat_ip_tcp()
435 struct tcphdr *tcph; in nf_flow_nat_ipv6_tcp() local
437 tcph = (void *)(skb_network_header(skb) + thoff); in nf_flow_nat_ipv6_tcp()
438 inet_proto_csum_replace16(&tcph->check, skb, addr->s6_addr32, in nf_flow_nat_ipv6_tcp()
Dnft_flow_offload.c292 struct tcphdr _tcph, *tcph = NULL; in nft_flow_offload_eval() local
309 tcph = skb_header_pointer(pkt->skb, nft_thoff(pkt), in nft_flow_offload_eval()
311 if (unlikely(!tcph || tcph->fin || tcph->rst || in nft_flow_offload_eval()
355 if (tcph) { in nft_flow_offload_eval()
Dnf_conntrack_pptp.c497 const struct tcphdr *tcph; in conntrack_pptp_help() local
522 tcph = skb_header_pointer(skb, nexthdr_off, sizeof(_tcph), &_tcph); in conntrack_pptp_help()
523 if (!tcph) in conntrack_pptp_help()
526 nexthdr_off += tcph->doff * 4; in conntrack_pptp_help()
527 datalen = tcplen - tcph->doff * 4; in conntrack_pptp_help()
/linux-6.1.9/net/ipv4/netfilter/
Dipt_ECN.c47 struct tcphdr _tcph, *tcph; in set_ect_tcp() local
51 tcph = skb_header_pointer(skb, ip_hdrlen(skb), sizeof(_tcph), &_tcph); in set_ect_tcp()
52 if (!tcph) in set_ect_tcp()
56 tcph->ece == einfo->proto.tcp.ece) && in set_ect_tcp()
58 tcph->cwr == einfo->proto.tcp.cwr)) in set_ect_tcp()
61 if (skb_ensure_writable(skb, ip_hdrlen(skb) + sizeof(*tcph))) in set_ect_tcp()
63 tcph = (void *)ip_hdr(skb) + ip_hdrlen(skb); in set_ect_tcp()
65 oldval = ((__be16 *)tcph)[6]; in set_ect_tcp()
67 tcph->ece = einfo->proto.tcp.ece; in set_ect_tcp()
69 tcph->cwr = einfo->proto.tcp.cwr; in set_ect_tcp()
[all …]
Dnf_reject_ipv4.c197 struct tcphdr *tcph; in nf_reject_ip_tcphdr_put() local
200 tcph = skb_put_zero(nskb, sizeof(struct tcphdr)); in nf_reject_ip_tcphdr_put()
201 tcph->source = oth->dest; in nf_reject_ip_tcphdr_put()
202 tcph->dest = oth->source; in nf_reject_ip_tcphdr_put()
203 tcph->doff = sizeof(struct tcphdr) / 4; in nf_reject_ip_tcphdr_put()
206 tcph->seq = oth->ack_seq; in nf_reject_ip_tcphdr_put()
208 tcph->ack_seq = htonl(ntohl(oth->seq) + oth->syn + oth->fin + in nf_reject_ip_tcphdr_put()
211 tcph->ack = 1; in nf_reject_ip_tcphdr_put()
214 tcph->rst = 1; in nf_reject_ip_tcphdr_put()
215 tcph->check = ~tcp_v4_check(sizeof(struct tcphdr), niph->saddr, in nf_reject_ip_tcphdr_put()
[all …]
/linux-6.1.9/net/netfilter/ipvs/
Dip_vs_proto_tcp.c105 tcp_fast_csum_update(int af, struct tcphdr *tcph, in tcp_fast_csum_update() argument
112 tcph->check = in tcp_fast_csum_update()
115 ~csum_unfold(tcph->check)))); in tcp_fast_csum_update()
118 tcph->check = in tcp_fast_csum_update()
121 ~csum_unfold(tcph->check)))); in tcp_fast_csum_update()
126 tcp_partial_csum_update(int af, struct tcphdr *tcph, in tcp_partial_csum_update() argument
133 tcph->check = in tcp_partial_csum_update()
136 csum_unfold(tcph->check)))); in tcp_partial_csum_update()
139 tcph->check = in tcp_partial_csum_update()
142 csum_unfold(tcph->check)))); in tcp_partial_csum_update()
[all …]
/linux-6.1.9/net/ipv6/netfilter/
Dnf_reject_ipv6.c225 struct tcphdr *tcph; in nf_reject_ip6_tcphdr_put() local
229 tcph = skb_put(nskb, sizeof(struct tcphdr)); in nf_reject_ip6_tcphdr_put()
231 tcph->doff = sizeof(struct tcphdr)/4; in nf_reject_ip6_tcphdr_put()
232 tcph->source = oth->dest; in nf_reject_ip6_tcphdr_put()
233 tcph->dest = oth->source; in nf_reject_ip6_tcphdr_put()
237 tcph->seq = oth->ack_seq; in nf_reject_ip6_tcphdr_put()
238 tcph->ack_seq = 0; in nf_reject_ip6_tcphdr_put()
241 tcph->ack_seq = htonl(ntohl(oth->seq) + oth->syn + oth->fin + in nf_reject_ip6_tcphdr_put()
243 tcph->seq = 0; in nf_reject_ip6_tcphdr_put()
247 ((u_int8_t *)tcph)[13] = 0; in nf_reject_ip6_tcphdr_put()
[all …]
/linux-6.1.9/tools/testing/selftests/net/
Dgro.c258 struct tcphdr *tcph = buf; in fill_transportlayer() local
260 memset(tcph, 0, sizeof(*tcph)); in fill_transportlayer()
262 tcph->source = htons(SPORT); in fill_transportlayer()
263 tcph->dest = htons(DPORT); in fill_transportlayer()
264 tcph->seq = ntohl(START_SEQ + seq_offset); in fill_transportlayer()
265 tcph->ack_seq = ntohl(START_ACK + ack_offset); in fill_transportlayer()
266 tcph->ack = 1; in fill_transportlayer()
267 tcph->fin = fin; in fill_transportlayer()
268 tcph->doff = 5; in fill_transportlayer()
269 tcph->window = htons(TCP_MAXWIN); in fill_transportlayer()
[all …]
/linux-6.1.9/drivers/infiniband/hw/irdma/
Dcm.c310 struct tcphdr *tcph; in irdma_form_ah_cm_frame() local
338 pktsize = sizeof(*tcph) + opts_len + hdr_len + pd_len; in irdma_form_ah_cm_frame()
343 sqbuf->tcphlen = sizeof(*tcph) + opts_len; in irdma_form_ah_cm_frame()
346 tcph = (struct tcphdr *)buf; in irdma_form_ah_cm_frame()
347 buf += sizeof(*tcph); in irdma_form_ah_cm_frame()
349 tcph->source = htons(cm_node->loc_port); in irdma_form_ah_cm_frame()
350 tcph->dest = htons(cm_node->rem_port); in irdma_form_ah_cm_frame()
351 tcph->seq = htonl(cm_node->tcp_cntxt.loc_seq_num); in irdma_form_ah_cm_frame()
355 tcph->ack_seq = htonl(cm_node->tcp_cntxt.loc_ack_num); in irdma_form_ah_cm_frame()
356 tcph->ack = 1; in irdma_form_ah_cm_frame()
[all …]
Dutils.c1400 struct tcphdr *tcph = (struct tcphdr *)buf->tcph; in irdma_ieq_get_qp() local
1410 loc_port = ntohs(tcph->dest); in irdma_ieq_get_qp()
1411 rem_port = ntohs(tcph->source); in irdma_ieq_get_qp()
1431 struct tcphdr *tcph = (struct tcphdr *)buf->tcph; in irdma_send_ieq_ack() local
1434 cm_node->tcp_cntxt.loc_seq_num = ntohl(tcph->ack_seq); in irdma_send_ieq_ack()
1493 struct tcphdr *tcph; in irdma_gen1_ieq_update_tcpip_info() local
1501 tcph = (struct tcphdr *)(addr + buf->maclen + iphlen); in irdma_gen1_ieq_update_tcpip_info()
1504 tcph->seq = htonl(seqnum); in irdma_gen1_ieq_update_tcpip_info()
1516 struct tcphdr *tcph; in irdma_ieq_update_tcpip_info() local
1523 tcph = (struct tcphdr *)addr; in irdma_ieq_update_tcpip_info()
[all …]
/linux-6.1.9/net/core/
Dtso.c36 struct tcphdr *tcph = (struct tcphdr *)hdr; in tso_build_hdr() local
38 put_unaligned_be32(tso->tcp_seq, &tcph->seq); in tso_build_hdr()
42 tcph->psh = 0; in tso_build_hdr()
43 tcph->fin = 0; in tso_build_hdr()
44 tcph->rst = 0; in tso_build_hdr()
/linux-6.1.9/net/sched/
Dact_csum.c207 struct tcphdr *tcph; in tcf_csum_ipv4_tcp() local
213 tcph = tcf_csum_skb_nextlayer(skb, ihl, ipl, sizeof(*tcph)); in tcf_csum_ipv4_tcp()
214 if (tcph == NULL) in tcf_csum_ipv4_tcp()
218 tcph->check = 0; in tcf_csum_ipv4_tcp()
219 skb->csum = csum_partial(tcph, ipl - ihl, 0); in tcf_csum_ipv4_tcp()
220 tcph->check = tcp_v4_check(ipl - ihl, in tcf_csum_ipv4_tcp()
231 struct tcphdr *tcph; in tcf_csum_ipv6_tcp() local
237 tcph = tcf_csum_skb_nextlayer(skb, ihl, ipl, sizeof(*tcph)); in tcf_csum_ipv6_tcp()
238 if (tcph == NULL) in tcf_csum_ipv6_tcp()
242 tcph->check = 0; in tcf_csum_ipv6_tcp()
[all …]
Dact_nat.c168 struct tcphdr *tcph; in tcf_nat_act() local
170 if (!pskb_may_pull(skb, ihl + sizeof(*tcph) + noff) || in tcf_nat_act()
171 skb_try_make_writable(skb, ihl + sizeof(*tcph) + noff)) in tcf_nat_act()
174 tcph = (void *)(skb_network_header(skb) + ihl); in tcf_nat_act()
175 inet_proto_csum_replace4(&tcph->check, skb, addr, new_addr, in tcf_nat_act()
Dsch_cake.c906 const struct tcphdr *tcph; in cake_get_tcphdr() local
945 tcph = skb_header_pointer(skb, offset, sizeof(_tcph), &_tcph); in cake_get_tcphdr()
946 if (!tcph || tcph->doff < 5) in cake_get_tcphdr()
950 min(__tcp_hdrlen(tcph), bufsize), buf); in cake_get_tcphdr()
953 static const void *cake_get_tcpopt(const struct tcphdr *tcph, in cake_get_tcpopt() argument
957 int length = __tcp_hdrlen(tcph) - sizeof(struct tcphdr); in cake_get_tcpopt()
958 const u8 *ptr = (const u8 *)(tcph + 1); in cake_get_tcpopt()
1065 static void cake_tcph_get_tstamp(const struct tcphdr *tcph, in cake_tcph_get_tstamp() argument
1071 ptr = cake_get_tcpopt(tcph, TCPOPT_TIMESTAMP, &opsize); in cake_tcph_get_tstamp()
1079 static bool cake_tcph_may_drop(const struct tcphdr *tcph, in cake_tcph_may_drop() argument
[all …]
Dact_ct.c451 struct tcphdr **tcph) in tcf_ct_flow_table_fill_tuple_ipv4() argument
494 *tcph = (void *)(skb_network_header(skb) + thoff); in tcf_ct_flow_table_fill_tuple_ipv4()
524 struct tcphdr **tcph) in tcf_ct_flow_table_fill_tuple_ipv6() argument
563 *tcph = (void *)(skb_network_header(skb) + thoff); in tcf_ct_flow_table_fill_tuple_ipv6()
598 struct tcphdr *tcph = NULL; in tcf_ct_flow_table_lookup() local
605 if (!tcf_ct_flow_table_fill_tuple_ipv4(skb, &tuple, &tcph)) in tcf_ct_flow_table_lookup()
609 if (!tcf_ct_flow_table_fill_tuple_ipv6(skb, &tuple, &tcph)) in tcf_ct_flow_table_lookup()
624 if (tcph && (unlikely(tcph->fin || tcph->rst))) { in tcf_ct_flow_table_lookup()
/linux-6.1.9/drivers/net/ethernet/chelsio/inline_crypto/chtls/
Dchtls_cm.c1012 const struct tcphdr *tcph; in chtls_pass_accept_rpl() local
1056 tcph = (struct tcphdr *)((u8 *)(req + 1) + in chtls_pass_accept_rpl()
1058 if (tcph->ece && tcph->cwr) in chtls_pass_accept_rpl()
1299 struct tcphdr *tcph; in chtls_pass_accept_request() local
1354 tcph = (struct tcphdr *)(iph + 1); in chtls_pass_accept_request()
1358 tcph = (struct tcphdr *)(ip6h + 1); in chtls_pass_accept_request()
1373 tcp_rsk(oreq)->rcv_isn = ntohl(tcph->seq); in chtls_pass_accept_request()
1374 chtls_set_req_port(oreq, tcph->source, tcph->dest); in chtls_pass_accept_request()
1391 th_ecn = tcph->ece && tcph->cwr; in chtls_pass_accept_request()

12