/linux-3.4.99/net/core/ |
D | skbuff.c | 2193 struct sk_buff* skb1, in skb_split_inside_header() argument 2198 skb_copy_from_linear_data_offset(skb, len, skb_put(skb1, pos - len), in skb_split_inside_header() 2202 skb_shinfo(skb1)->frags[i] = skb_shinfo(skb)->frags[i]; in skb_split_inside_header() 2204 skb_shinfo(skb1)->nr_frags = skb_shinfo(skb)->nr_frags; in skb_split_inside_header() 2206 skb1->data_len = skb->data_len; in skb_split_inside_header() 2207 skb1->len += skb1->data_len; in skb_split_inside_header() 2214 struct sk_buff* skb1, in skb_split_no_header() argument 2221 skb1->len = skb1->data_len = skb->len - len; in skb_split_no_header() 2229 skb_shinfo(skb1)->frags[k] = skb_shinfo(skb)->frags[i]; in skb_split_no_header() 2241 skb_shinfo(skb1)->frags[0].page_offset += len - pos; in skb_split_no_header() [all …]
|
/linux-3.4.99/net/sched/ |
D | sch_choke.c | 170 static bool choke_match_flow(struct sk_buff *skb1, in choke_match_flow() argument 173 if (skb1->protocol != skb2->protocol) in choke_match_flow() 176 if (!choke_skb_cb(skb1)->keys_valid) { in choke_match_flow() 177 choke_skb_cb(skb1)->keys_valid = 1; in choke_match_flow() 178 skb_flow_dissect(skb1, &choke_skb_cb(skb1)->keys); in choke_match_flow() 186 return !memcmp(&choke_skb_cb(skb1)->keys, in choke_match_flow()
|
/linux-3.4.99/net/batman-adv/ |
D | send.c | 266 struct sk_buff *skb1; in send_outstanding_bcast_packet() local 284 skb1 = skb_clone(forw_packet->skb, GFP_ATOMIC); in send_outstanding_bcast_packet() 285 if (skb1) in send_outstanding_bcast_packet() 286 send_skb_packet(skb1, hard_iface, broadcast_addr); in send_outstanding_bcast_packet()
|
/linux-3.4.99/net/llc/ |
D | llc_sap.c | 371 struct sk_buff *skb1; in llc_do_mcast() local 375 skb1 = skb_clone(skb, GFP_ATOMIC); in llc_do_mcast() 376 if (!skb1) { in llc_do_mcast() 381 llc_sap_rcv(sap, skb1, stack[i]); in llc_do_mcast()
|
/linux-3.4.99/net/ipx/ |
D | af_ipx.c | 400 struct sk_buff *skb1; in ipxitf_demux_socket() local 403 skb1 = skb_clone(skb, GFP_ATOMIC); in ipxitf_demux_socket() 405 if (!skb1) in ipxitf_demux_socket() 408 skb1 = skb; in ipxitf_demux_socket() 411 ipxitf_def_skb_handler(s, skb1); in ipxitf_demux_socket() 471 struct sk_buff *skb1 = NULL, *skb2 = NULL; in ipxitf_demux_socket() local 519 skb1 = skb_clone(skb, GFP_ATOMIC); in ipxitf_demux_socket() 521 skb1 = skb; in ipxitf_demux_socket() 524 if (!skb1) in ipxitf_demux_socket() 529 skb2 = skb_clone(skb1, GFP_ATOMIC); in ipxitf_demux_socket() [all …]
|
/linux-3.4.99/net/ipv4/ |
D | tcp_input.c | 4500 struct sk_buff *skb1; in tcp_data_queue_ofo() local 4518 skb1 = skb_peek_tail(&tp->out_of_order_queue); in tcp_data_queue_ofo() 4519 if (!skb1) { in tcp_data_queue_ofo() 4534 if (seq == TCP_SKB_CB(skb1)->end_seq) { in tcp_data_queue_ofo() 4540 if (skb->len <= skb_tailroom(skb1) && !tcp_hdr(skb)->fin) { in tcp_data_queue_ofo() 4544 skb_put(skb1, skb->len), in tcp_data_queue_ofo() 4546 TCP_SKB_CB(skb1)->end_seq = end_seq; in tcp_data_queue_ofo() 4547 TCP_SKB_CB(skb1)->ack_seq = TCP_SKB_CB(skb)->ack_seq; in tcp_data_queue_ofo() 4551 __skb_queue_after(&tp->out_of_order_queue, skb1, skb); in tcp_data_queue_ofo() 4565 if (!after(TCP_SKB_CB(skb1)->seq, seq)) in tcp_data_queue_ofo() [all …]
|
D | udp.c | 1501 struct sk_buff *skb1 = NULL; in flush_stack() local 1506 if (likely(skb1 == NULL)) in flush_stack() 1507 skb1 = (i == final) ? skb : skb_clone(skb, GFP_ATOMIC); in flush_stack() 1509 if (!skb1) { in flush_stack() 1517 if (skb1 && udp_queue_rcv_skb(sk, skb1) <= 0) in flush_stack() 1518 skb1 = NULL; in flush_stack() 1520 if (unlikely(skb1)) in flush_stack() 1521 kfree_skb(skb1); in flush_stack()
|
D | icmp.c | 311 struct sk_buff *skb1; in icmp_push_reply() local 313 skb_queue_walk(&sk->sk_write_queue, skb1) { in icmp_push_reply() 314 csum = csum_add(csum, skb1->csum); in icmp_push_reply()
|
/linux-3.4.99/net/ipv6/ |
D | udp.c | 604 struct sk_buff *skb1; in flush_stack() local 607 skb1 = (i == final) ? skb : skb_clone(skb, GFP_ATOMIC); in flush_stack() 610 if (skb1) { in flush_stack() 611 if (sk_rcvqueues_full(sk, skb1)) { in flush_stack() 612 kfree_skb(skb1); in flush_stack() 617 udpv6_queue_rcv_skb(sk, skb1); in flush_stack() 618 else if (sk_add_backlog(sk, skb1)) { in flush_stack() 619 kfree_skb(skb1); in flush_stack()
|
/linux-3.4.99/Documentation/networking/ |
D | x25-iface.txt | 78 preserved. Even if a device driver calls netif_rx(skb1) and later 80 earlier that skb1.
|
/linux-3.4.99/drivers/net/wireless/ath/ath6kl/ |
D | txrx.c | 1289 struct sk_buff *skb1 = NULL; in ath6kl_rx() local 1539 skb1 = skb_copy(skb, GFP_ATOMIC); in ath6kl_rx() 1550 skb1 = skb; in ath6kl_rx() 1557 if (skb1) in ath6kl_rx() 1558 ath6kl_data_tx(skb1, vif->ndev); in ath6kl_rx()
|
/linux-3.4.99/drivers/atm/ |
D | iphase.c | 639 struct sk_buff *skb = NULL, *skb1 = NULL; in ia_tx_poll() local 664 skb1 = skb_dequeue(&iavcc->txing_skb); in ia_tx_poll() 665 while (skb1 && (skb1 != skb)) { in ia_tx_poll() 666 if (!(IA_SKB_STATE(skb1) & IA_TX_DONE)) { in ia_tx_poll() 670 if ((vcc->pop) && (skb1->len != 0)) in ia_tx_poll() 672 vcc->pop(vcc, skb1); in ia_tx_poll() 674 (long)skb1);) in ia_tx_poll() 677 dev_kfree_skb_any(skb1); in ia_tx_poll() 678 skb1 = skb_dequeue(&iavcc->txing_skb); in ia_tx_poll() 680 if (!skb1) { in ia_tx_poll()
|
/linux-3.4.99/include/net/ |
D | tcp.h | 955 struct sk_buff *skb1; in tcp_prequeue() local 959 while ((skb1 = __skb_dequeue(&tp->ucopy.prequeue)) != NULL) { in tcp_prequeue() 960 sk_backlog_rcv(sk, skb1); in tcp_prequeue()
|
/linux-3.4.99/drivers/net/ethernet/amd/ |
D | ni65.c | 1103 struct sk_buff *skb1 = p->recv_skb[p->rmdnum]; in ni65_recv_intr() local 1107 skb = skb1; in ni65_recv_intr()
|
/linux-3.4.99/drivers/net/ethernet/qlogic/ |
D | qla3xxx.c | 2050 struct sk_buff *skb1 = NULL, *skb2; in ql_process_macip_rx_intr() local 2064 skb1 = lrg_buf_cb1->skb; in ql_process_macip_rx_intr() 2066 if (*((u16 *) skb1->data) != 0xFFFF) in ql_process_macip_rx_intr() 2087 skb_copy_from_linear_data_offset(skb1, VLAN_ID_LEN, in ql_process_macip_rx_intr()
|
/linux-3.4.99/include/linux/ |
D | skbuff.h | 2181 struct sk_buff *skb1, const u32 len);
|