Home
last modified time | relevance | path

Searched refs:skb1 (Results 1 – 16 of 16) sorted by relevance

/linux-3.4.99/net/core/
Dskbuff.c2193 struct sk_buff* skb1, in skb_split_inside_header() argument
2198 skb_copy_from_linear_data_offset(skb, len, skb_put(skb1, pos - len), in skb_split_inside_header()
2202 skb_shinfo(skb1)->frags[i] = skb_shinfo(skb)->frags[i]; in skb_split_inside_header()
2204 skb_shinfo(skb1)->nr_frags = skb_shinfo(skb)->nr_frags; in skb_split_inside_header()
2206 skb1->data_len = skb->data_len; in skb_split_inside_header()
2207 skb1->len += skb1->data_len; in skb_split_inside_header()
2214 struct sk_buff* skb1, in skb_split_no_header() argument
2221 skb1->len = skb1->data_len = skb->len - len; in skb_split_no_header()
2229 skb_shinfo(skb1)->frags[k] = skb_shinfo(skb)->frags[i]; in skb_split_no_header()
2241 skb_shinfo(skb1)->frags[0].page_offset += len - pos; in skb_split_no_header()
[all …]
/linux-3.4.99/net/sched/
Dsch_choke.c170 static bool choke_match_flow(struct sk_buff *skb1, in choke_match_flow() argument
173 if (skb1->protocol != skb2->protocol) in choke_match_flow()
176 if (!choke_skb_cb(skb1)->keys_valid) { in choke_match_flow()
177 choke_skb_cb(skb1)->keys_valid = 1; in choke_match_flow()
178 skb_flow_dissect(skb1, &choke_skb_cb(skb1)->keys); in choke_match_flow()
186 return !memcmp(&choke_skb_cb(skb1)->keys, in choke_match_flow()
/linux-3.4.99/net/batman-adv/
Dsend.c266 struct sk_buff *skb1; in send_outstanding_bcast_packet() local
284 skb1 = skb_clone(forw_packet->skb, GFP_ATOMIC); in send_outstanding_bcast_packet()
285 if (skb1) in send_outstanding_bcast_packet()
286 send_skb_packet(skb1, hard_iface, broadcast_addr); in send_outstanding_bcast_packet()
/linux-3.4.99/net/llc/
Dllc_sap.c371 struct sk_buff *skb1; in llc_do_mcast() local
375 skb1 = skb_clone(skb, GFP_ATOMIC); in llc_do_mcast()
376 if (!skb1) { in llc_do_mcast()
381 llc_sap_rcv(sap, skb1, stack[i]); in llc_do_mcast()
/linux-3.4.99/net/ipx/
Daf_ipx.c400 struct sk_buff *skb1; in ipxitf_demux_socket() local
403 skb1 = skb_clone(skb, GFP_ATOMIC); in ipxitf_demux_socket()
405 if (!skb1) in ipxitf_demux_socket()
408 skb1 = skb; in ipxitf_demux_socket()
411 ipxitf_def_skb_handler(s, skb1); in ipxitf_demux_socket()
471 struct sk_buff *skb1 = NULL, *skb2 = NULL; in ipxitf_demux_socket() local
519 skb1 = skb_clone(skb, GFP_ATOMIC); in ipxitf_demux_socket()
521 skb1 = skb; in ipxitf_demux_socket()
524 if (!skb1) in ipxitf_demux_socket()
529 skb2 = skb_clone(skb1, GFP_ATOMIC); in ipxitf_demux_socket()
[all …]
/linux-3.4.99/net/ipv4/
Dtcp_input.c4500 struct sk_buff *skb1; in tcp_data_queue_ofo() local
4518 skb1 = skb_peek_tail(&tp->out_of_order_queue); in tcp_data_queue_ofo()
4519 if (!skb1) { in tcp_data_queue_ofo()
4534 if (seq == TCP_SKB_CB(skb1)->end_seq) { in tcp_data_queue_ofo()
4540 if (skb->len <= skb_tailroom(skb1) && !tcp_hdr(skb)->fin) { in tcp_data_queue_ofo()
4544 skb_put(skb1, skb->len), in tcp_data_queue_ofo()
4546 TCP_SKB_CB(skb1)->end_seq = end_seq; in tcp_data_queue_ofo()
4547 TCP_SKB_CB(skb1)->ack_seq = TCP_SKB_CB(skb)->ack_seq; in tcp_data_queue_ofo()
4551 __skb_queue_after(&tp->out_of_order_queue, skb1, skb); in tcp_data_queue_ofo()
4565 if (!after(TCP_SKB_CB(skb1)->seq, seq)) in tcp_data_queue_ofo()
[all …]
Dudp.c1501 struct sk_buff *skb1 = NULL; in flush_stack() local
1506 if (likely(skb1 == NULL)) in flush_stack()
1507 skb1 = (i == final) ? skb : skb_clone(skb, GFP_ATOMIC); in flush_stack()
1509 if (!skb1) { in flush_stack()
1517 if (skb1 && udp_queue_rcv_skb(sk, skb1) <= 0) in flush_stack()
1518 skb1 = NULL; in flush_stack()
1520 if (unlikely(skb1)) in flush_stack()
1521 kfree_skb(skb1); in flush_stack()
Dicmp.c311 struct sk_buff *skb1; in icmp_push_reply() local
313 skb_queue_walk(&sk->sk_write_queue, skb1) { in icmp_push_reply()
314 csum = csum_add(csum, skb1->csum); in icmp_push_reply()
/linux-3.4.99/net/ipv6/
Dudp.c604 struct sk_buff *skb1; in flush_stack() local
607 skb1 = (i == final) ? skb : skb_clone(skb, GFP_ATOMIC); in flush_stack()
610 if (skb1) { in flush_stack()
611 if (sk_rcvqueues_full(sk, skb1)) { in flush_stack()
612 kfree_skb(skb1); in flush_stack()
617 udpv6_queue_rcv_skb(sk, skb1); in flush_stack()
618 else if (sk_add_backlog(sk, skb1)) { in flush_stack()
619 kfree_skb(skb1); in flush_stack()
/linux-3.4.99/Documentation/networking/
Dx25-iface.txt78 preserved. Even if a device driver calls netif_rx(skb1) and later
80 earlier that skb1.
/linux-3.4.99/drivers/net/wireless/ath/ath6kl/
Dtxrx.c1289 struct sk_buff *skb1 = NULL; in ath6kl_rx() local
1539 skb1 = skb_copy(skb, GFP_ATOMIC); in ath6kl_rx()
1550 skb1 = skb; in ath6kl_rx()
1557 if (skb1) in ath6kl_rx()
1558 ath6kl_data_tx(skb1, vif->ndev); in ath6kl_rx()
/linux-3.4.99/drivers/atm/
Diphase.c639 struct sk_buff *skb = NULL, *skb1 = NULL; in ia_tx_poll() local
664 skb1 = skb_dequeue(&iavcc->txing_skb); in ia_tx_poll()
665 while (skb1 && (skb1 != skb)) { in ia_tx_poll()
666 if (!(IA_SKB_STATE(skb1) & IA_TX_DONE)) { in ia_tx_poll()
670 if ((vcc->pop) && (skb1->len != 0)) in ia_tx_poll()
672 vcc->pop(vcc, skb1); in ia_tx_poll()
674 (long)skb1);) in ia_tx_poll()
677 dev_kfree_skb_any(skb1); in ia_tx_poll()
678 skb1 = skb_dequeue(&iavcc->txing_skb); in ia_tx_poll()
680 if (!skb1) { in ia_tx_poll()
/linux-3.4.99/include/net/
Dtcp.h955 struct sk_buff *skb1; in tcp_prequeue() local
959 while ((skb1 = __skb_dequeue(&tp->ucopy.prequeue)) != NULL) { in tcp_prequeue()
960 sk_backlog_rcv(sk, skb1); in tcp_prequeue()
/linux-3.4.99/drivers/net/ethernet/amd/
Dni65.c1103 struct sk_buff *skb1 = p->recv_skb[p->rmdnum]; in ni65_recv_intr() local
1107 skb = skb1; in ni65_recv_intr()
/linux-3.4.99/drivers/net/ethernet/qlogic/
Dqla3xxx.c2050 struct sk_buff *skb1 = NULL, *skb2; in ql_process_macip_rx_intr() local
2064 skb1 = lrg_buf_cb1->skb; in ql_process_macip_rx_intr()
2066 if (*((u16 *) skb1->data) != 0xFFFF) in ql_process_macip_rx_intr()
2087 skb_copy_from_linear_data_offset(skb1, VLAN_ID_LEN, in ql_process_macip_rx_intr()
/linux-3.4.99/include/linux/
Dskbuff.h2181 struct sk_buff *skb1, const u32 len);