/linux-6.1.9/net/unix/ |
D | garbage.c | 97 spin_lock(&x->sk_receive_queue.lock); in scan_inflight() 98 skb_queue_walk_safe(&x->sk_receive_queue, skb, next) { in scan_inflight() 125 __skb_unlink(skb, &x->sk_receive_queue); in scan_inflight() 130 spin_unlock(&x->sk_receive_queue.lock); in scan_inflight() 147 spin_lock(&x->sk_receive_queue.lock); in scan_children() 148 skb_queue_walk_safe(&x->sk_receive_queue, skb, next) { in scan_children() 157 spin_unlock(&x->sk_receive_queue.lock); in scan_children() 310 skb_queue_tail(&skb->sk->sk_receive_queue, skb); in unix_gc()
|
D | diag.c | 69 spin_lock(&sk->sk_receive_queue.lock); in sk_diag_dump_icons() 72 sk->sk_receive_queue.qlen * sizeof(u32)); in sk_diag_dump_icons() 78 skb_queue_walk(&sk->sk_receive_queue, skb) { in sk_diag_dump_icons() 92 spin_unlock(&sk->sk_receive_queue.lock); in sk_diag_dump_icons() 98 spin_unlock(&sk->sk_receive_queue.lock); in sk_diag_dump_icons() 107 rql.udiag_rqueue = sk->sk_receive_queue.qlen; in sk_diag_show_rqlen()
|
D | af_unix.c | 228 return skb_queue_len(&sk->sk_receive_queue) > sk->sk_max_ack_backlog; in unix_recvq_full() 233 return skb_queue_len_lockless(&sk->sk_receive_queue) > in unix_recvq_full_lockless() 550 if (!skb_queue_empty(&sk->sk_receive_queue)) { in unix_dgram_disconnected() 551 skb_queue_purge(&sk->sk_receive_queue); in unix_dgram_disconnected() 570 skb_queue_purge(&sk->sk_receive_queue); in unix_sock_destructor() 631 if (!skb_queue_empty(&sk->sk_receive_queue) || embrion) in unix_release_sock() 644 while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) { in unix_release_sock() 795 spin_lock(&sk->sk_receive_queue.lock); in unix_count_nr_fds() 796 skb = skb_peek(&sk->sk_receive_queue); in unix_count_nr_fds() 800 skb = skb_peek_next(skb, &sk->sk_receive_queue); in unix_count_nr_fds() [all …]
|
D | unix_bpf.c | 10 ({ !skb_queue_empty(&__sk->sk_receive_queue) || \ 62 if (!skb_queue_empty(&sk->sk_receive_queue) && in unix_bpf_recvmsg()
|
/linux-6.1.9/net/atm/ |
D | signaling.c | 33 skb_queue_tail(&sk_atm(sigd)->sk_receive_queue, skb); in sigd_put_skb() 113 skb_queue_tail(&sk->sk_receive_queue, skb); in sigd_send() 204 if (skb_peek(&sk_atm(vcc)->sk_receive_queue)) in sigd_close() 206 skb_queue_purge(&sk_atm(vcc)->sk_receive_queue); in sigd_close()
|
D | raw.c | 28 skb_queue_tail(&sk->sk_receive_queue, skb); in atm_push_raw()
|
/linux-6.1.9/net/bluetooth/ |
D | af_bluetooth.c | 312 if (!skb_queue_empty(&sk->sk_receive_queue)) in bt_sock_data_wait() 355 skb = skb_dequeue(&sk->sk_receive_queue); in bt_sock_stream_recvmsg() 381 skb_queue_head(&sk->sk_receive_queue, skb); in bt_sock_stream_recvmsg() 420 skb_queue_head(&sk->sk_receive_queue, skb); in bt_sock_stream_recvmsg() 427 skb_queue_head(&sk->sk_receive_queue, skb); in bt_sock_stream_recvmsg() 475 if (!skb_queue_empty_lockless(&sk->sk_receive_queue)) in bt_sock_poll() 520 skb = skb_peek(&sk->sk_receive_queue); in bt_sock_ioctl()
|
/linux-6.1.9/net/sctp/ |
D | ulpqueue.c | 138 &sk->sk_receive_queue); in sctp_clear_pd() 155 __skb_queue_tail(&sk->sk_receive_queue, in sctp_clear_pd() 216 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event() 230 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event() 239 queue = &sk->sk_receive_queue; in sctp_ulpq_tail_event() 254 if (queue == &sk->sk_receive_queue && !sp->data_ready_signalled) { in sctp_ulpq_tail_event() 1085 if (skb_queue_empty(&asoc->base.sk->sk_receive_queue)) { in sctp_ulpq_renege() 1125 __skb_queue_tail(&sk->sk_receive_queue, sctp_event2skb(ev)); in sctp_ulpq_abort_pd()
|
/linux-6.1.9/net/caif/ |
D | caif_socket.c | 123 struct sk_buff_head *list = &sk->sk_receive_queue; in caif_queue_rcv_skb() 312 if (!skb_queue_empty(&sk->sk_receive_queue) || in caif_stream_data_wait() 376 skb = skb_dequeue(&sk->sk_receive_queue); in caif_stream_recvmsg() 421 skb_queue_head(&sk->sk_receive_queue, skb); in caif_stream_recvmsg() 435 skb_queue_head(&sk->sk_receive_queue, skb); in caif_stream_recvmsg() 445 skb_queue_head(&sk->sk_receive_queue, skb); in caif_stream_recvmsg() 909 spin_lock_bh(&sk->sk_receive_queue.lock); in caif_release() 911 spin_unlock_bh(&sk->sk_receive_queue.lock); in caif_release() 952 if (!skb_queue_empty_lockless(&sk->sk_receive_queue) || in caif_poll()
|
/linux-6.1.9/net/phonet/ |
D | datagram.c | 39 skb = skb_peek(&sk->sk_receive_queue); in pn_ioctl() 64 skb_queue_purge(&sk->sk_receive_queue); in pn_destruct()
|
D | pep.c | 405 queue = &sk->sk_receive_queue; in pipe_do_rcv() 464 skb_queue_purge(&sk->sk_receive_queue); in pipe_destruct() 577 skb_queue_tail(&sk->sk_receive_queue, skb); in pipe_handler_do_rcv() 684 skb_queue_head(&sk->sk_receive_queue, skb); in pep_do_rcv() 937 else if (!skb_queue_empty(&sk->sk_receive_queue)) in pep_ioctl() 938 answ = skb_peek(&sk->sk_receive_queue)->len; in pep_ioctl() 1234 struct sk_buff *skb = skb_dequeue(&sk->sk_receive_queue); in pep_read()
|
/linux-6.1.9/net/kcm/ |
D | kcmproc.c | 119 kcm->sk.sk_receive_queue.qlen, in kcm_format_sock() 149 psock->sk->sk_receive_queue.qlen, in kcm_format_psock() 167 if (psock->sk->sk_receive_queue.qlen) { in kcm_format_psock()
|
/linux-6.1.9/net/rxrpc/ |
D | af_rxrpc.c | 814 spin_lock_bh(&sk->sk_receive_queue.lock); in rxrpc_shutdown() 821 spin_unlock_bh(&sk->sk_receive_queue.lock); in rxrpc_shutdown() 836 rxrpc_purge_queue(&sk->sk_receive_queue); in rxrpc_sock_destructor() 874 spin_lock_bh(&sk->sk_receive_queue.lock); in rxrpc_release_sock() 876 spin_unlock_bh(&sk->sk_receive_queue.lock); in rxrpc_release_sock() 888 rxrpc_purge_queue(&sk->sk_receive_queue); in rxrpc_release_sock()
|
/linux-6.1.9/net/tipc/ |
D | socket.c | 281 kfree_skb(__skb_dequeue(&sk->sk_receive_queue)); in tsk_advance_rx_queue() 312 while ((skb = __skb_dequeue(&sk->sk_receive_queue))) in tsk_rej_rx_queue() 569 skb = skb_peek(&sk->sk_receive_queue); in __tipc_shutdown() 571 __skb_unlink(skb, &sk->sk_receive_queue); in __tipc_shutdown() 587 skb = __skb_dequeue(&sk->sk_receive_queue); in __tipc_shutdown() 589 __skb_queue_purge(&sk->sk_receive_queue); in __tipc_shutdown() 605 __skb_queue_purge(&sk->sk_receive_queue); in __tipc_shutdown() 813 if (!skb_queue_empty_lockless(&sk->sk_receive_queue)) in tipc_poll() 821 if (skb_queue_empty_lockless(&sk->sk_receive_queue)) in tipc_poll() 1843 if (timeo && skb_queue_empty(&sk->sk_receive_queue)) { in tipc_wait_for_rcvmsg() [all …]
|
/linux-6.1.9/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
D | chtls_cm.h | 175 __skb_unlink(skb, &sk->sk_receive_queue); in chtls_free_skb() 182 __skb_unlink(skb, &sk->sk_receive_queue); in chtls_kfree_skb()
|
D | chtls_io.c | 1464 skb = skb_peek(&sk->sk_receive_queue); in chtls_pt_recvmsg() 1521 __skb_unlink(skb, &sk->sk_receive_queue); in chtls_pt_recvmsg() 1601 next_skb = skb_peek(&sk->sk_receive_queue); in chtls_pt_recvmsg() 1643 skb_queue_walk(&sk->sk_receive_queue, skb) { in peekmsg() 1759 skb_queue_empty_lockless(&sk->sk_receive_queue) && in chtls_recvmsg() 1789 skb = skb_peek(&sk->sk_receive_queue); in chtls_recvmsg() 1901 !skb_peek(&sk->sk_receive_queue)) in chtls_recvmsg()
|
/linux-6.1.9/drivers/xen/ |
D | pvcalls-back.c | 116 spin_lock_irqsave(&map->sock->sk->sk_receive_queue.lock, flags); in pvcalls_conn_back_read() 117 if (skb_queue_empty(&map->sock->sk->sk_receive_queue)) { in pvcalls_conn_back_read() 119 spin_unlock_irqrestore(&map->sock->sk->sk_receive_queue.lock, in pvcalls_conn_back_read() 123 spin_unlock_irqrestore(&map->sock->sk->sk_receive_queue.lock, flags); in pvcalls_conn_back_read() 148 spin_lock_irqsave(&map->sock->sk->sk_receive_queue.lock, flags); in pvcalls_conn_back_read() 149 if (ret > 0 && !skb_queue_empty(&map->sock->sk->sk_receive_queue)) in pvcalls_conn_back_read() 151 spin_unlock_irqrestore(&map->sock->sk->sk_receive_queue.lock, flags); in pvcalls_conn_back_read()
|
/linux-6.1.9/net/mptcp/ |
D | subflow.c | 942 if (skb_queue_is_last(&ssk->sk_receive_queue, skb)) { in validate_data_csum() 948 while ((skb = skb_peek(&ssk->sk_receive_queue))) in validate_data_csum() 989 skb = skb_peek(&ssk->sk_receive_queue); in get_mapping_status() 1151 if (skb_queue_empty(&ssk->sk_receive_queue) && in subflow_sched_work_if_closed() 1205 if (!skb_peek(&ssk->sk_receive_queue)) in subflow_check_data_avail() 1216 trace_subflow_check_data_avail(status, skb_peek(&ssk->sk_receive_queue)); in subflow_check_data_avail() 1224 skb = skb_peek(&ssk->sk_receive_queue); in subflow_check_data_avail() 1284 while ((skb = skb_peek(&ssk->sk_receive_queue))) in subflow_check_data_avail() 1294 skb = skb_peek(&ssk->sk_receive_queue); in subflow_check_data_avail()
|
/linux-6.1.9/net/l2tp/ |
D | l2tp_ip.c | 575 spin_lock_bh(&sk->sk_receive_queue.lock); in l2tp_ioctl() 576 skb = skb_peek(&sk->sk_receive_queue); in l2tp_ioctl() 578 spin_unlock_bh(&sk->sk_receive_queue.lock); in l2tp_ioctl()
|
/linux-6.1.9/net/nfc/ |
D | llcp_sock.c | 572 if (!skb_queue_empty_lockless(&sk->sk_receive_queue)) in llcp_sock_poll() 834 skb_queue_empty(&sk->sk_receive_queue)) { in llcp_sock_recvmsg() 861 skb_queue_head(&sk->sk_receive_queue, skb); in llcp_sock_recvmsg() 892 skb_queue_head(&sk->sk_receive_queue, skb); in llcp_sock_recvmsg() 957 skb_queue_purge(&sk->sk_receive_queue); in llcp_sock_destruct()
|
/linux-6.1.9/net/packet/ |
D | af_packet.c | 687 spin_lock(&po->sk.sk_receive_queue.lock); in prb_retire_rx_blk_timer_expired() 749 spin_unlock(&po->sk.sk_receive_queue.lock); in prb_retire_rx_blk_timer_expired() 2209 spin_lock(&sk->sk_receive_queue.lock); in packet_rcv() 2213 __skb_queue_tail(&sk->sk_receive_queue, skb); in packet_rcv() 2214 spin_unlock(&sk->sk_receive_queue.lock); in packet_rcv() 2356 spin_lock(&sk->sk_receive_queue.lock); in tpacket_rcv() 2394 __skb_queue_tail(&sk->sk_receive_queue, copy_skb); in tpacket_rcv() 2396 spin_unlock(&sk->sk_receive_queue.lock); in tpacket_rcv() 2483 spin_lock(&sk->sk_receive_queue.lock); in tpacket_rcv() 2486 spin_unlock(&sk->sk_receive_queue.lock); in tpacket_rcv() [all …]
|
/linux-6.1.9/net/core/ |
D | datagram.c | 315 return __skb_recv_datagram(sk, &sk->sk_receive_queue, flags, in skb_recv_datagram() 393 int err = __sk_queue_drop_skb(sk, &sk->sk_receive_queue, skb, flags, in skb_kill_datagram() 817 if (!skb_queue_empty_lockless(&sk->sk_receive_queue)) in datagram_poll()
|
/linux-6.1.9/net/llc/ |
D | af_llc.c | 711 if (skb_queue_empty(&sk->sk_receive_queue)) { in llc_ui_accept() 718 skb = skb_dequeue(&sk->sk_receive_queue); in llc_ui_accept() 804 skb = skb_peek(&sk->sk_receive_queue); in llc_ui_recvmsg() 888 skb_unlink(skb, &sk->sk_receive_queue); in llc_ui_recvmsg() 910 skb_unlink(skb, &sk->sk_receive_queue); in llc_ui_recvmsg()
|
/linux-6.1.9/net/ieee802154/ |
D | socket.c | 550 spin_lock_bh(&sk->sk_receive_queue.lock); in dgram_ioctl() 551 skb = skb_peek(&sk->sk_receive_queue); in dgram_ioctl() 559 spin_unlock_bh(&sk->sk_receive_queue.lock); in dgram_ioctl() 998 skb_queue_purge(&sk->sk_receive_queue); in ieee802154_sock_destruct()
|
/linux-6.1.9/net/dccp/ |
D | proto.c | 273 __skb_queue_purge(&sk->sk_receive_queue); in dccp_disconnect() 379 skb = skb_peek(&sk->sk_receive_queue); in dccp_ioctl() 809 struct sk_buff *skb = skb_peek(&sk->sk_receive_queue); in dccp_recvmsg() 999 while ((skb = __skb_dequeue(&sk->sk_receive_queue)) != NULL) { in dccp_close()
|