Lines Matching refs:unack
67 struct sk_buff_head unack; /* Unack'ed packets queue */ member
176 BT_DBG("hu %p retransmitting %u pkts", hu, h5->unack.qlen); in h5_timed_event()
178 spin_lock_irqsave_nested(&h5->unack.lock, flags, SINGLE_DEPTH_NESTING); in h5_timed_event()
180 while ((skb = __skb_dequeue_tail(&h5->unack)) != NULL) { in h5_timed_event()
185 spin_unlock_irqrestore(&h5->unack.lock, flags); in h5_timed_event()
203 skb_queue_purge(&h5->unack); in h5_peer_reset()
230 skb_queue_head_init(&h5->unack); in h5_open()
258 skb_queue_purge(&h5->unack); in h5_close()
291 spin_lock_irqsave(&h5->unack.lock, flags); in h5_pkt_cull()
293 to_remove = skb_queue_len(&h5->unack); in h5_pkt_cull()
311 skb_queue_walk_safe(&h5->unack, skb, tmp) { in h5_pkt_cull()
315 __skb_unlink(skb, &h5->unack); in h5_pkt_cull()
319 if (skb_queue_empty(&h5->unack)) in h5_pkt_cull()
323 spin_unlock_irqrestore(&h5->unack.lock, flags); in h5_pkt_cull()
769 spin_lock_irqsave_nested(&h5->unack.lock, flags, SINGLE_DEPTH_NESTING); in h5_dequeue()
771 if (h5->unack.qlen >= h5->tx_win) in h5_dequeue()
779 __skb_queue_tail(&h5->unack, skb); in h5_dequeue()
781 spin_unlock_irqrestore(&h5->unack.lock, flags); in h5_dequeue()
790 spin_unlock_irqrestore(&h5->unack.lock, flags); in h5_dequeue()