/linux-3.4.99/drivers/media/video/saa7164/ |
D | saa7164-cmd.c | 36 ret = dev->cmds[i].seqno; in saa7164_cmd_alloc_seqno() 45 void saa7164_cmd_free_seqno(struct saa7164_dev *dev, u8 seqno) in saa7164_cmd_free_seqno() argument 48 if ((dev->cmds[seqno].inuse == 1) && in saa7164_cmd_free_seqno() 49 (dev->cmds[seqno].seqno == seqno)) { in saa7164_cmd_free_seqno() 50 dev->cmds[seqno].inuse = 0; in saa7164_cmd_free_seqno() 51 dev->cmds[seqno].signalled = 0; in saa7164_cmd_free_seqno() 52 dev->cmds[seqno].timeout = 0; in saa7164_cmd_free_seqno() 57 void saa7164_cmd_timeout_seqno(struct saa7164_dev *dev, u8 seqno) in saa7164_cmd_timeout_seqno() argument 60 if ((dev->cmds[seqno].inuse == 1) && in saa7164_cmd_timeout_seqno() 61 (dev->cmds[seqno].seqno == seqno)) { in saa7164_cmd_timeout_seqno() [all …]
|
/linux-3.4.99/drivers/gpu/drm/radeon/ |
D | radeon_trace.h | 32 TP_PROTO(struct drm_device *dev, u32 seqno), 34 TP_ARGS(dev, seqno), 38 __field(u32, seqno) 43 __entry->seqno = seqno; 46 TP_printk("dev=%u, seqno=%u", __entry->dev, __entry->seqno) 51 TP_PROTO(struct drm_device *dev, u32 seqno), 53 TP_ARGS(dev, seqno) 58 TP_PROTO(struct drm_device *dev, u32 seqno), 60 TP_ARGS(dev, seqno) 65 TP_PROTO(struct drm_device *dev, u32 seqno), [all …]
|
/linux-3.4.99/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_irq.c | 63 static bool vmw_fifo_idle(struct vmw_private *dev_priv, uint32_t seqno) in vmw_fifo_idle() argument 78 uint32_t seqno = ioread32(fifo_mem + SVGA_FIFO_FENCE); in vmw_update_seqno() local 80 if (dev_priv->last_read_seqno != seqno) { in vmw_update_seqno() 81 dev_priv->last_read_seqno = seqno; in vmw_update_seqno() 82 vmw_marker_pull(&fifo_state->marker_queue, seqno); in vmw_update_seqno() 88 uint32_t seqno) in vmw_seqno_passed() argument 93 if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP)) in vmw_seqno_passed() 98 if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP)) in vmw_seqno_passed() 102 vmw_fifo_idle(dev_priv, seqno)) in vmw_seqno_passed() 110 ret = ((atomic_read(&dev_priv->marker_seq) - seqno) in vmw_seqno_passed() [all …]
|
D | vmwgfx_marker.c | 33 uint32_t seqno; member 57 uint32_t seqno) in vmw_marker_push() argument 64 marker->seqno = seqno; in vmw_marker_push() 91 if (signaled_seqno - marker->seqno > (1 << 30)) in vmw_marker_pull() 148 uint32_t seqno; in vmw_wait_lag() local 154 seqno = atomic_read(&dev_priv->marker_seq); in vmw_wait_lag() 158 seqno = marker->seqno; in vmw_wait_lag() 162 ret = vmw_wait_seqno(dev_priv, false, seqno, true, in vmw_wait_lag() 168 (void) vmw_marker_pull(queue, seqno); in vmw_wait_lag()
|
D | vmwgfx_fence.c | 211 u32 seqno, in vmw_fence_obj_init() argument 219 fence->seqno = seqno; in vmw_fence_obj_init() 329 iowrite32(fence->seqno, in vmw_fence_goal_new_locked() 365 goal_seqno - fence->seqno < VMW_FENCE_WRAP)) in vmw_fence_goal_check_locked() 368 iowrite32(fence->seqno, fifo_mem + SVGA_FIFO_FENCE_GOAL); in vmw_fence_goal_check_locked() 380 uint32_t seqno, new_seqno; in vmw_fences_update() local 383 seqno = ioread32(fifo_mem + SVGA_FIFO_FENCE); in vmw_fences_update() 387 if (seqno - fence->seqno < VMW_FENCE_WRAP) { in vmw_fences_update() 399 needs_rerun = vmw_fence_goal_new_locked(fman, seqno); in vmw_fences_update() 413 if (new_seqno != seqno) { in vmw_fences_update() [all …]
|
/linux-3.4.99/drivers/gpu/drm/i915/ |
D | i915_trace.h | 236 TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), 237 TP_ARGS(ring, seqno), 242 __field(u32, seqno) 248 __entry->seqno = seqno; 249 i915_trace_irq_get(ring, seqno); 253 __entry->dev, __entry->ring, __entry->seqno) 280 TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), 281 TP_ARGS(ring, seqno), 286 __field(u32, seqno) 292 __entry->seqno = seqno; [all …]
|
D | intel_ringbuffer.c | 441 u32 seqno, in update_mboxes() argument 448 intel_ring_emit(ring, seqno); in update_mboxes() 463 u32 *seqno) in gen6_add_request() argument 476 *seqno = i915_gem_next_request_seqno(ring); in gen6_add_request() 478 update_mboxes(ring, *seqno, mbox1_reg); in gen6_add_request() 479 update_mboxes(ring, *seqno, mbox2_reg); in gen6_add_request() 482 intel_ring_emit(ring, *seqno); in gen6_add_request() 500 u32 seqno) in intel_ring_sync() argument 512 intel_ring_emit(waiter, seqno); in intel_ring_sync() 524 u32 seqno) in render_ring_sync_to() argument [all …]
|
D | intel_ringbuffer.h | 77 u32 *seqno); 84 u32 seqno); 211 static inline void i915_trace_irq_get(struct intel_ring_buffer *ring, u32 seqno) in i915_trace_irq_get() argument 214 ring->trace_irq_seqno = seqno; in i915_trace_irq_get()
|
D | i915_gem_execbuffer.c | 843 u32 seqno; in i915_gem_execbuffer_sync_rings() local 855 seqno = obj->last_rendering_seqno; in i915_gem_execbuffer_sync_rings() 856 if (seqno <= from->sync_seqno[idx]) in i915_gem_execbuffer_sync_rings() 859 if (seqno == from->outstanding_lazy_request) { in i915_gem_execbuffer_sync_rings() 872 seqno = request->seqno; in i915_gem_execbuffer_sync_rings() 875 from->sync_seqno[idx] = seqno; in i915_gem_execbuffer_sync_rings() 877 return to->sync_to(to, from, seqno - 1); in i915_gem_execbuffer_sync_rings() 994 u32 seqno) in i915_gem_execbuffer_move_to_active() argument 1007 i915_gem_object_move_to_active(obj, ring, seqno); in i915_gem_execbuffer_move_to_active() 1088 u32 seqno; in i915_gem_do_execbuffer() local [all …]
|
/linux-3.4.99/net/batman-adv/ |
D | unicast.c | 62 tfp->seqno = 0; in frag_merge_packet() 90 tfp->seqno = ntohs(up->seqno); in frag_create_entry() 108 tfp->seqno = 0; in frag_create_buffer() 124 search_seqno = ntohs(up->seqno)+1; in frag_search_packet() 126 search_seqno = ntohs(up->seqno)-1; in frag_search_packet() 133 if (tfp->seqno == ntohs(up->seqno)) in frag_search_packet() 138 if (tfp->seqno == search_seqno) { in frag_search_packet() 230 uint16_t seqno; in frag_send_skb() local 267 seqno = atomic_add_return(2, &hard_iface->frag_seqno); in frag_send_skb() 268 frag1->seqno = htons(seqno - 1); in frag_send_skb() [all …]
|
D | packet.h | 102 uint32_t seqno; member 119 uint16_t seqno; member 133 uint16_t seqno; member 152 uint16_t seqno; member 158 uint32_t seqno; member 165 uint32_t seqno; /* sequence number */ member
|
D | bat_iv_ogm.c | 138 ntohl(batman_ogm_packet->seqno), in bat_iv_ogm_send_to_if() 197 ntohl(batman_ogm_packet->seqno), in bat_iv_ogm_emit() 512 batman_ogm_packet->seqno = htonl(batman_ogm_packet->seqno); in bat_iv_ogm_forward() 541 batman_ogm_packet->seqno = in bat_iv_ogm_schedule() 542 htonl((uint32_t)atomic_read(&hard_iface->seqno)); in bat_iv_ogm_schedule() 562 atomic_inc(&hard_iface->seqno); in bat_iv_ogm_schedule() 842 seq_diff = batman_ogm_packet->seqno - orig_node->last_real_seqno; in bat_iv_ogm_update_seqnos() 855 batman_ogm_packet->seqno); in bat_iv_ogm_update_seqnos() 876 orig_node->last_real_seqno, batman_ogm_packet->seqno); in bat_iv_ogm_update_seqnos() 877 orig_node->last_real_seqno = batman_ogm_packet->seqno; in bat_iv_ogm_update_seqnos() [all …]
|
/linux-3.4.99/net/dccp/ |
D | ackvec.c | 54 int dccp_ackvec_update_records(struct dccp_ackvec *av, u64 seqno, u8 nonce_sum) in dccp_ackvec_update_records() argument 62 avr->avr_ack_seqno = seqno; in dccp_ackvec_update_records() 135 u64 seqno, enum dccp_ackvec_states state) in dccp_ackvec_update_old() argument 166 (unsigned long long)seqno, state); in dccp_ackvec_update_old() 200 u64 seqno, enum dccp_ackvec_states state) in dccp_ackvec_add_new() argument 240 av->av_buf_ackno = seqno; in dccp_ackvec_add_new() 251 u64 seqno = DCCP_SKB_CB(skb)->dccpd_seq; in dccp_ackvec_input() local 255 dccp_ackvec_add_new(av, 1, seqno, state); in dccp_ackvec_input() 256 av->av_tail_ackno = seqno; in dccp_ackvec_input() 259 s64 num_packets = dccp_delta_seqno(av->av_buf_ackno, seqno); in dccp_ackvec_input() [all …]
|
D | input.c | 193 u64 lswl, lawl, seqno = DCCP_SKB_CB(skb)->dccpd_seq, in dccp_check_seqno() local 211 dccp_delta_seqno(dp->dccps_swl, seqno) >= 0) in dccp_check_seqno() 212 dccp_update_gsr(sk, seqno); in dccp_check_seqno() 238 if (between48(seqno, lswl, dp->dccps_swh) && in dccp_check_seqno() 241 dccp_update_gsr(sk, seqno); in dccp_check_seqno() 269 (unsigned long long) lswl, (unsigned long long) seqno, in dccp_check_seqno() 279 seqno = dp->dccps_gsr; in dccp_check_seqno() 280 dccp_send_sync(sk, seqno, DCCP_PKT_SYNC); in dccp_check_seqno()
|
/linux-3.4.99/net/dccp/ccids/lib/ |
D | packet_history.h | 51 u64 seqno; member 56 tfrc_tx_hist_find_entry(struct tfrc_tx_hist_entry *head, u64 seqno) in tfrc_tx_hist_find_entry() argument 58 while (head != NULL && head->seqno != seqno) in tfrc_tx_hist_find_entry() 63 extern int tfrc_tx_hist_add(struct tfrc_tx_hist_entry **headp, u64 seqno);
|
/linux-3.4.99/net/tipc/ |
D | bcast.c | 160 static void bclink_update_last_sent(struct tipc_node *node, u32 seqno) in bclink_update_last_sent() argument 162 node->bclink.last_sent = less_eq(node->bclink.last_sent, seqno) ? in bclink_update_last_sent() 163 seqno : node->bclink.last_sent; in bclink_update_last_sent() 398 static void bclink_accept_pkt(struct tipc_node *node, u32 seqno) in bclink_accept_pkt() argument 400 bclink_update_last_sent(node, seqno); in bclink_accept_pkt() 401 node->bclink.last_in = seqno; in bclink_accept_pkt() 410 if (((seqno - tipc_own_addr) % TIPC_MIN_LINK_WIN) == 0) { in bclink_accept_pkt() 429 u32 seqno; in tipc_bclink_recv_pkt() local 468 seqno = msg_seqno(msg); in tipc_bclink_recv_pkt() 471 if (likely(seqno == next_in)) { in tipc_bclink_recv_pkt() [all …]
|
/linux-3.4.99/drivers/net/ppp/ |
D | ppp_deflate.c | 28 int seqno; member 153 state->seqno = 0; in z_comp_init() 173 state->seqno = 0; in z_comp_reset() 216 put_unaligned_be16(state->seqno, wptr); in z_compress() 221 ++state->seqno; in z_compress() 368 state->seqno = 0; in z_decomp_init() 389 state->seqno = 0; in z_decomp_reset() 431 if (seq != (state->seqno & 0xffff)) { in z_decompress() 434 state->unit, seq, state->seqno & 0xffff); in z_decompress() 437 ++state->seqno; in z_decompress() [all …]
|
D | bsd_comp.c | 144 unsigned short seqno; /* sequence # of next packet */ member 317 db->seqno = 0; in bsd_reset() 645 *wptr++ = db->seqno >> 8; in bsd_compress() 646 *wptr++ = db->seqno; in bsd_compress() 745 ++db->seqno; in bsd_compress() 875 if (seq != db->seqno) in bsd_decompress() 880 db->unit, seq, db->seqno - 1); in bsd_decompress() 885 ++db->seqno; in bsd_decompress() 958 max_ent, explen, db->seqno); in bsd_decompress() 1122 db->unit, db->seqno - 1); in bsd_decompress()
|
/linux-3.4.99/security/selinux/ |
D | netlink.c | 61 msg->seqno = *((u32 *)data); in selnl_add_payload() 104 void selnl_notify_policyload(u32 seqno) in selnl_notify_policyload() argument 106 selnl_notify(SELNL_MSG_POLICYLOAD, &seqno); in selnl_notify_policyload()
|
D | avc.c | 352 static int avc_latest_notif_update(int seqno, int is_insert) in avc_latest_notif_update() argument 360 if (seqno < avc_cache.latest_notif) { in avc_latest_notif_update() 362 seqno, avc_cache.latest_notif); in avc_latest_notif_update() 366 if (seqno > avc_cache.latest_notif) in avc_latest_notif_update() 367 avc_cache.latest_notif = seqno; in avc_latest_notif_update() 397 if (avc_latest_notif_update(avd->seqno, 1)) in avc_insert() 619 u32 seqno) in avc_update_node() argument 646 seqno == pos->ae.avd.seqno){ in avc_update_node() 725 int avc_ss_reset(u32 seqno) in avc_ss_reset() argument 743 avc_latest_notif_update(seqno, 0); in avc_ss_reset() [all …]
|
/linux-3.4.99/drivers/isdn/i4l/ |
D | isdn_bsdcomp.c | 123 u16 seqno; /* sequence # of next packet */ member 269 db->seqno = 0; in bsd_reset() 505 v[0] = db->seqno >> 8; in bsd_compress() 506 v[1] = db->seqno; in bsd_compress() 588 ++db->seqno; in bsd_compress() 689 if (seq != db->seqno) { in bsd_decompress() 692 db->unit, seq, db->seqno - 1); in bsd_decompress() 697 ++db->seqno; in bsd_decompress() 753 max_ent, skb_out->len, db->seqno); in bsd_decompress() 888 db->unit, db->seqno - 1); in bsd_decompress()
|
/linux-3.4.99/drivers/tty/hvc/ |
D | hvsi_lib.c | 12 packet->seqno = atomic_inc_return(&pv->seqno); in hvsi_send_packet() 24 atomic_set(&pv->seqno, 0); in hvsi_start_handshake() 101 r.query_seqno = pkt->hdr.seqno; in hvsi_got_query() 268 q.hdr.seqno = atomic_inc_return(&pv->seqno); in hvsilib_read_mctrl()
|
D | hvsi.c | 89 atomic_t seqno; /* HVSI packet sequence number */ member 225 header->seqno); in dump_packet() 300 packet.hdr.seqno = atomic_inc_return(&hp->seqno); in hvsi_version_respond() 324 hvsi_version_respond(hp, query->hdr.seqno); in hvsi_recv_query() 584 packet.hdr.seqno = atomic_inc_return(&hp->seqno); in hvsi_query() 626 packet.hdr.seqno = atomic_inc_return(&hp->seqno); in hvsi_set_mctrl() 709 packet.hdr.seqno = atomic_inc_return(&hp->seqno); in hvsi_put_chars() 726 packet.hdr.seqno = atomic_inc_return(&hp->seqno); in hvsi_close_protocol() 755 atomic_set(&hp->seqno, 0); in hvsi_open()
|
/linux-3.4.99/security/selinux/ss/ |
D | status.c | 108 void selinux_status_update_policyload(int seqno) in selinux_status_update_policyload() argument 119 status->policyload = seqno; in selinux_status_update_policyload()
|
/linux-3.4.99/drivers/net/ethernet/sfc/ |
D | mcdi.c | 77 u32 xflags, seqno; in efx_mcdi_copyin() local 82 seqno = mcdi->seqno & SEQ_MASK; in efx_mcdi_copyin() 92 MCDI_HEADER_SEQ, seqno, in efx_mcdi_copyin() 173 } else if ((respseq ^ mcdi->seqno) & SEQ_MASK) { in efx_mcdi_poll() 176 respseq, mcdi->seqno); in efx_mcdi_poll() 294 static void efx_mcdi_ev_cpl(struct efx_nic *efx, unsigned int seqno, in efx_mcdi_ev_cpl() argument 302 if ((seqno ^ mcdi->seqno) & SEQ_MASK) { in efx_mcdi_ev_cpl() 309 "seq 0x%x\n", seqno, mcdi->seqno); in efx_mcdi_ev_cpl() 337 ++mcdi->seqno; in efx_mcdi_rpc() 353 ++mcdi->seqno; in efx_mcdi_rpc()
|