/linux-3.4.99/drivers/gpu/drm/via/ |
D | via_dmablit.c | 66 int num_desc = vsg->num_desc; in via_unmap_blit_from_device() local 67 unsigned cur_descriptor_page = num_desc / vsg->descriptors_per_page; in via_unmap_blit_from_device() 68 unsigned descriptor_this_page = num_desc % vsg->descriptors_per_page; in via_unmap_blit_from_device() 73 while (num_desc--) { in via_unmap_blit_from_device() 109 int num_desc = 0; in via_map_blit_for_device() local 148 num_desc++; in via_map_blit_for_device() 161 vsg->num_desc = num_desc; in via_map_blit_for_device() 272 vsg->num_desc_pages = (vsg->num_desc + vsg->descriptors_per_page - 1) / in via_alloc_desc_pages() 285 vsg->num_desc); in via_alloc_desc_pages()
|
D | via_dmablit.h | 45 int num_desc; member
|
/linux-3.4.99/drivers/net/ethernet/qlogic/netxen/ |
D | netxen_nic_init.c | 118 for (i = 0; i < rds_ring->num_desc; ++i) { in netxen_release_rx_buffers() 140 for (i = 0; i < tx_ring->num_desc; i++) { in netxen_release_tx_buffers() 215 tx_ring->num_desc = adapter->num_txd; in netxen_alloc_sw_resources() 241 rds_ring->num_desc = adapter->num_rxd; in netxen_alloc_sw_resources() 260 rds_ring->num_desc = adapter->num_jumbo_rxd; in netxen_alloc_sw_resources() 276 rds_ring->num_desc = adapter->num_lro_rxd; in netxen_alloc_sw_resources() 293 for (i = 0; i < rds_ring->num_desc; i++) { in netxen_alloc_sw_resources() 307 sds_ring->num_desc = adapter->num_rxd; in netxen_alloc_sw_resources() 1467 index = get_next_index(index, sds_ring->num_desc); in netxen_handle_fw_message() 1563 if (unlikely(index >= rds_ring->num_desc)) in netxen_process_rcv() [all …]
|
D | netxen_nic_ethtool.c | 430 u32 num_desc; in netxen_validate_ringparam() local 431 num_desc = max(val, min); in netxen_validate_ringparam() 432 num_desc = min(num_desc, max); in netxen_validate_ringparam() 433 num_desc = roundup_pow_of_two(num_desc); in netxen_validate_ringparam() 435 if (val != num_desc) { in netxen_validate_ringparam() 437 netxen_nic_driver_name, r_name, num_desc, val); in netxen_validate_ringparam() 440 return num_desc; in netxen_validate_ringparam()
|
D | netxen_nic_ctx.c | 353 prq_rds[i].ring_size = cpu_to_le32(rds_ring->num_desc); in nx_fw_cmd_create_rx_ctx() 366 prq_sds[i].ring_size = cpu_to_le32(sds_ring->num_desc); in nx_fw_cmd_create_rx_ctx() 495 prq_cds->ring_size = cpu_to_le32(tx_ring->num_desc); in nx_fw_cmd_create_tx_ctx() 721 hwctx->cmd_ring_size = cpu_to_le32(tx_ring->num_desc); in netxen_init_old_ctx() 730 cpu_to_le32(rds_ring->num_desc); in netxen_init_old_ctx() 738 hwctx->sts_ring_size = cpu_to_le32(sds_ring->num_desc); in netxen_init_old_ctx() 741 hwctx->sts_rings[ring].size = cpu_to_le32(sds_ring->num_desc); in netxen_init_old_ctx()
|
D | netxen_nic.h | 78 (sizeof(struct rcv_desc) * (rds_ring)->num_desc) 80 (sizeof(struct netxen_rx_buffer) * rds_ring->num_desc) 82 (sizeof(struct status_desc) * (sds_ring)->num_desc) 84 (sizeof(struct netxen_cmd_buffer) * tx_ring->num_desc) 86 (sizeof(struct cmd_desc_type0) * tx_ring->num_desc) 627 u32 num_desc; member 641 u32 num_desc; member 662 u32 num_desc; member 1860 tx_ring->sw_consumer, tx_ring->num_desc); in netxen_tx_avail()
|
D | netxen_nic_main.c | 1896 producer = get_next_index(producer, tx_ring->num_desc); in netxen_tso_check() 1913 producer = get_next_index(producer, tx_ring->num_desc); in netxen_tso_check() 1992 u32 num_txd = tx_ring->num_desc; in netxen_nic_xmit_frame()
|
D | netxen_nic_hw.c | 622 producer = get_next_index(producer, tx_ring->num_desc); in netxen_send_cmd_descs()
|
/linux-3.4.99/drivers/net/ethernet/qlogic/qlcnic/ |
D | qlcnic_init.c | 100 for (i = 0; i < rds_ring->num_desc; ++i) { in qlcnic_release_rx_buffers() 129 for (i = 0; i < rds_ring->num_desc; i++) { in qlcnic_reset_rx_buffers_list() 145 for (i = 0; i < tx_ring->num_desc; i++) { in qlcnic_release_tx_buffers() 219 tx_ring->num_desc = adapter->num_txd; in qlcnic_alloc_sw_resources() 243 rds_ring->num_desc = adapter->num_rxd; in qlcnic_alloc_sw_resources() 249 rds_ring->num_desc = adapter->num_jumbo_rxd; in qlcnic_alloc_sw_resources() 272 for (i = 0; i < rds_ring->num_desc; i++) { in qlcnic_alloc_sw_resources() 285 sds_ring->num_desc = adapter->num_rxd; in qlcnic_alloc_sw_resources() 1396 index = get_next_index(index, sds_ring->num_desc); in qlcnic_handle_fw_message() 1543 if (unlikely(index >= rds_ring->num_desc)) in qlcnic_process_rcv() [all …]
|
D | qlcnic.h | 65 (sizeof(struct rcv_desc) * (rds_ring)->num_desc) 67 (sizeof(struct qlcnic_rx_buffer) * rds_ring->num_desc) 69 (sizeof(struct status_desc) * (sds_ring)->num_desc) 71 (sizeof(struct qlcnic_cmd_buffer) * tx_ring->num_desc) 73 (sizeof(struct cmd_desc_type0) * tx_ring->num_desc) 494 u32 num_desc; member 506 u32 num_desc; member 524 u32 num_desc; member 1557 return tx_ring->sw_consumer + tx_ring->num_desc - in qlcnic_tx_avail()
|
D | qlcnic_ethtool.c | 426 u32 num_desc; in qlcnic_validate_ringparam() local 427 num_desc = max(val, min); in qlcnic_validate_ringparam() 428 num_desc = min(num_desc, max); in qlcnic_validate_ringparam() 429 num_desc = roundup_pow_of_two(num_desc); in qlcnic_validate_ringparam() 431 if (val != num_desc) { in qlcnic_validate_ringparam() 433 qlcnic_driver_name, r_name, num_desc, val); in qlcnic_validate_ringparam() 436 return num_desc; in qlcnic_validate_ringparam()
|
D | qlcnic_ctx.c | 267 prq_rds[i].ring_size = cpu_to_le32(rds_ring->num_desc); in qlcnic_fw_cmd_create_rx_ctx() 282 prq_sds[i].ring_size = cpu_to_le32(sds_ring->num_desc); in qlcnic_fw_cmd_create_rx_ctx() 411 prq_cds->ring_size = cpu_to_le32(tx_ring->num_desc); in qlcnic_fw_cmd_create_tx_ctx()
|
D | qlcnic_main.c | 1924 tx_ring->producer = get_next_index(producer, tx_ring->num_desc); in qlcnic_change_filter() 2067 producer = get_next_index(producer, tx_ring->num_desc); in qlcnic_tx_pkt() 2084 producer = get_next_index(producer, tx_ring->num_desc); in qlcnic_tx_pkt() 2206 u32 num_txd = tx_ring->num_desc; in qlcnic_xmit_frame() 2518 sw_consumer = get_next_index(sw_consumer, tx_ring->num_desc); in qlcnic_process_cmd_ring()
|
D | qlcnic_hw.c | 346 producer = get_next_index(producer, tx_ring->num_desc); in qlcnic_send_cmd_descs()
|
/linux-3.4.99/drivers/net/ethernet/ti/ |
D | davinci_cpdma.c | 83 int num_desc, used_desc; member 156 pool->num_desc = size / pool->desc_size; in cpdma_desc_pool_create() 158 bitmap_size = (pool->num_desc / BITS_PER_LONG) * sizeof(long); in cpdma_desc_pool_create() 219 cpdma_desc_alloc(struct cpdma_desc_pool *pool, int num_desc) in cpdma_desc_alloc() argument 227 index = bitmap_find_next_zero_area(pool->bitmap, pool->num_desc, 0, in cpdma_desc_alloc() 228 num_desc, 0); in cpdma_desc_alloc() 229 if (index < pool->num_desc) { in cpdma_desc_alloc() 230 bitmap_set(pool->bitmap, index, num_desc); in cpdma_desc_alloc() 240 struct cpdma_desc __iomem *desc, int num_desc) in cpdma_desc_free() argument 247 bitmap_clear(pool->bitmap, index, num_desc); in cpdma_desc_free()
|
/linux-3.4.99/drivers/net/ethernet/toshiba/ |
D | spider_net.c | 326 dma_free_coherent(&card->pdev->dev, chain->num_desc, in spider_net_free_chain() 350 alloc_size = chain->num_desc * sizeof(struct spider_net_hw_descr); in spider_net_init_chain() 358 memset(chain->ring, 0, chain->num_desc * sizeof(struct spider_net_descr)); in spider_net_init_chain() 364 for (i=0; i < chain->num_desc; i++, descr++, hwdescr++) { in spider_net_init_chain() 762 if (cnt < card->tx_chain.num_desc/4) in spider_net_set_low_watermark() 1034 chain->num_desc); in show_rx_chain() 1064 int from = (chain->num_desc + off - cnt) % chain->num_desc; in show_rx_chain() 1065 int to = (chain->num_desc + off - 1) % chain->num_desc; in show_rx_chain() 1133 for (i=0; i<chain->num_desc; i++) { in spider_net_resync_head_ptr() 1153 for (i=0; i<chain->num_desc; i++) { in spider_net_resync_tail_ptr() [all …]
|
D | spider_net_ethtool.c | 125 ering->tx_pending = card->tx_chain.num_desc; in spider_net_ethtool_get_ringparam() 127 ering->rx_pending = card->rx_chain.num_desc; in spider_net_ethtool_get_ringparam()
|
D | spider_net.h | 410 int num_desc; member
|
/linux-3.4.99/drivers/atm/ |
D | iphase.c | 614 int num_desc; in ia_que_tx() local 616 num_desc = ia_avail_descs(iadev); in ia_que_tx() 618 while (num_desc && (skb = skb_dequeue(&iadev->tx_backlog))) { in ia_que_tx() 632 num_desc--; in ia_que_tx()
|