Searched refs:QDIO_MAX_BUFFERS_PER_Q (Results 1 – 10 of 10) sorted by relevance
48 sbal_idx = i % QDIO_MAX_BUFFERS_PER_Q; in zfcp_qdio_zero_sbals()61 used = QDIO_MAX_BUFFERS_PER_Q - atomic_read(&qdio->req_q_free); in zfcp_qdio_account()98 if (atomic_read(&qdio->req_q_free) < QDIO_MAX_BUFFERS_PER_Q) in zfcp_qdio_request_tasklet()135 QDIO_MAX_BUFFERS_PER_Q; in zfcp_qdio_int_resp()149 sbal_idx = (idx + sbal_no) % QDIO_MAX_BUFFERS_PER_Q; in zfcp_qdio_int_resp()168 if (atomic_read(&qdio->req_q_free) < QDIO_MAX_BUFFERS_PER_Q) in zfcp_qdio_irq_tasklet()210 q_req->sbal_last %= QDIO_MAX_BUFFERS_PER_Q; in zfcp_qdio_sbal_chain()349 qdio->req_q_idx %= QDIO_MAX_BUFFERS_PER_Q; in zfcp_qdio_send()364 ret = qdio_alloc_buffers(qdio->req_q, QDIO_MAX_BUFFERS_PER_Q); in zfcp_qdio_allocate()368 ret = qdio_alloc_buffers(qdio->res_q, QDIO_MAX_BUFFERS_PER_Q); in zfcp_qdio_allocate()[all …]
41 struct qdio_buffer *res_q[QDIO_MAX_BUFFERS_PER_Q];42 struct qdio_buffer *req_q[QDIO_MAX_BUFFERS_PER_Q];129 % QDIO_MAX_BUFFERS_PER_Q; in zfcp_qdio_req_init()219 QDIO_MAX_BUFFERS_PER_Q; in zfcp_qdio_sbal_limit()
17 #define QDIO_MAX_BUFFERS_PER_Q 128 macro18 #define QDIO_MAX_BUFFERS_MASK (QDIO_MAX_BUFFERS_PER_Q - 1)170 struct slibe slibe[QDIO_MAX_BUFFERS_PER_Q];235 struct sl_element element[QDIO_MAX_BUFFERS_PER_Q];243 u8 val[QDIO_MAX_BUFFERS_PER_Q];
278 QDIO_MAX_BUFFERS_PER_Q); in qdio_init_buf_states()281 QDIO_MAX_BUFFERS_PER_Q); in qdio_init_buf_states()1227 if (bufnr >= QDIO_MAX_BUFFERS_PER_Q || count > QDIO_MAX_BUFFERS_PER_Q) in qdio_add_bufs_to_input_queue()1262 if (used == QDIO_MAX_BUFFERS_PER_Q) in handle_outbound()1272 } else if (count < QDIO_MAX_BUFFERS_PER_Q && in handle_outbound()1298 if (bufnr >= QDIO_MAX_BUFFERS_PER_Q || count > QDIO_MAX_BUFFERS_PER_Q) in qdio_add_bufs_to_output_queue()
167 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++) in setup_storage_lists()181 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++) in setup_storage_lists()
198 struct qdio_buffer *sbal[QDIO_MAX_BUFFERS_PER_Q] ____cacheline_aligned;
124 for (i = 0; i < QDIO_MAX_BUFFERS_PER_Q; i++) { in qstat_show()
153 param->rx_max_pending = QDIO_MAX_BUFFERS_PER_Q; in qeth_get_ringparam()156 param->tx_max_pending = QDIO_MAX_BUFFERS_PER_Q; in qeth_get_ringparam()161 param->tx_pending = QDIO_MAX_BUFFERS_PER_Q; in qeth_get_ringparam()
397 struct qdio_buffer *qdio_bufs[QDIO_MAX_BUFFERS_PER_Q];398 struct qeth_qdio_buffer bufs[QDIO_MAX_BUFFERS_PER_Q];492 struct qdio_buffer *qdio_bufs[QDIO_MAX_BUFFERS_PER_Q];493 struct qeth_qdio_out_buffer *bufs[QDIO_MAX_BUFFERS_PER_Q];537 return atomic_read(&queue->used_buffers) >= QDIO_MAX_BUFFERS_PER_Q; in qeth_out_queue_is_full()
322 qdio_free_buffers(q->qdio_bufs, QDIO_MAX_BUFFERS_PER_Q); in qeth_free_qdio_queue()334 if (qdio_alloc_buffers(q->qdio_bufs, QDIO_MAX_BUFFERS_PER_Q)) { in qeth_alloc_qdio_queue()339 for (i = 0; i < QDIO_MAX_BUFFERS_PER_Q; ++i) in qeth_alloc_qdio_queue()353 QDIO_MAX_BUFFERS_PER_Q); in qeth_cq_init()1426 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j) { in qeth_drain_output_queue()2552 qdio_free_buffers(q->qdio_bufs, QDIO_MAX_BUFFERS_PER_Q); in qeth_free_output_queue()2564 if (qdio_alloc_buffers(q->qdio_bufs, QDIO_MAX_BUFFERS_PER_Q)) in qeth_alloc_output_queue()2567 for (i = 0; i < QDIO_MAX_BUFFERS_PER_Q; i++) { in qeth_alloc_output_queue()2577 qdio_free_buffers(q->qdio_bufs, QDIO_MAX_BUFFERS_PER_Q); in qeth_alloc_output_queue()2658 for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j) { in qeth_free_qdio_queues()[all …]