Searched refs:size_mask (Results 1 – 14 of 14) sorted by relevance
335 #define size_mask(x) ((1U<<(x))-1) macro343 #define dde_count_mask size_mask(8)348 #define csb_v_mask size_mask(1)350 #define csb_f_mask size_mask(1)352 #define csb_cs_mask size_mask(8)354 #define csb_cc_mask size_mask(8)356 #define csb_ce_mask size_mask(8)361 #define ccb_cm_mask size_mask(3)366 #define vas_buf_num_mask size_mask(6)368 #define send_wc_id_mask size_mask(16)[all …]
66 ring->size_mask = size - 1; in mlx4_en_create_tx_ring()385 ring->cons & ring->size_mask, in mlx4_en_free_tx_buf()414 wqe_index = be16_to_cpu(err_cqe->wqe_index) & ring->size_mask; in mlx4_en_handle_err_cqe()441 u32 size_mask = ring->size_mask; in mlx4_en_process_tx_cq() local456 index = cons_index & size_mask; in mlx4_en_process_tx_cq()460 ring_index = ring_cons & size_mask; in mlx4_en_process_tx_cq()481 new_index = be16_to_cpu(cqe->wqe_index) & size_mask; in mlx4_en_process_tx_cq()487 ring_index = (ring_index + last_nr_txbb) & size_mask; in mlx4_en_process_tx_cq()509 index = cons_index & size_mask; in mlx4_en_process_tx_cq()935 index = ring->prod & ring->size_mask; in mlx4_en_xmit()[all …]
281 ring->size_mask = size - 1; in mlx4_en_create_rx_ring()370 ring->size_mask = ring->actual_size - 1; in mlx4_en_activate_rx_rings()557 ring->prod & ring->size_mask, in mlx4_en_refill_rx_buffers()740 index = cq->mcq.cons_index & ring->size_mask; in mlx4_en_process_rx_cq()959 index = (cq->mcq.cons_index) & ring->size_mask; in mlx4_en_process_rx_cq()
299 u32 size_mask; member334 u32 size_mask; member
537 u32 size_mask; in pcxhr_read_rmh_status() local571 size_mask = data; in pcxhr_read_rmh_status()572 while (size_mask) { in pcxhr_read_rmh_status()573 if (size_mask & 1) in pcxhr_read_rmh_status()575 size_mask >>= 1; in pcxhr_read_rmh_status()
150 u16 size_mask; member
173 f->size_mask = memsz - 1; in bdx_fifo_init()1708 f->m.rptr &= f->m.size_mask; in bdx_tx_cleanup()
1267 size_t size_cpus, size_mask; in cpu_map_data__alloc() local1284 size_mask = sizeof(u16) + sizeof(struct perf_record_mask_cpu_map32) + in cpu_map_data__alloc()1286 if (syn_data->has_any_cpu || size_cpus < size_mask) { in cpu_map_data__alloc()1294 syn_data->size = header_size + PERF_ALIGN(size_mask, sizeof(u64)); in cpu_map_data__alloc()
210 u32 size_mask; member
808 if (region->size & erase[i].size_mask) { in spi_nor_region_check_overlay()
1512 *remainder = (u32)dividend & erase->size_mask; in spi_nor_div_by_erase_size()2420 erase->size_mask = (1 << erase->size_shift) - 1; in spi_nor_set_erase_type()
152 idx &= htt->rx_ring.size_mask; in __ath10k_htt_rx_ring_fill_n()197 idx &= htt->rx_ring.size_mask; in __ath10k_htt_rx_ring_fill_n()335 idx &= htt->rx_ring.size_mask; in ath10k_htt_rx_netbuf_pop()796 htt->rx_ring.size_mask = htt->rx_ring.size - 1; in ath10k_htt_rx_alloc()827 htt->rx_ring.sw_rd_idx.msdu_payld = htt->rx_ring.size_mask; in ath10k_htt_rx_alloc()
1829 unsigned int size_mask; member
1945 static int atomic_size_to_mode(int size_mask) in atomic_size_to_mode() argument1950 int supported_size_mask = size_mask & 0x1ff; in atomic_size_to_mode()