/linux-5.19.10/tools/testing/selftests/bpf/map_tests/ |
D | map_in_map_batch_ops.c | 123 __u32 batch_size, bool delete_entries) in fetch_and_validate() argument 137 for (step_size = batch_size; in fetch_and_validate() 139 step_size += batch_size) { in fetch_and_validate() 180 delete_entries ? "LOOKUP_AND_DELETE" : "LOOKUP", batch_size); in fetch_and_validate() 193 int batch_size[2] = {5, 10}; in _map_in_map_batch_ops() local 222 batch_size[op_index], false); in _map_in_map_batch_ops()
|
/linux-5.19.10/tools/testing/selftests/bpf/ |
D | test_lru_map.c | 243 unsigned int batch_size; in test_lru_sanity1() local 256 batch_size = tgt_free / 2; in test_lru_sanity1() 257 assert(batch_size * 2 == tgt_free); in test_lru_sanity1() 259 map_size = tgt_free + batch_size; in test_lru_sanity1() 275 end_key = 1 + batch_size; in test_lru_sanity1() 320 unsigned int batch_size; in test_lru_sanity2() local 333 batch_size = tgt_free / 2; in test_lru_sanity2() 334 assert(batch_size * 2 == tgt_free); in test_lru_sanity2() 336 map_size = tgt_free + batch_size; in test_lru_sanity2() 375 end_key = 1 + batch_size; in test_lru_sanity2() [all …]
|
D | xdpxceiver.c | 795 static int complete_pkts(struct xsk_socket_info *xsk, int batch_size) in complete_pkts() argument 803 rcvd = xsk_ring_cons__peek(&xsk->umem->cq, batch_size, &idx); in complete_pkts()
|
/linux-5.19.10/net/bpf/ |
D | test_run.c | 112 u32 batch_size; member 156 .pool_size = xdp->batch_size, in xdp_test_run_setup() 162 xdp->frames = kvmalloc_array(xdp->batch_size, sizeof(void *), GFP_KERNEL); in xdp_test_run_setup() 166 xdp->skbs = kvmalloc_array(xdp->batch_size, sizeof(void *), GFP_KERNEL); in xdp_test_run_setup() 273 batch_sz = min_t(u32, repeat, xdp->batch_size); in xdp_test_run_batch() 345 u32 repeat, u32 batch_size, u32 *time) in bpf_test_run_xdp_live() argument 348 struct xdp_test_data xdp = { .batch_size = batch_size }; in bpf_test_run_xdp_live() 774 if (kattr->test.flags || kattr->test.cpu || kattr->test.batch_size) in bpf_prog_test_run_tracing() 838 kattr->test.repeat || kattr->test.batch_size) in bpf_prog_test_run_raw_tp() 1072 if (kattr->test.flags || kattr->test.cpu || kattr->test.batch_size) in bpf_prog_test_run_skb() [all …]
|
/linux-5.19.10/samples/bpf/ |
D | xdpsock_user.c | 1405 int batch_size) in complete_tx_only() argument 1418 rcvd = xsk_ring_cons__peek(&xsk->umem->cq, batch_size, &idx); in complete_tx_only() 1496 int batch_size, unsigned long tx_ns) in tx_only() argument 1501 while (xsk_ring_prod__reserve(&xsk->tx, batch_size, &idx) < in tx_only() 1502 batch_size) { in tx_only() 1503 complete_tx_only(xsk, batch_size); in tx_only() 1513 for (i = 0; i < batch_size; i++) { in tx_only() 1535 xsk_ring_prod__submit(&xsk->tx, batch_size); in tx_only() 1536 xsk->ring_stats.tx_npkts += batch_size; in tx_only() 1537 xsk->outstanding_tx += batch_size; in tx_only() [all …]
|
/linux-5.19.10/drivers/gpu/drm/i915/gt/ |
D | intel_renderstate.c | 90 so->batch_size = rodata->batch_items * sizeof(u32); in render_state_setup() 225 so->batch_offset, so->batch_size, in intel_renderstate_emit()
|
D | intel_renderstate.h | 40 u32 batch_size; member
|
/linux-5.19.10/drivers/net/ethernet/sfc/falcon/ |
D | rx.c | 334 unsigned int fill_level, batch_size; in ef4_fast_push_rx_descriptors() local 352 batch_size = efx->rx_pages_per_batch * efx->rx_bufs_per_page; in ef4_fast_push_rx_descriptors() 354 EF4_BUG_ON_PARANOID(space < batch_size); in ef4_fast_push_rx_descriptors() 371 } while ((space -= batch_size) >= batch_size); in ef4_fast_push_rx_descriptors()
|
/linux-5.19.10/drivers/net/ethernet/sfc/siena/ |
D | rx_common.c | 460 unsigned int fill_level, batch_size; in efx_siena_fast_push_rx_descriptors() local 478 batch_size = efx->rx_pages_per_batch * efx->rx_bufs_per_page; in efx_siena_fast_push_rx_descriptors() 480 EFX_WARN_ON_ONCE_PARANOID(space < batch_size); in efx_siena_fast_push_rx_descriptors() 495 } while ((space -= batch_size) >= batch_size); in efx_siena_fast_push_rx_descriptors()
|
/linux-5.19.10/drivers/net/ethernet/sfc/ |
D | rx_common.c | 456 unsigned int fill_level, batch_size; in efx_fast_push_rx_descriptors() local 474 batch_size = efx->rx_pages_per_batch * efx->rx_bufs_per_page; in efx_fast_push_rx_descriptors() 476 EFX_WARN_ON_ONCE_PARANOID(space < batch_size); in efx_fast_push_rx_descriptors() 491 } while ((space -= batch_size) >= batch_size); in efx_fast_push_rx_descriptors()
|
/linux-5.19.10/tools/lib/bpf/ |
D | bpf.h | 484 __u32 batch_size; member 486 #define bpf_test_run_opts__last_field batch_size
|
D | bpf.c | 979 attr.test.batch_size = OPTS_GET(opts, batch_size, 0); in bpf_prog_test_run_opts()
|
/linux-5.19.10/tools/testing/selftests/bpf/prog_tests/ |
D | xdp_do_redirect.c | 105 .batch_size = 64, in test_xdp_do_redirect()
|
/linux-5.19.10/drivers/dma/idxd/ |
D | sysfs.c | 948 u64 batch_size; in wq_max_batch_size_store() local 957 rc = __get_sysfs_u64(buf, &batch_size); in wq_max_batch_size_store() 961 if (batch_size > idxd->max_batch_size) in wq_max_batch_size_store() 964 wq->max_batch_size = (u32)batch_size; in wq_max_batch_size_store()
|
/linux-5.19.10/drivers/md/ |
D | raid5.c | 6406 int i, batch_size = 0, hash; in handle_active_stripes() local 6409 while (batch_size < MAX_STRIPE_BATCH && in handle_active_stripes() 6411 batch[batch_size++] = sh; in handle_active_stripes() 6413 if (batch_size == 0) { in handle_active_stripes() 6421 return batch_size; in handle_active_stripes() 6436 for (i = 0; i < batch_size; i++) in handle_active_stripes() 6443 for (i = 0; i < batch_size; i++) { in handle_active_stripes() 6447 return batch_size; in handle_active_stripes() 6466 int batch_size, released; in raid5_do_work() local 6470 batch_size = handle_active_stripes(conf, group_id, worker, in raid5_do_work() [all …]
|
/linux-5.19.10/Documentation/bpf/ |
D | bpf_prog_run.rst | 94 the ``batch_size`` parameter, up to a maximum of 256 packets. For each batch,
|
/linux-5.19.10/fs/btrfs/ |
D | tree-log.c | 3842 int batch_size = 0; in process_dir_items_leaf() local 3944 if (batch_size > 0) { in process_dir_items_leaf() 3946 batch_start, batch_size); in process_dir_items_leaf() 3949 batch_size = 0; in process_dir_items_leaf() 3953 if (batch_size == 0) in process_dir_items_leaf() 3955 batch_size++; in process_dir_items_leaf() 3958 if (batch_size > 0) { in process_dir_items_leaf() 3962 batch_start, batch_size); in process_dir_items_leaf()
|
/linux-5.19.10/drivers/media/pci/tw5864/ |
D | tw5864-video.c | 1322 int batch_size = H264_BUF_CNT; in tw5864_handle_frame_task() local 1325 while (dev->h264_buf_r_index != dev->h264_buf_w_index && batch_size--) { in tw5864_handle_frame_task()
|
/linux-5.19.10/drivers/net/wireless/mediatek/mt76/ |
D | mt76_connac_mcu.c | 2039 int i, n_chan, batch_size, idx = 0, tx_power, last_ch; in mt76_connac_mcu_rate_txpower_band() local 2059 batch_size = DIV_ROUND_UP(n_chan, batch_len); in mt76_connac_mcu_rate_txpower_band() 2068 for (i = 0; i < batch_size; i++) { in mt76_connac_mcu_rate_txpower_band() 2073 num_ch = i == batch_size - 1 ? n_chan % batch_len : batch_len; in mt76_connac_mcu_rate_txpower_band()
|
/linux-5.19.10/mm/ |
D | vmscan.c | 713 long batch_size = shrinker->batch ? shrinker->batch in do_shrink_slab() local 763 while (total_scan >= batch_size || in do_shrink_slab() 766 unsigned long nr_to_scan = min(batch_size, total_scan); in do_shrink_slab()
|
D | memory.c | 1925 const int batch_size = min_t(int, pages_to_write_in_pmd, 8); in insert_pages() local 1928 for (pte = start_pte; pte_idx < batch_size; ++pte, ++pte_idx) { in insert_pages() 1941 pages_to_write_in_pmd -= batch_size; in insert_pages() 1942 remaining_pages_total -= batch_size; in insert_pages()
|
/linux-5.19.10/tools/include/uapi/linux/ |
D | bpf.h | 1406 __u32 batch_size; member
|
/linux-5.19.10/include/uapi/linux/ |
D | bpf.h | 1406 __u32 batch_size; member
|
/linux-5.19.10/tools/perf/scripts/python/ |
D | exported-sql-viewer.py | 2620 def FetchBatch(self, batch_size): argument 2622 while batch_size > fetched: 2641 batch_size = min(glb_chunk_sz, target - self.fetched) 2642 self.FetchBatch(batch_size)
|
/linux-5.19.10/kernel/bpf/ |
D | syscall.c | 3567 #define BPF_PROG_TEST_RUN_LAST_FIELD test.batch_size
|