/linux-6.6.21/fs/xfs/libxfs/ |
D | xfs_iext_tree.c | 454 int *nr_entries) in xfs_iext_split_node() argument 466 *nr_entries = 0; in xfs_iext_split_node() 482 *nr_entries = nr_move; in xfs_iext_split_node() 484 *nr_entries = nr_keep; in xfs_iext_split_node() 500 int i, pos, nr_entries; in xfs_iext_insert_node() local 509 nr_entries = xfs_iext_node_nr_entries(node, pos); in xfs_iext_insert_node() 511 ASSERT(pos >= nr_entries || xfs_iext_key_cmp(node, pos, offset) != 0); in xfs_iext_insert_node() 512 ASSERT(nr_entries <= KEYS_PER_NODE); in xfs_iext_insert_node() 514 if (nr_entries == KEYS_PER_NODE) in xfs_iext_insert_node() 515 new = xfs_iext_split_node(&node, &pos, &nr_entries); in xfs_iext_insert_node() [all …]
|
/linux-6.6.21/lib/ |
D | stackdepot.c | 359 unsigned int nr_entries, in __stack_depot_save() argument 377 nr_entries = filter_irq_stacks(entries, nr_entries); in __stack_depot_save() 379 if (unlikely(nr_entries == 0) || stack_depot_disabled) in __stack_depot_save() 382 hash = hash_stack(entries, nr_entries); in __stack_depot_save() 390 found = find_stack(smp_load_acquire(bucket), entries, nr_entries, hash); in __stack_depot_save() 417 found = find_stack(*bucket, entries, nr_entries, hash); in __stack_depot_save() 420 depot_alloc_stack(entries, nr_entries, hash, &prealloc); in __stack_depot_save() 453 unsigned int nr_entries, in stack_depot_save() argument 456 return __stack_depot_save(entries, nr_entries, alloc_flags, true); in stack_depot_save() 501 unsigned int nr_entries; in stack_depot_print() local [all …]
|
/linux-6.6.21/drivers/dma/dw-edma/ |
D | dw-edma-v0-debugfs.c | 101 int nr_entries, struct dentry *dent) in dw_edma_debugfs_create_x32() argument 106 entries = devm_kcalloc(dw->chip->dev, nr_entries, sizeof(*entries), in dw_edma_debugfs_create_x32() 111 for (i = 0; i < nr_entries; i++) { in dw_edma_debugfs_create_x32() 133 int nr_entries; in dw_edma_debugfs_regs_ch() local 135 nr_entries = ARRAY_SIZE(debugfs_regs); in dw_edma_debugfs_regs_ch() 136 dw_edma_debugfs_create_x32(dw, debugfs_regs, nr_entries, dent); in dw_edma_debugfs_regs_ch() 178 int nr_entries, i; in dw_edma_debugfs_regs_wr() local 183 nr_entries = ARRAY_SIZE(debugfs_regs); in dw_edma_debugfs_regs_wr() 184 dw_edma_debugfs_create_x32(dw, debugfs_regs, nr_entries, regs_dent); in dw_edma_debugfs_regs_wr() 187 nr_entries = ARRAY_SIZE(debugfs_unroll_regs); in dw_edma_debugfs_regs_wr() [all …]
|
/linux-6.6.21/kernel/ |
D | stacktrace.c | 24 void stack_trace_print(const unsigned long *entries, unsigned int nr_entries, in stack_trace_print() argument 32 for (i = 0; i < nr_entries; i++) in stack_trace_print() 48 unsigned int nr_entries, int spaces) in stack_trace_snprint() argument 55 for (i = 0; i < nr_entries && size; i++) { in stack_trace_snprint() 277 return trace.nr_entries; in stack_trace_save() 302 return trace.nr_entries; in stack_trace_save_tsk() 324 return trace.nr_entries; in stack_trace_save_regs() 349 return ret ? ret : trace.nr_entries; in stack_trace_save_tsk_reliable() 369 return trace.nr_entries; in stack_trace_save_user() 390 unsigned int filter_irq_stacks(unsigned long *entries, unsigned int nr_entries) in filter_irq_stacks() argument [all …]
|
D | backtracetest.c | 48 unsigned int nr_entries; in backtrace_test_saved() local 53 nr_entries = stack_trace_save(entries, ARRAY_SIZE(entries), 0); in backtrace_test_saved() 54 stack_trace_print(entries, nr_entries, 0); in backtrace_test_saved()
|
/linux-6.6.21/drivers/md/persistent-data/ |
D | dm-btree-remove.c | 62 uint32_t nr_entries = le32_to_cpu(n->header.nr_entries); in node_shift() local 67 BUG_ON(shift > nr_entries); in node_shift() 71 (nr_entries - shift) * sizeof(__le64)); in node_shift() 74 (nr_entries - shift) * value_size); in node_shift() 76 BUG_ON(nr_entries + shift > le32_to_cpu(n->header.max_entries)); in node_shift() 79 nr_entries * sizeof(__le64)); in node_shift() 82 nr_entries * value_size); in node_shift() 88 uint32_t nr_left = le32_to_cpu(left->header.nr_entries); in node_copy() 131 unsigned int nr_entries = le32_to_cpu(n->header.nr_entries); in delete_at() local 132 unsigned int nr_to_copy = nr_entries - (index + 1); in delete_at() [all …]
|
D | dm-btree.c | 46 int lo = -1, hi = le32_to_cpu(n->header.nr_entries); in bsearch() 77 uint32_t nr_entries = le32_to_cpu(n->header.nr_entries); in inc_children() local 80 dm_tm_with_runs(tm, value_ptr(n, 0), nr_entries, dm_tm_inc_range); in inc_children() 83 vt->inc(vt->context, value_ptr(n, 0), nr_entries); in inc_children() 90 uint32_t nr_entries = le32_to_cpu(node->header.nr_entries); in insert_at() local 94 if (index > nr_entries || in insert_at() 96 nr_entries >= max_entries) { in insert_at() 104 array_insert(node->keys, sizeof(*node->keys), nr_entries, index, &key_le); in insert_at() 105 array_insert(value_base(node), value_size, nr_entries, index, value); in insert_at() 106 node->header.nr_entries = cpu_to_le32(nr_entries + 1); in insert_at() [all …]
|
D | dm-array.c | 27 __le32 nr_entries; member 114 unsigned int nr_entries = le32_to_cpu(ab->nr_entries); in on_entries() local 116 fn(info->value_type.context, element_at(info, ab, 0), nr_entries); in on_entries() 164 (*ab)->nr_entries = cpu_to_le32(0); in alloc_ablock() 178 uint32_t nr_entries, delta, i; in fill_ablock() local 182 BUG_ON(new_nr < le32_to_cpu(ab->nr_entries)); in fill_ablock() 184 nr_entries = le32_to_cpu(ab->nr_entries); in fill_ablock() 185 delta = new_nr - nr_entries; in fill_ablock() 188 for (i = nr_entries; i < new_nr; i++) in fill_ablock() 190 ab->nr_entries = cpu_to_le32(new_nr); in fill_ablock() [all …]
|
/linux-6.6.21/arch/x86/kernel/ |
D | e820.c | 82 for (i = 0; i < table->nr_entries; i++) { in _e820__mapped_any() 117 for (i = 0; i < e820_table->nr_entries; i++) { in __e820__mapped_all() 168 int x = table->nr_entries; in __e820__range_add() 179 table->nr_entries++; in __e820__range_add() 207 for (i = 0; i < e820_table->nr_entries; i++) { in e820__print_table() 332 if (table->nr_entries < 2) in e820__update_table() 335 BUG_ON(table->nr_entries > max_nr_entries); in e820__update_table() 338 for (i = 0; i < table->nr_entries; i++) { in e820__update_table() 344 for (i = 0; i < 2 * table->nr_entries; i++) in e820__update_table() 352 for (i = 0; i < table->nr_entries; i++) { in e820__update_table() [all …]
|
D | ldt.c | 81 set_ldt(ldt_slot_va(ldt->slot), ldt->nr_entries); in load_mm_ldt() 83 set_ldt(ldt->entries, ldt->nr_entries); in load_mm_ldt() 183 new_ldt->nr_entries = num_entries; in alloc_ldt_struct() 309 nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE); in map_ldt_struct() 361 nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE); in unmap_ldt_struct() 418 paravirt_alloc_ldt(ldt->entries, ldt->nr_entries); in finalize_ldt_struct() 439 paravirt_free_ldt(ldt->entries, ldt->nr_entries); in free_ldt_struct() 440 if (ldt->nr_entries * LDT_ENTRY_SIZE > PAGE_SIZE) in free_ldt_struct() 463 new_ldt = alloc_ldt_struct(old_mm->context.ldt->nr_entries); in ldt_dup_context() 470 new_ldt->nr_entries * LDT_ENTRY_SIZE); in ldt_dup_context() [all …]
|
/linux-6.6.21/mm/kmsan/ |
D | report.c | 87 unsigned int nr_entries, chained_nr_entries, skipnr; in kmsan_print_origin() local 98 nr_entries = stack_depot_fetch(origin, &entries); in kmsan_print_origin() 100 magic = nr_entries ? entries[0] : 0; in kmsan_print_origin() 101 if ((nr_entries == 4) && (magic == KMSAN_ALLOCA_MAGIC_ORIGIN)) { in kmsan_print_origin() 113 if ((nr_entries == 3) && (magic == KMSAN_CHAIN_MAGIC_ORIGIN)) { in kmsan_print_origin() 137 if (nr_entries) { in kmsan_print_origin() 138 skipnr = get_stack_skipnr(entries, nr_entries); in kmsan_print_origin() 139 stack_trace_print(entries + skipnr, nr_entries - skipnr, in kmsan_print_origin()
|
/linux-6.6.21/tools/perf/util/ |
D | syscalltbl.c | 69 int nr_entries = 0, i, j; in syscalltbl__init_native() local 74 ++nr_entries; in syscalltbl__init_native() 76 entries = tbl->syscalls.entries = malloc(sizeof(struct syscall) * nr_entries); in syscalltbl__init_native() 88 qsort(tbl->syscalls.entries, nr_entries, sizeof(struct syscall), syscallcmp); in syscalltbl__init_native() 89 tbl->syscalls.nr_entries = nr_entries; in syscalltbl__init_native() 120 tbl->syscalls.nr_entries, sizeof(*sc), in syscalltbl__id() 131 for (i = *idx + 1; i < tbl->syscalls.nr_entries; ++i) { in syscalltbl__strglobmatch_next()
|
D | rblist.h | 24 unsigned int nr_entries; member 42 return rblist->nr_entries == 0; in rblist__empty() 47 return rblist->nr_entries; in rblist__nr_entries()
|
/linux-6.6.21/tools/testing/selftests/bpf/benchs/ |
D | bench_bpf_hashmap_lookup.c | 24 __u32 nr_entries; member 30 .nr_entries = 500, 91 args.nr_entries = ret; in parse_arg() 121 if (args.nr_entries > args.max_entries) { in validate() 123 args.max_entries, args.nr_entries); in validate() 171 ctx.skel->bss->nr_entries = args.nr_entries; in setup() 172 ctx.skel->bss->nr_loops = args.nr_loops / args.nr_entries; in setup() 188 for (u64 i = 0; i < args.nr_entries; i++) { in setup()
|
/linux-6.6.21/tools/perf/ui/ |
D | browser.c | 108 if (browser->nr_entries == 0) in ui_browser__list_head_seek() 316 if (browser->nr_entries > 1) { in ui_browser__scrollbar_set() 318 (browser->nr_entries - 1)); in ui_browser__scrollbar_set() 348 if (browser->nr_entries == 0 && browser->no_samples_msg) in __ui_browser__refresh() 368 void ui_browser__update_nr_entries(struct ui_browser *browser, u32 nr_entries) in ui_browser__update_nr_entries() argument 370 off_t offset = nr_entries - browser->nr_entries; in ui_browser__update_nr_entries() 372 browser->nr_entries = nr_entries; in ui_browser__update_nr_entries() 424 if (browser->index == browser->nr_entries - 1) in ui_browser__run() 455 if (browser->top_idx + browser->rows > browser->nr_entries - 1) in ui_browser__run() 459 if (browser->index + offset > browser->nr_entries - 1) in ui_browser__run() [all …]
|
/linux-6.6.21/arch/mips/kernel/ |
D | stacktrace.c | 30 trace->entries[trace->nr_entries++] = addr; in save_raw_context_stack() 31 if (trace->nr_entries >= trace->max_entries) in save_raw_context_stack() 58 trace->entries[trace->nr_entries++] = pc; in save_context_stack() 59 if (trace->nr_entries >= trace->max_entries) in save_context_stack() 83 WARN_ON(trace->nr_entries || !trace->max_entries); in save_stack_trace_tsk()
|
/linux-6.6.21/tools/perf/trace/beauty/ |
D | ioctl.c | 41 if (nr < strarray__ioctl_tty_cmd.nr_entries && strarray__ioctl_tty_cmd.entries[nr] != NULL) in ioctl__scnprintf_tty_cmd() 52 if (nr < strarray__drm_ioctl_cmds.nr_entries && strarray__drm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_drm_cmd() 63 …if (nr < strarray__sndrv_pcm_ioctl_cmds.nr_entries && strarray__sndrv_pcm_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_pcm_cmd() 74 …if (nr < strarray__sndrv_ctl_ioctl_cmds.nr_entries && strarray__sndrv_ctl_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_ctl_cmd() 85 if (nr < strarray__kvm_ioctl_cmds.nr_entries && strarray__kvm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_kvm_cmd() 98 if (nr < s->nr_entries && s->entries[nr] != NULL) in ioctl__scnprintf_vhost_virtio_cmd() 109 if (nr < strarray__perf_ioctl_cmds.nr_entries && strarray__perf_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_perf_cmd() 120 …if (nr < strarray__usbdevfs_ioctl_cmds.nr_entries && strarray__usbdevfs_ioctl_cmds.entries[nr] != … in ioctl__scnprintf_usbdevfs_cmd()
|
/linux-6.6.21/drivers/md/ |
D | dm-cache-policy-internal.h | 117 static inline size_t bitset_size_in_bytes(unsigned int nr_entries) in bitset_size_in_bytes() argument 119 return sizeof(unsigned long) * dm_div_up(nr_entries, BITS_PER_LONG); in bitset_size_in_bytes() 122 static inline unsigned long *alloc_bitset(unsigned int nr_entries) in alloc_bitset() argument 124 size_t s = bitset_size_in_bytes(nr_entries); in alloc_bitset() 129 static inline void clear_bitset(void *bitset, unsigned int nr_entries) in clear_bitset() argument 131 size_t s = bitset_size_in_bytes(nr_entries); in clear_bitset()
|
/linux-6.6.21/arch/sh/kernel/ |
D | stacktrace.c | 33 if (trace->nr_entries < trace->max_entries) in save_stack_address() 34 trace->entries[trace->nr_entries++] = addr; in save_stack_address() 65 if (trace->nr_entries < trace->max_entries) in save_stack_address_nosched() 66 trace->entries[trace->nr_entries++] = addr; in save_stack_address_nosched()
|
/linux-6.6.21/include/linux/ |
D | stacktrace.h | 66 void stack_trace_print(const unsigned long *trace, unsigned int nr_entries, 69 unsigned int nr_entries, int spaces); 78 unsigned int filter_irq_stacks(unsigned long *entries, unsigned int nr_entries); 83 unsigned int nr_entries, max_entries; member
|
/linux-6.6.21/arch/openrisc/kernel/ |
D | stacktrace.c | 38 if (trace->nr_entries < trace->max_entries) in save_stack_address() 39 trace->entries[trace->nr_entries++] = addr; in save_stack_address() 64 if (trace->nr_entries < trace->max_entries) in save_stack_address_nosched() 65 trace->entries[trace->nr_entries++] = addr; in save_stack_address_nosched()
|
/linux-6.6.21/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum1_kvdl.c | 122 unsigned int entry_index, nr_entries; in mlxsw_sp1_kvdl_part_alloc() local 124 nr_entries = (info->end_index - info->start_index + 1) / in mlxsw_sp1_kvdl_part_alloc() 126 entry_index = find_first_zero_bit(part->usage, nr_entries); in mlxsw_sp1_kvdl_part_alloc() 127 if (entry_index == nr_entries) in mlxsw_sp1_kvdl_part_alloc() 215 unsigned int nr_entries; in mlxsw_sp1_kvdl_part_init() local 226 nr_entries = div_u64(resource_size, info->alloc_size); in mlxsw_sp1_kvdl_part_init() 227 part = kzalloc(struct_size(part, usage, BITS_TO_LONGS(nr_entries)), in mlxsw_sp1_kvdl_part_init() 280 unsigned int nr_entries; in mlxsw_sp1_kvdl_part_occ() local 284 nr_entries = (info->end_index - in mlxsw_sp1_kvdl_part_occ() 287 while ((bit = find_next_bit(part->usage, nr_entries, bit + 1)) in mlxsw_sp1_kvdl_part_occ() [all …]
|
/linux-6.6.21/arch/x86/xen/ |
D | setup.c | 211 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_find_pfn_range() 474 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_foreach_remap_area() 476 if (entry->type == E820_TYPE_RAM || i == xen_e820_table.nr_entries - 1) { in xen_foreach_remap_area() 609 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_ignore_unusable() 627 for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++) { in xen_is_e820_reserved() 652 for (mapcnt = 0; mapcnt < xen_e820_table.nr_entries; mapcnt++, entry++) { in xen_find_free_area() 746 memmap.nr_entries = ARRAY_SIZE(xen_e820_table.entries); in xen_memory_setup() 759 memmap.nr_entries = 1; in xen_memory_setup() 768 BUG_ON(memmap.nr_entries == 0); in xen_memory_setup() 769 xen_e820_table.nr_entries = memmap.nr_entries; in xen_memory_setup() [all …]
|
/linux-6.6.21/arch/sparc/kernel/ |
D | stacktrace.c | 58 trace->entries[trace->nr_entries++] = pc; in __save_stack_trace() 66 if (trace->nr_entries < in __save_stack_trace() 68 trace->entries[trace->nr_entries++] = pc; in __save_stack_trace() 74 } while (trace->nr_entries < trace->max_entries); in __save_stack_trace()
|
/linux-6.6.21/arch/ia64/kernel/ |
D | stacktrace.c | 19 trace->nr_entries = 0; in ia64_do_save_stack() 25 trace->entries[trace->nr_entries++] = ip; in ia64_do_save_stack() 26 if (trace->nr_entries == trace->max_entries) in ia64_do_save_stack()
|