/linux-6.6.21/tools/bpf/bpftool/ |
D | pids.c | 34 struct obj_refs *refs; in add_ref() local 40 refs = entry->pvalue; in add_ref() 42 for (i = 0; i < refs->ref_cnt; i++) { in add_ref() 43 if (refs->refs[i].pid == e->pid) in add_ref() 47 tmp = realloc(refs->refs, (refs->ref_cnt + 1) * sizeof(*ref)); in add_ref() 53 refs->refs = tmp; in add_ref() 54 ref = &refs->refs[refs->ref_cnt]; in add_ref() 57 refs->ref_cnt++; in add_ref() 63 refs = calloc(1, sizeof(*refs)); in add_ref() 64 if (!refs) { in add_ref() [all …]
|
/linux-6.6.21/include/linux/ |
D | freelist.h | 19 atomic_t refs; member 50 atomic_set_release(&node->refs, 1); in __freelist_add() 57 if (atomic_fetch_add_release(REFS_ON_FREELIST - 1, &node->refs) == 1) in __freelist_add() 70 if (!atomic_fetch_add_release(REFS_ON_FREELIST, &node->refs)) { in freelist_add() 82 unsigned int refs; in freelist_try_get() local 86 refs = atomic_read(&head->refs); in freelist_try_get() 87 if ((refs & REFS_MASK) == 0 || in freelist_try_get() 88 !atomic_try_cmpxchg_acquire(&head->refs, &refs, refs+1)) { in freelist_try_get() 106 WARN_ON_ONCE(atomic_read(&head->refs) & REFS_ON_FREELIST); in freelist_try_get() 112 atomic_fetch_add(-2, &head->refs); in freelist_try_get() [all …]
|
D | refcount.h | 112 atomic_t refs; member 115 #define REFCOUNT_INIT(n) { .refs = ATOMIC_INIT(n), } 136 atomic_set(&r->refs, n); in refcount_set() 147 return atomic_read(&r->refs); in refcount_read() 157 } while (!atomic_try_cmpxchg_relaxed(&r->refs, &old, old + i)); in __refcount_add_not_zero() 193 int old = atomic_fetch_add_relaxed(i, &r->refs); in __refcount_add() 272 int old = atomic_fetch_sub_release(i, &r->refs); in __refcount_sub_and_test() 338 int old = atomic_fetch_sub_release(1, &r->refs); in __refcount_dec()
|
/linux-6.6.21/drivers/media/v4l2-core/ |
D | v4l2-h264.c | 60 b->refs[i].longterm = true; in v4l2_h264_init_reflist_builder() 68 if (!b->refs[i].longterm && dpb[i].frame_num > cur_frame_num) in v4l2_h264_init_reflist_builder() 69 b->refs[i].frame_num = (int)dpb[i].frame_num - in v4l2_h264_init_reflist_builder() 72 b->refs[i].frame_num = dpb[i].frame_num; in v4l2_h264_init_reflist_builder() 74 b->refs[i].top_field_order_cnt = dpb[i].top_field_order_cnt; in v4l2_h264_init_reflist_builder() 75 b->refs[i].bottom_field_order_cnt = dpb[i].bottom_field_order_cnt; in v4l2_h264_init_reflist_builder() 113 return min(b->refs[ref->index].top_field_order_cnt, in v4l2_h264_get_poc() 114 b->refs[ref->index].bottom_field_order_cnt); in v4l2_h264_get_poc() 116 return b->refs[ref->index].top_field_order_cnt; in v4l2_h264_get_poc() 118 return b->refs[ref->index].bottom_field_order_cnt; in v4l2_h264_get_poc() [all …]
|
/linux-6.6.21/tools/include/linux/ |
D | refcount.h | 53 atomic_t refs; member 56 #define REFCOUNT_INIT(n) { .refs = ATOMIC_INIT(n), } 60 atomic_set(&r->refs, n); in refcount_set() 65 return atomic_read(&r->refs); in refcount_read() 78 unsigned int old, new, val = atomic_read(&r->refs); in refcount_inc_not_zero() 89 old = atomic_cmpxchg_relaxed(&r->refs, val, new); in refcount_inc_not_zero() 123 unsigned int old, new, val = atomic_read(&r->refs); in refcount_sub_and_test() 135 old = atomic_cmpxchg_release(&r->refs, val, new); in refcount_sub_and_test()
|
/linux-6.6.21/drivers/xen/ |
D | gntdev-dmabuf.c | 44 grant_ref_t *refs; member 437 int count, u32 domid, u32 *refs, u32 *fd) in dmabuf_exp_from_refs() argument 449 map->grants[i].ref = refs[i]; in dmabuf_exp_from_refs() 488 dmabuf_imp_grant_foreign_access(unsigned long *gfns, u32 *refs, in dmabuf_imp_grant_foreign_access() argument 512 refs[i] = cur_ref; in dmabuf_imp_grant_foreign_access() 522 static void dmabuf_imp_end_foreign_access(u32 *refs, int count) in dmabuf_imp_end_foreign_access() argument 527 if (refs[i] != INVALID_GRANT_REF) in dmabuf_imp_end_foreign_access() 528 gnttab_end_foreign_access(refs[i], NULL); in dmabuf_imp_end_foreign_access() 533 kfree(gntdev_dmabuf->u.imp.refs); in dmabuf_imp_free_storage() 546 gntdev_dmabuf->u.imp.refs = kcalloc(count, in dmabuf_imp_alloc_storage() [all …]
|
/linux-6.6.21/io_uring/ |
D | refs.h | 12 ((unsigned int) atomic_read(&(req->refs)) + 127u <= 127u) 17 return atomic_inc_not_zero(&req->refs); in req_ref_inc_not_zero() 26 return atomic_dec_and_test(&req->refs); in req_ref_put_and_test() 33 atomic_inc(&req->refs); in req_ref_get() 40 atomic_set(&req->refs, nr); in __io_req_set_refcount()
|
/linux-6.6.21/drivers/gpu/drm/nouveau/nvkm/core/ |
D | event.c | 34 if (--event->refs[index * event->types_nr + type] == 0) { in nvkm_event_put() 51 if (++event->refs[index * event->types_nr + type] == 1) { in nvkm_event_get() 175 if (!event->refs || WARN_ON(id >= event->index_nr)) in nvkm_event_ntfy() 194 if (event->refs) { in nvkm_event_fini() 195 kfree(event->refs); in nvkm_event_fini() 196 event->refs = NULL; in nvkm_event_fini() 204 event->refs = kzalloc(array3_size(index_nr, types_nr, sizeof(*event->refs)), GFP_KERNEL); in __nvkm_event_init() 205 if (!event->refs) in __nvkm_event_init()
|
/linux-6.6.21/fs/smb/client/ |
D | dfs_cache.c | 250 static inline void dump_refs(const struct dfs_info3_param *refs, int numrefs) in dump_refs() argument 256 const struct dfs_info3_param *ref = &refs[i]; in dump_refs() 382 static int copy_ref_data(const struct dfs_info3_param *refs, int numrefs, in copy_ref_data() argument 388 ce->ttl = max_t(int, refs[0].ttl, CACHE_MIN_TTL); in copy_ref_data() 390 ce->srvtype = refs[0].server_type; in copy_ref_data() 391 ce->hdr_flags = refs[0].flags; in copy_ref_data() 392 ce->ref_flags = refs[0].ref_flag; in copy_ref_data() 393 ce->path_consumed = refs[0].path_consumed; in copy_ref_data() 398 t = alloc_target(refs[i].node_name, refs[i].path_consumed); in copy_ref_data() 420 static struct cache_entry *alloc_cache_entry(struct dfs_info3_param *refs, int numrefs) in alloc_cache_entry() argument [all …]
|
D | dfs.h | 26 struct dfs_ref refs[MAX_NESTED_LINKS]; member 29 #define ref_walk_start(w) ((w)->refs) 30 #define ref_walk_end(w) (&(w)->refs[ARRAY_SIZE((w)->refs) - 1])
|
/linux-6.6.21/mm/ |
D | gup.c | 71 static inline struct folio *try_get_folio(struct page *page, int refs) in try_get_folio() argument 79 if (unlikely(!folio_ref_try_add_rcu(folio, refs))) in try_get_folio() 92 if (!put_devmap_managed_page_refs(&folio->page, refs)) in try_get_folio() 93 folio_put_refs(folio, refs); in try_get_folio() 126 struct folio *try_grab_folio(struct page *page, int refs, unsigned int flags) in try_grab_folio() argument 137 return try_get_folio(page, refs); in try_grab_folio() 148 folio = try_get_folio(page, refs); in try_grab_folio() 159 if (!put_devmap_managed_page_refs(&folio->page, refs)) in try_grab_folio() 160 folio_put_refs(folio, refs); in try_grab_folio() 173 atomic_add(refs, &folio->_pincount); in try_grab_folio() [all …]
|
D | workingset.c | 240 int refs = folio_lru_refs(folio); in lru_gen_eviction() local 241 int tier = lru_tier_from_refs(refs); in lru_gen_eviction() 250 token = (min_seq << LRU_REFS_WIDTH) | max(refs - 1, 0); in lru_gen_eviction() 255 return pack_shadow(mem_cgroup_id(memcg), pgdat, token, refs); in lru_gen_eviction() 282 int hist, tier, refs; in lru_gen_refault() local 305 refs = (token & (BIT(LRU_REFS_WIDTH) - 1)) + workingset; in lru_gen_refault() 306 tier = lru_tier_from_refs(refs); in lru_gen_refault() 318 if (lru_gen_in_fault() || refs >= BIT(LRU_REFS_WIDTH) - 1) { in lru_gen_refault()
|
/linux-6.6.21/drivers/iommu/arm/arm-smmu-v3/ |
D | arm-smmu-v3-sva.c | 20 refcount_t refs; member 32 refcount_t refs; member 61 refcount_inc(&cd->refs); in arm_smmu_share_asid() 115 refcount_set(&cd->refs, 1); in arm_smmu_alloc_shared_cd() 281 refcount_inc(&smmu_mn->refs); in arm_smmu_mmu_notifier_get() 296 refcount_set(&smmu_mn->refs, 1); in arm_smmu_mmu_notifier_get() 328 if (!refcount_dec_and_test(&smmu_mn->refs)) in arm_smmu_mmu_notifier_put() 363 refcount_inc(&bond->refs); in __arm_smmu_sva_bind() 374 refcount_set(&bond->refs, 1); in __arm_smmu_sva_bind() 553 if (!WARN_ON(!bond) && refcount_dec_and_test(&bond->refs)) { in arm_smmu_sva_remove_dev_pasid()
|
/linux-6.6.21/drivers/thermal/ |
D | k3_j72xx_bandgap.c | 78 int refs[4]; member 90 idx1 = err_vals->refs[seg]; in create_table_segments() 92 idx2 = err_vals->refs[seg + 1]; in create_table_segments() 95 ref1 = err_vals->refs[seg]; in create_table_segments() 96 ref2 = err_vals->refs[seg + 1]; in create_table_segments() 461 err_vals.refs[0] = MINUS40CREF; in k3_j72xx_bandgap_probe() 462 err_vals.refs[1] = PLUS30CREF; in k3_j72xx_bandgap_probe() 463 err_vals.refs[2] = PLUS125CREF; in k3_j72xx_bandgap_probe() 464 err_vals.refs[3] = PLUS150CREF; in k3_j72xx_bandgap_probe()
|
/linux-6.6.21/fs/btrfs/ |
D | delayed-inode.c | 52 refcount_set(&delayed_node->refs, 0); in btrfs_init_delayed_node() 69 refcount_inc(&node->refs); in btrfs_get_delayed_node() 78 refcount_inc(&node->refs); /* can be accessed */ in btrfs_get_delayed_node() 100 if (refcount_inc_not_zero(&node->refs)) { in btrfs_get_delayed_node() 101 refcount_inc(&node->refs); in btrfs_get_delayed_node() 135 refcount_set(&node->refs, 2); in btrfs_get_or_create_delayed_node() 176 refcount_inc(&node->refs); /* inserted into list */ in btrfs_queue_delayed_node() 190 refcount_dec(&node->refs); /* not in the list */ in btrfs_dequeue_delayed_node() 211 refcount_inc(&node->refs); in btrfs_first_delayed_node() 238 refcount_inc(&next->refs); in btrfs_next_delayed_node() [all …]
|
D | delayed-ref.h | 36 refcount_t refs; member 82 refcount_t refs; member 341 WARN_ON(refcount_read(&ref->refs) == 0); in btrfs_put_delayed_ref() 342 if (refcount_dec_and_test(&ref->refs)) { in btrfs_put_delayed_ref() 371 if (refcount_dec_and_test(&head->refs)) in btrfs_put_delayed_ref_head()
|
/linux-6.6.21/lib/ |
D | refcount.c | 59 return atomic_try_cmpxchg_release(&r->refs, &val, 0); in refcount_dec_if_one() 76 unsigned int new, val = atomic_read(&r->refs); in refcount_dec_not_one() 91 } while (!atomic_try_cmpxchg_release(&r->refs, &val, new)); in refcount_dec_not_one()
|
/linux-6.6.21/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
D | cgrp.c | 39 if (refcount_dec_and_test(&ectx->refs)) { in nvkm_cgrp_ectx_put() 65 refcount_inc(&ectx->refs); in nvkm_cgrp_ectx_get() 76 refcount_set(&ectx->refs, 1); in nvkm_cgrp_ectx_get() 100 if (refcount_dec_and_test(&vctx->refs)) { in nvkm_cgrp_vctx_put() 130 refcount_inc(&vctx->refs); in nvkm_cgrp_vctx_get() 151 refcount_set(&vctx->refs, 1); in nvkm_cgrp_vctx_get()
|
/linux-6.6.21/include/uapi/xen/ |
D | gntdev.h | 70 struct ioctl_gntdev_grant_ref refs[1]; member 256 __u32 refs[1]; member 298 __u32 refs[1]; member
|
/linux-6.6.21/tools/testing/selftests/bpf/progs/ |
D | map_kptr.c | 308 if (p_st->cnt.refs.counter != ref) { in test_map_kptr_ref_pre() 318 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_pre() 326 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_pre() 338 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_pre() 355 if (!p_st || p_st->cnt.refs.counter != ref) in test_map_kptr_ref_post() 361 if (p_st->cnt.refs.counter != ref) { in test_map_kptr_ref_post() 371 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_post() 471 if (p->cnt.refs.counter != ref) { in test_map_kptr_ref3()
|
/linux-6.6.21/drivers/net/ethernet/mellanox/mlx4/ |
D | port.c | 69 table->refs[i] = 0; in mlx4_init_mac_table() 83 table->refs[i] = 0; in mlx4_init_vlan_table() 118 if (table->refs[i] && in find_index() 157 if (!table->refs[i]) in mlx4_find_cached_mac() 232 dup_table->refs[index_at_port]) { in __mlx4_register_mac() 242 if (!table->refs[index_at_dup_port] || in __mlx4_register_mac() 251 if (!table->refs[i]) { in __mlx4_register_mac() 255 if (!dup_table->refs[i]) in __mlx4_register_mac() 265 ++table->refs[i]; in __mlx4_register_mac() 308 table->refs[free] = 1; in __mlx4_register_mac() [all …]
|
/linux-6.6.21/drivers/misc/sgi-xp/ |
D | xpc.h | 684 s32 refs = atomic_dec_return(&ch->references); in xpc_msgqueue_deref() local 686 DBUG_ON(refs < 0); in xpc_msgqueue_deref() 687 if (refs == 0) in xpc_msgqueue_deref() 701 s32 refs = atomic_dec_return(&part->references); in xpc_part_deref() local 703 DBUG_ON(refs < 0); in xpc_part_deref() 704 if (refs == 0 && part->setup_state == XPC_P_SS_WTEARDOWN) in xpc_part_deref()
|
/linux-6.6.21/kernel/trace/ |
D | trace_events_user.c | 170 struct user_event_refs *refs; member 2074 struct user_event_refs *refs; in user_events_write_core() local 2088 refs = rcu_dereference_sched(info->refs); in user_events_write_core() 2095 if (likely(refs && idx < refs->count)) in user_events_write_core() 2096 user = refs->events[idx]; in user_events_write_core() 2194 struct user_event_refs *refs, *new_refs; in user_events_ref_add() local 2197 refs = rcu_dereference_protected(info->refs, in user_events_ref_add() 2200 if (refs) { in user_events_ref_add() 2201 count = refs->count; in user_events_ref_add() 2204 if (refs->events[i] == user) in user_events_ref_add() [all …]
|
/linux-6.6.21/drivers/gpu/drm/nouveau/ |
D | nouveau_vmm.c | 65 if (vma && --vma->refs <= 0) { in nouveau_vma_del() 86 vma->refs++; in nouveau_vma_new() 93 vma->refs = 1; in nouveau_vma_new()
|
/linux-6.6.21/drivers/most/ |
D | core.c | 34 int refs; member 852 if (c->pipe0.refs && c->pipe0.comp->tx_completion) in arm_mbo() 855 if (c->pipe1.refs && c->pipe1.comp->tx_completion) in arm_mbo() 963 if (c->pipe0.refs && c->pipe1.refs && in channel_has_mbo() 996 if (c->pipe0.refs && c->pipe1.refs && in most_get_mbo() 1068 if (c->pipe0.refs && c->pipe0.comp->rx_completion && in most_read_completion() 1072 if (c->pipe1.refs && c->pipe1.comp->rx_completion && in most_read_completion() 1101 if (c->pipe0.refs + c->pipe1.refs > 0) in most_start_channel() 1141 c->pipe0.refs++; in most_start_channel() 1143 c->pipe1.refs++; in most_start_channel() [all …]
|