Home
last modified time | relevance | path

Searched refs:waiters (Results 1 – 25 of 45) sorted by relevance

12

/linux-6.1.9/drivers/gpu/drm/radeon/
Dradeon_semaphore.c50 (*semaphore)->waiters = 0; in radeon_semaphore_create()
66 --semaphore->waiters; in radeon_semaphore_emit_signal()
83 ++semaphore->waiters; in radeon_semaphore_emit_wait()
99 if ((*semaphore)->waiters > 0) { in radeon_semaphore_free()
Dradeon_trace.h176 __field(signed, waiters)
182 __entry->waiters = sem->waiters;
187 __entry->waiters, __entry->gpu_addr)
/linux-6.1.9/kernel/locking/
Dpercpu-rwsem.c23 init_waitqueue_head(&sem->waiters); in __percpu_init_rwsem()
146 spin_lock_irq(&sem->waiters.lock); in percpu_rwsem_wait()
154 __add_wait_queue_entry_tail(&sem->waiters, &wq_entry); in percpu_rwsem_wait()
156 spin_unlock_irq(&sem->waiters.lock); in percpu_rwsem_wait()
273 __wake_up(&sem->waiters, TASK_NORMAL, 1, sem); in percpu_up_write()
Drtmutex_common.h95 return !RB_EMPTY_ROOT(&lock->waiters.rb_root); in rt_mutex_has_waiters()
106 struct rb_node *leftmost = rb_first_cached(&lock->waiters); in rt_mutex_waiter_is_top_waiter()
113 struct rb_node *leftmost = rb_first_cached(&lock->waiters); in rt_mutex_top_waiter()
161 lock->waiters = RB_ROOT_CACHED; in __rt_mutex_base_init()
Dww_mutex.h96 struct rb_node *n = rb_first(&lock->rtmutex.waiters.rb_root); in __ww_waiter_first()
123 struct rb_node *n = rb_last(&lock->rtmutex.waiters.rb_root); in __ww_waiter_last()
/linux-6.1.9/Documentation/locking/
Drt-mutex.rst32 The enqueueing of the waiters into the rtmutex waiter tree is done in
35 priority waiters tree. This tree too queues in priority order. Whenever
42 without waiters. The optimized fastpath operations require cmpxchg
50 keep track of the "lock has waiters" state:
56 NULL 1 lock is free and has waiters and the top waiter
59 taskpointer 1 lock is held and has waiters [2]_
71 waiters. This can happen when grabbing the lock in the slow path.
Drt-mutex-design.rst107 place the task in the waiters rbtree of a mutex as well as the
113 waiters
199 Every mutex keeps track of all the waiters that are blocked on itself. The
200 mutex has a rbtree to store these waiters by priority. This tree is protected
209 a tree of all top waiters of the mutexes that are owned by the process.
210 Note that this tree only holds the top waiters and not all waiters that are
309 flag. It's set whenever there are waiters on a mutex.
358 the pi_waiters of a task holds an order by priority of all the top waiters
410 in the pi_waiters and waiters trees that the task is blocked on. This function
434 the task on the waiters tree of the mutex, and if need be, the pi_waiters
[all …]
Dfutex-requeue-pi.rst7 left without an owner if it has waiters; doing so would break the PI
48 has waiters. Note that pthread_cond_wait() attempts to lock the
50 underlying rt_mutex with waiters, and no owner, breaking the
89 In order to ensure the rt_mutex has an owner if it has waiters, it
/linux-6.1.9/drivers/md/persistent-data/
Ddm-block-manager.c46 struct list_head waiters; member
153 list_for_each_entry_safe(w, tmp, &lock->waiters, list) { in __wake_many()
179 INIT_LIST_HEAD(&lock->waiters); in bl_init()
188 list_empty(&lock->waiters); in __available_for_read()
214 list_add_tail(&w.list, &lock->waiters); in bl_down_read()
249 if (!list_empty(&lock->waiters)) in bl_up_read()
266 if (lock->count == 0 && list_empty(&lock->waiters)) { in bl_down_write()
281 list_add(&w.list, &lock->waiters); in bl_down_write()
295 if (!list_empty(&lock->waiters)) in bl_up_write()
/linux-6.1.9/tools/testing/selftests/futex/include/
Dfutex2test.h18 static inline int futex_waitv(volatile struct futex_waitv *waiters, unsigned long nr_waiters, in futex_waitv() argument
21 return syscall(__NR_futex_waitv, waiters, nr_waiters, flags, timo, clockid); in futex_waitv()
/linux-6.1.9/include/linux/
Dpercpu-rwsem.h16 wait_queue_head_t waiters; member
35 .waiters = __WAIT_QUEUE_HEAD_INITIALIZER(name.waiters), \
Drtmutex.h25 struct rb_root_cached waiters; member
32 .waiters = RB_ROOT_CACHED, \
/linux-6.1.9/kernel/futex/
Dfutex.h46 atomic_t waiters; member
187 atomic_inc(&hb->waiters); in futex_hb_waiters_inc()
202 atomic_dec(&hb->waiters); in futex_hb_waiters_dec()
213 return atomic_read(&hb->waiters); in futex_hb_waiters_pending()
Dsyscalls.c246 SYSCALL_DEFINE5(futex_waitv, struct futex_waitv __user *, waiters, in SYSCALL_DEFINE5() argument
260 if (!nr_futexes || nr_futexes > FUTEX_WAITV_MAX || !waiters) in SYSCALL_DEFINE5()
294 ret = futex_parse_waitv(futexv, waiters, nr_futexes); in SYSCALL_DEFINE5()
/linux-6.1.9/drivers/gpu/host1x/
Ddebug.c93 unsigned int waiters = 0; in show_syncpts() local
97 waiters++; in show_syncpts()
103 if (!show_all && !min && !max && !waiters) in show_syncpts()
108 i, m->syncpt[i].name, min, max, waiters); in show_syncpts()
/linux-6.1.9/Documentation/userspace-api/
Dfutex2.rst25 futex_waitv(struct futex_waitv *waiters, unsigned int nr_futexes,
39 pointer for the first item of the array is passed as ``waiters``. An invalid
40 address for ``waiters`` or for any ``uaddr`` returns ``-EFAULT``.
52 For each entry in ``waiters`` array, the current value at ``uaddr`` is compared
/linux-6.1.9/drivers/greybus/
Dgreybus_trace.h102 __field(int, waiters)
112 __entry->waiters = atomic_read(&operation->waiters);
118 __entry->active, __entry->waiters, __entry->errno)
Doperation.c90 if (atomic_read(&operation->waiters)) in gb_operation_put_active()
554 atomic_set(&operation->waiters, 0); in gb_operation_create_common()
1082 atomic_inc(&operation->waiters); in gb_operation_cancel()
1085 atomic_dec(&operation->waiters); in gb_operation_cancel()
1109 atomic_inc(&operation->waiters); in gb_operation_cancel_incoming()
1112 atomic_dec(&operation->waiters); in gb_operation_cancel_incoming()
/linux-6.1.9/drivers/gpu/drm/amd/amdkfd/
Dkfd_events.c896 static void free_waiters(uint32_t num_events, struct kfd_event_waiter *waiters, in free_waiters() argument
902 if (waiters[i].event) { in free_waiters()
903 spin_lock(&waiters[i].event->lock); in free_waiters()
904 remove_wait_queue(&waiters[i].event->wq, in free_waiters()
905 &waiters[i].wait); in free_waiters()
906 if (undo_auto_reset && waiters[i].activated && in free_waiters()
907 waiters[i].event && waiters[i].event->auto_reset) in free_waiters()
908 set_event(waiters[i].event); in free_waiters()
909 spin_unlock(&waiters[i].event->lock); in free_waiters()
912 kfree(waiters); in free_waiters()
/linux-6.1.9/drivers/tee/optee/
Dcall.c62 list_add_tail(&w->list_node, &cq->waiters); in optee_cq_wait_init()
77 list_add_tail(&w->list_node, &cq->waiters); in optee_cq_wait_for_completion()
86 list_for_each_entry(w, &cq->waiters, list_node) { in optee_cq_complete_one()
/linux-6.1.9/Documentation/scheduler/
Dcompletion.rst247 achieved calls complete() to signal exactly one of the waiters that it can
252 ... or calls complete_all() to signal all current and future waiters::
262 of waiters to continue - each call to complete() will simply increment the
287 completions that were not yet consumed by waiters (implying that there are
288 waiters) and true otherwise::
/linux-6.1.9/include/linux/greybus/
Doperation.h108 atomic_t waiters; member
/linux-6.1.9/Documentation/translations/zh_CN/scheduler/
Dcompletion.rst223 ... or calls complete_all() to signal all current and future waiters::
/linux-6.1.9/tools/testing/selftests/filesystems/epoll/
Depoll_wakeup_test.c3162 int waiters; member
3192 __atomic_fetch_add(&ctx->waiters, 1, __ATOMIC_ACQUIRE); in epoll60_wait_thread()
3206 __atomic_fetch_sub(&ctx->waiters, 1, __ATOMIC_RELEASE); in epoll60_wait_thread()
3226 return __atomic_load_n(&ctx->waiters, __ATOMIC_ACQUIRE); in count_waiters()
3232 pthread_t waiters[ARRAY_SIZE(ctx.evfd)]; in TEST() local
3252 for (i = 0; i < ARRAY_SIZE(waiters); i++) in TEST()
3253 ASSERT_EQ(pthread_create(&waiters[i], NULL, in TEST()
3290 for (i = 0; i < ARRAY_SIZE(waiters); i++) in TEST()
3291 ret = pthread_kill(waiters[i], SIGUSR1); in TEST()
3292 for (i = 0; i < ARRAY_SIZE(waiters); i++) in TEST()
[all …]
/linux-6.1.9/fs/xfs/
Dxfs_log.c204 INIT_LIST_HEAD(&head->waiters); in xlog_grant_head_init()
215 list_for_each_entry(tic, &head->waiters, t_queue) in xlog_grant_head_wake_all()
247 list_for_each_entry(tic, &head->waiters, t_queue) { in xlog_grant_head_wake()
294 list_add_tail(&tic->t_queue, &head->waiters); in xlog_grant_head_wait()
359 if (!list_empty_careful(&head->waiters)) { in xlog_grant_head_check()
1155 if (!list_empty_careful(&log->l_write_head.waiters)) { in xfs_log_space_wake()
1164 if (!list_empty_careful(&log->l_reserve_head.waiters)) { in xfs_log_space_wake()

12