Home
last modified time | relevance | path

Searched refs:arch_spin_unlock (Results 1 – 25 of 47) sorted by relevance

12

/linux-6.1.9/arch/parisc/include/asm/
Dspinlock.h26 static inline void arch_spin_unlock(arch_spinlock_t *x) in arch_spin_unlock() function
69 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_trylock()
94 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_trylock()
119 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_unlock()
130 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_unlock()
Dfutex.h30 arch_spin_unlock(s); in _futex_spin_unlock_irqrestore()
/linux-6.1.9/arch/arc/include/asm/
Dspinlock.h67 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
261 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
315 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_trylock()
340 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_trylock()
365 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_unlock()
376 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_unlock()
Dsmp.h115 arch_spin_unlock(&smp_atomic_ops_lock); \
/linux-6.1.9/include/linux/
Dspinlock_up.h45 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
65 # define arch_spin_unlock(lock) do { barrier(); (void)(lock); } while (0) macro
/linux-6.1.9/kernel/locking/
Dqrwlock.c56 arch_spin_unlock(&lock->wait_lock); in queued_read_lock_slowpath()
88 arch_spin_unlock(&lock->wait_lock); in queued_write_lock_slowpath()
/linux-6.1.9/arch/arm/common/
Dmcpm_entry.c232 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_power_up()
268 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_power_down()
274 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_power_down()
335 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_suspend()
365 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_powered_up()
/linux-6.1.9/kernel/kcsan/
Dselftest.c173 KCSAN_CHECK_READ_BARRIER(arch_spin_unlock(&arch_spinlock)); in test_barrier()
202 KCSAN_CHECK_WRITE_BARRIER(arch_spin_unlock(&arch_spinlock)); in test_barrier()
234 KCSAN_CHECK_RW_BARRIER(arch_spin_unlock(&arch_spinlock)); in test_barrier()
/linux-6.1.9/kernel/trace/
Dtrace_stack.c282 arch_spin_unlock(&stack_trace_max_lock); in check_stack()
360 arch_spin_unlock(&stack_trace_max_lock); in stack_max_size_write()
410 arch_spin_unlock(&stack_trace_max_lock); in t_stop()
Dtrace_clock.c139 arch_spin_unlock(&trace_clock_struct.lock); in trace_clock_global()
/linux-6.1.9/arch/powerpc/kvm/
Dbook3s_xive.c575 arch_spin_unlock(&sb->lock); in xive_vm_h_eoi()
583 arch_spin_unlock(&sb->lock); in xive_vm_h_eoi()
1105 arch_spin_unlock(&sb->lock); in xive_lock_and_mask()
1139 arch_spin_unlock(&sb->lock); in xive_lock_for_unmask()
1365 arch_spin_unlock(&sb->lock); in kvmppc_xive_set_xive()
1387 arch_spin_unlock(&sb->lock); in kvmppc_xive_get_xive()
1426 arch_spin_unlock(&sb->lock); in kvmppc_xive_int_on()
1452 arch_spin_unlock(&sb->lock); in kvmppc_xive_int_off()
1643 arch_spin_unlock(&sb->lock); in kvmppc_xive_set_mapped()
1723 arch_spin_unlock(&sb->lock); in kvmppc_xive_clr_mapped()
[all …]
Dbook3s_xics.c160 arch_spin_unlock(&ics->lock); in write_xive()
216 arch_spin_unlock(&ics->lock); in kvmppc_xics_get_xive()
473 arch_spin_unlock(&ics->lock); in icp_deliver_irq()
502 arch_spin_unlock(&ics->lock); in icp_deliver_irq()
509 arch_spin_unlock(&ics->lock); in icp_deliver_irq()
1009 arch_spin_unlock(&ics->lock); in xics_debug_show()
1217 arch_spin_unlock(&ics->lock); in xics_get_source()
1275 arch_spin_unlock(&ics->lock); in xics_set_source()
Dbook3s_xive_native.c273 arch_spin_unlock(&sb->lock); in xive_native_esb_fault()
412 arch_spin_unlock(&sb->lock); in kvmppc_xive_native_set_source()
460 arch_spin_unlock(&sb->lock); in kvmppc_xive_native_update_source_config()
538 arch_spin_unlock(&sb->lock); in kvmppc_xive_native_sync_source()
848 arch_spin_unlock(&sb->lock); in kvmppc_xive_reset()
930 arch_spin_unlock(&sb->lock); in kvmppc_xive_native_eq_sync()
1251 arch_spin_unlock(&sb->lock); in xive_native_debug_show()
/linux-6.1.9/tools/include/linux/
Dspinlock.h30 static inline void arch_spin_unlock(arch_spinlock_t *mutex) in arch_spin_unlock() function
/linux-6.1.9/arch/hexagon/include/asm/
Dspinlock.h130 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
/linux-6.1.9/include/asm-generic/
Dspinlock.h63 static __always_inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
Dqspinlock.h148 #define arch_spin_unlock(l) queued_spin_unlock(l) macro
/linux-6.1.9/arch/alpha/include/asm/
Dspinlock.h24 static inline void arch_spin_unlock(arch_spinlock_t * lock) in arch_spin_unlock() function
/linux-6.1.9/arch/x86/kernel/
Dtsc_sync.c288 arch_spin_unlock(&sync_lock); in check_tsc_warp()
318 arch_spin_unlock(&sync_lock); in check_tsc_warp()
/linux-6.1.9/arch/s390/lib/
Dspinlock.c286 arch_spin_unlock(&rw->wait); in arch_read_lock_wait()
309 arch_spin_unlock(&rw->wait); in arch_write_lock_wait()
/linux-6.1.9/arch/sh/include/asm/
Dspinlock-cas.h33 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
Dspinlock-llsc.h46 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
/linux-6.1.9/arch/sparc/include/asm/
Dspinlock_32.h48 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
/linux-6.1.9/arch/s390/include/asm/
Dspinlock.h77 static inline void arch_spin_unlock(arch_spinlock_t *lp) in arch_spin_unlock() function
/linux-6.1.9/arch/powerpc/platforms/pasemi/
Dsetup.c112 arch_spin_unlock(&timebase_lock); in pas_give_timebase()
128 arch_spin_unlock(&timebase_lock); in pas_take_timebase()

12