Home
last modified time | relevance | path

Searched refs:arch_spin_unlock (Results 1 – 25 of 48) sorted by relevance

12

/linux-6.6.21/arch/parisc/include/asm/
Dspinlock.h49 static inline void arch_spin_unlock(arch_spinlock_t *x) in arch_spin_unlock() function
96 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_trylock()
121 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_trylock()
146 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_unlock()
157 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_unlock()
/linux-6.6.21/arch/arc/include/asm/
Dspinlock.h67 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
261 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
315 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_trylock()
340 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_trylock()
365 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_unlock()
376 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_unlock()
Dsmp.h117 arch_spin_unlock(&smp_atomic_ops_lock); \
/linux-6.6.21/include/linux/
Dspinlock_up.h45 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
65 # define arch_spin_unlock(lock) do { barrier(); (void)(lock); } while (0) macro
/linux-6.6.21/arch/arm/common/
Dmcpm_entry.c232 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_power_up()
268 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_power_down()
274 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_power_down()
335 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_suspend()
365 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_powered_up()
/linux-6.6.21/kernel/locking/
Dqrwlock.c56 arch_spin_unlock(&lock->wait_lock); in queued_read_lock_slowpath()
88 arch_spin_unlock(&lock->wait_lock); in queued_write_lock_slowpath()
/linux-6.6.21/kernel/kcsan/
Dselftest.c166 KCSAN_CHECK_READ_BARRIER(arch_spin_unlock(&arch_spinlock)); in test_barrier()
195 KCSAN_CHECK_WRITE_BARRIER(arch_spin_unlock(&arch_spinlock)); in test_barrier()
227 KCSAN_CHECK_RW_BARRIER(arch_spin_unlock(&arch_spinlock)); in test_barrier()
/linux-6.6.21/arch/x86/mm/
Dkmmio.c356 arch_spin_unlock(&kmmio_lock); in post_kmmio_handler()
467 arch_spin_unlock(&kmmio_lock); in register_kmmio_probe()
516 arch_spin_unlock(&kmmio_lock); in remove_kmmio_fault_pages()
559 arch_spin_unlock(&kmmio_lock); in unregister_kmmio_probe()
/linux-6.6.21/kernel/trace/
Dtrace_stack.c282 arch_spin_unlock(&stack_trace_max_lock); in check_stack()
360 arch_spin_unlock(&stack_trace_max_lock); in stack_max_size_write()
410 arch_spin_unlock(&stack_trace_max_lock); in t_stop()
Dtrace_clock.c139 arch_spin_unlock(&trace_clock_struct.lock); in trace_clock_global()
/linux-6.6.21/arch/powerpc/kvm/
Dbook3s_xive.c575 arch_spin_unlock(&sb->lock); in xive_vm_h_eoi()
583 arch_spin_unlock(&sb->lock); in xive_vm_h_eoi()
1105 arch_spin_unlock(&sb->lock); in xive_lock_and_mask()
1139 arch_spin_unlock(&sb->lock); in xive_lock_for_unmask()
1365 arch_spin_unlock(&sb->lock); in kvmppc_xive_set_xive()
1387 arch_spin_unlock(&sb->lock); in kvmppc_xive_get_xive()
1426 arch_spin_unlock(&sb->lock); in kvmppc_xive_int_on()
1452 arch_spin_unlock(&sb->lock); in kvmppc_xive_int_off()
1643 arch_spin_unlock(&sb->lock); in kvmppc_xive_set_mapped()
1723 arch_spin_unlock(&sb->lock); in kvmppc_xive_clr_mapped()
[all …]
Dbook3s_xics.c160 arch_spin_unlock(&ics->lock); in write_xive()
216 arch_spin_unlock(&ics->lock); in kvmppc_xics_get_xive()
473 arch_spin_unlock(&ics->lock); in icp_deliver_irq()
502 arch_spin_unlock(&ics->lock); in icp_deliver_irq()
509 arch_spin_unlock(&ics->lock); in icp_deliver_irq()
1009 arch_spin_unlock(&ics->lock); in xics_debug_show()
1217 arch_spin_unlock(&ics->lock); in xics_get_source()
1275 arch_spin_unlock(&ics->lock); in xics_set_source()
Dbook3s_xive_native.c272 arch_spin_unlock(&sb->lock); in xive_native_esb_fault()
411 arch_spin_unlock(&sb->lock); in kvmppc_xive_native_set_source()
459 arch_spin_unlock(&sb->lock); in kvmppc_xive_native_update_source_config()
537 arch_spin_unlock(&sb->lock); in kvmppc_xive_native_sync_source()
847 arch_spin_unlock(&sb->lock); in kvmppc_xive_reset()
929 arch_spin_unlock(&sb->lock); in kvmppc_xive_native_eq_sync()
1250 arch_spin_unlock(&sb->lock); in xive_native_debug_show()
/linux-6.6.21/tools/include/linux/
Dspinlock.h30 static inline void arch_spin_unlock(arch_spinlock_t *mutex) in arch_spin_unlock() function
/linux-6.6.21/arch/hexagon/include/asm/
Dspinlock.h130 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
/linux-6.6.21/include/asm-generic/
Dspinlock.h63 static __always_inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
Dqspinlock.h148 #define arch_spin_unlock(l) queued_spin_unlock(l) macro
/linux-6.6.21/arch/alpha/include/asm/
Dspinlock.h24 static inline void arch_spin_unlock(arch_spinlock_t * lock) in arch_spin_unlock() function
/linux-6.6.21/arch/x86/kernel/
Dtsc_sync.c288 arch_spin_unlock(&sync_lock); in check_tsc_warp()
318 arch_spin_unlock(&sync_lock); in check_tsc_warp()
/linux-6.6.21/arch/s390/lib/
Dspinlock.c286 arch_spin_unlock(&rw->wait); in arch_read_lock_wait()
309 arch_spin_unlock(&rw->wait); in arch_write_lock_wait()
/linux-6.6.21/arch/sh/include/asm/
Dspinlock-cas.h33 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
Dspinlock-llsc.h46 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
/linux-6.6.21/arch/sparc/include/asm/
Dspinlock_32.h48 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
/linux-6.6.21/arch/powerpc/include/asm/
Dqspinlock.h166 #define arch_spin_unlock(l) queued_spin_unlock(l) macro
/linux-6.6.21/arch/s390/include/asm/
Dspinlock.h77 static inline void arch_spin_unlock(arch_spinlock_t *lp) in arch_spin_unlock() function

12