12#ifndef ZEPHYR_INCLUDE_SPINLOCK_H_
13#define ZEPHYR_INCLUDE_SPINLOCK_H_
34struct z_spinlock_key {
50#ifdef CONFIG_TICKET_SPINLOCKS
69#ifdef CONFIG_SPIN_VALIDATE
74#ifdef CONFIG_SPIN_LOCK_TIME_LIMIT
81#if defined(CONFIG_NONZERO_SPINLOCK_SIZE) && !defined(CONFIG_SMP) && !defined(CONFIG_SPIN_VALIDATE)
103#ifdef CONFIG_SPIN_VALIDATE
105bool z_spin_unlock_valid(
struct k_spinlock *l);
106void z_spin_lock_set_owner(
struct k_spinlock *l);
107BUILD_ASSERT(CONFIG_MP_MAX_NUM_CPUS <= 4,
"Too many CPUs for mask");
109# ifdef CONFIG_KERNEL_COHERENCE
110bool z_spin_lock_mem_coherent(
struct k_spinlock *l);
131#ifdef CONFIG_SPIN_VALIDATE
132 __ASSERT(z_spin_lock_valid(l),
"Invalid spinlock %p", l);
133#ifdef CONFIG_KERNEL_COHERENCE
134 __ASSERT_NO_MSG(z_spin_lock_mem_coherent(l));
142#ifdef CONFIG_SPIN_VALIDATE
143 z_spin_lock_set_owner(l);
144#if defined(CONFIG_SPIN_LOCK_TIME_LIMIT) && (CONFIG_SPIN_LOCK_TIME_LIMIT != 0)
192 z_spinlock_validate_pre(l);
194#ifdef CONFIG_TICKET_SPINLOCKS
210 z_spinlock_validate_post(l);
233 z_spinlock_validate_pre(l);
235#ifdef CONFIG_TICKET_SPINLOCKS
256 if (!
atomic_cas(&l->tail, ticket_val, ticket_val + 1)) {
265 z_spinlock_validate_post(l);
303#ifdef CONFIG_SPIN_VALIDATE
304 __ASSERT(z_spin_unlock_valid(l),
"Not my spinlock %p", l);
306#if defined(CONFIG_SPIN_LOCK_TIME_LIMIT) && (CONFIG_SPIN_LOCK_TIME_LIMIT != 0)
309 __ASSERT(delta < CONFIG_SPIN_LOCK_TIME_LIMIT,
310 "Spin lock %p held %u cycles, longer than limit of %u cycles",
311 l, delta, CONFIG_SPIN_LOCK_TIME_LIMIT);
316#ifdef CONFIG_TICKET_SPINLOCKS
337#if defined(CONFIG_SMP) && defined(CONFIG_TEST)
347#ifdef CONFIG_TICKET_SPINLOCKS
350 return !
atomic_cas(&l->tail, ticket_val, ticket_val);
361#ifdef CONFIG_SPIN_VALIDATE
362 __ASSERT(z_spin_unlock_valid(l),
"Not my spinlock %p", l);
365#ifdef CONFIG_TICKET_SPINLOCKS
373#if defined(CONFIG_SPIN_VALIDATE) && defined(__GNUC__)
376 __ASSERT(k->key,
"K_SPINLOCK exited with goto, break or return, "
377 "use K_SPINLOCK_BREAK instead.");
379#define K_SPINLOCK_ONEXIT __attribute__((__cleanup__(z_spin_onexit)))
381#define K_SPINLOCK_ONEXIT
394#define K_SPINLOCK_BREAK continue
437#define K_SPINLOCK(lck) \
438 for (k_spinlock_key_t __i K_SPINLOCK_ONEXIT = {}, __key = k_spin_lock(lck); !__i.key; \
439 k_spin_unlock((lck), __key), __i.key = 1)
uint32_t sys_clock_cycle_get_32(void)
static ALWAYS_INLINE unsigned int arch_irq_lock(void)
Disable all interrupts on the local CPU.
Definition irq.h:168
static ALWAYS_INLINE void arch_irq_unlock(unsigned int key)
Definition irq.h:176
void arch_spin_relax(void)
Perform architecture specific processing within spin loops.
long atomic_t
Definition atomic_types.h:15
atomic_t atomic_val_t
Definition atomic_types.h:16
atomic_val_t atomic_get(const atomic_t *target)
Atomic get.
atomic_val_t atomic_clear(atomic_t *target)
Atomic clear.
atomic_val_t atomic_inc(atomic_t *target)
Atomic increment.
bool atomic_cas(atomic_t *target, atomic_val_t old_value, atomic_val_t new_value)
Atomic compare-and-set.
static ALWAYS_INLINE int k_spin_trylock(struct k_spinlock *l, k_spinlock_key_t *k)
Attempt to lock a spinlock.
Definition spinlock.h:229
static ALWAYS_INLINE void k_spin_unlock(struct k_spinlock *l, k_spinlock_key_t key)
Unlock a spin lock.
Definition spinlock.h:299
static ALWAYS_INLINE k_spinlock_key_t k_spin_lock(struct k_spinlock *l)
Lock a spinlock.
Definition spinlock.h:181
struct z_spinlock_key k_spinlock_key_t
Spinlock key type.
Definition spinlock.h:126
#define EBUSY
Mount device busy.
Definition errno.h:54
__UINT32_TYPE__ uint32_t
Definition stdint.h:90
__UINTPTR_TYPE__ uintptr_t
Definition stdint.h:105
Kernel Spin Lock.
Definition spinlock.h:45