13#ifndef ZEPHYR_INCLUDE_KERNEL_H_
14#define ZEPHYR_INCLUDE_KERNEL_H_
16#if !defined(_ASMLANGUAGE)
38#if CONFIG_NUM_COOP_PRIORITIES + CONFIG_NUM_PREEMPT_PRIORITIES == 0
39#error Zero available thread priorities defined!
42#define K_PRIO_COOP(x) (-(CONFIG_NUM_COOP_PRIORITIES - (x)))
43#define K_PRIO_PREEMPT(x) (x)
45#define K_HIGHEST_THREAD_PRIO (-CONFIG_NUM_COOP_PRIORITIES)
46#define K_LOWEST_THREAD_PRIO CONFIG_NUM_PREEMPT_PRIORITIES
47#define K_IDLE_PRIO K_LOWEST_THREAD_PRIO
48#define K_HIGHEST_APPLICATION_THREAD_PRIO (K_HIGHEST_THREAD_PRIO)
49#define K_LOWEST_APPLICATION_THREAD_PRIO (K_LOWEST_THREAD_PRIO - 1)
52#define _POLL_EVENT_OBJ_INIT(obj) \
53 .poll_events = SYS_DLIST_STATIC_INIT(&obj.poll_events),
54#define _POLL_EVENT sys_dlist_t poll_events
56#define _POLL_EVENT_OBJ_INIT(obj)
164#define K_ESSENTIAL (BIT(0))
166#if defined(CONFIG_FPU_SHARING)
176#define K_FP_REGS (BIT(1))
185#define K_USER (BIT(2))
195#define K_INHERIT_PERMS (BIT(3))
206#define K_CALLBACK_STATE (BIT(4))
211#if defined(CONFIG_FPU_SHARING) && defined(CONFIG_X86_SSE)
221#define K_SSE_REGS (BIT(7))
227#if !defined(_ASMLANGUAGE)
280 void *p1,
void *p2,
void *p3,
321#define k_thread_access_grant(thread, ...) \
322 FOR_EACH_FIXED_ARG(k_object_access_grant, (;), thread, __VA_ARGS__)
344#if defined(CONFIG_INIT_STACKS) && defined(CONFIG_THREAD_STACK_INFO)
365__syscall
int k_thread_stack_space_get(
const struct k_thread *
thread,
369#if (CONFIG_HEAP_MEM_POOL_SIZE > 0)
495__syscall
k_tid_t z_current_get(
void);
497#ifdef CONFIG_THREAD_LOCAL_STORAGE
499extern __thread
k_tid_t z_tls_current;
511#ifdef CONFIG_THREAD_LOCAL_STORAGE
512 return z_tls_current;
514 return z_current_get();
558#ifdef CONFIG_SYS_CLOCK_EXISTS
569static inline k_ticks_t z_impl_k_thread_timeout_expires_ticks(
572 return z_timeout_expires(&
t->
base.timeout);
584static inline k_ticks_t z_impl_k_thread_timeout_remaining_ticks(
587 return z_timeout_remaining(&
t->
base.timeout);
599struct _static_thread_data {
602 unsigned int init_stack_size;
610 void (*init_abort)(void);
611 const char *init_name;
614#define Z_THREAD_INITIALIZER(thread, stack, stack_size, \
616 prio, options, delay, abort, tname) \
618 .init_thread = (thread), \
619 .init_stack = (stack), \
620 .init_stack_size = (stack_size), \
621 .init_entry = (k_thread_entry_t)entry, \
622 .init_p1 = (void *)p1, \
623 .init_p2 = (void *)p2, \
624 .init_p3 = (void *)p3, \
625 .init_prio = (prio), \
626 .init_options = (options), \
627 .init_delay = (delay), \
628 .init_abort = (abort), \
629 .init_name = STRINGIFY(tname), \
665#define K_THREAD_DEFINE(name, stack_size, \
667 prio, options, delay) \
668 K_THREAD_STACK_DEFINE(_k_thread_stack_##name, stack_size); \
669 struct k_thread _k_thread_obj_##name; \
670 STRUCT_SECTION_ITERABLE(_static_thread_data, _k_thread_data_##name) = \
671 Z_THREAD_INITIALIZER(&_k_thread_obj_##name, \
672 _k_thread_stack_##name, stack_size, \
673 entry, p1, p2, p3, prio, options, delay, \
675 const k_tid_t name = (k_tid_t)&_k_thread_obj_##name
716#ifdef CONFIG_SCHED_DEADLINE
752#ifdef CONFIG_SCHED_CPU_MASK
918 extern bool z_sys_post_kernel;
920 return !z_sys_post_kernel;
1048#define K_NO_WAIT Z_TIMEOUT_NO_WAIT
1062#define K_NSEC(t) Z_TIMEOUT_NS(t)
1076#define K_USEC(t) Z_TIMEOUT_US(t)
1088#define K_CYC(t) Z_TIMEOUT_CYC(t)
1100#define K_TICKS(t) Z_TIMEOUT_TICKS(t)
1112#define K_MSEC(ms) Z_TIMEOUT_MS(ms)
1124#define K_SECONDS(s) K_MSEC((s) * MSEC_PER_SEC)
1136#define K_MINUTES(m) K_SECONDS((m) * 60)
1148#define K_HOURS(h) K_MINUTES((h) * 60)
1158#define K_FOREVER Z_FOREVER
1160#ifdef CONFIG_TIMEOUT_64BIT
1173#define K_TIMEOUT_ABS_TICKS(t) \
1174 Z_TIMEOUT_TICKS(Z_TICK_ABS((k_ticks_t)MAX(t, 0)))
1187#define K_TIMEOUT_ABS_MS(t) K_TIMEOUT_ABS_TICKS(k_ms_to_ticks_ceil64(t))
1201#define K_TIMEOUT_ABS_US(t) K_TIMEOUT_ABS_TICKS(k_us_to_ticks_ceil64(t))
1215#define K_TIMEOUT_ABS_NS(t) K_TIMEOUT_ABS_TICKS(k_ns_to_ticks_ceil64(t))
1229#define K_TIMEOUT_ABS_CYC(t) K_TIMEOUT_ABS_TICKS(k_cyc_to_ticks_ceil64(t))
1253 void (*expiry_fn)(
struct k_timer *
timer);
1256 void (*stop_fn)(
struct k_timer *
timer);
1270#define Z_TIMER_INITIALIZER(obj, expiry, stop) \
1274 .fn = z_timer_expiration_handler, \
1277 .wait_q = Z_WAIT_Q_INIT(&obj.wait_q), \
1278 .expiry_fn = expiry, \
1333#define K_TIMER_DEFINE(name, expiry_fn, stop_fn) \
1334 STRUCT_SECTION_ITERABLE(k_timer, name) = \
1335 Z_TIMER_INITIALIZER(name, expiry_fn, stop_fn)
1418#ifdef CONFIG_SYS_CLOCK_EXISTS
1432static inline k_ticks_t z_impl_k_timer_expires_ticks(
1433 const struct k_timer *
timer)
1435 return z_timeout_expires(&
timer->timeout);
1447static inline k_ticks_t z_impl_k_timer_remaining_ticks(
1448 const struct k_timer *
timer)
1450 return z_timeout_remaining(&
timer->timeout);
1487static inline void z_impl_k_timer_user_data_set(
struct k_timer *
timer,
1502static inline void *z_impl_k_timer_user_data_get(
const struct k_timer *
timer)
1504 return timer->user_data;
1584 delta = uptime - *reftime;
1615 if (!
IS_ENABLED(CONFIG_TIMER_HAS_64BIT_CYCLE_COUNTER)) {
1616 __ASSERT(0,
"64-bit cycle counter not enabled on this platform. "
1617 "See CONFIG_TIMER_HAS_64BIT_CYCLE_COUNTER");
1642#define Z_QUEUE_INITIALIZER(obj) \
1644 .data_q = SYS_SFLIST_STATIC_INIT(&obj.data_q), \
1646 .wait_q = Z_WAIT_Q_INIT(&obj.wait_q), \
1647 _POLL_EVENT_OBJ_INIT(obj) \
1650extern void *z_queue_node_peek(
sys_sfnode_t *node,
bool needs_free);
1871static inline int z_impl_k_queue_is_empty(
struct k_queue *
queue)
1907#define K_QUEUE_DEFINE(name) \
1908 STRUCT_SECTION_ITERABLE(k_queue, name) = \
1909 Z_QUEUE_INITIALIZER(name)
1913#ifdef CONFIG_USERSPACE
1934struct z_futex_data {
1939#define Z_FUTEX_DATA_INITIALIZER(obj) \
1941 .wait_q = Z_WAIT_Q_INIT(&obj.wait_q) \
2008#define Z_EVENT_INITIALIZER(obj) \
2010 .wait_q = Z_WAIT_Q_INIT(&obj.wait_q), \
2110#define K_EVENT_DEFINE(name) \
2111 STRUCT_SECTION_ITERABLE(k_event, name) = \
2112 Z_EVENT_INITIALIZER(name);
2117 struct k_queue _queue;
2123#define Z_FIFO_INITIALIZER(obj) \
2125 ._queue = Z_QUEUE_INITIALIZER(obj._queue) \
2145#define k_fifo_init(fifo) \
2147 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_fifo, init, fifo); \
2148 k_queue_init(&(fifo)->_queue); \
2149 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_fifo, init, fifo); \
2163#define k_fifo_cancel_wait(fifo) \
2165 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_fifo, cancel_wait, fifo); \
2166 k_queue_cancel_wait(&(fifo)->_queue); \
2167 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_fifo, cancel_wait, fifo); \
2182#define k_fifo_put(fifo, data) \
2184 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_fifo, put, fifo, data); \
2185 k_queue_append(&(fifo)->_queue, data); \
2186 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_fifo, put, fifo, data); \
2205#define k_fifo_alloc_put(fifo, data) \
2207 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_fifo, alloc_put, fifo, data); \
2208 int ret = k_queue_alloc_append(&(fifo)->_queue, data); \
2209 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_fifo, alloc_put, fifo, data, ret); \
2227#define k_fifo_put_list(fifo, head, tail) \
2229 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_fifo, put_list, fifo, head, tail); \
2230 k_queue_append_list(&(fifo)->_queue, head, tail); \
2231 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_fifo, put_list, fifo, head, tail); \
2247#define k_fifo_put_slist(fifo, list) \
2249 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_fifo, put_slist, fifo, list); \
2250 k_queue_merge_slist(&(fifo)->_queue, list); \
2251 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_fifo, put_slist, fifo, list); \
2271#define k_fifo_get(fifo, timeout) \
2273 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_fifo, get, fifo, timeout); \
2274 void *ret = k_queue_get(&(fifo)->_queue, timeout); \
2275 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_fifo, get, fifo, timeout, ret); \
2292#define k_fifo_is_empty(fifo) \
2293 k_queue_is_empty(&(fifo)->_queue)
2308#define k_fifo_peek_head(fifo) \
2310 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_fifo, peek_head, fifo); \
2311 void *ret = k_queue_peek_head(&(fifo)->_queue); \
2312 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_fifo, peek_head, fifo, ret); \
2327#define k_fifo_peek_tail(fifo) \
2329 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_fifo, peek_tail, fifo); \
2330 void *ret = k_queue_peek_tail(&(fifo)->_queue); \
2331 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_fifo, peek_tail, fifo, ret); \
2344#define K_FIFO_DEFINE(name) \
2345 STRUCT_SECTION_ITERABLE_ALTERNATE(k_queue, k_fifo, name) = \
2346 Z_FIFO_INITIALIZER(name)
2351 struct k_queue _queue;
2358#define Z_LIFO_INITIALIZER(obj) \
2360 ._queue = Z_QUEUE_INITIALIZER(obj._queue) \
2380#define k_lifo_init(lifo) \
2382 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_lifo, init, lifo); \
2383 k_queue_init(&(lifo)->_queue); \
2384 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_lifo, init, lifo); \
2399#define k_lifo_put(lifo, data) \
2401 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_lifo, put, lifo, data); \
2402 k_queue_prepend(&(lifo)->_queue, data); \
2403 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_lifo, put, lifo, data); \
2422#define k_lifo_alloc_put(lifo, data) \
2424 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_lifo, alloc_put, lifo, data); \
2425 int ret = k_queue_alloc_prepend(&(lifo)->_queue, data); \
2426 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_lifo, alloc_put, lifo, data, ret); \
2447#define k_lifo_get(lifo, timeout) \
2449 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_lifo, get, lifo, timeout); \
2450 void *ret = k_queue_get(&(lifo)->_queue, timeout); \
2451 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_lifo, get, lifo, timeout, ret); \
2464#define K_LIFO_DEFINE(name) \
2465 STRUCT_SECTION_ITERABLE_ALTERNATE(k_queue, k_lifo, name) = \
2466 Z_LIFO_INITIALIZER(name)
2473#define K_STACK_FLAG_ALLOC ((uint8_t)1)
2480 stack_data_t *base, *next, *top;
2487#define Z_STACK_INITIALIZER(obj, stack_buffer, stack_num_entries) \
2489 .wait_q = Z_WAIT_Q_INIT(&obj.wait_q), \
2490 .base = stack_buffer, \
2491 .next = stack_buffer, \
2492 .top = stack_buffer + stack_num_entries, \
2596#define K_STACK_DEFINE(name, stack_num_entries) \
2597 stack_data_t __noinit \
2598 _k_stack_buf_##name[stack_num_entries]; \
2599 STRUCT_SECTION_ITERABLE(k_stack, name) = \
2600 Z_STACK_INITIALIZER(name, _k_stack_buf_##name, \
2613extern struct k_work_q k_sys_work_q;
2647#define Z_MUTEX_INITIALIZER(obj) \
2649 .wait_q = Z_WAIT_Q_INIT(&obj.wait_q), \
2652 .owner_orig_prio = K_LOWEST_APPLICATION_THREAD_PRIO, \
2668#define K_MUTEX_DEFINE(name) \
2669 STRUCT_SECTION_ITERABLE(k_mutex, name) = \
2670 Z_MUTEX_INITIALIZER(name)
2741#define Z_CONDVAR_INITIALIZER(obj) \
2743 .wait_q = Z_WAIT_Q_INIT(&obj.wait_q), \
2807#define K_CONDVAR_DEFINE(name) \
2808 STRUCT_SECTION_ITERABLE(k_condvar, name) = \
2809 Z_CONDVAR_INITIALIZER(name)
2829#define Z_SEM_INITIALIZER(obj, initial_count, count_limit) \
2831 .wait_q = Z_WAIT_Q_INIT(&obj.wait_q), \
2832 .count = initial_count, \
2833 .limit = count_limit, \
2834 _POLL_EVENT_OBJ_INIT(obj) \
2855#define K_SEM_MAX_LIMIT UINT_MAX
2873 unsigned int limit);
2932static inline unsigned int z_impl_k_sem_count_get(
struct k_sem *
sem)
2948#define K_SEM_DEFINE(name, initial_count, count_limit) \
2949 STRUCT_SECTION_ITERABLE(k_sem, name) = \
2950 Z_SEM_INITIALIZER(name, initial_count, count_limit); \
2951 BUILD_ASSERT(((count_limit) != 0) && \
2952 ((initial_count) <= (count_limit)) && \
2953 ((count_limit) <= K_SEM_MAX_LIMIT));
3512 K_WORK_RUNNING_BIT = 0,
3513 K_WORK_CANCELING_BIT = 1,
3514 K_WORK_QUEUED_BIT = 2,
3515 K_WORK_DELAYED_BIT = 3,
3517 K_WORK_MASK =
BIT(K_WORK_DELAYED_BIT) |
BIT(K_WORK_QUEUED_BIT)
3518 |
BIT(K_WORK_RUNNING_BIT) |
BIT(K_WORK_CANCELING_BIT),
3521 K_WORK_DELAYABLE_BIT = 8,
3522 K_WORK_DELAYABLE =
BIT(K_WORK_DELAYABLE_BIT),
3525 K_WORK_QUEUE_STARTED_BIT = 0,
3526 K_WORK_QUEUE_STARTED =
BIT(K_WORK_QUEUE_STARTED_BIT),
3527 K_WORK_QUEUE_BUSY_BIT = 1,
3528 K_WORK_QUEUE_BUSY =
BIT(K_WORK_QUEUE_BUSY_BIT),
3529 K_WORK_QUEUE_DRAIN_BIT = 2,
3530 K_WORK_QUEUE_DRAIN =
BIT(K_WORK_QUEUE_DRAIN_BIT),
3531 K_WORK_QUEUE_PLUGGED_BIT = 3,
3532 K_WORK_QUEUE_PLUGGED =
BIT(K_WORK_QUEUE_PLUGGED_BIT),
3535 K_WORK_QUEUE_NO_YIELD_BIT = 8,
3536 K_WORK_QUEUE_NO_YIELD =
BIT(K_WORK_QUEUE_NO_YIELD_BIT),
3595#define Z_WORK_INITIALIZER(work_handler) { \
3596 .handler = work_handler, \
3611#define Z_WORK_DELAYABLE_INITIALIZER(work_handler) { \
3613 .handler = work_handler, \
3614 .flags = K_WORK_DELAYABLE, \
3634#define K_WORK_DELAYABLE_DEFINE(work, work_handler) \
3635 struct k_work_delayable work \
3636 = Z_WORK_DELAYABLE_INITIALIZER(work_handler)
3649struct z_work_flusher {
3660struct z_work_canceller {
3773 return &
queue->thread;
3787 size_t stack_size,
int prio)
3797#define Z_DELAYED_WORK_INITIALIZER(work_handler) __DEPRECATED_MACRO { \
3798 .work = Z_WORK_DELAYABLE_INITIALIZER(work_handler), \
3816 return (rc >= 0) ? 0 : rc;
3826 return (rc >= 0) ? 0 : rc;
3864 return (rc == 0) ? 0 : -
EINVAL;
3923struct k_work_user_q {
3924 struct k_queue
queue;
3929 K_WORK_USER_STATE_PENDING,
3942#if defined(__cplusplus) && ((__cplusplus - 0) < 202002L)
3943#define Z_WORK_USER_INITIALIZER(work_handler) { NULL, work_handler, 0 }
3945#define Z_WORK_USER_INITIALIZER(work_handler) \
3947 ._reserved = NULL, \
3948 .handler = work_handler, \
3964#define K_WORK_USER_DEFINE(work, work_handler) \
3965 struct k_work_user work = Z_WORK_USER_INITIALIZER(work_handler)
3979 *
work = (
struct k_work_user)Z_WORK_USER_INITIALIZER(
handler);
4022 struct k_work_user *
work)
4027 K_WORK_USER_STATE_PENDING)) {
4035 K_WORK_USER_STATE_PENDING);
4063 size_t stack_size,
int prio,
4075 struct z_poller poller;
4103#define K_WORK_DEFINE(work, work_handler) \
4104 struct k_work work = Z_WORK_INITIALIZER(work_handler)
4117#define K_DELAYED_WORK_DEFINE(work, work_handler) __DEPRECATED_MACRO \
4118 struct k_delayed_work work = Z_DELAYED_WORK_INITIALIZER(work_handler)
4167 struct k_work_poll *
work,
4267#define Z_MSGQ_INITIALIZER(obj, q_buffer, q_msg_size, q_max_msgs) \
4269 .wait_q = Z_WAIT_Q_INIT(&obj.wait_q), \
4270 .msg_size = q_msg_size, \
4271 .max_msgs = q_max_msgs, \
4272 .buffer_start = q_buffer, \
4273 .buffer_end = q_buffer + (q_max_msgs * q_msg_size), \
4274 .read_ptr = q_buffer, \
4275 .write_ptr = q_buffer, \
4277 _POLL_EVENT_OBJ_INIT(obj) \
4285#define K_MSGQ_FLAG_ALLOC BIT(0)
4320#define K_MSGQ_DEFINE(q_name, q_msg_size, q_max_msgs, q_align) \
4321 static char __noinit __aligned(q_align) \
4322 _k_fifo_buf_##q_name[(q_max_msgs) * (q_msg_size)]; \
4323 STRUCT_SECTION_ITERABLE(k_msgq, q_name) = \
4324 Z_MSGQ_INITIALIZER(q_name, _k_fifo_buf_##q_name, \
4325 q_msg_size, q_max_msgs)
4528#if (CONFIG_NUM_MBOX_ASYNC_MSGS > 0)
4530 struct k_sem *_async_sem;
4550#define Z_MBOX_INITIALIZER(obj) \
4552 .tx_msg_queue = Z_WAIT_Q_INIT(&obj.tx_msg_queue), \
4553 .rx_msg_queue = Z_WAIT_Q_INIT(&obj.rx_msg_queue), \
4569#define K_MBOX_DEFINE(name) \
4570 STRUCT_SECTION_ITERABLE(k_mbox, name) = \
4571 Z_MBOX_INITIALIZER(name) \
4685#define K_PIPE_FLAG_ALLOC BIT(0)
4687#define Z_PIPE_INITIALIZER(obj, pipe_buffer, pipe_buffer_size) \
4689 .buffer = pipe_buffer, \
4690 .size = pipe_buffer_size, \
4696 .readers = Z_WAIT_Q_INIT(&obj.wait_q.readers), \
4697 .writers = Z_WAIT_Q_INIT(&obj.wait_q.writers) \
4719#define K_PIPE_DEFINE(name, pipe_buffer_size, pipe_align) \
4720 static unsigned char __noinit __aligned(pipe_align) \
4721 _k_pipe_buf_##name[pipe_buffer_size]; \
4722 STRUCT_SECTION_ITERABLE(k_pipe, name) = \
4723 Z_PIPE_INITIALIZER(name, _k_pipe_buf_##name, pipe_buffer_size)
4787 size_t bytes_to_write,
size_t *bytes_written,
4810 size_t bytes_to_read,
size_t *bytes_read,
4876#ifdef CONFIG_MEM_SLAB_TRACE_MAX_UTILIZATION
4883#define Z_MEM_SLAB_INITIALIZER(obj, slab_buffer, slab_block_size, \
4886 .wait_q = Z_WAIT_Q_INIT(&obj.wait_q), \
4888 .num_blocks = slab_num_blocks, \
4889 .block_size = slab_block_size, \
4890 .buffer = slab_buffer, \
4891 .free_list = NULL, \
4929#define K_MEM_SLAB_DEFINE(name, slab_block_size, slab_num_blocks, slab_align) \
4930 char __noinit_named(k_mem_slab_buf_##name) \
4931 __aligned(WB_UP(slab_align)) \
4932 _k_mem_slab_buf_##name[(slab_num_blocks) * WB_UP(slab_block_size)]; \
4933 STRUCT_SECTION_ITERABLE(k_mem_slab, name) = \
4934 Z_MEM_SLAB_INITIALIZER(name, _k_mem_slab_buf_##name, \
4935 WB_UP(slab_block_size), slab_num_blocks)
4951#define K_MEM_SLAB_DEFINE_STATIC(name, slab_block_size, slab_num_blocks, slab_align) \
4952 static char __noinit_named(k_mem_slab_buf_##name) \
4953 __aligned(WB_UP(slab_align)) \
4954 _k_mem_slab_buf_##name[(slab_num_blocks) * WB_UP(slab_block_size)]; \
4955 static STRUCT_SECTION_ITERABLE(k_mem_slab, name) = \
4956 Z_MEM_SLAB_INITIALIZER(name, _k_mem_slab_buf_##name, \
4957 WB_UP(slab_block_size), slab_num_blocks)
4981 size_t block_size,
uint32_t num_blocks);
5031 return slab->num_used;
5046#ifdef CONFIG_MEM_SLAB_TRACE_MAX_UTILIZATION
5047 return slab->max_used;
5066 return slab->num_blocks - slab->num_used;
5159#define Z_HEAP_MIN_SIZE (sizeof(void *) > 4 ? 56 : 44)
5177#define Z_HEAP_DEFINE_IN_SECT(name, bytes, in_section) \
5180 kheap_##name[MAX(bytes, Z_HEAP_MIN_SIZE)]; \
5181 STRUCT_SECTION_ITERABLE(k_heap, name) = { \
5183 .init_mem = kheap_##name, \
5184 .init_bytes = MAX(bytes, Z_HEAP_MIN_SIZE), \
5202#define K_HEAP_DEFINE(name, bytes) \
5203 Z_HEAP_DEFINE_IN_SECT(name, bytes, \
5204 __noinit_named(kheap_buf_##name))
5220#define K_HEAP_DEFINE_NOCACHE(name, bytes) \
5221 Z_HEAP_DEFINE_IN_SECT(name, bytes, __nocache)
5296#define _INIT_OBJ_POLL_EVENT(obj) do { (obj)->poll_event = NULL; } while (false)
5298#define _INIT_OBJ_POLL_EVENT(obj) do { } while (false)
5302enum _poll_types_bits {
5310 _POLL_TYPE_SEM_AVAILABLE,
5313 _POLL_TYPE_DATA_AVAILABLE,
5316 _POLL_TYPE_MSGQ_DATA_AVAILABLE,
5321#define Z_POLL_TYPE_BIT(type) (1U << ((type) - 1U))
5324enum _poll_states_bits {
5326 _POLL_STATE_NOT_READY,
5329 _POLL_STATE_SIGNALED,
5332 _POLL_STATE_SEM_AVAILABLE,
5335 _POLL_STATE_DATA_AVAILABLE,
5338 _POLL_STATE_CANCELLED,
5341 _POLL_STATE_MSGQ_DATA_AVAILABLE,
5346#define Z_POLL_STATE_BIT(state) (1U << ((state) - 1U))
5348#define _POLL_EVENT_NUM_UNUSED_BITS \
5352 + _POLL_NUM_STATES \
5368#define K_POLL_TYPE_IGNORE 0
5369#define K_POLL_TYPE_SIGNAL Z_POLL_TYPE_BIT(_POLL_TYPE_SIGNAL)
5370#define K_POLL_TYPE_SEM_AVAILABLE Z_POLL_TYPE_BIT(_POLL_TYPE_SEM_AVAILABLE)
5371#define K_POLL_TYPE_DATA_AVAILABLE Z_POLL_TYPE_BIT(_POLL_TYPE_DATA_AVAILABLE)
5372#define K_POLL_TYPE_FIFO_DATA_AVAILABLE K_POLL_TYPE_DATA_AVAILABLE
5373#define K_POLL_TYPE_MSGQ_DATA_AVAILABLE Z_POLL_TYPE_BIT(_POLL_TYPE_MSGQ_DATA_AVAILABLE)
5384#define K_POLL_STATE_NOT_READY 0
5385#define K_POLL_STATE_SIGNALED Z_POLL_STATE_BIT(_POLL_STATE_SIGNALED)
5386#define K_POLL_STATE_SEM_AVAILABLE Z_POLL_STATE_BIT(_POLL_STATE_SEM_AVAILABLE)
5387#define K_POLL_STATE_DATA_AVAILABLE Z_POLL_STATE_BIT(_POLL_STATE_DATA_AVAILABLE)
5388#define K_POLL_STATE_FIFO_DATA_AVAILABLE K_POLL_STATE_DATA_AVAILABLE
5389#define K_POLL_STATE_MSGQ_DATA_AVAILABLE Z_POLL_STATE_BIT(_POLL_STATE_MSGQ_DATA_AVAILABLE)
5390#define K_POLL_STATE_CANCELLED Z_POLL_STATE_BIT(_POLL_STATE_CANCELLED)
5407#define K_POLL_SIGNAL_INITIALIZER(obj) \
5409 .poll_events = SYS_DLIST_STATIC_INIT(&obj.poll_events), \
5450#define K_POLL_EVENT_INITIALIZER(_event_type, _event_mode, _event_obj) \
5453 .type = _event_type, \
5454 .state = K_POLL_STATE_NOT_READY, \
5455 .mode = _event_mode, \
5458 .obj = _event_obj, \
5462#define K_POLL_EVENT_STATIC_INITIALIZER(_event_type, _event_mode, _event_obj, \
5466 .type = _event_type, \
5467 .state = K_POLL_STATE_NOT_READY, \
5468 .mode = _event_mode, \
5471 .obj = _event_obj, \
5491 int mode,
void *obj);
5567 unsigned int *signaled,
int *
result);
5653#define z_except_reason(reason) ARCH_EXCEPT(reason)
5656#if !defined(CONFIG_ASSERT_NO_FILE_INFO)
5657#define __EXCEPT_LOC() __ASSERT_PRINT("@ %s:%d\n", __FILE__, __LINE__)
5659#define __EXCEPT_LOC()
5669#define z_except_reason(reason) do { \
5671 z_fatal_error(reason, NULL); \
5687#define k_oops() z_except_reason(K_ERR_KERNEL_OOPS)
5697#define k_panic() z_except_reason(K_ERR_KERNEL_PANIC)
5706extern void z_init_thread_base(
struct _thread_base *thread_base,
5707 int priority,
uint32_t initial_state,
5708 unsigned int options);
5710#ifdef CONFIG_MULTITHREADING
5714extern void z_init_static_threads(
void);
5719#define z_init_static_threads() do { } while (false)
5725extern bool z_is_thread_essential(
void);
5729void z_smp_thread_swap(
void);
5735extern void z_timer_expiration_handler(
struct _timeout *
t);
5745__syscall
void k_str_out(
char *
c,
size_t n);
static uint32_t arch_k_cycle_get_32(void)
Definition: misc.h:26
static uint64_t arch_k_cycle_get_64(void)
Definition: misc.h:33
struct z_thread_stack_element k_thread_stack_t
Typedef of struct z_thread_stack_element.
Definition: arch_interface.h:44
void(* k_thread_entry_t)(void *p1, void *p2, void *p3)
Thread entry point function type.
Definition: arch_interface.h:46
static struct k_thread thread[2]
Definition: atomic.c:26
long atomic_t
Definition: atomic.h:22
ZTEST_BMEM int timeout
Definition: main.c:31
ZTEST_BMEM int count
Definition: main.c:33
void arch_cpu_atomic_idle(unsigned int key)
Atomically re-enable interrupts and enter low power mode.
void arch_cpu_idle(void)
Power save idle routine.
static bool atomic_test_bit(const atomic_t *target, int bit)
Atomically test a bit.
Definition: atomic.h:128
static void atomic_clear_bit(atomic_t *target, int bit)
Atomically clear a bit.
Definition: atomic.h:186
static bool atomic_test_and_set_bit(atomic_t *target, int bit)
Atomically set a bit.
Definition: atomic.h:167
static uint32_t k_cycle_get_32(void)
Read the hardware clock.
Definition: kernel.h:1598
int64_t k_uptime_ticks(void)
Get system uptime, in system ticks.
static uint32_t k_uptime_get_32(void)
Get system uptime (32-bit version).
Definition: kernel.h:1563
uint32_t k_ticks_t
Tick precision used in timeout APIs.
Definition: sys_clock.h:48
static int64_t k_uptime_delta(int64_t *reftime)
Get elapsed time.
Definition: kernel.h:1579
static uint64_t k_cycle_get_64(void)
Read the 64-bit hardware clock.
Definition: kernel.h:1613
static int64_t k_uptime_get(void)
Get system uptime.
Definition: kernel.h:1539
int k_condvar_signal(struct k_condvar *condvar)
Signals one thread that is pending on the condition variable.
int k_condvar_wait(struct k_condvar *condvar, struct k_mutex *mutex, k_timeout_t timeout)
Waits on the condition variable releasing the mutex lock.
int k_condvar_init(struct k_condvar *condvar)
Initialize a condition variable.
int k_condvar_broadcast(struct k_condvar *condvar)
Unblock all threads that are pending on the condition variable.
static void k_cpu_idle(void)
Make the CPU idle.
Definition: kernel.h:5620
static void k_cpu_atomic_idle(unsigned int key)
Make the CPU idle in an atomic fashion.
Definition: kernel.h:5639
struct _dnode sys_dnode_t
Definition: dlist.h:49
struct _dnode sys_dlist_t
Definition: dlist.h:48
uint32_t k_event_wait(struct k_event *event, uint32_t events, bool reset, k_timeout_t timeout)
Wait for any of the specified events.
void k_event_set(struct k_event *event, uint32_t events)
Set the events in an event object.
void k_event_post(struct k_event *event, uint32_t events)
Post one or more events to an event object.
void k_event_init(struct k_event *event)
Initialize an event object.
uint32_t k_event_wait_all(struct k_event *event, uint32_t events, bool reset, k_timeout_t timeout)
Wait for any of the specified events.
static bool sys_sflist_is_empty(sys_sflist_t *list)
Test if the given list is empty.
Definition: sflist.h:323
int k_futex_wait(struct k_futex *futex, int expected, k_timeout_t timeout)
Pend the current thread on a futex.
int k_futex_wake(struct k_futex *futex, bool wake_all)
Wake one/all threads pending on a futex.
void * k_heap_alloc(struct k_heap *h, size_t bytes, k_timeout_t timeout)
Allocate memory from a k_heap.
void k_heap_free(struct k_heap *h, void *mem)
Free memory allocated by k_heap_alloc()
void k_free(void *ptr)
Free memory allocated from heap.
void k_heap_init(struct k_heap *h, void *mem, size_t bytes)
Initialize a k_heap.
void * k_malloc(size_t size)
Allocate memory from the heap.
void * k_calloc(size_t nmemb, size_t size)
Allocate memory from heap, array style.
void * k_aligned_alloc(size_t align, size_t size)
Allocate memory from the heap with a specified alignment.
void * k_heap_aligned_alloc(struct k_heap *h, size_t align, size_t bytes, k_timeout_t timeout)
Allocate aligned memory from a k_heap.
bool k_is_in_isr(void)
Determine if code is running at interrupt level.
int k_is_preempt_thread(void)
Determine if code is running in a preemptible thread.
static bool k_is_pre_kernel(void)
Test whether startup is in the before-main-task phase.
Definition: kernel.h:916
static ZTEST_BMEM char buffer[8]
Test mailbox enhance capabilities.
Definition: test_mbox_api.c:566
int k_mbox_get(struct k_mbox *mbox, struct k_mbox_msg *rx_msg, void *buffer, k_timeout_t timeout)
Receive a mailbox message.
void k_mbox_data_get(struct k_mbox_msg *rx_msg, void *buffer)
Retrieve mailbox message data into a buffer.
void k_mbox_init(struct k_mbox *mbox)
Initialize a mailbox.
int k_mbox_put(struct k_mbox *mbox, struct k_mbox_msg *tx_msg, k_timeout_t timeout)
Send a mailbox message in a synchronous manner.
void k_mbox_async_put(struct k_mbox *mbox, struct k_mbox_msg *tx_msg, struct k_sem *sem)
Send a mailbox message in an asynchronous manner.
int k_mem_slab_init(struct k_mem_slab *slab, void *buffer, size_t block_size, uint32_t num_blocks)
Initialize a memory slab.
void k_mem_slab_free(struct k_mem_slab *slab, void **mem)
Free memory allocated from a memory slab.
int k_mem_slab_alloc(struct k_mem_slab *slab, void **mem, k_timeout_t timeout)
Allocate memory from a memory slab.
static uint32_t k_mem_slab_num_used_get(struct k_mem_slab *slab)
Get the number of used blocks in a memory slab.
Definition: kernel.h:5029
static uint32_t k_mem_slab_max_used_get(struct k_mem_slab *slab)
Get the number of maximum used blocks so far in a memory slab.
Definition: kernel.h:5044
static uint32_t k_mem_slab_num_free_get(struct k_mem_slab *slab)
Get the number of unused blocks in a memory slab.
Definition: kernel.h:5064
int k_msgq_peek(struct k_msgq *msgq, void *data)
Peek/read a message from a message queue.
uint32_t k_msgq_num_used_get(struct k_msgq *msgq)
Get the number of messages in a message queue.
void k_msgq_init(struct k_msgq *msgq, char *buffer, size_t msg_size, uint32_t max_msgs)
Initialize a message queue.
int k_msgq_put(struct k_msgq *msgq, const void *data, k_timeout_t timeout)
Send a message to a message queue.
uint32_t k_msgq_num_free_get(struct k_msgq *msgq)
Get the amount of free space in a message queue.
void k_msgq_get_attrs(struct k_msgq *msgq, struct k_msgq_attrs *attrs)
Get basic attributes of a message queue.
void k_msgq_purge(struct k_msgq *msgq)
Purge a message queue.
int k_msgq_alloc_init(struct k_msgq *msgq, size_t msg_size, uint32_t max_msgs)
Initialize a message queue.
int k_msgq_get(struct k_msgq *msgq, void *data, k_timeout_t timeout)
Receive a message from a message queue.
int k_msgq_cleanup(struct k_msgq *msgq)
Release allocated buffer for a queue.
int k_mutex_unlock(struct k_mutex *mutex)
Unlock a mutex.
int k_mutex_init(struct k_mutex *mutex)
Initialize a mutex.
int k_mutex_lock(struct k_mutex *mutex, k_timeout_t timeout)
Lock a mutex.
size_t k_pipe_read_avail(struct k_pipe *pipe)
Query the number of bytes that may be read from pipe.
int k_pipe_alloc_init(struct k_pipe *pipe, size_t size)
Initialize a pipe and allocate a buffer for it.
void k_pipe_flush(struct k_pipe *pipe)
Flush the pipe of write data.
int k_pipe_put(struct k_pipe *pipe, void *data, size_t bytes_to_write, size_t *bytes_written, size_t min_xfer, k_timeout_t timeout)
Write data to a pipe.
void k_pipe_buffer_flush(struct k_pipe *pipe)
Flush the pipe's internal buffer.
int k_pipe_cleanup(struct k_pipe *pipe)
Release a pipe's allocated buffer.
int k_pipe_get(struct k_pipe *pipe, void *data, size_t bytes_to_read, size_t *bytes_read, size_t min_xfer, k_timeout_t timeout)
Read data from a pipe.
void k_pipe_init(struct k_pipe *pipe, unsigned char *buffer, size_t size)
Initialize a pipe.
size_t k_pipe_write_avail(struct k_pipe *pipe)
Query the number of bytes that may be written to pipe.
void k_poll_signal_reset(struct k_poll_signal *sig)
k_poll_modes
Definition: kernel.h:5376
void k_poll_signal_check(struct k_poll_signal *sig, unsigned int *signaled, int *result)
Fetch the signaled state and result value of a poll signal.
void k_poll_event_init(struct k_poll_event *event, uint32_t type, int mode, void *obj)
Initialize one struct k_poll_event instance.
int k_poll(struct k_poll_event *events, int num_events, k_timeout_t timeout)
Wait for one or many of multiple poll events to occur.
int k_poll_signal_raise(struct k_poll_signal *sig, int result)
Signal a poll signal object.
void k_poll_signal_init(struct k_poll_signal *sig)
Initialize a poll signal object.
@ K_POLL_MODE_NOTIFY_ONLY
Definition: kernel.h:5378
@ K_POLL_NUM_MODES
Definition: kernel.h:5380
void k_queue_init(struct k_queue *queue)
Initialize a queue.
void * k_queue_get(struct k_queue *queue, k_timeout_t timeout)
Get an element from a queue.
void * k_queue_peek_tail(struct k_queue *queue)
Peek element at the tail of queue.
bool k_queue_unique_append(struct k_queue *queue, void *data)
Append an element to a queue only if it's not present already.
bool k_queue_remove(struct k_queue *queue, void *data)
Remove an element from a queue.
int k_queue_merge_slist(struct k_queue *queue, sys_slist_t *list)
Atomically add a list of elements to a queue.
int32_t k_queue_alloc_append(struct k_queue *queue, void *data)
Append an element to a queue.
void k_queue_cancel_wait(struct k_queue *queue)
Cancel waiting on a queue.
void * k_queue_peek_head(struct k_queue *queue)
Peek element at the head of queue.
void k_queue_prepend(struct k_queue *queue, void *data)
Prepend an element to a queue.
int k_queue_append_list(struct k_queue *queue, void *head, void *tail)
Atomically append a list of elements to a queue.
void k_queue_append(struct k_queue *queue, void *data)
Append an element to the end of a queue.
int32_t k_queue_alloc_prepend(struct k_queue *queue, void *data)
Prepend an element to a queue.
void k_queue_insert(struct k_queue *queue, void *prev, void *data)
Inserts an element to a queue.
int k_queue_is_empty(struct k_queue *queue)
Query a queue to see if it has data available.
void k_sem_reset(struct k_sem *sem)
Resets a semaphore's count to zero.
unsigned int k_sem_count_get(struct k_sem *sem)
Get a semaphore's count.
void k_sem_give(struct k_sem *sem)
Give a semaphore.
int k_sem_take(struct k_sem *sem, k_timeout_t timeout)
Take a semaphore.
int k_sem_init(struct k_sem *sem, unsigned int initial_count, unsigned int limit)
Initialize a semaphore.
int k_stack_pop(struct k_stack *stack, stack_data_t *data, k_timeout_t timeout)
Pop an element from a stack.
void k_stack_init(struct k_stack *stack, stack_data_t *buffer, uint32_t num_entries)
Initialize a stack.
int k_stack_cleanup(struct k_stack *stack)
Release a stack's allocated buffer.
int k_stack_push(struct k_stack *stack, stack_data_t data)
Push an element onto a stack.
int32_t k_stack_alloc_init(struct k_stack *stack, uint32_t num_entries)
Initialize a stack.
#define SYS_PORT_TRACING_TRACKING_FIELD(type)
Field added to kernel objects so they are tracked.
Definition: tracing_macros.h:345
#define IS_ENABLED(config_macro)
Check for macro definition in compiler-visible expressions.
Definition: util_macro.h:101
#define BIT(n)
Unsigned integer with bit position n set (signed in assembly language).
Definition: util_macro.h:44
#define CONTAINER_OF(ptr, type, field)
Get a pointer to a structure containing the element.
Definition: util.h:147
#define EINVAL
Definition: errno.h:61
#define EBUSY
Definition: errno.h:55
#define EALREADY
Definition: errno.h:105
int k_thread_name_copy(k_tid_t thread, char *buf, size_t size)
Copy the thread name into a supplied buffer.
void k_yield(void)
Yield the current thread.
void k_thread_resume(k_tid_t thread)
Resume a suspended thread.
void * k_thread_custom_data_get(void)
Get current thread's custom data.
void k_thread_abort(k_tid_t thread)
Abort a thread.
void k_thread_system_pool_assign(struct k_thread *thread)
Assign the system heap as a thread's resource pool.
int k_thread_name_set(k_tid_t thread, const char *str)
Set current thread name.
void k_thread_priority_set(k_tid_t thread, int prio)
Set a thread's priority.
int k_thread_cpu_mask_enable(k_tid_t thread, int cpu)
Enable thread to run on specified CPU.
void k_thread_foreach_unlocked(k_thread_user_cb_t user_cb, void *user_data)
Iterate over all the threads in the system without locking.
int k_thread_priority_get(k_tid_t thread)
Get a thread's priority.
static void k_thread_heap_assign(struct k_thread *thread, struct k_heap *heap)
Assign a resource memory pool to a thread.
Definition: kernel.h:338
FUNC_NORETURN void k_thread_user_mode_enter(k_thread_entry_t entry, void *p1, void *p2, void *p3)
Drop a thread's privileges permanently to user mode.
int k_thread_join(struct k_thread *thread, k_timeout_t timeout)
Sleep until a thread exits.
void k_thread_custom_data_set(void *value)
Set current thread's custom data.
int32_t k_sleep(k_timeout_t timeout)
Put the current thread to sleep.
k_ticks_t k_thread_timeout_remaining_ticks(const struct k_thread *t)
Get time remaining before a thread wakes up, in system ticks.
void k_sched_lock(void)
Lock the scheduler.
static int32_t k_msleep(int32_t ms)
Put the current thread to sleep.
Definition: kernel.h:431
void k_busy_wait(uint32_t usec_to_wait)
Cause the current thread to busy wait.
void k_thread_suspend(k_tid_t thread)
Suspend a thread.
void k_sched_unlock(void)
Unlock the scheduler.
static __attribute_const__ k_tid_t k_current_get(void)
Get thread ID of the current thread.
Definition: kernel.h:509
k_ticks_t k_thread_timeout_expires_ticks(const struct k_thread *t)
Get time when a thread wakes up, in system ticks.
int k_thread_cpu_mask_clear(k_tid_t thread)
Sets all CPU enable masks to zero.
void k_sched_time_slice_set(int32_t slice, int prio)
Set time-slicing period and scope.
void k_thread_start(k_tid_t thread)
Start an inactive thread.
int k_thread_cpu_mask_disable(k_tid_t thread, int cpu)
Prevent thread to run on specified CPU.
void k_wakeup(k_tid_t thread)
Wake up a sleeping thread.
k_tid_t k_thread_create(struct k_thread *new_thread, k_thread_stack_t *stack, size_t stack_size, k_thread_entry_t entry, void *p1, void *p2, void *p3, int prio, uint32_t options, k_timeout_t delay)
Create a thread.
void k_thread_deadline_set(k_tid_t thread, int deadline)
Set deadline expiration time for scheduler.
const char * k_thread_name_get(k_tid_t thread)
Get thread name.
void k_thread_foreach(k_thread_user_cb_t user_cb, void *user_data)
Iterate over all the threads in the system.
const char * k_thread_state_str(k_tid_t thread_id)
Get thread state string.
int32_t k_usleep(int32_t us)
Put the current thread to sleep with microsecond resolution.
int k_thread_cpu_mask_enable_all(k_tid_t thread)
Sets all CPU enable masks to one.
void(* k_thread_user_cb_t)(const struct k_thread *thread, void *user_data)
Definition: kernel.h:95
k_ticks_t k_timer_expires_ticks(const struct k_timer *timer)
Get next expiration time of a timer, in system ticks.
k_ticks_t k_timer_remaining_ticks(const struct k_timer *timer)
Get time remaining before a timer next expires, in system ticks.
void(* k_timer_stop_t)(struct k_timer *timer)
Timer stop function type.
Definition: kernel.h:1320
void * k_timer_user_data_get(const struct k_timer *timer)
Retrieve the user-specific data from a timer.
void k_timer_init(struct k_timer *timer, k_timer_expiry_t expiry_fn, k_timer_stop_t stop_fn)
Initialize a timer.
void(* k_timer_expiry_t)(struct k_timer *timer)
Timer expiry function type.
Definition: kernel.h:1304
void k_timer_start(struct k_timer *timer, k_timeout_t duration, k_timeout_t period)
Start a timer.
static uint32_t k_timer_remaining_get(struct k_timer *timer)
Get time remaining before a timer next expires.
Definition: kernel.h:1463
uint32_t k_timer_status_sync(struct k_timer *timer)
Synchronize thread to timer expiration.
void k_timer_stop(struct k_timer *timer)
Stop a timer.
uint32_t k_timer_status_get(struct k_timer *timer)
Read timer status.
void k_timer_user_data_set(struct k_timer *timer, void *user_data)
Associate user-specific data with a timer.
static int k_delayed_work_cancel(struct k_delayed_work *work)
Definition: kernel.h:3830
int k_work_poll_submit_to_queue(struct k_work_q *work_q, struct k_work_poll *work, struct k_poll_event *events, int num_events, k_timeout_t timeout)
Submit a triggered work item.
static k_tid_t k_work_queue_thread_get(struct k_work_q *queue)
Access the thread that animates a work queue.
Definition: kernel.h:3771
static bool k_work_is_pending(const struct k_work *work)
Test whether a work item is currently pending.
Definition: kernel.h:3742
int k_work_queue_drain(struct k_work_q *queue, bool plug)
Wait until the work queue has drained, optionally plugging it.
static k_ticks_t k_work_delayable_expires_get(const struct k_work_delayable *dwork)
Get the absolute tick count at which a scheduled delayable work will be submitted.
Definition: kernel.h:3759
int k_work_schedule_for_queue(struct k_work_q *queue, struct k_work_delayable *dwork, k_timeout_t delay)
Submit an idle work item to a queue after a delay.
int k_work_delayable_busy_get(const struct k_work_delayable *dwork)
Busy state flags from the delayable work item.
static k_ticks_t k_delayed_work_remaining_ticks(struct k_delayed_work *work)
Definition: kernel.h:3893
void k_work_init_delayable(struct k_work_delayable *dwork, k_work_handler_t handler)
Initialize a delayable work structure.
int k_work_poll_cancel(struct k_work_poll *work)
Cancel a triggered work item.
void k_work_user_queue_start(struct k_work_user_q *work_q, k_thread_stack_t *stack, size_t stack_size, int prio, const char *name)
Start a workqueue in user mode.
void k_work_poll_init(struct k_work_poll *work, k_work_handler_t handler)
Initialize a triggered work item.
int k_work_cancel(struct k_work *work)
Cancel a work item.
static int32_t k_delayed_work_remaining_get(struct k_delayed_work *work)
Definition: kernel.h:3877
static bool k_delayed_work_pending(struct k_delayed_work *work)
Definition: kernel.h:3871
static int k_work_user_submit_to_queue(struct k_work_user_q *work_q, struct k_work_user *work)
Submit a work item to a user mode workqueue.
Definition: kernel.h:4021
int k_work_submit_to_queue(struct k_work_q *queue, struct k_work *work)
Submit a work item to a queue.
static bool k_work_user_is_pending(struct k_work_user *work)
Check if a userspace work item is pending.
Definition: kernel.h:3998
void(* k_work_handler_t)(struct k_work *work)
The signature for a work item handler function.
Definition: kernel.h:2980
int k_work_schedule(struct k_work_delayable *dwork, k_timeout_t delay)
Submit an idle work item to the system work queue after a delay.
static bool k_work_delayable_is_pending(const struct k_work_delayable *dwork)
Test whether a delayed work item is currently pending.
Definition: kernel.h:3753
static bool k_work_pending(const struct k_work *work)
Definition: kernel.h:3779
bool k_work_cancel_delayable_sync(struct k_work_delayable *dwork, struct k_work_sync *sync)
Cancel delayable work and wait.
int k_work_cancel_delayable(struct k_work_delayable *dwork)
Cancel delayable work.
static int k_delayed_work_submit_to_queue(struct k_work_q *work_q, struct k_delayed_work *work, k_timeout_t delay)
Definition: kernel.h:3809
static void k_work_user_init(struct k_work_user *work, k_work_user_handler_t handler)
Initialize a userspace work item.
Definition: kernel.h:3976
int k_work_queue_unplug(struct k_work_q *queue)
Release a work queue to accept new submissions.
int k_work_reschedule(struct k_work_delayable *dwork, k_timeout_t delay)
Reschedule a work item to the system work queue after a delay.
bool k_work_cancel_sync(struct k_work *work, struct k_work_sync *sync)
Cancel a work item and wait for it to complete.
int k_work_busy_get(const struct k_work *work)
Busy state flags from the work item.
static int k_delayed_work_submit(struct k_delayed_work *work, k_timeout_t delay)
Definition: kernel.h:3820
static struct k_work_delayable * k_work_delayable_from_work(struct k_work *work)
Get the parent delayable work structure from a work pointer.
Definition: kernel.h:3748
static k_ticks_t k_work_delayable_remaining_get(const struct k_work_delayable *dwork)
Get the number of ticks until a scheduled delayable work will be submitted.
Definition: kernel.h:3765
bool k_work_flush(struct k_work *work, struct k_work_sync *sync)
Wait for last-submitted instance to complete.
int k_work_reschedule_for_queue(struct k_work_q *queue, struct k_work_delayable *dwork, k_timeout_t delay)
Reschedule a work item to a queue after a delay.
static k_ticks_t k_delayed_work_expires_ticks(struct k_delayed_work *work)
Definition: kernel.h:3886
int k_work_submit(struct k_work *work)
Submit a work item to the system queue.
bool k_work_flush_delayable(struct k_work_delayable *dwork, struct k_work_sync *sync)
Flush delayable work.
static void k_work_q_start(struct k_work_q *work_q, k_thread_stack_t *stack, size_t stack_size, int prio)
Definition: kernel.h:3785
int k_work_poll_submit(struct k_work_poll *work, struct k_poll_event *events, int num_events, k_timeout_t timeout)
Submit a triggered work item to the system workqueue.
void k_work_queue_init(struct k_work_q *queue)
Initialize a work queue structure.
void k_work_queue_start(struct k_work_q *queue, k_thread_stack_t *stack, size_t stack_size, int prio, const struct k_work_queue_config *cfg)
Initialize a work queue.
void k_work_init(struct k_work *work, k_work_handler_t handler)
Initialize a (non-delayable) work structure.
static void k_delayed_work_init(struct k_delayed_work *work, k_work_handler_t handler)
Definition: kernel.h:3802
void(* k_work_user_handler_t)(struct k_work_user *work)
Work item handler function type for user work queues.
Definition: kernel.h:3917
@ K_WORK_CANCELING
Flag indicating a work item that is being canceled.
Definition: kernel.h:3554
@ K_WORK_QUEUED
Flag indicating a work item that has been submitted to a queue but has not started running.
Definition: kernel.h:3561
@ K_WORK_DELAYED
Flag indicating a delayed work item that is scheduled for submission to a queue.
Definition: kernel.h:3568
@ K_WORK_RUNNING
Flag indicating a work item that is running under a work queue thread.
Definition: kernel.h:3548
flags
Definition: http_parser.h:131
state
Definition: http_parser_state.h:30
int k_float_disable(struct k_thread *thread)
Disable preservation of floating point context information.
void k_sys_runtime_stats_disable(void)
Disable gathering of system runtime statistics.
int k_thread_runtime_stats_enable(k_tid_t thread)
Enable gathering of runtime statistics for specified thread.
void k_sys_runtime_stats_enable(void)
Enable gathering of system runtime statistics.
int k_float_enable(struct k_thread *thread, unsigned int options)
Enable preservation of floating point context information.
int k_thread_runtime_stats_get(k_tid_t thread, k_thread_runtime_stats_t *stats)
Get the runtime statistics of a thread.
execution_context_types
Definition: kernel.h:80
@ K_ISR
Definition: kernel.h:81
@ K_COOP_THREAD
Definition: kernel.h:82
@ K_PREEMPT_THREAD
Definition: kernel.h:83
int k_thread_runtime_stats_all_get(k_thread_runtime_stats_t *stats)
Get the runtime statistics of all threads.
int k_thread_runtime_stats_disable(k_tid_t thread)
Disable gathering of runtime statistics for specified thread.
static ZTEST_BMEM volatile int ret
Definition: k_float_disable.c:28
Header files included by kernel.h.
struct k_mem_slab ms
Definition: kobject.c:1313
struct k_mutex mutex
Definition: kobject.c:1315
struct k_thread t
Definition: kobject.c:1321
struct k_msgq msgq
Definition: test_msgq_contexts.c:12
char c
Definition: printk.c:71
void * ptr
Definition: printk.c:79
static struct k_work work[2]
Definition: main.c:16
struct _sfnode sys_sfnode_t
Definition: sflist.h:39
struct _sflist sys_sflist_t
Definition: sflist.h:46
struct _slist sys_slist_t
Definition: slist.h:40
struct _snode sys_snode_t
Definition: slist.h:33
char stack[2048]
Definition: main.c:22
static struct k_spinlock lock
Definition: spinlock_error_case.c:12
static k_spinlock_key_t key
Definition: spinlock_error_case.c:14
__UINT32_TYPE__ uint32_t
Definition: stdint.h:60
__INT32_TYPE__ int32_t
Definition: stdint.h:44
__UINT64_TYPE__ uint64_t
Definition: stdint.h:61
__UINT8_TYPE__ uint8_t
Definition: stdint.h:58
__UINTPTR_TYPE__ uintptr_t
Definition: stdint.h:75
__INT64_TYPE__ int64_t
Definition: stdint.h:45
Static init entry structure for each device driver or services.
Definition: init.h:43
Definition: kernel.h:2737
_wait_q_t wait_q
Definition: kernel.h:2738
Definition: kernel.h:3793
struct k_work_delayable work
Definition: kernel.h:3794
Definition: kernel.h:2002
struct k_spinlock lock
Definition: kernel.h:2005
uint32_t events
Definition: kernel.h:2004
_wait_q_t wait_q
Definition: kernel.h:2003
Definition: kernel.h:2116
futex structure
Definition: kernel.h:1923
atomic_t val
Definition: kernel.h:1924
Definition: kernel.h:5078
struct k_spinlock lock
Definition: kernel.h:5081
struct sys_heap heap
Definition: kernel.h:5079
_wait_q_t wait_q
Definition: kernel.h:5080
Definition: kernel.h:2350
Mailbox Message Structure.
Definition: kernel.h:4509
struct k_mem_block tx_block
Definition: kernel.h:4521
k_tid_t tx_target_thread
Definition: kernel.h:4525
void * tx_data
Definition: kernel.h:4517
k_tid_t rx_source_thread
Definition: kernel.h:4523
uint32_t info
Definition: kernel.h:4515
size_t size
Definition: kernel.h:4513
Mailbox Structure.
Definition: kernel.h:4537
_wait_q_t tx_msg_queue
Definition: kernel.h:4539
struct k_spinlock lock
Definition: kernel.h:4542
_wait_q_t rx_msg_queue
Definition: kernel.h:4541
Definition: mempool_heap.h:24
Memory Domain.
Definition: mem_domain.h:80
Memory Partition.
Definition: mem_domain.h:55
Message Queue Attributes.
Definition: kernel.h:4290
uint32_t used_msgs
Definition: kernel.h:4296
size_t msg_size
Definition: kernel.h:4292
uint32_t max_msgs
Definition: kernel.h:4294
Message Queue Structure.
Definition: kernel.h:4235
size_t msg_size
Definition: kernel.h:4241
char * read_ptr
Definition: kernel.h:4249
uint32_t used_msgs
Definition: kernel.h:4253
char * buffer_end
Definition: kernel.h:4247
struct k_spinlock lock
Definition: kernel.h:4239
char * write_ptr
Definition: kernel.h:4251
char * buffer_start
Definition: kernel.h:4245
uint8_t flags
Definition: kernel.h:4255
_wait_q_t wait_q
Definition: kernel.h:4237
uint32_t max_msgs
Definition: kernel.h:4243
Definition: kernel.h:2629
uint32_t lock_count
Definition: kernel.h:2636
_wait_q_t wait_q
Definition: kernel.h:2631
int owner_orig_prio
Definition: kernel.h:2639
struct k_thread * owner
Definition: kernel.h:2633
Definition: kernel.h:4664
uint8_t flags
Definition: kernel.h:4677
_wait_q_t readers
Definition: kernel.h:4673
size_t write_index
Definition: kernel.h:4669
size_t bytes_used
Definition: kernel.h:4667
struct k_spinlock lock
Definition: kernel.h:4670
struct k_pipe::@122 wait_q
_wait_q_t writers
Definition: kernel.h:4674
size_t size
Definition: kernel.h:4666
unsigned char * buffer
Definition: kernel.h:4665
size_t read_index
Definition: kernel.h:4668
Poll Event.
Definition: kernel.h:5417
struct k_poll_signal * signal
Definition: kernel.h:5442
uint32_t tag
Definition: kernel.h:5425
struct k_fifo * fifo
Definition: kernel.h:5444
struct k_msgq * msgq
Definition: kernel.h:5446
struct k_queue * queue
Definition: kernel.h:5445
uint32_t unused
Definition: kernel.h:5437
uint32_t type
Definition: kernel.h:5428
struct k_sem * sem
Definition: kernel.h:5443
uint32_t state
Definition: kernel.h:5431
uint32_t mode
Definition: kernel.h:5434
struct z_poller * poller
Definition: kernel.h:5422
void * obj
Definition: kernel.h:5441
Definition: kernel.h:5393
sys_dlist_t poll_events
Definition: kernel.h:5395
int result
Definition: kernel.h:5404
unsigned int signaled
Definition: kernel.h:5401
Kernel Spin Lock.
Definition: spinlock.h:42
struct _thread_base base
Definition: thread.h:218
struct k_heap * resource_pool
Definition: thread.h:296
struct __thread_entry entry
Definition: thread.h:242
Kernel timeout type.
Definition: sys_clock.h:65
A structure used to submit work after a delay.
Definition: kernel.h:3600
struct _timeout timeout
Definition: kernel.h:3605
struct k_work_q * queue
Definition: kernel.h:3608
struct k_work work
Definition: kernel.h:3602
A structure used to hold work until it can be processed.
Definition: kernel.h:3719
sys_slist_t pending
Definition: kernel.h:3728
_wait_q_t drainq
Definition: kernel.h:3734
_wait_q_t notifyq
Definition: kernel.h:3731
uint32_t flags
Definition: kernel.h:3737
struct k_thread thread
Definition: kernel.h:3721
A structure holding optional configuration items for a work queue.
Definition: kernel.h:3696
const char * name
Definition: kernel.h:3701
bool no_yield
Definition: kernel.h:3715
A structure holding internal state for a pending synchronous operation on a work item or queue.
Definition: kernel.h:3683
struct z_work_canceller canceller
Definition: kernel.h:3686
struct z_work_flusher flusher
Definition: kernel.h:3685
A structure used to submit work.
Definition: kernel.h:3572
k_work_handler_t handler
Definition: kernel.h:3581
uint32_t flags
Definition: kernel.h:3592
struct k_work_q * queue
Definition: kernel.h:3584
sys_snode_t node
Definition: kernel.h:3578
Definition: sys_heap.h:55
static fdata_t data[2]
Definition: test_fifo_contexts.c:15
static struct k_mbox mbox
Definition: test_mbox_api.c:28
static struct k_pipe pipe
Definition: test_mutex_error.c:18
struct k_queue queue
Definition: test_queue_contexts.c:17
static int init_prio
Definition: test_sched_timeslice_and_lock.c:15
static ZTEST_BMEM struct thread_data expected
static uint64_t k_ticks_to_ms_floor64(uint64_t t)
Convert ticks to milliseconds.
Definition: time_units.h:1102
static uint32_t k_ticks_to_ms_floor32(uint32_t t)
Convert ticks to milliseconds.
Definition: time_units.h:1088
static struct k_timer timer[3]
Definition: timeout_order.c:13
static struct k_sem sem[3]
Definition: timeout_order.c:14
static void handler(struct k_timer *timer)
Definition: main.c:19
static const intptr_t user_data[5]
Definition: main.c:590
static struct k_work_delayable dwork
Definition: main.c:48