28#ifndef ZEPHYR_INCLUDE_RTIO_RTIO_H_
29#define ZEPHYR_INCLUDE_RTIO_RTIO_H_
72#ifdef CONFIG_RTIO_SUBMIT_SEM
76 struct k_sem *submit_sem;
81#ifdef CONFIG_RTIO_CONSUME_SEM
86 struct k_sem *consume_sem;
103#ifdef CONFIG_RTIO_SYS_MEM_BLOCKS
105 struct sys_mem_blocks *block_pool;
116#define Z_RTIO_DEFINE(name, _sqe_pool, _cqe_pool, _block_pool) \
117 IF_ENABLED(CONFIG_RTIO_SUBMIT_SEM, \
118 (static K_SEM_DEFINE(CONCAT(_submit_sem_, name), 0, K_SEM_MAX_LIMIT))) \
119 IF_ENABLED(CONFIG_RTIO_CONSUME_SEM, \
120 (static K_SEM_DEFINE(CONCAT(_consume_sem_, name), 0, K_SEM_MAX_LIMIT))) \
121 STRUCT_SECTION_ITERABLE(rtio, name) = { \
122 IF_ENABLED(CONFIG_RTIO_SUBMIT_SEM, (.submit_sem = &CONCAT(_submit_sem_, name),)) \
123 IF_ENABLED(CONFIG_RTIO_SUBMIT_SEM, (.submit_count = 0,)) \
124 IF_ENABLED(CONFIG_RTIO_CONSUME_SEM, (.consume_sem = &CONCAT(_consume_sem_, name),))\
125 .cq_count = ATOMIC_INIT(0), \
126 .xcqcnt = ATOMIC_INIT(0), \
127 .sqe_pool = _sqe_pool, \
128 .cqe_pool = _cqe_pool, \
129 IF_ENABLED(CONFIG_RTIO_SYS_MEM_BLOCKS, (.block_pool = _block_pool,)) \
130 .sq = MPSC_INIT((name.sq)), \
131 .cq = MPSC_INIT((name.cq)), \
142#define RTIO_DEFINE(name, sq_sz, cq_sz) \
143 Z_RTIO_SQE_POOL_DEFINE(CONCAT(name, _sqe_pool), sq_sz); \
144 Z_RTIO_CQE_POOL_DEFINE(CONCAT(name, _cqe_pool), cq_sz); \
145 Z_RTIO_DEFINE(name, &CONCAT(name, _sqe_pool), \
146 &CONCAT(name, _cqe_pool), NULL)
158#ifndef CONFIG_RTIO_SYS_MEM_BLOCKS
165 return BIT(
r->block_pool->info.blk_sz_shift);
176#if defined(CONFIG_RTIO_SYS_MEM_BLOCKS) || defined(__DOXYGEN__)
177static inline uint16_t __rtio_compute_mempool_block_index(
const struct rtio *
r,
const void *ptr)
180 struct sys_mem_blocks *mem_pool =
r->block_pool;
184 uint32_t buff_size = mem_pool->info.num_blocks * block_size;
186 if (addr < buff || addr >= buff + buff_size) {
189 return (addr - buff) / block_size;
196#ifndef CONFIG_RTIO_SYS_MEM_BLOCKS
210 if (block_size == 0) {
219 *buf_len = num_blks * block_size;
223 if (bytes <= block_size) {
228 }
while (bytes >= min_sz);
236#ifndef CONFIG_RTIO_SYS_MEM_BLOCKS
241 size_t num_blks = buf_len >>
r->block_pool->info.blk_sz_shift;
263#define RTIO_BMEM COND_CODE_1(CONFIG_USERSPACE, (K_APP_BMEM(rtio_partition) static), (static))
274#define RTIO_DMEM COND_CODE_1(CONFIG_USERSPACE, (K_APP_DMEM(rtio_partition) static), (static))
278#define Z_RTIO_BLOCK_POOL_DEFINE(name, blk_sz, blk_cnt, blk_align) \
279 RTIO_BMEM uint8_t __aligned(WB_UP(blk_align)) \
280 CONCAT(_block_pool_, name)[blk_cnt*WB_UP(blk_sz)]; \
281 _SYS_MEM_BLOCKS_DEFINE_WITH_EXT_BUF(name, WB_UP(blk_sz), blk_cnt, \
282 CONCAT(_block_pool_, name), RTIO_DMEM)
299#define RTIO_DEFINE_WITH_MEMPOOL(name, sq_sz, cq_sz, num_blks, blk_size, balign) \
300 Z_RTIO_SQE_POOL_DEFINE(name##_sqe_pool, sq_sz); \
301 Z_RTIO_CQE_POOL_DEFINE(name##_cqe_pool, cq_sz); \
302 Z_RTIO_BLOCK_POOL_DEFINE(name##_block_pool, blk_size, num_blks, balign); \
303 Z_RTIO_DEFINE(name, &name##_sqe_pool, &name##_cqe_pool, &name##_block_pool)
316 return r->sqe_pool->pool_free;
332 if (iodev_sqe ==
NULL) {
340 return &iodev_sqe->
sqe;
360 for (i = 0; i < n; i++) {
361 iodev_sqe = rtio_sqe_pool_alloc(
r->sqe_pool);
362 if (iodev_sqe ==
NULL) {
365 sqes[i] = &iodev_sqe->
sqe;
373 rtio_sqe_pool_free(
r->sqe_pool, iodev_sqe);
380 for (i = 0; i < n; i++) {
398 while (node !=
NULL) {
400 rtio_sqe_pool_free(
r->sqe_pool, iodev_sqe);
411 struct rtio_cqe *cqe = rtio_cqe_pool_alloc(
r->cqe_pool);
449#ifdef CONFIG_RTIO_CONSUME_SEM
482#ifdef CONFIG_RTIO_CONSUME_SEM
486 while (node ==
NULL) {
504 rtio_cqe_pool_free(
r->cqe_pool, cqe);
523 if ((cqe->
result < 0) && (res == 0)) {
528 }
while (cqe !=
NULL);
546static inline void z_impl_rtio_sqe_signal(
struct rtio_sqe *sqe)
565#ifdef CONFIG_RTIO_SYS_MEM_BLOCKS
567 struct rtio *
r = iodev_sqe->
r;
568 struct sys_mem_blocks *mem_pool =
r->block_pool;
569 unsigned int blk_index = 0;
570 unsigned int blk_count = 0;
573 blk_index = (iodev_sqe->
sqe.
rx.
buf - mem_pool->buffer) >>
574 mem_pool->info.blk_sz_shift;
575 blk_count = iodev_sqe->
sqe.
rx.
buf_len >> mem_pool->info.blk_sz_shift;
580 ARG_UNUSED(iodev_sqe);
604static inline int z_impl_rtio_cqe_get_mempool_buffer(
const struct rtio *
r,
struct rtio_cqe *cqe,
607#ifdef CONFIG_RTIO_SYS_MEM_BLOCKS
613 *buff_len = blk_count * blk_size;
616 *buff =
r->block_pool->buffer + blk_idx * blk_size;
618 __ASSERT_NO_MSG(*buff >=
r->block_pool->buffer);
619 __ASSERT_NO_MSG(*buff <
620 r->block_pool->buffer + blk_size *
r->block_pool->info.num_blocks);
631 ARG_UNUSED(buff_len);
690#ifdef CONFIG_RTIO_CONSUME_SEM
705#ifdef CONFIG_RTIO_SUBMIT_SEM
706 if (
r->submit_count > 0) {
708 if (
r->submit_count == 0) {
716#define __RTIO_MEMPOOL_GET_NUM_BLKS(num_bytes, blk_size) (((num_bytes) + (blk_size)-1) / (blk_size))
735#ifdef CONFIG_RTIO_SYS_MEM_BLOCKS
737 struct rtio *
r = iodev_sqe->
r;
758 ARG_UNUSED(max_buf_len);
786static inline void z_impl_rtio_release_buffer(
struct rtio *
r,
void *buff,
uint32_t buff_len)
788#ifdef CONFIG_RTIO_SYS_MEM_BLOCKS
789 if (
r ==
NULL || buff ==
NULL ||
r->block_pool ==
NULL || buff_len == 0) {
797 ARG_UNUSED(buff_len);
811#ifdef CONFIG_RTIO_SUBMIT_SEM
815#ifdef CONFIG_RTIO_CONSUME_SEM
831#ifdef CONFIG_RTIO_SUBMIT_SEM
835#ifdef CONFIG_RTIO_CONSUME_SEM
852static inline int z_impl_rtio_sqe_cancel(
struct rtio_sqe *sqe)
860 }
while (iodev_sqe !=
NULL);
881 struct rtio_sqe **handle,
size_t sqe_count);
883static inline int z_impl_rtio_sqe_copy_in_get_handles(
struct rtio *
r,
const struct rtio_sqe *sqes,
890 if (acquirable < sqe_count) {
894 for (
unsigned long i = 0; i < sqe_count; i++) {
896 __ASSERT_NO_MSG(sqe !=
NULL);
897 if (handle !=
NULL && i == 0) {
946static inline int z_impl_rtio_cqe_copy_out(
struct rtio *
r,
962 cqes[copied++] = *cqe;
986#ifdef CONFIG_RTIO_SUBMIT_SEM
987static inline int z_impl_rtio_submit(
struct rtio *
r,
uint32_t wait_count)
992 if (wait_count > 0) {
994 "expected rtio submit with wait count to be called from a thread");
997 r->submit_count = wait_count;
1002 if (wait_count > 0) {
1005 "semaphore was reset or timed out while waiting on completions!");
1012static inline int z_impl_rtio_submit(
struct rtio *
r,
uint32_t wait_count)
1018 uintptr_t cq_complete_count = cq_count + wait_count;
1019 bool wraps = cq_complete_count < cq_count;
1064static inline struct rtio *z_impl_rtio_pool_acquire(
struct rtio_pool *pool)
1068 for (
size_t i = 0; i < pool->
pool_size; i++) {
1090static inline void z_impl_rtio_pool_release(
struct rtio_pool *pool,
struct rtio *
r)
1097 for (
size_t i = 0; i < pool->
pool_size; i++) {
1109#define Z_RTIO_POOL_NAME_N(n, name) \
1112#define Z_RTIO_POOL_DEFINE_N(n, name, sq_sz, cq_sz) \
1113 RTIO_DEFINE(Z_RTIO_POOL_NAME_N(n, name), sq_sz, cq_sz)
1115#define Z_RTIO_POOL_REF_N(n, name) \
1116 &Z_RTIO_POOL_NAME_N(n, name)
1128#define RTIO_POOL_DEFINE(name, pool_sz, sq_sz, cq_sz) \
1129 LISTIFY(pool_sz, Z_RTIO_POOL_DEFINE_N, (;), name, sq_sz, cq_sz); \
1130 static struct rtio *name##_contexts[] = { \
1131 LISTIFY(pool_sz, Z_RTIO_POOL_REF_N, (,), name) \
1133 ATOMIC_DEFINE(name##_used, pool_sz); \
1134 STRUCT_SECTION_ITERABLE(rtio_pool, name) = { \
1135 .pool_size = pool_sz, \
1136 .contexts = name##_contexts, \
1137 .used = name##_used, \
1150#include <zephyr/syscalls/rtio.h>
workaround assembler barfing for ST r
Definition asm-macro-32-bit-gnu.h:24
long atomic_t
Definition atomic_types.h:15
RTIO Completion Queue Events and Related Functions.
static _Bool atomic_test_and_set_bit(atomic_t *target, int bit)
Atomically set a bit and test it.
Definition atomic.h:172
static void atomic_clear_bit(atomic_t *target, int bit)
Atomically clear a bit.
Definition atomic.h:193
atomic_val_t atomic_get(const atomic_t *target)
Atomic get.
atomic_val_t atomic_inc(atomic_t *target)
Atomic increment.
_Bool atomic_cas(atomic_t *target, atomic_val_t old_value, atomic_val_t new_value)
Atomic compare-and-set.
#define K_FOREVER
Generate infinite timeout delay.
Definition kernel.h:1666
#define K_NO_WAIT
Generate null timeout delay.
Definition kernel.h:1556
k_timepoint_t sys_timepoint_calc(k_timeout_t timeout)
Calculate a timepoint value.
static bool sys_timepoint_expired(k_timepoint_t timepoint)
Indicates if timepoint is expired.
Definition clock.h:388
#define K_TIMEOUT_EQ(a, b)
Compare timeouts for equality.
Definition clock.h:80
bool k_is_in_isr(void)
Determine if code is running at interrupt level.
int sys_mem_blocks_free_contiguous(sys_mem_blocks_t *mem_block, void *block, size_t count)
Free contiguous multiple memory blocks.
int sys_mem_blocks_alloc_contiguous(sys_mem_blocks_t *mem_block, size_t count, void **out_block)
Allocate a contiguous set of memory blocks.
static ALWAYS_INLINE void mpsc_push(struct mpsc *q, struct mpsc_node *n)
Push a node.
Definition mpsc_lockfree.h:126
static struct mpsc_node * mpsc_pop(struct mpsc *q)
Pop a node off of the list.
Definition mpsc_lockfree.h:145
#define RTIO_CQE_FLAG_MEMPOOL_GET_BLK_CNT(flags)
Get the block count of a mempool flags.
Definition cqe.h:63
#define RTIO_CQE_FLAG_MEMPOOL_GET_BLK_IDX(flags)
Get the block index of a mempool flags.
Definition cqe.h:55
#define RTIO_CQE_FLAG_MEMPOOL_BUFFER
The entry's buffer was allocated from the RTIO's mempool.
Definition cqe.h:45
#define RTIO_CQE_FLAG_PREP_MEMPOOL(blk_idx, blk_cnt)
Prepare CQE flags for a mempool read.
Definition cqe.h:72
#define RTIO_CQE_FLAG_GET(flags)
Definition cqe.h:47
#define RTIO_OP_RX
An operation that receives (reads).
Definition sqe.h:144
#define RTIO_SQE_MEMPOOL_BUFFER
The buffer should be allocated by the RTIO mempool.
Definition sqe.h:105
#define RTIO_SQE_CANCELED
The SQE should not execute if possible.
Definition sqe.h:113
void rtio_pool_release(struct rtio_pool *pool, struct rtio *r)
Return an RTIO context to a pool.
void rtio_executor_err(struct rtio_iodev_sqe *iodev_sqe, int result)
static uint32_t rtio_sqe_acquirable(struct rtio *r)
Count of acquirable submission queue events.
Definition rtio.h:314
struct rtio * rtio_pool_acquire(struct rtio_pool *pool)
Obtain an RTIO context from a pool.
static size_t rtio_mempool_block_size(const struct rtio *r)
Get the mempool block size of the RTIO context.
Definition rtio.h:156
static void rtio_cqe_submit(struct rtio *r, int result, void *userdata, uint32_t flags)
Submit a completion queue event with a given result and userdata.
Definition rtio.h:678
void rtio_release_buffer(struct rtio *r, void *buff, uint32_t buff_len)
Release memory that was allocated by the RTIO's memory pool.
static int rtio_sqe_copy_in(struct rtio *r, const struct rtio_sqe *sqes, size_t sqe_count)
Copy an array of SQEs into the queue.
Definition rtio.h:922
static void rtio_cqe_produce(struct rtio *r, struct rtio_cqe *cqe)
Produce a complete queue event if available.
Definition rtio.h:427
static uint32_t rtio_cqe_compute_flags(struct rtio_iodev_sqe *iodev_sqe)
Compute the CQE flags from the rtio_iodev_sqe entry.
Definition rtio.h:561
void rtio_executor_ok(struct rtio_iodev_sqe *iodev_sqe, int result)
static int rtio_block_pool_alloc(struct rtio *r, size_t min_sz, size_t max_sz, uint8_t **buf, uint32_t *buf_len)
Definition rtio.h:193
int rtio_sqe_copy_in_get_handles(struct rtio *r, const struct rtio_sqe *sqes, struct rtio_sqe **handle, size_t sqe_count)
Copy an array of SQEs into the queue and get resulting handles back.
struct k_mem_partition rtio_partition
The memory partition associated with all RTIO context information.
static struct rtio_sqe * rtio_sqe_acquire(struct rtio *r)
Acquire a single submission queue event if available.
Definition rtio.h:327
static void rtio_sqe_drop_all(struct rtio *r)
Drop all previously acquired sqe.
Definition rtio.h:393
int rtio_cqe_copy_out(struct rtio *r, struct rtio_cqe *cqes, size_t cqe_count, k_timeout_t timeout)
Copy an array of CQEs from the queue.
static int rtio_flush_completion_queue(struct rtio *r)
Flush completion queue.
Definition rtio.h:515
static void rtio_access_revoke(struct rtio *r, struct k_thread *t)
Revoke access to an RTIO context from a user thread.
Definition rtio.h:827
static void rtio_access_grant(struct rtio *r, struct k_thread *t)
Grant access to an RTIO context to a user thread.
Definition rtio.h:807
static void rtio_cqe_release(struct rtio *r, struct rtio_cqe *cqe)
Release consumed completion queue event.
Definition rtio.h:501
static int rtio_sqe_rx_buf(const struct rtio_iodev_sqe *iodev_sqe, uint32_t min_buf_len, uint32_t max_buf_len, uint8_t **buf, uint32_t *buf_len)
Get the buffer associate with the RX submission.
Definition rtio.h:730
static void rtio_iodev_sqe_err(struct rtio_iodev_sqe *iodev_sqe, int result)
Inform the executor of a submissions completion with error.
Definition rtio.h:662
int rtio_sqe_cancel(struct rtio_sqe *sqe)
Attempt to cancel an SQE.
static void rtio_iodev_sqe_ok(struct rtio_iodev_sqe *iodev_sqe, int result)
Inform the executor of a submission completion with success.
Definition rtio.h:649
static struct rtio_cqe * rtio_cqe_acquire(struct rtio *r)
Acquire a complete queue event if available.
Definition rtio.h:408
static struct rtio_cqe * rtio_cqe_consume(struct rtio *r)
Consume a single completion queue event if available.
Definition rtio.h:443
void rtio_sqe_signal(struct rtio_sqe *sqe)
Signal an AWAIT SQE.
static struct rtio_iodev_sqe * rtio_iodev_sqe_next(const struct rtio_iodev_sqe *iodev_sqe)
Get the next sqe in the chain or transaction.
Definition sqe.h:713
int rtio_cqe_get_mempool_buffer(const struct rtio *r, struct rtio_cqe *cqe, uint8_t **buff, uint32_t *buff_len)
Retrieve the mempool buffer that was allocated for the CQE.
void rtio_executor_submit(struct rtio *r)
static struct rtio_cqe * rtio_cqe_consume_block(struct rtio *r)
Wait for and consume a single completion queue event.
Definition rtio.h:477
static int rtio_sqe_acquire_array(struct rtio *r, size_t n, struct rtio_sqe **sqes)
Acquire a number of submission queue events if available.
Definition rtio.h:355
static void rtio_block_pool_free(struct rtio *r, void *buf, uint32_t buf_len)
Definition rtio.h:234
int rtio_submit(struct rtio *r, uint32_t wait_count)
Submit I/O requests to the underlying executor.
void k_sem_reset(struct k_sem *sem)
Resets a semaphore's count to zero.
void k_sem_give(struct k_sem *sem)
Give a semaphore.
int k_sem_take(struct k_sem *sem, k_timeout_t timeout)
Take a semaphore.
#define SYS_PORT_TRACING_FUNC_ENTER(type, func,...)
Tracing macro for the entry into a function that might or might not return a value.
Definition tracing_macros.h:257
#define SYS_PORT_TRACING_FUNC_EXIT(type, func,...)
Tracing macro for when a function ends its execution.
Definition tracing_macros.h:283
#define SYS_PORT_TRACING_FUNC(type, func,...)
Tracing macro for function calls which are not directly associated with a specific type of object.
Definition tracing_macros.h:244
#define BIT(n)
Unsigned integer with bit position n set (signed in assembly language).
Definition util_macro.h:44
#define CONTAINER_OF(ptr, type, field)
Get a pointer to a structure containing the element.
Definition util.h:281
#define DIV_ROUND_UP(n, d)
Divide and round up.
Definition util.h:348
#define EINVAL
Invalid argument.
Definition errno.h:60
#define ENOMEM
Not enough core.
Definition errno.h:50
#define ENOTSUP
Unsupported value.
Definition errno.h:114
void k_yield(void)
Yield the current thread.
static __attribute_const__ k_tid_t k_current_get(void)
Get thread ID of the current thread.
Definition kernel.h:836
void k_object_access_grant(const void *object, struct k_thread *thread)
Grant a thread access to a kernel object.
void k_object_access_revoke(const void *object, struct k_thread *thread)
Revoke a thread's access to a kernel object.
#define NULL
Definition iar_missing_defs.h:20
RTIO I/O Device and Related Functions.
flags
Definition parser.h:97
RTIO Submission Queue Events and Related Functions.
__UINT32_TYPE__ uint32_t
Definition stdint.h:90
__UINT8_TYPE__ uint8_t
Definition stdint.h:88
#define UINT16_MAX
Definition stdint.h:28
__UINTPTR_TYPE__ uintptr_t
Definition stdint.h:105
__UINT16_TYPE__ uint16_t
Definition stdint.h:89
void * memset(void *buf, int c, size_t n)
Memory Partition.
Definition mem_domain.h:55
Semaphore structure.
Definition kernel.h:3607
Thread Structure.
Definition thread.h:259
Kernel timeout type.
Definition clock.h:65
Kernel timepoint type.
Definition clock.h:291
Queue member.
Definition mpsc_lockfree.h:79
MPSC Queue.
Definition mpsc_lockfree.h:86
A completion queue event.
Definition cqe.h:83
void * userdata
Associated userdata with operation.
Definition cqe.h:87
struct mpsc_node q
Definition cqe.h:84
uint32_t flags
Flags associated with the operation.
Definition cqe.h:88
int32_t result
Result from operation.
Definition cqe.h:86
IO device submission queue entry.
Definition sqe.h:394
struct rtio_sqe sqe
Definition sqe.h:395
struct rtio * r
Definition sqe.h:398
struct mpsc_node q
Definition sqe.h:396
Pool of RTIO contexts to use with dynamically created threads.
Definition rtio.h:1043
struct rtio ** contexts
Array containing contexts of the pool.
Definition rtio.h:1048
atomic_t * used
Atomic bitmap to signal a member is used/unused.
Definition rtio.h:1051
size_t pool_size
Size of the pool.
Definition rtio.h:1045
A submission queue event.
Definition sqe.h:304
void * userdata
User provided data which is returned upon operation completion.
Definition sqe.h:322
uint8_t op
Op code.
Definition sqe.h:305
struct rtio_sqe::@126267262255374054123217063150244034155174062054::@222067021034074304061254367152327164076222165070 rx
OP_RX.
atomic_t ok
Definition sqe.h:381
struct rtio_sqe::@126267262255374054123217063150244034155174062054::@236333123355174166163204241333175337261032350217 await
OP_AWAIT.
uint32_t buf_len
Length of buffer.
Definition sqe.h:328
uint16_t flags
Op Flags.
Definition sqe.h:309
const uint8_t * buf
Buffer to write from.
Definition sqe.h:329
rtio_callback_t callback
Definition sqe.h:346
An RTIO context containing what can be viewed as a pair of queues.
Definition rtio.h:71
struct rtio_cqe_pool * cqe_pool
Definition rtio.h:101
struct mpsc sq
Definition rtio.h:109
atomic_t cq_count
Definition rtio.h:90
struct rtio_sqe_pool * sqe_pool
Definition rtio.h:98
atomic_t xcqcnt
Definition rtio.h:95
struct mpsc cq
Definition rtio.h:112
static __pinned_func bool k_is_user_context(void)
Indicate whether the CPU is currently in user mode.
Definition syscall.h:121