Zephyr API Documentation 4.4.99
A Scalable Open Source RTOS
Loading...
Searching...
No Matches
rtio.h
Go to the documentation of this file.
1/*
2 * SPDX-FileCopyrightText: Copyright (c) 2022 Intel Corporation
3 * SPDX-FileCopyrightText: Copyright (c) 2026 Infineon Technologies AG,
4 * or an affiliate of Infineon Technologies AG.
5 *
6 * SPDX-License-Identifier: Apache-2.0
7 */
8
27
28#ifndef ZEPHYR_INCLUDE_RTIO_RTIO_H_
29#define ZEPHYR_INCLUDE_RTIO_RTIO_H_
30
31#include <string.h>
32
34#include <zephyr/device.h>
35#include <zephyr/kernel.h>
37#include <zephyr/sys/__assert.h>
38#include <zephyr/sys/atomic.h>
40#include <zephyr/sys/util.h>
42#include <zephyr/rtio/sqe.h>
43#include <zephyr/rtio/cqe.h>
44#include <zephyr/rtio/iodev.h>
45
46#ifdef __cplusplus
47extern "C" {
48#endif
49
50
59
71struct rtio {
72#ifdef CONFIG_RTIO_SUBMIT_SEM
73 /* A wait semaphore which may suspend the calling thread
74 * to wait for some number of completions when calling submit
75 */
76 struct k_sem *submit_sem;
77
78 uint32_t submit_count;
79#endif
80
81#ifdef CONFIG_RTIO_CONSUME_SEM
82 /* A wait semaphore which may suspend the calling thread
83 * to wait for some number of completions while consuming
84 * them from the completion queue
85 */
86 struct k_sem *consume_sem;
87#endif
88
89 /* Total number of completions */
91
92 /* Number of completions that were unable to be submitted with results
93 * due to the cq spsc being full
94 */
96
97 /* Submission queue object pool with free list */
98 struct rtio_sqe_pool *sqe_pool;
99
100 /* Complete queue object pool with free list */
101 struct rtio_cqe_pool *cqe_pool;
102
103#ifdef CONFIG_RTIO_SYS_MEM_BLOCKS
104 /* Mem block pool */
105 struct sys_mem_blocks *block_pool;
106#endif
107
108 /* Submission queue */
109 struct mpsc sq;
110
111 /* Completion queue */
112 struct mpsc cq;
113};
114
115/* @cond ignore */
116#define Z_RTIO_DEFINE(name, _sqe_pool, _cqe_pool, _block_pool) \
117 IF_ENABLED(CONFIG_RTIO_SUBMIT_SEM, \
118 (static K_SEM_DEFINE(CONCAT(_submit_sem_, name), 0, K_SEM_MAX_LIMIT))) \
119 IF_ENABLED(CONFIG_RTIO_CONSUME_SEM, \
120 (static K_SEM_DEFINE(CONCAT(_consume_sem_, name), 0, K_SEM_MAX_LIMIT))) \
121 STRUCT_SECTION_ITERABLE(rtio, name) = { \
122 IF_ENABLED(CONFIG_RTIO_SUBMIT_SEM, (.submit_sem = &CONCAT(_submit_sem_, name),)) \
123 IF_ENABLED(CONFIG_RTIO_SUBMIT_SEM, (.submit_count = 0,)) \
124 IF_ENABLED(CONFIG_RTIO_CONSUME_SEM, (.consume_sem = &CONCAT(_consume_sem_, name),))\
125 .cq_count = ATOMIC_INIT(0), \
126 .xcqcnt = ATOMIC_INIT(0), \
127 .sqe_pool = _sqe_pool, \
128 .cqe_pool = _cqe_pool, \
129 IF_ENABLED(CONFIG_RTIO_SYS_MEM_BLOCKS, (.block_pool = _block_pool,)) \
130 .sq = MPSC_INIT((name.sq)), \
131 .cq = MPSC_INIT((name.cq)), \
132 }
133/* @endcond */
134
142#define RTIO_DEFINE(name, sq_sz, cq_sz) \
143 Z_RTIO_SQE_POOL_DEFINE(CONCAT(name, _sqe_pool), sq_sz); \
144 Z_RTIO_CQE_POOL_DEFINE(CONCAT(name, _cqe_pool), cq_sz); \
145 Z_RTIO_DEFINE(name, &CONCAT(name, _sqe_pool), \
146 &CONCAT(name, _cqe_pool), NULL)
147
148
156static inline size_t rtio_mempool_block_size(const struct rtio *r)
157{
158#ifndef CONFIG_RTIO_SYS_MEM_BLOCKS
159 ARG_UNUSED(r);
160 return 0;
161#else
162 if (r == NULL || r->block_pool == NULL) {
163 return 0;
164 }
165 return BIT(r->block_pool->info.blk_sz_shift);
166#endif
167}
168
176#if defined(CONFIG_RTIO_SYS_MEM_BLOCKS) || defined(__DOXYGEN__)
177static inline uint16_t __rtio_compute_mempool_block_index(const struct rtio *r, const void *ptr)
178{
179 uintptr_t addr = (uintptr_t)ptr;
180 struct sys_mem_blocks *mem_pool = r->block_pool;
181 uint32_t block_size = rtio_mempool_block_size(r);
182
183 uintptr_t buff = (uintptr_t)mem_pool->buffer;
184 uint32_t buff_size = mem_pool->info.num_blocks * block_size;
185
186 if (addr < buff || addr >= buff + buff_size) {
187 return UINT16_MAX;
188 }
189 return (addr - buff) / block_size;
190}
191#endif
192
193static inline int rtio_block_pool_alloc(struct rtio *r, size_t min_sz,
194 size_t max_sz, uint8_t **buf, uint32_t *buf_len)
195{
196#ifndef CONFIG_RTIO_SYS_MEM_BLOCKS
197 ARG_UNUSED(r);
198 ARG_UNUSED(min_sz);
199 ARG_UNUSED(max_sz);
200 ARG_UNUSED(buf);
201 ARG_UNUSED(buf_len);
202 return -ENOTSUP;
203#else
204 const uint32_t block_size = rtio_mempool_block_size(r);
205 uint32_t bytes = max_sz;
206
207 /* Not every context has a block pool and the block size may return 0 in
208 * that case
209 */
210 if (block_size == 0) {
211 return -ENOMEM;
212 }
213
214 do {
215 size_t num_blks = DIV_ROUND_UP(bytes, block_size);
216 int rc = sys_mem_blocks_alloc_contiguous(r->block_pool, num_blks, (void **)buf);
217
218 if (rc == 0) {
219 *buf_len = num_blks * block_size;
220 return 0;
221 }
222
223 if (bytes <= block_size) {
224 break;
225 }
226
227 bytes -= block_size;
228 } while (bytes >= min_sz);
229
230 return -ENOMEM;
231#endif
232}
233
234static inline void rtio_block_pool_free(struct rtio *r, void *buf, uint32_t buf_len)
235{
236#ifndef CONFIG_RTIO_SYS_MEM_BLOCKS
237 ARG_UNUSED(r);
238 ARG_UNUSED(buf);
239 ARG_UNUSED(buf_len);
240#else
241 size_t num_blks = buf_len >> r->block_pool->info.blk_sz_shift;
242
243 sys_mem_blocks_free_contiguous(r->block_pool, buf, num_blks);
244#endif
245}
246
247
249extern struct k_mem_partition rtio_partition;
250
251
252/* Do not try and reformat the macros */
253
263#if CONFIG_RTIO_BLOCK_POOL_PLACEMENT_DTCM
264#define RTIO_BMEM Z_GENERIC_SECTION(".dtcm_bss") static
265#elif defined(CONFIG_RTIO_BLOCK_POOL_PLACEMENT_NOCACHE)
266#define RTIO_BMEM __nocache static
267#else
268#define RTIO_BMEM COND_CODE_1(CONFIG_USERSPACE, (K_APP_BMEM(rtio_partition) static), (static))
269#endif
270
280#if CONFIG_RTIO_BLOCK_POOL_PLACEMENT_DTCM
281#define RTIO_DMEM Z_GENERIC_SECTION(".dtcm_data") static
282#elif defined(CONFIG_RTIO_BLOCK_POOL_PLACEMENT_NOCACHE)
283#define RTIO_DMEM __nocache_load static
284#else
285#define RTIO_DMEM COND_CODE_1(CONFIG_USERSPACE, (K_APP_DMEM(rtio_partition) static), (static))
286#endif
287
288/* clang-format off */
289/* @cond ignore */
290#define Z_RTIO_BLOCK_POOL_DEFINE(name, blk_sz, blk_cnt, blk_align) \
291 RTIO_BMEM uint8_t __aligned(WB_UP(blk_align)) \
292 CONCAT(_block_pool_, name)[blk_cnt*WB_UP(blk_sz)]; \
293 _SYS_MEM_BLOCKS_DEFINE_WITH_EXT_BUF(name, WB_UP(blk_sz), blk_cnt, \
294 CONCAT(_block_pool_, name), RTIO_DMEM)
295
296/* @endcond */
297
298
299/* clang-format on */
300
311#define RTIO_DEFINE_WITH_MEMPOOL(name, sq_sz, cq_sz, num_blks, blk_size, balign) \
312 Z_RTIO_SQE_POOL_DEFINE(name##_sqe_pool, sq_sz); \
313 Z_RTIO_CQE_POOL_DEFINE(name##_cqe_pool, cq_sz); \
314 Z_RTIO_BLOCK_POOL_DEFINE(name##_block_pool, blk_size, num_blks, balign); \
315 Z_RTIO_DEFINE(name, &name##_sqe_pool, &name##_cqe_pool, &name##_block_pool)
316
317/* clang-format on */
318
326static inline uint32_t rtio_sqe_acquirable(struct rtio *r)
327{
328 return r->sqe_pool->pool_free;
329}
330
339static inline struct rtio_sqe *rtio_sqe_acquire(struct rtio *r)
340{
341 SYS_PORT_TRACING_FUNC_ENTER(rtio, sqe_acquire, r);
342 struct rtio_iodev_sqe *iodev_sqe = rtio_sqe_pool_alloc(r->sqe_pool);
343
344 if (iodev_sqe == NULL) {
345 SYS_PORT_TRACING_FUNC_EXIT(rtio, sqe_acquire, r, NULL);
346 return NULL;
347 }
348
349 mpsc_push(&r->sq, &iodev_sqe->q);
350
351 SYS_PORT_TRACING_FUNC_EXIT(rtio, sqe_acquire, r, &iodev_sqe->sqe);
352 return &iodev_sqe->sqe;
353}
354
367static inline int rtio_sqe_acquire_array(struct rtio *r, size_t n, struct rtio_sqe **sqes)
368{
369 struct rtio_iodev_sqe *iodev_sqe;
370 size_t i;
371
372 for (i = 0; i < n; i++) {
373 iodev_sqe = rtio_sqe_pool_alloc(r->sqe_pool);
374 if (iodev_sqe == NULL) {
375 break;
376 }
377 sqes[i] = &iodev_sqe->sqe;
378 }
379
380 /* Not enough SQEs in the pool */
381 if (i < n) {
382 while (i > 0) {
383 i--;
384 iodev_sqe = CONTAINER_OF(sqes[i], struct rtio_iodev_sqe, sqe);
385 rtio_sqe_pool_free(r->sqe_pool, iodev_sqe);
386 sqes[i] = NULL;
387 }
388
389 return -ENOMEM;
390 }
391
392 for (i = 0; i < n; i++) {
393 iodev_sqe = CONTAINER_OF(sqes[i], struct rtio_iodev_sqe, sqe);
394 mpsc_push(&r->sq, &iodev_sqe->q);
395 }
396
397 return 0;
398}
399
405static inline void rtio_sqe_drop_all(struct rtio *r)
406{
407 struct rtio_iodev_sqe *iodev_sqe;
408 struct mpsc_node *node = mpsc_pop(&r->sq);
409
410 while (node != NULL) {
411 iodev_sqe = CONTAINER_OF(node, struct rtio_iodev_sqe, q);
412 rtio_sqe_pool_free(r->sqe_pool, iodev_sqe);
413 node = mpsc_pop(&r->sq);
414 }
415}
416
420static inline struct rtio_cqe *rtio_cqe_acquire(struct rtio *r)
421{
422 SYS_PORT_TRACING_FUNC_ENTER(rtio, cqe_acquire, r);
423 struct rtio_cqe *cqe = rtio_cqe_pool_alloc(r->cqe_pool);
424
425 if (cqe == NULL) {
426 SYS_PORT_TRACING_FUNC_EXIT(rtio, cqe_acquire, r, NULL);
427 return NULL;
428 }
429
430 memset(cqe, 0, sizeof(struct rtio_cqe));
431
432 SYS_PORT_TRACING_FUNC_EXIT(rtio, cqe_acquire, r, cqe);
433 return cqe;
434}
435
439static inline void rtio_cqe_produce(struct rtio *r, struct rtio_cqe *cqe)
440{
441 mpsc_push(&r->cq, &cqe->q);
442}
443
455static inline struct rtio_cqe *rtio_cqe_consume(struct rtio *r)
456{
457 SYS_PORT_TRACING_FUNC_ENTER(rtio, cqe_consume, r);
458 struct mpsc_node *node;
459 struct rtio_cqe *cqe = NULL;
460
461#ifdef CONFIG_RTIO_CONSUME_SEM
462 if (k_sem_take(r->consume_sem, K_NO_WAIT) != 0) {
463 SYS_PORT_TRACING_FUNC_EXIT(rtio, cqe_consume, r, NULL);
464 return NULL;
465 }
466#endif
467
468 node = mpsc_pop(&r->cq);
469 if (node == NULL) {
470 SYS_PORT_TRACING_FUNC_EXIT(rtio, cqe_consume, r, NULL);
471 return NULL;
472 }
473 cqe = CONTAINER_OF(node, struct rtio_cqe, q);
474
475 SYS_PORT_TRACING_FUNC_EXIT(rtio, cqe_consume, r, cqe);
476 return cqe;
477}
478
489static inline struct rtio_cqe *rtio_cqe_consume_block(struct rtio *r)
490{
491 struct mpsc_node *node;
492 struct rtio_cqe *cqe;
493
494#ifdef CONFIG_RTIO_CONSUME_SEM
495 k_sem_take(r->consume_sem, K_FOREVER);
496#endif
497 node = mpsc_pop(&r->cq);
498 while (node == NULL) {
499 Z_SPIN_DELAY(1);
500 node = mpsc_pop(&r->cq);
501 }
502 cqe = CONTAINER_OF(node, struct rtio_cqe, q);
503
504 return cqe;
505}
506
513static inline void rtio_cqe_release(struct rtio *r, struct rtio_cqe *cqe)
514{
515 SYS_PORT_TRACING_FUNC(rtio, cqe_release, r, cqe);
516 rtio_cqe_pool_free(r->cqe_pool, cqe);
517}
518
527static inline int rtio_flush_completion_queue(struct rtio *r)
528{
529 struct rtio_cqe *cqe;
530 int res = 0;
531
532 do {
533 cqe = rtio_cqe_consume(r);
534 if (cqe != NULL) {
535 if ((cqe->result < 0) && (res == 0)) {
536 res = cqe->result;
537 }
538 rtio_cqe_release(r, cqe);
539 }
540 } while (cqe != NULL);
541
542 return res;
543}
544
556__syscall void rtio_sqe_signal(struct rtio_sqe *sqe);
557
558static inline void z_impl_rtio_sqe_signal(struct rtio_sqe *sqe)
559{
560 struct rtio_iodev_sqe *iodev_sqe = CONTAINER_OF(sqe, struct rtio_iodev_sqe, sqe);
561
562 if (!atomic_cas(&iodev_sqe->sqe.await.ok, 0, 1)) {
563 iodev_sqe->sqe.await.callback(iodev_sqe, iodev_sqe->sqe.await.userdata);
564 }
565}
566
573static inline uint32_t rtio_cqe_compute_flags(struct rtio_iodev_sqe *iodev_sqe)
574{
575 uint32_t flags = 0;
576
577#ifdef CONFIG_RTIO_SYS_MEM_BLOCKS
578 if (iodev_sqe->sqe.op == RTIO_OP_RX && iodev_sqe->sqe.flags & RTIO_SQE_MEMPOOL_BUFFER) {
579 struct rtio *r = iodev_sqe->r;
580 struct sys_mem_blocks *mem_pool = r->block_pool;
581 unsigned int blk_index = 0;
582 unsigned int blk_count = 0;
583
584 if (iodev_sqe->sqe.rx.buf) {
585 blk_index = (iodev_sqe->sqe.rx.buf - mem_pool->buffer) >>
586 mem_pool->info.blk_sz_shift;
587 blk_count = iodev_sqe->sqe.rx.buf_len >> mem_pool->info.blk_sz_shift;
588 }
589 flags = RTIO_CQE_FLAG_PREP_MEMPOOL(blk_index, blk_count);
590 }
591#else
592 ARG_UNUSED(iodev_sqe);
593#endif
594
595 return flags;
596}
597
613__syscall int rtio_cqe_get_mempool_buffer(const struct rtio *r, struct rtio_cqe *cqe,
614 uint8_t **buff, uint32_t *buff_len);
615
616static inline int z_impl_rtio_cqe_get_mempool_buffer(const struct rtio *r, struct rtio_cqe *cqe,
617 uint8_t **buff, uint32_t *buff_len)
618{
619#ifdef CONFIG_RTIO_SYS_MEM_BLOCKS
621 unsigned int blk_idx = RTIO_CQE_FLAG_MEMPOOL_GET_BLK_IDX(cqe->flags);
622 unsigned int blk_count = RTIO_CQE_FLAG_MEMPOOL_GET_BLK_CNT(cqe->flags);
624
625 *buff_len = blk_count * blk_size;
626
627 if (blk_count > 0) {
628 *buff = r->block_pool->buffer + blk_idx * blk_size;
629
630 __ASSERT_NO_MSG(*buff >= r->block_pool->buffer);
631 __ASSERT_NO_MSG(*buff <
632 r->block_pool->buffer + blk_size * r->block_pool->info.num_blocks);
633 } else {
634 *buff = NULL;
635 }
636 return 0;
637 }
638 return -EINVAL;
639#else
640 ARG_UNUSED(r);
641 ARG_UNUSED(cqe);
642 ARG_UNUSED(buff);
643 ARG_UNUSED(buff_len);
644
645 return -ENOTSUP;
646#endif
647}
648
650void rtio_executor_ok(struct rtio_iodev_sqe *iodev_sqe, int result);
651void rtio_executor_err(struct rtio_iodev_sqe *iodev_sqe, int result);
652
661static inline void rtio_iodev_sqe_ok(struct rtio_iodev_sqe *iodev_sqe, int result)
662{
663 rtio_executor_ok(iodev_sqe, result);
664}
665
674static inline void rtio_iodev_sqe_err(struct rtio_iodev_sqe *iodev_sqe, int result)
675{
676 rtio_executor_err(iodev_sqe, result);
677}
678
690static inline void rtio_cqe_submit(struct rtio *r, int result, void *userdata, uint32_t flags)
691{
692 SYS_PORT_TRACING_FUNC_ENTER(rtio, cqe_submit, r, result, flags);
693 struct rtio_cqe *cqe = rtio_cqe_acquire(r);
694
695 if (cqe == NULL) {
696 atomic_inc(&r->xcqcnt);
697 } else {
698 cqe->result = result;
699 cqe->userdata = userdata;
700 cqe->flags = flags;
701 rtio_cqe_produce(r, cqe);
702#ifdef CONFIG_RTIO_CONSUME_SEM
703 k_sem_give(r->consume_sem);
704#endif
705 }
706
707 /* atomic_t isn't guaranteed to wrap correctly as it could be signed, so
708 * we must resort to a cas loop.
709 */
710 atomic_t val, new_val;
711
712 do {
713 val = atomic_get(&r->cq_count);
714 new_val = (atomic_t)((uintptr_t)val + 1);
715 } while (!atomic_cas(&r->cq_count, val, new_val));
716
717#ifdef CONFIG_RTIO_SUBMIT_SEM
718 if (r->submit_count > 0) {
719 r->submit_count--;
720 if (r->submit_count == 0) {
721 k_sem_give(r->submit_sem);
722 }
723 }
724#endif
725 SYS_PORT_TRACING_FUNC_EXIT(rtio, cqe_submit, r);
726}
727
728#define __RTIO_MEMPOOL_GET_NUM_BLKS(num_bytes, blk_size) (((num_bytes) + (blk_size)-1) / (blk_size))
729
742static inline int rtio_sqe_rx_buf(const struct rtio_iodev_sqe *iodev_sqe, uint32_t min_buf_len,
743 uint32_t max_buf_len, uint8_t **buf, uint32_t *buf_len)
744{
745 struct rtio_sqe *sqe = (struct rtio_sqe *)&iodev_sqe->sqe;
746
747#ifdef CONFIG_RTIO_SYS_MEM_BLOCKS
748 if (sqe->op == RTIO_OP_RX && sqe->flags & RTIO_SQE_MEMPOOL_BUFFER) {
749 struct rtio *r = iodev_sqe->r;
750
751 if (sqe->rx.buf != NULL) {
752 if (sqe->rx.buf_len < min_buf_len) {
753 return -ENOMEM;
754 }
755 *buf = sqe->rx.buf;
756 *buf_len = sqe->rx.buf_len;
757 return 0;
758 }
759
760 int rc = rtio_block_pool_alloc(r, min_buf_len, max_buf_len, buf, buf_len);
761 if (rc == 0) {
762 sqe->rx.buf = *buf;
763 sqe->rx.buf_len = *buf_len;
764 return 0;
765 }
766
767 return -ENOMEM;
768 }
769#else
770 ARG_UNUSED(max_buf_len);
771#endif
772
773 if (sqe->rx.buf_len < min_buf_len) {
774 return -ENOMEM;
775 }
776
777 *buf = sqe->rx.buf;
778 *buf_len = sqe->rx.buf_len;
779 return 0;
780}
781
796__syscall void rtio_release_buffer(struct rtio *r, void *buff, uint32_t buff_len);
797
798static inline void z_impl_rtio_release_buffer(struct rtio *r, void *buff, uint32_t buff_len)
799{
800#ifdef CONFIG_RTIO_SYS_MEM_BLOCKS
801 if (r == NULL || buff == NULL || r->block_pool == NULL || buff_len == 0) {
802 return;
803 }
804
805 rtio_block_pool_free(r, buff, buff_len);
806#else
807 ARG_UNUSED(r);
808 ARG_UNUSED(buff);
809 ARG_UNUSED(buff_len);
810#endif
811}
812
819static inline void rtio_access_grant(struct rtio *r, struct k_thread *t)
820{
822
823#ifdef CONFIG_RTIO_SUBMIT_SEM
824 k_object_access_grant(r->submit_sem, t);
825#endif
826
827#ifdef CONFIG_RTIO_CONSUME_SEM
828 k_object_access_grant(r->consume_sem, t);
829#endif
830}
831
832
839static inline void rtio_access_revoke(struct rtio *r, struct k_thread *t)
840{
842
843#ifdef CONFIG_RTIO_SUBMIT_SEM
844 k_object_access_revoke(r->submit_sem, t);
845#endif
846
847#ifdef CONFIG_RTIO_CONSUME_SEM
848 k_object_access_revoke(r->consume_sem, t);
849#endif
850}
851
862__syscall int rtio_sqe_cancel(struct rtio_sqe *sqe);
863
864static inline int z_impl_rtio_sqe_cancel(struct rtio_sqe *sqe)
865{
866 SYS_PORT_TRACING_FUNC(rtio, sqe_cancel, sqe);
867 struct rtio_iodev_sqe *iodev_sqe = CONTAINER_OF(sqe, struct rtio_iodev_sqe, sqe);
868
869 do {
870 iodev_sqe->sqe.flags |= RTIO_SQE_CANCELED;
871 iodev_sqe = rtio_iodev_sqe_next(iodev_sqe);
872 } while (iodev_sqe != NULL);
873
874 return 0;
875}
876
892__syscall int rtio_sqe_copy_in_get_handles(struct rtio *r, const struct rtio_sqe *sqes,
893 struct rtio_sqe **handle, size_t sqe_count);
894
895static inline int z_impl_rtio_sqe_copy_in_get_handles(struct rtio *r, const struct rtio_sqe *sqes,
896 struct rtio_sqe **handle,
897 size_t sqe_count)
898{
899 struct rtio_sqe *sqe;
900 uint32_t acquirable = rtio_sqe_acquirable(r);
901
902 if (acquirable < sqe_count) {
903 return -ENOMEM;
904 }
905
906 for (unsigned long i = 0; i < sqe_count; i++) {
907 sqe = rtio_sqe_acquire(r);
908 __ASSERT_NO_MSG(sqe != NULL);
909 if (handle != NULL && i == 0) {
910 *handle = sqe;
911 }
912 *sqe = sqes[i];
913 }
914
915 return 0;
916}
917
934static inline int rtio_sqe_copy_in(struct rtio *r, const struct rtio_sqe *sqes, size_t sqe_count)
935{
936 return rtio_sqe_copy_in_get_handles(r, sqes, NULL, sqe_count);
937}
938
954__syscall int rtio_cqe_copy_out(struct rtio *r,
955 struct rtio_cqe *cqes,
956 size_t cqe_count,
957 k_timeout_t timeout);
958static inline int z_impl_rtio_cqe_copy_out(struct rtio *r,
959 struct rtio_cqe *cqes,
960 size_t cqe_count,
961 k_timeout_t timeout)
962{
963 size_t copied = 0;
964 struct rtio_cqe *cqe;
965 k_timepoint_t end = sys_timepoint_calc(timeout);
966
967 do {
970 if (cqe == NULL) {
971 Z_SPIN_DELAY(25);
972 continue;
973 }
974 cqes[copied++] = *cqe;
975 rtio_cqe_release(r, cqe);
976 } while (copied < cqe_count && !sys_timepoint_expired(end));
977
978 return copied;
979}
980
996__syscall int rtio_submit(struct rtio *r, uint32_t wait_count);
997
998#ifdef CONFIG_RTIO_SUBMIT_SEM
999static inline int z_impl_rtio_submit(struct rtio *r, uint32_t wait_count)
1000{
1001 SYS_PORT_TRACING_FUNC_ENTER(rtio, submit, r, wait_count);
1002 int res = 0;
1003
1004 if (wait_count > 0) {
1005 __ASSERT(!k_is_in_isr(),
1006 "expected rtio submit with wait count to be called from a thread");
1007
1008 k_sem_reset(r->submit_sem);
1009 r->submit_count = wait_count;
1010 }
1011
1013
1014 if (wait_count > 0) {
1015 res = k_sem_take(r->submit_sem, K_FOREVER);
1016 __ASSERT(res == 0,
1017 "semaphore was reset or timed out while waiting on completions!");
1018 }
1019
1021 return res;
1022}
1023#else
1024static inline int z_impl_rtio_submit(struct rtio *r, uint32_t wait_count)
1025{
1026
1027 SYS_PORT_TRACING_FUNC_ENTER(rtio, submit, r, wait_count);
1028 int res = 0;
1029 uintptr_t cq_count = (uintptr_t)atomic_get(&r->cq_count);
1030 uintptr_t cq_complete_count = cq_count + wait_count;
1031 bool wraps = cq_complete_count < cq_count;
1032
1034
1035 if (wraps) {
1036 while ((uintptr_t)atomic_get(&r->cq_count) >= cq_count) {
1037 Z_SPIN_DELAY(10);
1038 k_yield();
1039 }
1040 }
1041
1042 while ((uintptr_t)atomic_get(&r->cq_count) < cq_complete_count) {
1043 Z_SPIN_DELAY(10);
1044 k_yield();
1045 }
1046
1048 return res;
1049}
1050#endif /* CONFIG_RTIO_SUBMIT_SEM */
1051
1058
1060 struct rtio **contexts;
1061
1064};
1065
1074__syscall struct rtio *rtio_pool_acquire(struct rtio_pool *pool);
1075
1076static inline struct rtio *z_impl_rtio_pool_acquire(struct rtio_pool *pool)
1077{
1078 struct rtio *r = NULL;
1079
1080 for (size_t i = 0; i < pool->pool_size; i++) {
1081 if (atomic_test_and_set_bit(pool->used, i) == 0) {
1082 r = pool->contexts[i];
1083 break;
1084 }
1085 }
1086
1087 if (r != NULL) {
1089 }
1090
1091 return r;
1092}
1093
1100__syscall void rtio_pool_release(struct rtio_pool *pool, struct rtio *r);
1101
1102static inline void z_impl_rtio_pool_release(struct rtio_pool *pool, struct rtio *r)
1103{
1104
1105 if (k_is_user_context()) {
1107 }
1108
1109 for (size_t i = 0; i < pool->pool_size; i++) {
1110 if (pool->contexts[i] == r) {
1111 atomic_clear_bit(pool->used, i);
1112 break;
1113 }
1114 }
1115}
1116
1117/* clang-format off */
1118
1120
1121#define Z_RTIO_POOL_NAME_N(n, name) \
1122 name##_##n
1123
1124#define Z_RTIO_POOL_DEFINE_N(n, name, sq_sz, cq_sz) \
1125 RTIO_DEFINE(Z_RTIO_POOL_NAME_N(n, name), sq_sz, cq_sz)
1126
1127#define Z_RTIO_POOL_REF_N(n, name) \
1128 &Z_RTIO_POOL_NAME_N(n, name)
1129
1131
1140#define RTIO_POOL_DEFINE(name, pool_sz, sq_sz, cq_sz) \
1141 LISTIFY(pool_sz, Z_RTIO_POOL_DEFINE_N, (;), name, sq_sz, cq_sz); \
1142 static struct rtio *name##_contexts[] = { \
1143 LISTIFY(pool_sz, Z_RTIO_POOL_REF_N, (,), name) \
1144 }; \
1145 ATOMIC_DEFINE(name##_used, pool_sz); \
1146 STRUCT_SECTION_ITERABLE(rtio_pool, name) = { \
1147 .pool_size = pool_sz, \
1148 .contexts = name##_contexts, \
1149 .used = name##_used, \
1150 }
1151
1152/* clang-format on */
1153
1157
1158#ifdef __cplusplus
1159}
1160#endif
1161
1162#include <zephyr/syscalls/rtio.h>
1163
1164#endif /* ZEPHYR_INCLUDE_RTIO_RTIO_H_ */
workaround assembler barfing for ST r
Definition asm-macro-32-bit-gnu.h:24
long atomic_t
Definition atomic_types.h:15
RTIO Completion Queue Events and Related Functions.
static _Bool atomic_test_and_set_bit(atomic_t *target, int bit)
Atomically set a bit and test it.
Definition atomic.h:172
static void atomic_clear_bit(atomic_t *target, int bit)
Atomically clear a bit.
Definition atomic.h:193
atomic_val_t atomic_get(const atomic_t *target)
Atomic get.
atomic_val_t atomic_inc(atomic_t *target)
Atomic increment.
_Bool atomic_cas(atomic_t *target, atomic_val_t old_value, atomic_val_t new_value)
Atomic compare-and-set.
#define K_FOREVER
Generate infinite timeout delay.
Definition kernel.h:1664
#define K_NO_WAIT
Generate null timeout delay.
Definition kernel.h:1554
k_timepoint_t sys_timepoint_calc(k_timeout_t timeout)
Calculate a timepoint value.
static bool sys_timepoint_expired(k_timepoint_t timepoint)
Indicates if timepoint is expired.
Definition clock.h:388
#define K_TIMEOUT_EQ(a, b)
Compare timeouts for equality.
Definition clock.h:80
bool k_is_in_isr(void)
Determine if code is running at interrupt level.
int sys_mem_blocks_free_contiguous(sys_mem_blocks_t *mem_block, void *block, size_t count)
Free contiguous multiple memory blocks.
int sys_mem_blocks_alloc_contiguous(sys_mem_blocks_t *mem_block, size_t count, void **out_block)
Allocate a contiguous set of memory blocks.
static ALWAYS_INLINE void mpsc_push(struct mpsc *q, struct mpsc_node *n)
Push a node.
Definition mpsc_lockfree.h:126
static struct mpsc_node * mpsc_pop(struct mpsc *q)
Pop a node off of the list.
Definition mpsc_lockfree.h:145
#define RTIO_CQE_FLAG_MEMPOOL_GET_BLK_CNT(flags)
Get the block count of a mempool flags.
Definition cqe.h:63
#define RTIO_CQE_FLAG_MEMPOOL_GET_BLK_IDX(flags)
Get the block index of a mempool flags.
Definition cqe.h:55
#define RTIO_CQE_FLAG_MEMPOOL_BUFFER
The entry's buffer was allocated from the RTIO's mempool.
Definition cqe.h:45
#define RTIO_CQE_FLAG_PREP_MEMPOOL(blk_idx, blk_cnt)
Prepare CQE flags for a mempool read.
Definition cqe.h:72
#define RTIO_CQE_FLAG_GET(flags)
Definition cqe.h:47
#define RTIO_OP_RX
An operation that receives (reads).
Definition sqe.h:144
#define RTIO_SQE_MEMPOOL_BUFFER
The buffer should be allocated by the RTIO mempool.
Definition sqe.h:105
#define RTIO_SQE_CANCELED
The SQE should not execute if possible.
Definition sqe.h:113
void rtio_pool_release(struct rtio_pool *pool, struct rtio *r)
Return an RTIO context to a pool.
void rtio_executor_err(struct rtio_iodev_sqe *iodev_sqe, int result)
static uint32_t rtio_sqe_acquirable(struct rtio *r)
Count of acquirable submission queue events.
Definition rtio.h:326
struct rtio * rtio_pool_acquire(struct rtio_pool *pool)
Obtain an RTIO context from a pool.
static size_t rtio_mempool_block_size(const struct rtio *r)
Get the mempool block size of the RTIO context.
Definition rtio.h:156
static void rtio_cqe_submit(struct rtio *r, int result, void *userdata, uint32_t flags)
Submit a completion queue event with a given result and userdata.
Definition rtio.h:690
void rtio_release_buffer(struct rtio *r, void *buff, uint32_t buff_len)
Release memory that was allocated by the RTIO's memory pool.
static int rtio_sqe_copy_in(struct rtio *r, const struct rtio_sqe *sqes, size_t sqe_count)
Copy an array of SQEs into the queue.
Definition rtio.h:934
static void rtio_cqe_produce(struct rtio *r, struct rtio_cqe *cqe)
Produce a complete queue event if available.
Definition rtio.h:439
static uint32_t rtio_cqe_compute_flags(struct rtio_iodev_sqe *iodev_sqe)
Compute the CQE flags from the rtio_iodev_sqe entry.
Definition rtio.h:573
void rtio_executor_ok(struct rtio_iodev_sqe *iodev_sqe, int result)
static int rtio_block_pool_alloc(struct rtio *r, size_t min_sz, size_t max_sz, uint8_t **buf, uint32_t *buf_len)
Definition rtio.h:193
int rtio_sqe_copy_in_get_handles(struct rtio *r, const struct rtio_sqe *sqes, struct rtio_sqe **handle, size_t sqe_count)
Copy an array of SQEs into the queue and get resulting handles back.
struct k_mem_partition rtio_partition
The memory partition associated with all RTIO context information.
static struct rtio_sqe * rtio_sqe_acquire(struct rtio *r)
Acquire a single submission queue event if available.
Definition rtio.h:339
static void rtio_sqe_drop_all(struct rtio *r)
Drop all previously acquired sqe.
Definition rtio.h:405
int rtio_cqe_copy_out(struct rtio *r, struct rtio_cqe *cqes, size_t cqe_count, k_timeout_t timeout)
Copy an array of CQEs from the queue.
static int rtio_flush_completion_queue(struct rtio *r)
Flush completion queue.
Definition rtio.h:527
static void rtio_access_revoke(struct rtio *r, struct k_thread *t)
Revoke access to an RTIO context from a user thread.
Definition rtio.h:839
static void rtio_access_grant(struct rtio *r, struct k_thread *t)
Grant access to an RTIO context to a user thread.
Definition rtio.h:819
static void rtio_cqe_release(struct rtio *r, struct rtio_cqe *cqe)
Release consumed completion queue event.
Definition rtio.h:513
static int rtio_sqe_rx_buf(const struct rtio_iodev_sqe *iodev_sqe, uint32_t min_buf_len, uint32_t max_buf_len, uint8_t **buf, uint32_t *buf_len)
Get the buffer associate with the RX submission.
Definition rtio.h:742
static void rtio_iodev_sqe_err(struct rtio_iodev_sqe *iodev_sqe, int result)
Inform the executor of a submissions completion with error.
Definition rtio.h:674
int rtio_sqe_cancel(struct rtio_sqe *sqe)
Attempt to cancel an SQE.
static void rtio_iodev_sqe_ok(struct rtio_iodev_sqe *iodev_sqe, int result)
Inform the executor of a submission completion with success.
Definition rtio.h:661
static struct rtio_cqe * rtio_cqe_acquire(struct rtio *r)
Acquire a complete queue event if available.
Definition rtio.h:420
static struct rtio_cqe * rtio_cqe_consume(struct rtio *r)
Consume a single completion queue event if available.
Definition rtio.h:455
void rtio_sqe_signal(struct rtio_sqe *sqe)
Signal an AWAIT SQE.
static struct rtio_iodev_sqe * rtio_iodev_sqe_next(const struct rtio_iodev_sqe *iodev_sqe)
Get the next sqe in the chain or transaction.
Definition sqe.h:713
int rtio_cqe_get_mempool_buffer(const struct rtio *r, struct rtio_cqe *cqe, uint8_t **buff, uint32_t *buff_len)
Retrieve the mempool buffer that was allocated for the CQE.
void rtio_executor_submit(struct rtio *r)
static struct rtio_cqe * rtio_cqe_consume_block(struct rtio *r)
Wait for and consume a single completion queue event.
Definition rtio.h:489
static int rtio_sqe_acquire_array(struct rtio *r, size_t n, struct rtio_sqe **sqes)
Acquire a number of submission queue events if available.
Definition rtio.h:367
static void rtio_block_pool_free(struct rtio *r, void *buf, uint32_t buf_len)
Definition rtio.h:234
int rtio_submit(struct rtio *r, uint32_t wait_count)
Submit I/O requests to the underlying executor.
void k_sem_reset(struct k_sem *sem)
Resets a semaphore's count to zero.
void k_sem_give(struct k_sem *sem)
Give a semaphore.
int k_sem_take(struct k_sem *sem, k_timeout_t timeout)
Take a semaphore.
#define SYS_PORT_TRACING_FUNC_ENTER(type, func,...)
Tracing macro for the entry into a function that might or might not return a value.
Definition tracing_macros.h:257
#define SYS_PORT_TRACING_FUNC_EXIT(type, func,...)
Tracing macro for when a function ends its execution.
Definition tracing_macros.h:283
#define SYS_PORT_TRACING_FUNC(type, func,...)
Tracing macro for function calls which are not directly associated with a specific type of object.
Definition tracing_macros.h:244
#define BIT(n)
Unsigned integer with bit position n set (signed in assembly language).
Definition util_macro.h:44
#define CONTAINER_OF(ptr, type, field)
Get a pointer to a structure containing the element.
Definition util.h:281
#define DIV_ROUND_UP(n, d)
Divide and round up.
Definition util.h:348
#define EINVAL
Invalid argument.
Definition errno.h:60
#define ENOMEM
Not enough core.
Definition errno.h:50
#define ENOTSUP
Unsupported value.
Definition errno.h:114
void k_yield(void)
Yield the current thread.
static __attribute_const__ k_tid_t k_current_get(void)
Get thread ID of the current thread.
Definition kernel.h:836
void k_object_access_grant(const void *object, struct k_thread *thread)
Grant a thread access to a kernel object.
void k_object_access_revoke(const void *object, struct k_thread *thread)
Revoke a thread's access to a kernel object.
#define NULL
Definition iar_missing_defs.h:20
RTIO I/O Device and Related Functions.
Public kernel APIs.
Memory Blocks Allocator.
flags
Definition parser.h:97
RTIO Submission Queue Events and Related Functions.
__UINT32_TYPE__ uint32_t
Definition stdint.h:90
__UINT8_TYPE__ uint8_t
Definition stdint.h:88
#define UINT16_MAX
Definition stdint.h:28
__UINTPTR_TYPE__ uintptr_t
Definition stdint.h:105
__UINT16_TYPE__ uint16_t
Definition stdint.h:89
void * memset(void *buf, int c, size_t n)
Memory Partition.
Definition mem_domain.h:55
Semaphore structure.
Definition kernel.h:3663
Thread Structure.
Definition thread.h:259
Kernel timeout type.
Definition clock.h:65
Kernel timepoint type.
Definition clock.h:291
Queue member.
Definition mpsc_lockfree.h:79
MPSC Queue.
Definition mpsc_lockfree.h:86
A completion queue event.
Definition cqe.h:83
void * userdata
Associated userdata with operation.
Definition cqe.h:87
struct mpsc_node q
Definition cqe.h:84
uint32_t flags
Flags associated with the operation.
Definition cqe.h:88
int32_t result
Result from operation.
Definition cqe.h:86
IO device submission queue entry.
Definition sqe.h:394
struct rtio_sqe sqe
Definition sqe.h:395
struct rtio * r
Definition sqe.h:398
struct mpsc_node q
Definition sqe.h:396
Pool of RTIO contexts to use with dynamically created threads.
Definition rtio.h:1055
struct rtio ** contexts
Array containing contexts of the pool.
Definition rtio.h:1060
atomic_t * used
Atomic bitmap to signal a member is used/unused.
Definition rtio.h:1063
size_t pool_size
Size of the pool.
Definition rtio.h:1057
A submission queue event.
Definition sqe.h:304
void * userdata
User provided data which is returned upon operation completion.
Definition sqe.h:322
uint8_t op
Op code.
Definition sqe.h:305
struct rtio_sqe::@126267262255374054123217063150244034155174062054::@222067021034074304061254367152327164076222165070 rx
OP_RX.
atomic_t ok
Definition sqe.h:381
struct rtio_sqe::@126267262255374054123217063150244034155174062054::@236333123355174166163204241333175337261032350217 await
OP_AWAIT.
uint32_t buf_len
Length of buffer.
Definition sqe.h:328
uint16_t flags
Op Flags.
Definition sqe.h:309
const uint8_t * buf
Buffer to write from.
Definition sqe.h:329
rtio_callback_t callback
Definition sqe.h:346
An RTIO context containing what can be viewed as a pair of queues.
Definition rtio.h:71
struct rtio_cqe_pool * cqe_pool
Definition rtio.h:101
struct mpsc sq
Definition rtio.h:109
atomic_t cq_count
Definition rtio.h:90
struct rtio_sqe_pool * sqe_pool
Definition rtio.h:98
atomic_t xcqcnt
Definition rtio.h:95
struct mpsc cq
Definition rtio.h:112
Misc utilities.
static __pinned_func bool k_is_user_context(void)
Indicate whether the CPU is currently in user mode.
Definition syscall.h:121