Zephyr API Documentation 4.3.99
A Scalable Open Source RTOS
Loading...
Searching...
No Matches
sqe.h
Go to the documentation of this file.
1/*
2 * SPDX-FileCopyrightText: Copyright (c) 2022 Intel Corporation
3 * SPDX-FileCopyrightText: Copyright (c) 2026 Infineon Technologies AG,
4 * or an affiliate of Infineon Technologies AG.
5 *
6 * SPDX-License-Identifier: Apache-2.0
7 */
8
13
14
15#ifndef ZEPHYR_INCLUDE_RTIO_SQE_H_
16#define ZEPHYR_INCLUDE_RTIO_SQE_H_
17
18#include <stdint.h>
19#include <string.h>
20#include <zephyr/kernel.h>
21#include <zephyr/sys/atomic.h>
23#include <zephyr/sys/util.h>
24
25#ifdef __cplusplus
26extern "C" {
27#endif
28
33
46
50#define RTIO_PRIO_LOW 0U
51
55#define RTIO_PRIO_NORM 127U
56
60#define RTIO_PRIO_HIGH 255U
61
65
66
73
81#define RTIO_SQE_CHAINED BIT(0)
82
93#define RTIO_SQE_TRANSACTION BIT(1)
94
95
105#define RTIO_SQE_MEMPOOL_BUFFER BIT(2)
106
113#define RTIO_SQE_CANCELED BIT(3)
114
121#define RTIO_SQE_MULTISHOT BIT(4)
122
126#define RTIO_SQE_NO_RESPONSE BIT(5)
127
131
132
139
141#define RTIO_OP_NOP 0
142
144#define RTIO_OP_RX (RTIO_OP_NOP+1)
145
147#define RTIO_OP_TX (RTIO_OP_RX+1)
148
150#define RTIO_OP_TINY_TX (RTIO_OP_TX+1)
151
153#define RTIO_OP_CALLBACK (RTIO_OP_TINY_TX+1)
154
156#define RTIO_OP_TXRX (RTIO_OP_CALLBACK+1)
157
159#define RTIO_OP_DELAY (RTIO_OP_TXRX+1)
160
162#define RTIO_OP_I2C_RECOVER (RTIO_OP_DELAY+1)
163
165#define RTIO_OP_I2C_CONFIGURE (RTIO_OP_I2C_RECOVER+1)
166
168#define RTIO_OP_I3C_RECOVER (RTIO_OP_I2C_CONFIGURE+1)
169
171#define RTIO_OP_I3C_CONFIGURE (RTIO_OP_I3C_RECOVER+1)
172
174#define RTIO_OP_I3C_CCC (RTIO_OP_I3C_CONFIGURE+1)
175
177#define RTIO_OP_AWAIT (RTIO_OP_I3C_CCC+1)
178
182
183
190
194#define RTIO_IODEV_I2C_STOP BIT(1)
195
199#define RTIO_IODEV_I2C_RESTART BIT(2)
200
204#define RTIO_IODEV_I2C_10_BITS BIT(3)
205
209
216
220#define RTIO_IODEV_I3C_STOP BIT(1)
221
225#define RTIO_IODEV_I3C_RESTART BIT(2)
226
230#define RTIO_IODEV_I3C_HDR BIT(3)
231
235#define RTIO_IODEV_I3C_NBCH BIT(4)
236
240#define RTIO_IODEV_I3C_HDR_MODE_MASK GENMASK(15, 8)
241
245#define RTIO_IODEV_I3C_HDR_MODE_SET(flags) \
246 FIELD_PREP(RTIO_IODEV_I3C_HDR_MODE_MASK, flags)
247
251#define RTIO_IODEV_I3C_HDR_MODE_GET(flags) \
252 FIELD_GET(RTIO_IODEV_I3C_HDR_MODE_MASK, flags)
253
257#define RTIO_IODEV_I3C_HDR_CMD_CODE_MASK GENMASK(22, 16)
258
262#define RTIO_IODEV_I3C_HDR_CMD_CODE_SET(flags) \
263 FIELD_PREP(RTIO_IODEV_I3C_HDR_CMD_CODE_MASK, flags)
264
268#define RTIO_IODEV_I3C_HDR_CMD_CODE_GET(flags) \
269 FIELD_GET(RTIO_IODEV_I3C_HDR_CMD_CODE_MASK, flags)
270
274
275
277struct rtio_sqe;
278struct rtio_iodev_sqe;
279struct rtio_iodev;
280struct rtio;
282
291typedef void (*rtio_callback_t)(struct rtio *r, const struct rtio_sqe *sqe, int res, void *arg0);
292
299typedef void (*rtio_signaled_t)(struct rtio_iodev_sqe *iodev_sqe, void *userdata);
300
304struct rtio_sqe {
306
308
310
312
313 const struct rtio_iodev *iodev;
314
322 void *userdata;
323
324 union {
325
327 struct {
329 const uint8_t *buf;
330 } tx;
331
333 struct {
334 uint32_t buf_len;
336 } rx;
337
339 struct {
342 } tiny_tx;
343
345 struct {
347 void *arg0;
348 } callback;
349
351 struct {
352 uint32_t buf_len;
355 } txrx;
356
357#ifdef CONFIG_RTIO_OP_DELAY
359 struct {
360 k_timeout_t timeout;
361 struct _timeout to;
362 } delay;
363#endif
364
367
369 struct {
370 /* enum i3c_config_type type; */
371 int type;
372 void *config;
373 } i3c_config;
374
376 /* struct i3c_ccc_payload *ccc_payload; */
378
380 struct {
383 void *userdata;
384 } await;
385 };
386};
387
388
395 struct rtio_sqe sqe;
396 struct mpsc_node q;
398 struct rtio *r;
399};
400
401
403/* Ensure the rtio_iodev_sqe never grows beyond a common cacheline size of 64 bytes */
404#if CONFIG_RTIO_SQE_CACHELINE_CHECK
405#ifdef CONFIG_DCACHE_LINE_SIZE
406#define RTIO_CACHE_LINE_SIZE CONFIG_DCACHE_LINE_SIZE
407#else
408#define RTIO_CACHE_LINE_SIZE 64
409#endif
410BUILD_ASSERT(sizeof(struct rtio_iodev_sqe) <= RTIO_CACHE_LINE_SIZE,
411 "RTIO performs best when the submissions queue entries are less than a cache line")
412#endif
414
418static inline void rtio_sqe_prep_nop(struct rtio_sqe *sqe,
419 const struct rtio_iodev *iodev,
420 void *userdata)
421{
422 memset(sqe, 0, sizeof(struct rtio_sqe));
423 sqe->op = RTIO_OP_NOP;
424 sqe->iodev = iodev;
425 sqe->userdata = userdata;
426}
427
431static inline void rtio_sqe_prep_read(struct rtio_sqe *sqe,
432 const struct rtio_iodev *iodev,
433 int8_t prio,
434 uint8_t *buf,
435 uint32_t len,
436 void *userdata)
437{
438 memset(sqe, 0, sizeof(struct rtio_sqe));
439 sqe->op = RTIO_OP_RX;
440 sqe->prio = prio;
441 sqe->iodev = iodev;
442 sqe->rx.buf_len = len;
443 sqe->rx.buf = buf;
444 sqe->userdata = userdata;
445}
446
452static inline void rtio_sqe_prep_read_with_pool(struct rtio_sqe *sqe,
453 const struct rtio_iodev *iodev, int8_t prio,
454 void *userdata)
455{
456 rtio_sqe_prep_read(sqe, iodev, prio, NULL, 0, userdata);
458}
459
460static inline void rtio_sqe_prep_read_multishot(struct rtio_sqe *sqe,
461 const struct rtio_iodev *iodev, int8_t prio,
462 void *userdata)
463{
464 rtio_sqe_prep_read_with_pool(sqe, iodev, prio, userdata);
466}
467
471static inline void rtio_sqe_prep_write(struct rtio_sqe *sqe,
472 const struct rtio_iodev *iodev,
473 int8_t prio,
474 const uint8_t *buf,
475 uint32_t len,
476 void *userdata)
477{
478 memset(sqe, 0, sizeof(struct rtio_sqe));
479 sqe->op = RTIO_OP_TX;
480 sqe->prio = prio;
481 sqe->iodev = iodev;
482 sqe->tx.buf_len = len;
483 sqe->tx.buf = buf;
484 sqe->userdata = userdata;
485}
486
497static inline void rtio_sqe_prep_tiny_write(struct rtio_sqe *sqe,
498 const struct rtio_iodev *iodev,
499 int8_t prio,
500 const uint8_t *tiny_write_data,
501 uint8_t tiny_write_len,
502 void *userdata)
503{
504 __ASSERT_NO_MSG(tiny_write_len <= sizeof(sqe->tiny_tx.buf));
505
506 memset(sqe, 0, sizeof(struct rtio_sqe));
507 sqe->op = RTIO_OP_TINY_TX;
508 sqe->prio = prio;
509 sqe->iodev = iodev;
510 sqe->tiny_tx.buf_len = tiny_write_len;
511 memcpy(sqe->tiny_tx.buf, tiny_write_data, tiny_write_len);
512 sqe->userdata = userdata;
513}
514
523static inline void rtio_sqe_prep_callback(struct rtio_sqe *sqe,
524 rtio_callback_t callback,
525 void *arg0,
526 void *userdata)
527{
528 memset(sqe, 0, sizeof(struct rtio_sqe));
529 sqe->op = RTIO_OP_CALLBACK;
530 sqe->prio = 0;
531 sqe->iodev = NULL;
532 sqe->callback.callback = callback;
533 sqe->callback.arg0 = arg0;
534 sqe->userdata = userdata;
535}
536
547static inline void rtio_sqe_prep_callback_no_cqe(struct rtio_sqe *sqe,
548 rtio_callback_t callback,
549 void *arg0,
550 void *userdata)
551{
552 rtio_sqe_prep_callback(sqe, callback, arg0, userdata);
554}
555
559static inline void rtio_sqe_prep_transceive(struct rtio_sqe *sqe,
560 const struct rtio_iodev *iodev,
561 int8_t prio,
562 const uint8_t *tx_buf,
563 uint8_t *rx_buf,
564 uint32_t buf_len,
565 void *userdata)
566{
567 memset(sqe, 0, sizeof(struct rtio_sqe));
568 sqe->op = RTIO_OP_TXRX;
569 sqe->prio = prio;
570 sqe->iodev = iodev;
571 sqe->txrx.buf_len = buf_len;
572 sqe->txrx.tx_buf = tx_buf;
573 sqe->txrx.rx_buf = rx_buf;
574 sqe->userdata = userdata;
575}
576
591static inline void rtio_sqe_prep_await(struct rtio_sqe *sqe,
592 const struct rtio_iodev *iodev,
593 int8_t prio,
594 void *userdata)
595{
596 memset(sqe, 0, sizeof(struct rtio_sqe));
597 sqe->op = RTIO_OP_AWAIT;
598 sqe->prio = prio;
599 sqe->iodev = iodev;
600 sqe->userdata = userdata;
601}
602
614static inline void rtio_sqe_prep_await_iodev(struct rtio_sqe *sqe, const struct rtio_iodev *iodev,
615 int8_t prio, void *userdata)
616{
617 __ASSERT_NO_MSG(iodev != NULL);
618 rtio_sqe_prep_await(sqe, iodev, prio, userdata);
619}
620
631static inline void rtio_sqe_prep_await_executor(struct rtio_sqe *sqe, int8_t prio, void *userdata)
632{
633 rtio_sqe_prep_await(sqe, NULL, prio, userdata);
634}
635
647#ifdef CONFIG_RTIO_OP_DELAY
648static inline void rtio_sqe_prep_delay(struct rtio_sqe *sqe,
649 k_timeout_t timeout,
650 void *userdata)
651{
652 memset(sqe, 0, sizeof(struct rtio_sqe));
653 sqe->op = RTIO_OP_DELAY;
654 sqe->prio = 0;
655 sqe->iodev = NULL;
656 sqe->delay.timeout = timeout;
657 sqe->userdata = userdata;
658}
659#else
660#define rtio_sqe_prep_delay(sqe, timeout, userdata) \
661 BUILD_ASSERT(false, "CONFIG_RTIO_OP_DELAY not enabled")
662#endif
663
672static inline struct rtio_iodev_sqe *rtio_txn_next(const struct rtio_iodev_sqe *iodev_sqe)
673{
674 struct rtio_iodev_sqe *next = NULL;
675
676 SYS_PORT_TRACING_FUNC_ENTER(rtio, txn_next, iodev_sqe->r, iodev_sqe);
677 if (iodev_sqe->sqe.flags & RTIO_SQE_TRANSACTION) {
678 next = iodev_sqe->next;
679 }
680 SYS_PORT_TRACING_FUNC_EXIT(rtio, txn_next, iodev_sqe->r, next);
681 return next;
682}
683
684
693static inline struct rtio_iodev_sqe *rtio_chain_next(const struct rtio_iodev_sqe *iodev_sqe)
694{
695 struct rtio_iodev_sqe *next = NULL;
696
697 SYS_PORT_TRACING_FUNC_ENTER(rtio, txn_next, iodev_sqe->r, iodev_sqe);
698 if (iodev_sqe->sqe.flags & RTIO_SQE_CHAINED) {
699 next = iodev_sqe->next;
700 }
701 SYS_PORT_TRACING_FUNC_EXIT(rtio, txn_next, iodev_sqe->r, next);
702 return next;
703}
704
713static inline struct rtio_iodev_sqe *rtio_iodev_sqe_next(const struct rtio_iodev_sqe *iodev_sqe)
714{
715 return iodev_sqe->next;
716}
717
728static inline void rtio_iodev_sqe_await_signal(struct rtio_iodev_sqe *iodev_sqe,
729 rtio_signaled_t callback,
730 void *userdata)
731{
732 iodev_sqe->sqe.await.callback = callback;
733 iodev_sqe->sqe.await.userdata = userdata;
734
735 if (!atomic_cas(&iodev_sqe->sqe.await.ok, 0, 1)) {
736 callback(iodev_sqe, userdata);
737 }
738}
739
740/* Private structures and functions used for the pool of sqe structures */
742
743struct rtio_sqe_pool {
744 struct mpsc free_q;
745 const uint16_t pool_size;
746 uint16_t pool_free;
747 struct rtio_iodev_sqe *pool;
748};
749
750static inline struct rtio_iodev_sqe *rtio_sqe_pool_alloc(struct rtio_sqe_pool *pool)
751{
752 struct mpsc_node *node = mpsc_pop(&pool->free_q);
753
754 if (node == NULL) {
755 return NULL;
756 }
757
758 struct rtio_iodev_sqe *iodev_sqe = CONTAINER_OF(node, struct rtio_iodev_sqe, q);
759
760 pool->pool_free--;
761
762 return iodev_sqe;
763}
764
765static inline void rtio_sqe_pool_free(struct rtio_sqe_pool *pool, struct rtio_iodev_sqe *iodev_sqe)
766{
767 mpsc_push(&pool->free_q, &iodev_sqe->q);
768
769 pool->pool_free++;
770}
771
772
773/* Do not try and reformat the macros */
774/* clang-format off */
775
776#define Z_RTIO_SQE_POOL_DEFINE(name, sz) \
777 static struct rtio_iodev_sqe CONCAT(_sqe_pool_, name)[sz]; \
778 STRUCT_SECTION_ITERABLE(rtio_sqe_pool, name) = { \
779 .free_q = MPSC_INIT((name.free_q)), \
780 .pool_size = sz, \
781 .pool_free = sz, \
782 .pool = CONCAT(_sqe_pool_, name), \
783 }
784
785/* clang-format on */
787
791
792#ifdef __cplusplus
793}
794#endif
795
796#endif /* ZEPHYR_INCLUDE_RTIO_SQE_H_ */
workaround assembler barfing for ST r
Definition asm-macro-32-bit-gnu.h:24
long atomic_t
Definition atomic_types.h:15
_Bool atomic_cas(atomic_t *target, atomic_val_t old_value, atomic_val_t new_value)
Atomic compare-and-set.
static ALWAYS_INLINE void mpsc_push(struct mpsc *q, struct mpsc_node *n)
Push a node.
Definition mpsc_lockfree.h:126
static struct mpsc_node * mpsc_pop(struct mpsc *q)
Pop a node off of the list.
Definition mpsc_lockfree.h:145
#define RTIO_OP_CALLBACK
An operation that calls a given function (callback).
Definition sqe.h:153
#define RTIO_OP_TINY_TX
An operation that transmits tiny writes by copying the data to write.
Definition sqe.h:150
#define RTIO_OP_TX
An operation that transmits (writes).
Definition sqe.h:147
#define RTIO_OP_TXRX
An operation that transceives (reads and writes simultaneously).
Definition sqe.h:156
#define RTIO_OP_NOP
An operation that does nothing and will complete immediately.
Definition sqe.h:141
#define RTIO_OP_AWAIT
An operation to await a signal while blocking the iodev (if one is provided).
Definition sqe.h:177
#define RTIO_OP_DELAY
An operation that takes a specified amount of time (asynchronously) before completing.
Definition sqe.h:159
#define RTIO_OP_RX
An operation that receives (reads).
Definition sqe.h:144
#define RTIO_SQE_MULTISHOT
The SQE should continue producing CQEs until canceled.
Definition sqe.h:121
#define RTIO_SQE_TRANSACTION
The next request in the queue is part of a transaction.
Definition sqe.h:93
#define RTIO_SQE_MEMPOOL_BUFFER
The buffer should be allocated by the RTIO mempool.
Definition sqe.h:105
#define RTIO_SQE_NO_RESPONSE
The SQE does not produce a CQE.
Definition sqe.h:126
#define RTIO_SQE_CHAINED
The next request in the queue should wait on this one.
Definition sqe.h:81
static void rtio_sqe_prep_read_with_pool(struct rtio_sqe *sqe, const struct rtio_iodev *iodev, int8_t prio, void *userdata)
Prepare a read op submission with context's mempool.
Definition sqe.h:452
static void rtio_sqe_prep_tiny_write(struct rtio_sqe *sqe, const struct rtio_iodev *iodev, int8_t prio, const uint8_t *tiny_write_data, uint8_t tiny_write_len, void *userdata)
Prepare a tiny write op submission.
Definition sqe.h:497
static void rtio_sqe_prep_nop(struct rtio_sqe *sqe, const struct rtio_iodev *iodev, void *userdata)
Prepare a nop (no op) submission.
Definition sqe.h:418
void(* rtio_callback_t)(struct rtio *r, const struct rtio_sqe *sqe, int res, void *arg0)
Callback signature for RTIO_OP_CALLBACK.
Definition sqe.h:291
static void rtio_sqe_prep_await_iodev(struct rtio_sqe *sqe, const struct rtio_iodev *iodev, int8_t prio, void *userdata)
Prepare an await op submission which blocks an rtio_iodev until completion.
Definition sqe.h:614
static void rtio_iodev_sqe_await_signal(struct rtio_iodev_sqe *iodev_sqe, rtio_signaled_t callback, void *userdata)
Await an AWAIT SQE signal from RTIO IODEV.
Definition sqe.h:728
static void rtio_sqe_prep_write(struct rtio_sqe *sqe, const struct rtio_iodev *iodev, int8_t prio, const uint8_t *buf, uint32_t len, void *userdata)
Prepare a write op submission.
Definition sqe.h:471
static void rtio_sqe_prep_read(struct rtio_sqe *sqe, const struct rtio_iodev *iodev, int8_t prio, uint8_t *buf, uint32_t len, void *userdata)
Prepare a read op submission.
Definition sqe.h:431
static void rtio_sqe_prep_read_multishot(struct rtio_sqe *sqe, const struct rtio_iodev *iodev, int8_t prio, void *userdata)
Definition sqe.h:460
static void rtio_sqe_prep_callback(struct rtio_sqe *sqe, rtio_callback_t callback, void *arg0, void *userdata)
Prepare a callback op submission.
Definition sqe.h:523
static void rtio_sqe_prep_await_executor(struct rtio_sqe *sqe, int8_t prio, void *userdata)
Prepare an await op submission which completes the sqe after being signaled.
Definition sqe.h:631
void(* rtio_signaled_t)(struct rtio_iodev_sqe *iodev_sqe, void *userdata)
Callback signature for RTIO_OP_AWAIT signaled.
Definition sqe.h:299
static void rtio_sqe_prep_transceive(struct rtio_sqe *sqe, const struct rtio_iodev *iodev, int8_t prio, const uint8_t *tx_buf, uint8_t *rx_buf, uint32_t buf_len, void *userdata)
Prepare a transceive op submission.
Definition sqe.h:559
#define rtio_sqe_prep_delay(sqe, timeout, userdata)
Prepare a delay operation submission which completes after the given timeout.
Definition sqe.h:660
static struct rtio_iodev_sqe * rtio_chain_next(const struct rtio_iodev_sqe *iodev_sqe)
Get the next sqe in the chain.
Definition sqe.h:693
static struct rtio_iodev_sqe * rtio_iodev_sqe_next(const struct rtio_iodev_sqe *iodev_sqe)
Get the next sqe in the chain or transaction.
Definition sqe.h:713
static void rtio_sqe_prep_callback_no_cqe(struct rtio_sqe *sqe, rtio_callback_t callback, void *arg0, void *userdata)
Prepare a callback op submission that does not create a CQE.
Definition sqe.h:547
static struct rtio_iodev_sqe * rtio_txn_next(const struct rtio_iodev_sqe *iodev_sqe)
Get the next sqe in the transaction.
Definition sqe.h:672
static void rtio_sqe_prep_await(struct rtio_sqe *sqe, const struct rtio_iodev *iodev, int8_t prio, void *userdata)
Prepare an await op submission.
Definition sqe.h:591
#define SYS_PORT_TRACING_FUNC_ENTER(type, func,...)
Tracing macro for the entry into a function that might or might not return a value.
Definition tracing_macros.h:257
#define SYS_PORT_TRACING_FUNC_EXIT(type, func,...)
Tracing macro for when a function ends its execution.
Definition tracing_macros.h:283
#define CONTAINER_OF(ptr, type, field)
Get a pointer to a structure containing the element.
Definition util.h:281
#define NULL
Definition iar_missing_defs.h:20
#define BUILD_ASSERT(EXPR, MSG...)
Definition llvm.h:51
Public kernel APIs.
A wait-free intrusive multi producer single consumer (MPSC) queue using a singly linked list.
__UINT32_TYPE__ uint32_t
Definition stdint.h:90
__UINT8_TYPE__ uint8_t
Definition stdint.h:88
__UINT16_TYPE__ uint16_t
Definition stdint.h:89
__INT8_TYPE__ int8_t
Definition stdint.h:72
void * memset(void *buf, int c, size_t n)
void * memcpy(void *ZRESTRICT d, const void *ZRESTRICT s, size_t n)
Kernel timeout type.
Definition clock.h:65
Queue member.
Definition mpsc_lockfree.h:79
IO device submission queue entry.
Definition sqe.h:394
struct rtio_iodev_sqe * next
Definition sqe.h:397
struct rtio_sqe sqe
Definition sqe.h:395
struct rtio * r
Definition sqe.h:398
struct mpsc_node q
Definition sqe.h:396
An IO device with a function table for submitting requests.
Definition iodev.h:48
A submission queue event.
Definition sqe.h:304
uint32_t i2c_config
OP_I2C_CONFIGURE.
Definition sqe.h:366
void * userdata
User provided data which is returned upon operation completion.
Definition sqe.h:322
const uint8_t * tx_buf
Buffer to write from.
Definition sqe.h:353
uint8_t op
Op code.
Definition sqe.h:305
struct rtio_sqe::@126267262255374054123217063150244034155174062054::@222067021034074304061254367152327164076222165070 rx
OP_RX.
struct rtio_sqe::@126267262255374054123217063150244034155174062054::@036347012270137154075057170156115300365157200061 tiny_tx
OP_TINY_TX.
void * arg0
Last argument given to callback.
Definition sqe.h:347
atomic_t ok
Definition sqe.h:381
uint8_t * rx_buf
Buffer to read into.
Definition sqe.h:354
uint8_t prio
Op priority.
Definition sqe.h:307
struct rtio_sqe::@126267262255374054123217063150244034155174062054::@236333123355174166163204241333175337261032350217 await
OP_AWAIT.
uint32_t buf_len
Length of buffer.
Definition sqe.h:328
const struct rtio_iodev * iodev
Device to operation on.
Definition sqe.h:313
uint32_t iodev_flags
Op iodev flags.
Definition sqe.h:311
void * ccc_payload
OP_I3C_CCC.
Definition sqe.h:377
struct rtio_sqe::@126267262255374054123217063150244034155174062054::@366052340324346171254040100167341103343101366005 txrx
OP_TXRX.
int type
Definition sqe.h:371
uint16_t flags
Op Flags.
Definition sqe.h:309
const uint8_t * buf
Buffer to write from.
Definition sqe.h:329
void * config
Definition sqe.h:372
struct rtio_sqe::@126267262255374054123217063150244034155174062054::@344335000264002157174077227357232142206143107046 tx
OP_TX.
rtio_callback_t callback
Definition sqe.h:346
An RTIO context containing what can be viewed as a pair of queues.
Definition rtio.h:71
Misc utilities.