@@ -1053,9 +1053,12 @@ static inline uint32_t rtio_sqe_acquirable(struct rtio *r)
10531053 */
10541054static inline struct rtio_iodev_sqe * rtio_txn_next (const struct rtio_iodev_sqe * iodev_sqe )
10551055{
1056+ SYS_PORT_TRACING_FUNC_ENTER (rtio , txn_next , iodev_sqe -> r , iodev_sqe );
10561057if (iodev_sqe -> sqe .flags & RTIO_SQE_TRANSACTION ) {
1058+ SYS_PORT_TRACING_FUNC_EXIT (rtio , txn_next , iodev_sqe -> r , iodev_sqe -> next );
10571059return iodev_sqe -> next ;
10581060} else {
1061+ SYS_PORT_TRACING_FUNC_EXIT (rtio , txn_next , iodev_sqe -> r , NULL );
10591062return NULL ;
10601063}
10611064}
@@ -1071,9 +1074,12 @@ static inline struct rtio_iodev_sqe *rtio_txn_next(const struct rtio_iodev_sqe *
10711074 */
10721075static inline struct rtio_iodev_sqe * rtio_chain_next (const struct rtio_iodev_sqe * iodev_sqe )
10731076{
1077+ SYS_PORT_TRACING_FUNC_ENTER (rtio , txn_next , iodev_sqe -> r , iodev_sqe );
10741078if (iodev_sqe -> sqe .flags & RTIO_SQE_CHAINED ) {
1079+ SYS_PORT_TRACING_FUNC_EXIT (rtio , txn_next , iodev_sqe -> r , iodev_sqe -> next );
10751080return iodev_sqe -> next ;
10761081} else {
1082+ SYS_PORT_TRACING_FUNC_EXIT (rtio , txn_next , iodev_sqe -> r , NULL );
10771083return NULL ;
10781084}
10791085}
@@ -1101,14 +1107,17 @@ static inline struct rtio_iodev_sqe *rtio_iodev_sqe_next(const struct rtio_iodev
11011107 */
11021108static inline struct rtio_sqe * rtio_sqe_acquire (struct rtio * r )
11031109{
1110+ SYS_PORT_TRACING_FUNC_ENTER (rtio , sqe_acquire , r );
11041111struct rtio_iodev_sqe * iodev_sqe = rtio_sqe_pool_alloc (r -> sqe_pool );
11051112
11061113if (iodev_sqe == NULL ) {
1114+ SYS_PORT_TRACING_FUNC_EXIT (rtio , sqe_acquire , r , NULL );
11071115return NULL ;
11081116}
11091117
11101118mpsc_push (& r -> sq , & iodev_sqe -> q );
11111119
1120+ SYS_PORT_TRACING_FUNC_EXIT (rtio , sqe_acquire , r , & iodev_sqe -> sqe );
11121121return & iodev_sqe -> sqe ;
11131122}
11141123
@@ -1134,14 +1143,17 @@ static inline void rtio_sqe_drop_all(struct rtio *r)
11341143 */
11351144static inline struct rtio_cqe * rtio_cqe_acquire (struct rtio * r )
11361145{
1146+ SYS_PORT_TRACING_FUNC_ENTER (rtio , cqe_acquire , r );
11371147struct rtio_cqe * cqe = rtio_cqe_pool_alloc (r -> cqe_pool );
11381148
11391149if (cqe == NULL ) {
1150+ SYS_PORT_TRACING_FUNC_EXIT (rtio , cqe_acquire , r , NULL );
11401151return NULL ;
11411152}
11421153
11431154memset (cqe , 0 , sizeof (struct rtio_cqe ));
11441155
1156+ SYS_PORT_TRACING_FUNC_EXIT (rtio , cqe_acquire , r , cqe );
11451157return cqe ;
11461158}
11471159
@@ -1166,21 +1178,25 @@ static inline void rtio_cqe_produce(struct rtio *r, struct rtio_cqe *cqe)
11661178 */
11671179static inline struct rtio_cqe * rtio_cqe_consume (struct rtio * r )
11681180{
1181+ SYS_PORT_TRACING_FUNC_ENTER (rtio , cqe_consume , r );
11691182struct mpsc_node * node ;
11701183struct rtio_cqe * cqe = NULL ;
11711184
11721185#ifdef CONFIG_RTIO_CONSUME_SEM
11731186if (k_sem_take (r -> consume_sem , K_NO_WAIT ) != 0 ) {
1187+ SYS_PORT_TRACING_FUNC_EXIT (rtio , cqe_consume , r , NULL );
11741188return NULL ;
11751189}
11761190#endif
11771191
11781192node = mpsc_pop (& r -> cq );
11791193if (node == NULL ) {
1194+ SYS_PORT_TRACING_FUNC_EXIT (rtio , cqe_consume , r , NULL );
11801195return NULL ;
11811196}
11821197cqe = CONTAINER_OF (node , struct rtio_cqe , q );
11831198
1199+ SYS_PORT_TRACING_FUNC_EXIT (rtio , cqe_consume , r , cqe );
11841200return cqe ;
11851201}
11861202
@@ -1220,6 +1236,7 @@ static inline struct rtio_cqe *rtio_cqe_consume_block(struct rtio *r)
12201236 */
12211237static inline void rtio_cqe_release (struct rtio * r , struct rtio_cqe * cqe )
12221238{
1239+ SYS_PORT_TRACING_FUNC (rtio , cqe_release , r , cqe );
12231240rtio_cqe_pool_free (r -> cqe_pool , cqe );
12241241}
12251242
@@ -1374,6 +1391,7 @@ static inline void rtio_iodev_sqe_err(struct rtio_iodev_sqe *iodev_sqe, int resu
13741391 */
13751392static inline void rtio_cqe_submit (struct rtio * r , int result , void * userdata , uint32_t flags )
13761393{
1394+ SYS_PORT_TRACING_FUNC_ENTER (rtio , cqe_submit , r , result , flags );
13771395struct rtio_cqe * cqe = rtio_cqe_acquire (r );
13781396
13791397if (cqe == NULL ) {
@@ -1406,6 +1424,7 @@ static inline void rtio_cqe_submit(struct rtio *r, int result, void *userdata, u
14061424}
14071425}
14081426#endif
1427+ SYS_PORT_TRACING_FUNC_EXIT (rtio , cqe_submit , r );
14091428}
14101429
14111430#define __RTIO_MEMPOOL_GET_NUM_BLKS (num_bytes , blk_size ) (((num_bytes) + (blk_size)-1) / (blk_size))
@@ -1546,6 +1565,7 @@ __syscall int rtio_sqe_cancel(struct rtio_sqe *sqe);
15461565
15471566static inline int z_impl_rtio_sqe_cancel (struct rtio_sqe * sqe )
15481567{
1568+ SYS_PORT_TRACING_FUNC (rtio , sqe_cancel , sqe );
15491569struct rtio_iodev_sqe * iodev_sqe = CONTAINER_OF (sqe , struct rtio_iodev_sqe , sqe );
15501570
15511571do {
@@ -1724,6 +1744,7 @@ __syscall int rtio_submit(struct rtio *r, uint32_t wait_count);
17241744#ifdef CONFIG_RTIO_SUBMIT_SEM
17251745static inline int z_impl_rtio_submit (struct rtio * r , uint32_t wait_count )
17261746{
1747+ SYS_PORT_TRACING_FUNC_ENTER (rtio , submit , r , wait_count );
17271748int res = 0 ;
17281749
17291750if (wait_count > 0 ) {
@@ -1742,12 +1763,14 @@ static inline int z_impl_rtio_submit(struct rtio *r, uint32_t wait_count)
17421763 "semaphore was reset or timed out while waiting on completions!" );
17431764}
17441765
1766+ SYS_PORT_TRACING_FUNC_EXIT (rtio , submit , r );
17451767return res ;
17461768}
17471769#else
17481770static inline int z_impl_rtio_submit (struct rtio * r , uint32_t wait_count )
17491771{
17501772
1773+ SYS_PORT_TRACING_FUNC_ENTER (rtio , submit , rtio , wait_count );
17511774int res = 0 ;
17521775uintptr_t cq_count = (uintptr_t )atomic_get (& r -> cq_count );
17531776uintptr_t cq_complete_count = cq_count + wait_count ;
@@ -1767,6 +1790,7 @@ static inline int z_impl_rtio_submit(struct rtio *r, uint32_t wait_count)
17671790k_yield ();
17681791}
17691792
1793+ SYS_PORT_TRACING_FUNC_EXIT (rtio , submit , r );
17701794return res ;
17711795}
17721796#endif /* CONFIG_RTIO_SUBMIT_SEM */
0 commit comments