forked from Qortal/Brooklyn
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
187 lines
4.3 KiB
187 lines
4.3 KiB
#ifndef LIB_URING_H |
|
#define LIB_URING_H |
|
|
|
#ifdef __cplusplus |
|
extern "C" { |
|
#endif |
|
|
|
#include <sys/uio.h> |
|
#include <signal.h> |
|
#include <string.h> |
|
#include "../../include/uapi/linux/io_uring.h" |
|
#include <inttypes.h> |
|
#include <linux/swab.h> |
|
#include "barrier.h" |
|
|
|
/* |
|
* Library interface to io_uring |
|
*/ |
|
struct io_uring_sq { |
|
unsigned *khead; |
|
unsigned *ktail; |
|
unsigned *kring_mask; |
|
unsigned *kring_entries; |
|
unsigned *kflags; |
|
unsigned *kdropped; |
|
unsigned *array; |
|
struct io_uring_sqe *sqes; |
|
|
|
unsigned sqe_head; |
|
unsigned sqe_tail; |
|
|
|
size_t ring_sz; |
|
}; |
|
|
|
struct io_uring_cq { |
|
unsigned *khead; |
|
unsigned *ktail; |
|
unsigned *kring_mask; |
|
unsigned *kring_entries; |
|
unsigned *koverflow; |
|
struct io_uring_cqe *cqes; |
|
|
|
size_t ring_sz; |
|
}; |
|
|
|
struct io_uring { |
|
struct io_uring_sq sq; |
|
struct io_uring_cq cq; |
|
int ring_fd; |
|
}; |
|
|
|
/* |
|
* System calls |
|
*/ |
|
extern int io_uring_setup(unsigned entries, struct io_uring_params *p); |
|
extern int io_uring_enter(int fd, unsigned to_submit, |
|
unsigned min_complete, unsigned flags, sigset_t *sig); |
|
extern int io_uring_register(int fd, unsigned int opcode, void *arg, |
|
unsigned int nr_args); |
|
|
|
/* |
|
* Library interface |
|
*/ |
|
extern int io_uring_queue_init(unsigned entries, struct io_uring *ring, |
|
unsigned flags); |
|
extern int io_uring_queue_mmap(int fd, struct io_uring_params *p, |
|
struct io_uring *ring); |
|
extern void io_uring_queue_exit(struct io_uring *ring); |
|
extern int io_uring_peek_cqe(struct io_uring *ring, |
|
struct io_uring_cqe **cqe_ptr); |
|
extern int io_uring_wait_cqe(struct io_uring *ring, |
|
struct io_uring_cqe **cqe_ptr); |
|
extern int io_uring_submit(struct io_uring *ring); |
|
extern struct io_uring_sqe *io_uring_get_sqe(struct io_uring *ring); |
|
|
|
/* |
|
* Must be called after io_uring_{peek,wait}_cqe() after the cqe has |
|
* been processed by the application. |
|
*/ |
|
static inline void io_uring_cqe_seen(struct io_uring *ring, |
|
struct io_uring_cqe *cqe) |
|
{ |
|
if (cqe) { |
|
struct io_uring_cq *cq = &ring->cq; |
|
|
|
(*cq->khead)++; |
|
/* |
|
* Ensure that the kernel sees our new head, the kernel has |
|
* the matching read barrier. |
|
*/ |
|
write_barrier(); |
|
} |
|
} |
|
|
|
/* |
|
* Command prep helpers |
|
*/ |
|
static inline void io_uring_sqe_set_data(struct io_uring_sqe *sqe, void *data) |
|
{ |
|
sqe->user_data = (unsigned long) data; |
|
} |
|
|
|
static inline void *io_uring_cqe_get_data(struct io_uring_cqe *cqe) |
|
{ |
|
return (void *) (uintptr_t) cqe->user_data; |
|
} |
|
|
|
static inline void io_uring_prep_rw(int op, struct io_uring_sqe *sqe, int fd, |
|
const void *addr, unsigned len, |
|
off_t offset) |
|
{ |
|
memset(sqe, 0, sizeof(*sqe)); |
|
sqe->opcode = op; |
|
sqe->fd = fd; |
|
sqe->off = offset; |
|
sqe->addr = (unsigned long) addr; |
|
sqe->len = len; |
|
} |
|
|
|
static inline void io_uring_prep_readv(struct io_uring_sqe *sqe, int fd, |
|
const struct iovec *iovecs, |
|
unsigned nr_vecs, off_t offset) |
|
{ |
|
io_uring_prep_rw(IORING_OP_READV, sqe, fd, iovecs, nr_vecs, offset); |
|
} |
|
|
|
static inline void io_uring_prep_read_fixed(struct io_uring_sqe *sqe, int fd, |
|
void *buf, unsigned nbytes, |
|
off_t offset) |
|
{ |
|
io_uring_prep_rw(IORING_OP_READ_FIXED, sqe, fd, buf, nbytes, offset); |
|
} |
|
|
|
static inline void io_uring_prep_writev(struct io_uring_sqe *sqe, int fd, |
|
const struct iovec *iovecs, |
|
unsigned nr_vecs, off_t offset) |
|
{ |
|
io_uring_prep_rw(IORING_OP_WRITEV, sqe, fd, iovecs, nr_vecs, offset); |
|
} |
|
|
|
static inline void io_uring_prep_write_fixed(struct io_uring_sqe *sqe, int fd, |
|
const void *buf, unsigned nbytes, |
|
off_t offset) |
|
{ |
|
io_uring_prep_rw(IORING_OP_WRITE_FIXED, sqe, fd, buf, nbytes, offset); |
|
} |
|
|
|
static inline void io_uring_prep_poll_add(struct io_uring_sqe *sqe, int fd, |
|
unsigned poll_mask) |
|
{ |
|
memset(sqe, 0, sizeof(*sqe)); |
|
sqe->opcode = IORING_OP_POLL_ADD; |
|
sqe->fd = fd; |
|
#if __BYTE_ORDER == __BIG_ENDIAN |
|
poll_mask = __swahw32(poll_mask); |
|
#endif |
|
sqe->poll_events = poll_mask; |
|
} |
|
|
|
static inline void io_uring_prep_poll_remove(struct io_uring_sqe *sqe, |
|
void *user_data) |
|
{ |
|
memset(sqe, 0, sizeof(*sqe)); |
|
sqe->opcode = IORING_OP_POLL_REMOVE; |
|
sqe->addr = (unsigned long) user_data; |
|
} |
|
|
|
static inline void io_uring_prep_fsync(struct io_uring_sqe *sqe, int fd, |
|
unsigned fsync_flags) |
|
{ |
|
memset(sqe, 0, sizeof(*sqe)); |
|
sqe->opcode = IORING_OP_FSYNC; |
|
sqe->fd = fd; |
|
sqe->fsync_flags = fsync_flags; |
|
} |
|
|
|
static inline void io_uring_prep_nop(struct io_uring_sqe *sqe) |
|
{ |
|
memset(sqe, 0, sizeof(*sqe)); |
|
sqe->opcode = IORING_OP_NOP; |
|
} |
|
|
|
#ifdef __cplusplus |
|
} |
|
#endif |
|
|
|
#endif
|
|
|