mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-09-29 22:02:02 +00:00
io_uring: name rsrc bits consistently
Keep resource related structs' and functions' naming consistent, in particular use "io_rsrc" prefix for everything. Signed-off-by: Pavel Begunkov <asml.silence@gmail.com> Link: https://lore.kernel.org/r/962f5acdf810f3a62831e65da3932cde24f6d9df.1617287883.git.asml.silence@gmail.com Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
parent
c80ca4707d
commit
b895c9a632
1 changed files with 71 additions and 79 deletions
150
fs/io_uring.c
150
fs/io_uring.c
|
@ -219,22 +219,22 @@ struct fixed_rsrc_table {
|
|||
struct file **files;
|
||||
};
|
||||
|
||||
struct fixed_rsrc_ref_node {
|
||||
struct io_rsrc_node {
|
||||
struct percpu_ref refs;
|
||||
struct list_head node;
|
||||
struct list_head rsrc_list;
|
||||
struct fixed_rsrc_data *rsrc_data;
|
||||
struct io_rsrc_data *rsrc_data;
|
||||
void (*rsrc_put)(struct io_ring_ctx *ctx,
|
||||
struct io_rsrc_put *prsrc);
|
||||
struct llist_node llist;
|
||||
bool done;
|
||||
};
|
||||
|
||||
struct fixed_rsrc_data {
|
||||
struct io_rsrc_data {
|
||||
struct fixed_rsrc_table *table;
|
||||
struct io_ring_ctx *ctx;
|
||||
|
||||
struct fixed_rsrc_ref_node *node;
|
||||
struct io_rsrc_node *node;
|
||||
struct percpu_ref refs;
|
||||
struct completion done;
|
||||
bool quiesce;
|
||||
|
@ -393,7 +393,7 @@ struct io_ring_ctx {
|
|||
* readers must ensure that ->refs is alive as long as the file* is
|
||||
* used. Only updated through io_uring_register(2).
|
||||
*/
|
||||
struct fixed_rsrc_data *file_data;
|
||||
struct io_rsrc_data *file_data;
|
||||
unsigned nr_user_files;
|
||||
|
||||
/* if used, fixed mapped user buffers */
|
||||
|
@ -447,7 +447,7 @@ struct io_ring_ctx {
|
|||
struct llist_head rsrc_put_llist;
|
||||
struct list_head rsrc_ref_list;
|
||||
spinlock_t rsrc_ref_lock;
|
||||
struct fixed_rsrc_ref_node *rsrc_backup_node;
|
||||
struct io_rsrc_node *rsrc_backup_node;
|
||||
|
||||
struct io_restriction restrictions;
|
||||
|
||||
|
@ -1027,9 +1027,8 @@ static void io_uring_try_cancel_requests(struct io_ring_ctx *ctx,
|
|||
struct task_struct *task,
|
||||
struct files_struct *files);
|
||||
static void io_uring_cancel_sqpoll(struct io_ring_ctx *ctx);
|
||||
static void destroy_fixed_rsrc_ref_node(struct fixed_rsrc_ref_node *ref_node);
|
||||
static struct fixed_rsrc_ref_node *alloc_fixed_rsrc_ref_node(
|
||||
struct io_ring_ctx *ctx);
|
||||
static void io_rsrc_node_destroy(struct io_rsrc_node *ref_node);
|
||||
static struct io_rsrc_node *io_rsrc_node_alloc(struct io_ring_ctx *ctx);
|
||||
static void io_ring_file_put(struct io_ring_ctx *ctx, struct io_rsrc_put *prsrc);
|
||||
|
||||
static void io_cqring_fill_event(struct io_kiocb *req, long res);
|
||||
|
@ -1075,7 +1074,7 @@ EXPORT_SYMBOL(io_uring_get_socket);
|
|||
#define io_for_each_link(pos, head) \
|
||||
for (pos = (head); pos; pos = pos->link)
|
||||
|
||||
static inline void io_set_resource_node(struct io_kiocb *req)
|
||||
static inline void io_req_set_rsrc_node(struct io_kiocb *req)
|
||||
{
|
||||
struct io_ring_ctx *ctx = req->ctx;
|
||||
|
||||
|
@ -6287,7 +6286,7 @@ static void io_wq_submit_work(struct io_wq_work *work)
|
|||
#endif
|
||||
#define FFS_MASK ~(FFS_ASYNC_READ|FFS_ASYNC_WRITE|FFS_ISREG)
|
||||
|
||||
static inline struct file **io_fixed_file_slot(struct fixed_rsrc_data *file_data,
|
||||
static inline struct file **io_fixed_file_slot(struct io_rsrc_data *file_data,
|
||||
unsigned i)
|
||||
{
|
||||
struct fixed_rsrc_table *table;
|
||||
|
@ -6321,7 +6320,7 @@ static struct file *io_file_get(struct io_submit_state *state,
|
|||
file_ptr &= ~FFS_MASK;
|
||||
/* mask in overlapping REQ_F and FFS bits */
|
||||
req->flags |= (file_ptr << REQ_F_ASYNC_READ_BIT);
|
||||
io_set_resource_node(req);
|
||||
io_req_set_rsrc_node(req);
|
||||
} else {
|
||||
trace_io_uring_file_get(ctx, fd);
|
||||
file = __io_file_get(state, fd);
|
||||
|
@ -7078,9 +7077,8 @@ static void __io_sqe_files_unregister(struct io_ring_ctx *ctx)
|
|||
|
||||
static void io_rsrc_data_ref_zero(struct percpu_ref *ref)
|
||||
{
|
||||
struct fixed_rsrc_data *data;
|
||||
struct io_rsrc_data *data = container_of(ref, struct io_rsrc_data, refs);
|
||||
|
||||
data = container_of(ref, struct fixed_rsrc_data, refs);
|
||||
complete(&data->done);
|
||||
}
|
||||
|
||||
|
@ -7094,20 +7092,20 @@ static inline void io_rsrc_ref_unlock(struct io_ring_ctx *ctx)
|
|||
spin_unlock_bh(&ctx->rsrc_ref_lock);
|
||||
}
|
||||
|
||||
static void io_sqe_rsrc_set_node(struct io_ring_ctx *ctx,
|
||||
struct fixed_rsrc_data *rsrc_data,
|
||||
struct fixed_rsrc_ref_node *ref_node)
|
||||
static void io_rsrc_node_set(struct io_ring_ctx *ctx,
|
||||
struct io_rsrc_data *rsrc_data,
|
||||
struct io_rsrc_node *rsrc_node)
|
||||
{
|
||||
io_rsrc_ref_lock(ctx);
|
||||
rsrc_data->node = ref_node;
|
||||
list_add_tail(&ref_node->node, &ctx->rsrc_ref_list);
|
||||
rsrc_data->node = rsrc_node;
|
||||
list_add_tail(&rsrc_node->node, &ctx->rsrc_ref_list);
|
||||
io_rsrc_ref_unlock(ctx);
|
||||
percpu_ref_get(&rsrc_data->refs);
|
||||
}
|
||||
|
||||
static void io_sqe_rsrc_kill_node(struct io_ring_ctx *ctx, struct fixed_rsrc_data *data)
|
||||
static void io_rsrc_node_kill(struct io_ring_ctx *ctx, struct io_rsrc_data *data)
|
||||
{
|
||||
struct fixed_rsrc_ref_node *ref_node = NULL;
|
||||
struct io_rsrc_node *ref_node = NULL;
|
||||
|
||||
io_rsrc_ref_lock(ctx);
|
||||
ref_node = data->node;
|
||||
|
@ -7117,21 +7115,21 @@ static void io_sqe_rsrc_kill_node(struct io_ring_ctx *ctx, struct fixed_rsrc_dat
|
|||
percpu_ref_kill(&ref_node->refs);
|
||||
}
|
||||
|
||||
static int io_rsrc_refnode_prealloc(struct io_ring_ctx *ctx)
|
||||
static int io_rsrc_node_prealloc(struct io_ring_ctx *ctx)
|
||||
{
|
||||
if (ctx->rsrc_backup_node)
|
||||
return 0;
|
||||
ctx->rsrc_backup_node = alloc_fixed_rsrc_ref_node(ctx);
|
||||
ctx->rsrc_backup_node = io_rsrc_node_alloc(ctx);
|
||||
return ctx->rsrc_backup_node ? 0 : -ENOMEM;
|
||||
}
|
||||
|
||||
static struct fixed_rsrc_ref_node *
|
||||
io_rsrc_refnode_get(struct io_ring_ctx *ctx,
|
||||
struct fixed_rsrc_data *rsrc_data,
|
||||
void (*rsrc_put)(struct io_ring_ctx *ctx,
|
||||
struct io_rsrc_put *prsrc))
|
||||
static struct io_rsrc_node *
|
||||
io_rsrc_node_get(struct io_ring_ctx *ctx,
|
||||
struct io_rsrc_data *rsrc_data,
|
||||
void (*rsrc_put)(struct io_ring_ctx *ctx,
|
||||
struct io_rsrc_put *prsrc))
|
||||
{
|
||||
struct fixed_rsrc_ref_node *node = ctx->rsrc_backup_node;
|
||||
struct io_rsrc_node *node = ctx->rsrc_backup_node;
|
||||
|
||||
WARN_ON_ONCE(!node);
|
||||
|
||||
|
@ -7141,12 +7139,12 @@ io_rsrc_refnode_get(struct io_ring_ctx *ctx,
|
|||
return node;
|
||||
}
|
||||
|
||||
static int io_rsrc_ref_quiesce(struct fixed_rsrc_data *data,
|
||||
static int io_rsrc_ref_quiesce(struct io_rsrc_data *data,
|
||||
struct io_ring_ctx *ctx,
|
||||
void (*rsrc_put)(struct io_ring_ctx *ctx,
|
||||
struct io_rsrc_put *prsrc))
|
||||
{
|
||||
struct fixed_rsrc_ref_node *node;
|
||||
struct io_rsrc_node *node;
|
||||
int ret;
|
||||
|
||||
if (data->quiesce)
|
||||
|
@ -7154,10 +7152,10 @@ static int io_rsrc_ref_quiesce(struct fixed_rsrc_data *data,
|
|||
|
||||
data->quiesce = true;
|
||||
do {
|
||||
ret = io_rsrc_refnode_prealloc(ctx);
|
||||
ret = io_rsrc_node_prealloc(ctx);
|
||||
if (ret)
|
||||
break;
|
||||
io_sqe_rsrc_kill_node(ctx, data);
|
||||
io_rsrc_node_kill(ctx, data);
|
||||
percpu_ref_kill(&data->refs);
|
||||
flush_delayed_work(&ctx->rsrc_put_work);
|
||||
|
||||
|
@ -7166,8 +7164,8 @@ static int io_rsrc_ref_quiesce(struct fixed_rsrc_data *data,
|
|||
break;
|
||||
|
||||
percpu_ref_resurrect(&data->refs);
|
||||
node = io_rsrc_refnode_get(ctx, data, rsrc_put);
|
||||
io_sqe_rsrc_set_node(ctx, data, node);
|
||||
node = io_rsrc_node_get(ctx, data, rsrc_put);
|
||||
io_rsrc_node_set(ctx, data, node);
|
||||
reinit_completion(&data->done);
|
||||
|
||||
mutex_unlock(&ctx->uring_lock);
|
||||
|
@ -7179,9 +7177,9 @@ static int io_rsrc_ref_quiesce(struct fixed_rsrc_data *data,
|
|||
return ret;
|
||||
}
|
||||
|
||||
static struct fixed_rsrc_data *alloc_fixed_rsrc_data(struct io_ring_ctx *ctx)
|
||||
static struct io_rsrc_data *io_rsrc_data_alloc(struct io_ring_ctx *ctx)
|
||||
{
|
||||
struct fixed_rsrc_data *data;
|
||||
struct io_rsrc_data *data;
|
||||
|
||||
data = kzalloc(sizeof(*data), GFP_KERNEL);
|
||||
if (!data)
|
||||
|
@ -7197,7 +7195,7 @@ static struct fixed_rsrc_data *alloc_fixed_rsrc_data(struct io_ring_ctx *ctx)
|
|||
return data;
|
||||
}
|
||||
|
||||
static void free_fixed_rsrc_data(struct fixed_rsrc_data *data)
|
||||
static void io_rsrc_data_free(struct io_rsrc_data *data)
|
||||
{
|
||||
percpu_ref_exit(&data->refs);
|
||||
kfree(data->table);
|
||||
|
@ -7206,7 +7204,7 @@ static void free_fixed_rsrc_data(struct fixed_rsrc_data *data)
|
|||
|
||||
static int io_sqe_files_unregister(struct io_ring_ctx *ctx)
|
||||
{
|
||||
struct fixed_rsrc_data *data = ctx->file_data;
|
||||
struct io_rsrc_data *data = ctx->file_data;
|
||||
unsigned nr_tables, i;
|
||||
int ret;
|
||||
|
||||
|
@ -7225,7 +7223,7 @@ static int io_sqe_files_unregister(struct io_ring_ctx *ctx)
|
|||
nr_tables = DIV_ROUND_UP(ctx->nr_user_files, IORING_MAX_FILES_TABLE);
|
||||
for (i = 0; i < nr_tables; i++)
|
||||
kfree(data->table[i].files);
|
||||
free_fixed_rsrc_data(data);
|
||||
io_rsrc_data_free(data);
|
||||
ctx->file_data = NULL;
|
||||
ctx->nr_user_files = 0;
|
||||
return 0;
|
||||
|
@ -7454,7 +7452,7 @@ static int io_sqe_files_scm(struct io_ring_ctx *ctx)
|
|||
}
|
||||
#endif
|
||||
|
||||
static int io_sqe_alloc_file_tables(struct fixed_rsrc_data *file_data,
|
||||
static int io_sqe_alloc_file_tables(struct io_rsrc_data *file_data,
|
||||
unsigned nr_tables, unsigned nr_files)
|
||||
{
|
||||
int i;
|
||||
|
@ -7544,9 +7542,9 @@ static void io_ring_file_put(struct io_ring_ctx *ctx, struct io_rsrc_put *prsrc)
|
|||
#endif
|
||||
}
|
||||
|
||||
static void __io_rsrc_put_work(struct fixed_rsrc_ref_node *ref_node)
|
||||
static void __io_rsrc_put_work(struct io_rsrc_node *ref_node)
|
||||
{
|
||||
struct fixed_rsrc_data *rsrc_data = ref_node->rsrc_data;
|
||||
struct io_rsrc_data *rsrc_data = ref_node->rsrc_data;
|
||||
struct io_ring_ctx *ctx = rsrc_data->ctx;
|
||||
struct io_rsrc_put *prsrc, *tmp;
|
||||
|
||||
|
@ -7570,10 +7568,10 @@ static void io_rsrc_put_work(struct work_struct *work)
|
|||
node = llist_del_all(&ctx->rsrc_put_llist);
|
||||
|
||||
while (node) {
|
||||
struct fixed_rsrc_ref_node *ref_node;
|
||||
struct io_rsrc_node *ref_node;
|
||||
struct llist_node *next = node->next;
|
||||
|
||||
ref_node = llist_entry(node, struct fixed_rsrc_ref_node, llist);
|
||||
ref_node = llist_entry(node, struct io_rsrc_node, llist);
|
||||
__io_rsrc_put_work(ref_node);
|
||||
node = next;
|
||||
}
|
||||
|
@ -7581,27 +7579,23 @@ static void io_rsrc_put_work(struct work_struct *work)
|
|||
|
||||
static void io_rsrc_node_ref_zero(struct percpu_ref *ref)
|
||||
{
|
||||
struct fixed_rsrc_ref_node *ref_node;
|
||||
struct fixed_rsrc_data *data;
|
||||
struct io_ring_ctx *ctx;
|
||||
struct io_rsrc_node *node = container_of(ref, struct io_rsrc_node, refs);
|
||||
struct io_rsrc_data *data = node->rsrc_data;
|
||||
struct io_ring_ctx *ctx = data->ctx;
|
||||
bool first_add = false;
|
||||
int delay = HZ;
|
||||
|
||||
ref_node = container_of(ref, struct fixed_rsrc_ref_node, refs);
|
||||
data = ref_node->rsrc_data;
|
||||
ctx = data->ctx;
|
||||
|
||||
io_rsrc_ref_lock(ctx);
|
||||
ref_node->done = true;
|
||||
node->done = true;
|
||||
|
||||
while (!list_empty(&ctx->rsrc_ref_list)) {
|
||||
ref_node = list_first_entry(&ctx->rsrc_ref_list,
|
||||
struct fixed_rsrc_ref_node, node);
|
||||
node = list_first_entry(&ctx->rsrc_ref_list,
|
||||
struct io_rsrc_node, node);
|
||||
/* recycle ref nodes in order */
|
||||
if (!ref_node->done)
|
||||
if (!node->done)
|
||||
break;
|
||||
list_del(&ref_node->node);
|
||||
first_add |= llist_add(&ref_node->llist, &ctx->rsrc_put_llist);
|
||||
list_del(&node->node);
|
||||
first_add |= llist_add(&node->llist, &ctx->rsrc_put_llist);
|
||||
}
|
||||
io_rsrc_ref_unlock(ctx);
|
||||
|
||||
|
@ -7614,10 +7608,9 @@ static void io_rsrc_node_ref_zero(struct percpu_ref *ref)
|
|||
queue_delayed_work(system_wq, &ctx->rsrc_put_work, delay);
|
||||
}
|
||||
|
||||
static struct fixed_rsrc_ref_node *alloc_fixed_rsrc_ref_node(
|
||||
struct io_ring_ctx *ctx)
|
||||
static struct io_rsrc_node *io_rsrc_node_alloc(struct io_ring_ctx *ctx)
|
||||
{
|
||||
struct fixed_rsrc_ref_node *ref_node;
|
||||
struct io_rsrc_node *ref_node;
|
||||
|
||||
ref_node = kzalloc(sizeof(*ref_node), GFP_KERNEL);
|
||||
if (!ref_node)
|
||||
|
@ -7635,19 +7628,18 @@ static struct fixed_rsrc_ref_node *alloc_fixed_rsrc_ref_node(
|
|||
}
|
||||
|
||||
static void init_fixed_file_ref_node(struct io_ring_ctx *ctx,
|
||||
struct fixed_rsrc_ref_node *ref_node)
|
||||
struct io_rsrc_node *ref_node)
|
||||
{
|
||||
ref_node->rsrc_data = ctx->file_data;
|
||||
ref_node->rsrc_put = io_ring_file_put;
|
||||
}
|
||||
|
||||
static void destroy_fixed_rsrc_ref_node(struct fixed_rsrc_ref_node *ref_node)
|
||||
static void io_rsrc_node_destroy(struct io_rsrc_node *ref_node)
|
||||
{
|
||||
percpu_ref_exit(&ref_node->refs);
|
||||
kfree(ref_node);
|
||||
}
|
||||
|
||||
|
||||
static int io_sqe_files_register(struct io_ring_ctx *ctx, void __user *arg,
|
||||
unsigned nr_args)
|
||||
{
|
||||
|
@ -7655,8 +7647,8 @@ static int io_sqe_files_register(struct io_ring_ctx *ctx, void __user *arg,
|
|||
unsigned nr_tables, i;
|
||||
struct file *file;
|
||||
int fd, ret = -ENOMEM;
|
||||
struct fixed_rsrc_ref_node *ref_node;
|
||||
struct fixed_rsrc_data *file_data;
|
||||
struct io_rsrc_node *ref_node;
|
||||
struct io_rsrc_data *file_data;
|
||||
|
||||
if (ctx->file_data)
|
||||
return -EBUSY;
|
||||
|
@ -7665,7 +7657,7 @@ static int io_sqe_files_register(struct io_ring_ctx *ctx, void __user *arg,
|
|||
if (nr_args > IORING_MAX_FIXED_FILES)
|
||||
return -EMFILE;
|
||||
|
||||
file_data = alloc_fixed_rsrc_data(ctx);
|
||||
file_data = io_rsrc_data_alloc(ctx);
|
||||
if (!file_data)
|
||||
return -ENOMEM;
|
||||
ctx->file_data = file_data;
|
||||
|
@ -7722,14 +7714,14 @@ static int io_sqe_files_register(struct io_ring_ctx *ctx, void __user *arg,
|
|||
return ret;
|
||||
}
|
||||
|
||||
ref_node = alloc_fixed_rsrc_ref_node(ctx);
|
||||
ref_node = io_rsrc_node_alloc(ctx);
|
||||
if (!ref_node) {
|
||||
io_sqe_files_unregister(ctx);
|
||||
return -ENOMEM;
|
||||
}
|
||||
init_fixed_file_ref_node(ctx, ref_node);
|
||||
|
||||
io_sqe_rsrc_set_node(ctx, file_data, ref_node);
|
||||
io_rsrc_node_set(ctx, file_data, ref_node);
|
||||
return ret;
|
||||
out_fput:
|
||||
for (i = 0; i < ctx->nr_user_files; i++) {
|
||||
|
@ -7741,7 +7733,7 @@ static int io_sqe_files_register(struct io_ring_ctx *ctx, void __user *arg,
|
|||
kfree(file_data->table[i].files);
|
||||
ctx->nr_user_files = 0;
|
||||
out_free:
|
||||
free_fixed_rsrc_data(ctx->file_data);
|
||||
io_rsrc_data_free(ctx->file_data);
|
||||
ctx->file_data = NULL;
|
||||
return ret;
|
||||
}
|
||||
|
@ -7789,10 +7781,10 @@ static int io_sqe_file_register(struct io_ring_ctx *ctx, struct file *file,
|
|||
#endif
|
||||
}
|
||||
|
||||
static int io_queue_rsrc_removal(struct fixed_rsrc_data *data, void *rsrc)
|
||||
static int io_queue_rsrc_removal(struct io_rsrc_data *data, void *rsrc)
|
||||
{
|
||||
struct io_rsrc_put *prsrc;
|
||||
struct fixed_rsrc_ref_node *ref_node = data->node;
|
||||
struct io_rsrc_node *ref_node = data->node;
|
||||
|
||||
prsrc = kzalloc(sizeof(*prsrc), GFP_KERNEL);
|
||||
if (!prsrc)
|
||||
|
@ -7804,7 +7796,7 @@ static int io_queue_rsrc_removal(struct fixed_rsrc_data *data, void *rsrc)
|
|||
return 0;
|
||||
}
|
||||
|
||||
static inline int io_queue_file_removal(struct fixed_rsrc_data *data,
|
||||
static inline int io_queue_file_removal(struct io_rsrc_data *data,
|
||||
struct file *file)
|
||||
{
|
||||
return io_queue_rsrc_removal(data, (void *)file);
|
||||
|
@ -7814,8 +7806,8 @@ static int __io_sqe_files_update(struct io_ring_ctx *ctx,
|
|||
struct io_uring_rsrc_update *up,
|
||||
unsigned nr_args)
|
||||
{
|
||||
struct fixed_rsrc_data *data = ctx->file_data;
|
||||
struct fixed_rsrc_ref_node *ref_node;
|
||||
struct io_rsrc_data *data = ctx->file_data;
|
||||
struct io_rsrc_node *ref_node;
|
||||
struct file *file, **file_slot;
|
||||
__s32 __user *fds;
|
||||
int fd, i, err;
|
||||
|
@ -7826,7 +7818,7 @@ static int __io_sqe_files_update(struct io_ring_ctx *ctx,
|
|||
return -EOVERFLOW;
|
||||
if (done > ctx->nr_user_files)
|
||||
return -EINVAL;
|
||||
err = io_rsrc_refnode_prealloc(ctx);
|
||||
err = io_rsrc_node_prealloc(ctx);
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
|
@ -7882,8 +7874,8 @@ static int __io_sqe_files_update(struct io_ring_ctx *ctx,
|
|||
|
||||
if (needs_switch) {
|
||||
percpu_ref_kill(&data->node->refs);
|
||||
ref_node = io_rsrc_refnode_get(ctx, data, io_ring_file_put);
|
||||
io_sqe_rsrc_set_node(ctx, data, ref_node);
|
||||
ref_node = io_rsrc_node_get(ctx, data, io_ring_file_put);
|
||||
io_rsrc_node_set(ctx, data, ref_node);
|
||||
}
|
||||
return done ? done : err;
|
||||
}
|
||||
|
@ -8559,7 +8551,7 @@ static void io_ring_ctx_free(struct io_ring_ctx *ctx)
|
|||
io_destroy_buffers(ctx);
|
||||
|
||||
if (ctx->rsrc_backup_node)
|
||||
destroy_fixed_rsrc_ref_node(ctx->rsrc_backup_node);
|
||||
io_rsrc_node_destroy(ctx->rsrc_backup_node);
|
||||
|
||||
#if defined(CONFIG_UNIX)
|
||||
if (ctx->ring_sock) {
|
||||
|
|
Loading…
Reference in a new issue