return 0;
}
-int io_uring_fill_params(unsigned entries, struct io_uring_params *p)
+int io_uring_fill_params(struct io_uring_params *p)
{
+ unsigned entries = p->sq_entries;
+
if (!entries)
return -EINVAL;
if (entries > IORING_MAX_ENTRIES) {
return 0;
}
-static __cold int io_uring_create(unsigned entries, struct io_uring_params *p,
+static __cold int io_uring_create(struct io_uring_params *p,
struct io_uring_params __user *params)
{
struct io_ring_ctx *ctx;
if (ret)
return ret;
- ret = io_uring_fill_params(entries, p);
+ ret = io_uring_fill_params(p);
if (unlikely(ret))
return ret;
if (p.flags & ~IORING_SETUP_FLAGS)
return -EINVAL;
- return io_uring_create(entries, &p, params);
+ p.sq_entries = entries;
+ return io_uring_create(&p, params);
}
static inline int io_uring_allowed(void)
unsigned long rings_size(unsigned int flags, unsigned int sq_entries,
unsigned int cq_entries, size_t *sq_offset);
-int io_uring_fill_params(unsigned entries, struct io_uring_params *p);
+int io_uring_fill_params(struct io_uring_params *p);
bool io_cqe_cache_refill(struct io_ring_ctx *ctx, bool overflow, bool cqe32);
int io_run_task_work_sig(struct io_ring_ctx *ctx);
int io_run_local_work(struct io_ring_ctx *ctx, int min_events, int max_events);
/* properties that are always inherited */
p.flags |= (ctx->flags & COPY_FLAGS);
- ret = io_uring_fill_params(p.sq_entries, &p);
+ ret = io_uring_fill_params(&p);
if (unlikely(ret))
return ret;