1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/kernel.h>
3 #include <linux/errno.h>
5 #include <linux/slab.h>
6 #include <linux/eventfd.h>
7 #include <linux/eventpoll.h>
8 #include <linux/io_uring.h>
9 #include <linux/io_uring_types.h>
15 struct eventfd_ctx
*cq_ev_fd
;
16 unsigned int eventfd_async
: 1;
23 IO_EVENTFD_OP_SIGNAL_BIT
,
26 static void io_eventfd_free(struct rcu_head
*rcu
)
28 struct io_ev_fd
*ev_fd
= container_of(rcu
, struct io_ev_fd
, rcu
);
30 eventfd_ctx_put(ev_fd
->cq_ev_fd
);
34 static void io_eventfd_do_signal(struct rcu_head
*rcu
)
36 struct io_ev_fd
*ev_fd
= container_of(rcu
, struct io_ev_fd
, rcu
);
38 eventfd_signal_mask(ev_fd
->cq_ev_fd
, EPOLL_URING_WAKE
);
40 if (refcount_dec_and_test(&ev_fd
->refs
))
44 void io_eventfd_signal(struct io_ring_ctx
*ctx
)
46 struct io_ev_fd
*ev_fd
= NULL
;
48 if (READ_ONCE(ctx
->rings
->cq_flags
) & IORING_CQ_EVENTFD_DISABLED
)
54 * rcu_dereference ctx->io_ev_fd once and use it for both for checking
57 ev_fd
= rcu_dereference(ctx
->io_ev_fd
);
60 * Check again if ev_fd exists incase an io_eventfd_unregister call
61 * completed between the NULL check of ctx->io_ev_fd at the start of
62 * the function and rcu_read_lock.
66 if (!refcount_inc_not_zero(&ev_fd
->refs
))
68 if (ev_fd
->eventfd_async
&& !io_wq_current_is_worker())
71 if (likely(eventfd_signal_allowed())) {
72 eventfd_signal_mask(ev_fd
->cq_ev_fd
, EPOLL_URING_WAKE
);
74 if (!atomic_fetch_or(BIT(IO_EVENTFD_OP_SIGNAL_BIT
), &ev_fd
->ops
)) {
75 call_rcu_hurry(&ev_fd
->rcu
, io_eventfd_do_signal
);
80 if (refcount_dec_and_test(&ev_fd
->refs
))
81 call_rcu(&ev_fd
->rcu
, io_eventfd_free
);
84 void io_eventfd_flush_signal(struct io_ring_ctx
*ctx
)
88 spin_lock(&ctx
->completion_lock
);
91 * Eventfd should only get triggered when at least one event has been
92 * posted. Some applications rely on the eventfd notification count
93 * only changing IFF a new CQE has been added to the CQ ring. There's
94 * no depedency on 1:1 relationship between how many times this
95 * function is called (and hence the eventfd count) and number of CQEs
96 * posted to the CQ ring.
98 skip
= ctx
->cached_cq_tail
== ctx
->evfd_last_cq_tail
;
99 ctx
->evfd_last_cq_tail
= ctx
->cached_cq_tail
;
100 spin_unlock(&ctx
->completion_lock
);
104 io_eventfd_signal(ctx
);
107 int io_eventfd_register(struct io_ring_ctx
*ctx
, void __user
*arg
,
108 unsigned int eventfd_async
)
110 struct io_ev_fd
*ev_fd
;
111 __s32 __user
*fds
= arg
;
114 ev_fd
= rcu_dereference_protected(ctx
->io_ev_fd
,
115 lockdep_is_held(&ctx
->uring_lock
));
119 if (copy_from_user(&fd
, fds
, sizeof(*fds
)))
122 ev_fd
= kmalloc(sizeof(*ev_fd
), GFP_KERNEL
);
126 ev_fd
->cq_ev_fd
= eventfd_ctx_fdget(fd
);
127 if (IS_ERR(ev_fd
->cq_ev_fd
)) {
128 int ret
= PTR_ERR(ev_fd
->cq_ev_fd
);
134 spin_lock(&ctx
->completion_lock
);
135 ctx
->evfd_last_cq_tail
= ctx
->cached_cq_tail
;
136 spin_unlock(&ctx
->completion_lock
);
138 ev_fd
->eventfd_async
= eventfd_async
;
139 ctx
->has_evfd
= true;
140 refcount_set(&ev_fd
->refs
, 1);
141 atomic_set(&ev_fd
->ops
, 0);
142 rcu_assign_pointer(ctx
->io_ev_fd
, ev_fd
);
146 int io_eventfd_unregister(struct io_ring_ctx
*ctx
)
148 struct io_ev_fd
*ev_fd
;
150 ev_fd
= rcu_dereference_protected(ctx
->io_ev_fd
,
151 lockdep_is_held(&ctx
->uring_lock
));
153 ctx
->has_evfd
= false;
154 rcu_assign_pointer(ctx
->io_ev_fd
, NULL
);
155 if (refcount_dec_and_test(&ev_fd
->refs
))
156 call_rcu(&ev_fd
->rcu
, io_eventfd_free
);