1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/kernel.h>
3 #include <linux/errno.h>
5 #include <linux/slab.h>
6 #include <linux/eventfd.h>
7 #include <linux/eventpoll.h>
8 #include <linux/io_uring.h>
9 #include <linux/io_uring_types.h>
15 struct eventfd_ctx
*cq_ev_fd
;
16 unsigned int eventfd_async
;
17 /* protected by ->completion_lock */
18 unsigned last_cq_tail
;
25 IO_EVENTFD_OP_SIGNAL_BIT
,
28 static void io_eventfd_free(struct rcu_head
*rcu
)
30 struct io_ev_fd
*ev_fd
= container_of(rcu
, struct io_ev_fd
, rcu
);
32 eventfd_ctx_put(ev_fd
->cq_ev_fd
);
36 static void io_eventfd_do_signal(struct rcu_head
*rcu
)
38 struct io_ev_fd
*ev_fd
= container_of(rcu
, struct io_ev_fd
, rcu
);
40 eventfd_signal_mask(ev_fd
->cq_ev_fd
, EPOLL_URING_WAKE
);
42 if (refcount_dec_and_test(&ev_fd
->refs
))
46 static void io_eventfd_put(struct io_ev_fd
*ev_fd
)
48 if (refcount_dec_and_test(&ev_fd
->refs
))
49 call_rcu(&ev_fd
->rcu
, io_eventfd_free
);
52 static void io_eventfd_release(struct io_ev_fd
*ev_fd
, bool put_ref
)
55 io_eventfd_put(ev_fd
);
60 * Returns true if the caller should put the ev_fd reference, false if not.
62 static bool __io_eventfd_signal(struct io_ev_fd
*ev_fd
)
64 if (eventfd_signal_allowed()) {
65 eventfd_signal_mask(ev_fd
->cq_ev_fd
, EPOLL_URING_WAKE
);
68 if (!atomic_fetch_or(BIT(IO_EVENTFD_OP_SIGNAL_BIT
), &ev_fd
->ops
)) {
69 call_rcu_hurry(&ev_fd
->rcu
, io_eventfd_do_signal
);
76 * Trigger if eventfd_async isn't set, or if it's set and the caller is
77 * an async worker. If ev_fd isn't valid, obviously return false.
79 static bool io_eventfd_trigger(struct io_ev_fd
*ev_fd
)
82 return !ev_fd
->eventfd_async
|| io_wq_current_is_worker();
87 * On success, returns with an ev_fd reference grabbed and the RCU read
90 static struct io_ev_fd
*io_eventfd_grab(struct io_ring_ctx
*ctx
)
92 struct io_ev_fd
*ev_fd
;
94 if (READ_ONCE(ctx
->rings
->cq_flags
) & IORING_CQ_EVENTFD_DISABLED
)
100 * rcu_dereference ctx->io_ev_fd once and use it for both for checking
103 ev_fd
= rcu_dereference(ctx
->io_ev_fd
);
106 * Check again if ev_fd exists in case an io_eventfd_unregister call
107 * completed between the NULL check of ctx->io_ev_fd at the start of
108 * the function and rcu_read_lock.
110 if (io_eventfd_trigger(ev_fd
) && refcount_inc_not_zero(&ev_fd
->refs
))
117 void io_eventfd_signal(struct io_ring_ctx
*ctx
)
119 struct io_ev_fd
*ev_fd
;
121 ev_fd
= io_eventfd_grab(ctx
);
123 io_eventfd_release(ev_fd
, __io_eventfd_signal(ev_fd
));
126 void io_eventfd_flush_signal(struct io_ring_ctx
*ctx
)
128 struct io_ev_fd
*ev_fd
;
130 ev_fd
= io_eventfd_grab(ctx
);
132 bool skip
, put_ref
= true;
135 * Eventfd should only get triggered when at least one event
136 * has been posted. Some applications rely on the eventfd
137 * notification count only changing IFF a new CQE has been
138 * added to the CQ ring. There's no dependency on 1:1
139 * relationship between how many times this function is called
140 * (and hence the eventfd count) and number of CQEs posted to
143 spin_lock(&ctx
->completion_lock
);
144 skip
= ctx
->cached_cq_tail
== ev_fd
->last_cq_tail
;
145 ev_fd
->last_cq_tail
= ctx
->cached_cq_tail
;
146 spin_unlock(&ctx
->completion_lock
);
149 put_ref
= __io_eventfd_signal(ev_fd
);
151 io_eventfd_release(ev_fd
, put_ref
);
155 int io_eventfd_register(struct io_ring_ctx
*ctx
, void __user
*arg
,
156 unsigned int eventfd_async
)
158 struct io_ev_fd
*ev_fd
;
159 __s32 __user
*fds
= arg
;
162 ev_fd
= rcu_dereference_protected(ctx
->io_ev_fd
,
163 lockdep_is_held(&ctx
->uring_lock
));
167 if (copy_from_user(&fd
, fds
, sizeof(*fds
)))
170 ev_fd
= kmalloc(sizeof(*ev_fd
), GFP_KERNEL
);
174 ev_fd
->cq_ev_fd
= eventfd_ctx_fdget(fd
);
175 if (IS_ERR(ev_fd
->cq_ev_fd
)) {
176 int ret
= PTR_ERR(ev_fd
->cq_ev_fd
);
182 spin_lock(&ctx
->completion_lock
);
183 ev_fd
->last_cq_tail
= ctx
->cached_cq_tail
;
184 spin_unlock(&ctx
->completion_lock
);
186 ev_fd
->eventfd_async
= eventfd_async
;
187 ctx
->has_evfd
= true;
188 refcount_set(&ev_fd
->refs
, 1);
189 atomic_set(&ev_fd
->ops
, 0);
190 rcu_assign_pointer(ctx
->io_ev_fd
, ev_fd
);
194 int io_eventfd_unregister(struct io_ring_ctx
*ctx
)
196 struct io_ev_fd
*ev_fd
;
198 ev_fd
= rcu_dereference_protected(ctx
->io_ev_fd
,
199 lockdep_is_held(&ctx
->uring_lock
));
201 ctx
->has_evfd
= false;
202 rcu_assign_pointer(ctx
->io_ev_fd
, NULL
);
203 io_eventfd_put(ev_fd
);