1 /*-------------------------------------------------------------------------
4 * Routines for inter-process latches
6 * The poll() implementation uses the so-called self-pipe trick to overcome the
7 * race condition involved with poll() and setting a global flag in the signal
8 * handler. When a latch is set and the current process is waiting for it, the
9 * signal handler wakes up the poll() in WaitLatch by writing a byte to a pipe.
10 * A signal by itself doesn't interrupt poll() on all platforms, and even on
11 * platforms where it does, a signal that arrives just before the poll() call
12 * does not prevent poll() from entering sleep. An incoming byte on a pipe
13 * however reliably interrupts the sleep, and causes poll() to return
14 * immediately even if the signal arrives before poll() begins.
16 * The epoll() implementation overcomes the race with a different technique: it
17 * keeps SIGURG blocked and consumes from a signalfd() descriptor instead. We
18 * don't need to register a signal handler or create our own self-pipe. We
19 * assume that any system that has Linux epoll() also has Linux signalfd().
21 * The kqueue() implementation waits for SIGURG with EVFILT_SIGNAL.
23 * The Windows implementation uses Windows events that are inherited by all
24 * postmaster child processes. There's no need for the self-pipe trick there.
26 * Portions Copyright (c) 1996-2025, PostgreSQL Global Development Group
27 * Portions Copyright (c) 1994, Regents of the University of California
30 * src/backend/storage/ipc/latch.c
32 *-------------------------------------------------------------------------
40 #ifdef HAVE_SYS_EPOLL_H
41 #include <sys/epoll.h>
43 #ifdef HAVE_SYS_EVENT_H
44 #include <sys/event.h>
46 #ifdef HAVE_SYS_SIGNALFD_H
47 #include <sys/signalfd.h>
53 #include "libpq/pqsignal.h"
54 #include "miscadmin.h"
56 #include "port/atomics.h"
57 #include "portability/instr_time.h"
58 #include "postmaster/postmaster.h"
59 #include "storage/fd.h"
60 #include "storage/ipc.h"
61 #include "storage/latch.h"
62 #include "storage/pmsignal.h"
63 #include "utils/memutils.h"
64 #include "utils/resowner.h"
67 * Select the fd readiness primitive to use. Normally the "most modern"
68 * primitive supported by the OS will be used, but for testing it can be
69 * useful to manually specify the used primitive. If desired, just add a
70 * define somewhere before this block.
72 #if defined(WAIT_USE_EPOLL) || defined(WAIT_USE_POLL) || \
73 defined(WAIT_USE_KQUEUE) || defined(WAIT_USE_WIN32)
74 /* don't overwrite manual choice */
75 #elif defined(HAVE_SYS_EPOLL_H)
76 #define WAIT_USE_EPOLL
77 #elif defined(HAVE_KQUEUE)
78 #define WAIT_USE_KQUEUE
79 #elif defined(HAVE_POLL)
82 #define WAIT_USE_WIN32
84 #error "no wait set implementation available"
88 * By default, we use a self-pipe with poll() and a signalfd with epoll(), if
89 * available. For testing the choice can also be manually specified.
91 #if defined(WAIT_USE_POLL) || defined(WAIT_USE_EPOLL)
92 #if defined(WAIT_USE_SELF_PIPE) || defined(WAIT_USE_SIGNALFD)
93 /* don't overwrite manual choice */
94 #elif defined(WAIT_USE_EPOLL) && defined(HAVE_SYS_SIGNALFD_H)
95 #define WAIT_USE_SIGNALFD
97 #define WAIT_USE_SELF_PIPE
101 /* typedef in latch.h */
106 int nevents
; /* number of registered events */
107 int nevents_space
; /* maximum number of events in this set */
110 * Array, of nevents_space length, storing the definition of events this
111 * set is waiting for.
116 * If WL_LATCH_SET is specified in any wait event, latch is a pointer to
117 * said latch, and latch_pos the offset in the ->events array. This is
118 * useful because we check the state of the latch before performing doing
119 * syscalls related to waiting.
125 * WL_EXIT_ON_PM_DEATH is converted to WL_POSTMASTER_DEATH, but this flag
126 * is set so that we'll exit immediately if postmaster death is detected,
127 * instead of returning.
129 bool exit_on_postmaster_death
;
131 #if defined(WAIT_USE_EPOLL)
133 /* epoll_wait returns events in a user provided arrays, allocate once */
134 struct epoll_event
*epoll_ret_events
;
135 #elif defined(WAIT_USE_KQUEUE)
137 /* kevent returns events in a user provided arrays, allocate once */
138 struct kevent
*kqueue_ret_events
;
139 bool report_postmaster_not_running
;
140 #elif defined(WAIT_USE_POLL)
141 /* poll expects events to be waited on every poll() call, prepare once */
142 struct pollfd
*pollfds
;
143 #elif defined(WAIT_USE_WIN32)
146 * Array of windows events. The first element always contains
147 * pgwin32_signal_event, so the remaining elements are offset by one (i.e.
154 /* A common WaitEventSet used to implement WaitLatch() */
155 static WaitEventSet
*LatchWaitSet
;
157 /* The position of the latch in LatchWaitSet. */
158 #define LatchWaitSetLatchPos 0
161 /* Are we currently in WaitLatch? The signal handler would like to know. */
162 static volatile sig_atomic_t waiting
= false;
165 #ifdef WAIT_USE_SIGNALFD
166 /* On Linux, we'll receive SIGURG via a signalfd file descriptor. */
167 static int signal_fd
= -1;
170 #ifdef WAIT_USE_SELF_PIPE
171 /* Read and write ends of the self-pipe */
172 static int selfpipe_readfd
= -1;
173 static int selfpipe_writefd
= -1;
175 /* Process owning the self-pipe --- needed for checking purposes */
176 static int selfpipe_owner_pid
= 0;
178 /* Private function prototypes */
179 static void latch_sigurg_handler(SIGNAL_ARGS
);
180 static void sendSelfPipeByte(void);
183 #if defined(WAIT_USE_SELF_PIPE) || defined(WAIT_USE_SIGNALFD)
184 static void drain(void);
187 #if defined(WAIT_USE_EPOLL)
188 static void WaitEventAdjustEpoll(WaitEventSet
*set
, WaitEvent
*event
, int action
);
189 #elif defined(WAIT_USE_KQUEUE)
190 static void WaitEventAdjustKqueue(WaitEventSet
*set
, WaitEvent
*event
, int old_events
);
191 #elif defined(WAIT_USE_POLL)
192 static void WaitEventAdjustPoll(WaitEventSet
*set
, WaitEvent
*event
);
193 #elif defined(WAIT_USE_WIN32)
194 static void WaitEventAdjustWin32(WaitEventSet
*set
, WaitEvent
*event
);
197 static inline int WaitEventSetWaitBlock(WaitEventSet
*set
, int cur_timeout
,
198 WaitEvent
*occurred_events
, int nevents
);
200 /* ResourceOwner support to hold WaitEventSets */
201 static void ResOwnerReleaseWaitEventSet(Datum res
);
203 static const ResourceOwnerDesc wait_event_set_resowner_desc
=
205 .name
= "WaitEventSet",
206 .release_phase
= RESOURCE_RELEASE_AFTER_LOCKS
,
207 .release_priority
= RELEASE_PRIO_WAITEVENTSETS
,
208 .ReleaseResource
= ResOwnerReleaseWaitEventSet
,
212 /* Convenience wrappers over ResourceOwnerRemember/Forget */
214 ResourceOwnerRememberWaitEventSet(ResourceOwner owner
, WaitEventSet
*set
)
216 ResourceOwnerRemember(owner
, PointerGetDatum(set
), &wait_event_set_resowner_desc
);
219 ResourceOwnerForgetWaitEventSet(ResourceOwner owner
, WaitEventSet
*set
)
221 ResourceOwnerForget(owner
, PointerGetDatum(set
), &wait_event_set_resowner_desc
);
226 * Initialize the process-local latch infrastructure.
228 * This must be called once during startup of any process that can wait on
229 * latches, before it issues any InitLatch() or OwnLatch() calls.
232 InitializeLatchSupport(void)
234 #if defined(WAIT_USE_SELF_PIPE)
237 if (IsUnderPostmaster
)
240 * We might have inherited connections to a self-pipe created by the
241 * postmaster. It's critical that child processes create their own
242 * self-pipes, of course, and we really want them to close the
243 * inherited FDs for safety's sake.
245 if (selfpipe_owner_pid
!= 0)
247 /* Assert we go through here but once in a child process */
248 Assert(selfpipe_owner_pid
!= MyProcPid
);
249 /* Release postmaster's pipe FDs; ignore any error */
250 (void) close(selfpipe_readfd
);
251 (void) close(selfpipe_writefd
);
252 /* Clean up, just for safety's sake; we'll set these below */
253 selfpipe_readfd
= selfpipe_writefd
= -1;
254 selfpipe_owner_pid
= 0;
255 /* Keep fd.c's accounting straight */
262 * Postmaster didn't create a self-pipe ... or else we're in an
263 * EXEC_BACKEND build, in which case it doesn't matter since the
264 * postmaster's pipe FDs were closed by the action of FD_CLOEXEC.
265 * fd.c won't have state to clean up, either.
267 Assert(selfpipe_readfd
== -1);
272 /* In postmaster or standalone backend, assert we do this but once */
273 Assert(selfpipe_readfd
== -1);
274 Assert(selfpipe_owner_pid
== 0);
278 * Set up the self-pipe that allows a signal handler to wake up the
279 * poll()/epoll_wait() in WaitLatch. Make the write-end non-blocking, so
280 * that SetLatch won't block if the event has already been set many times
281 * filling the kernel buffer. Make the read-end non-blocking too, so that
282 * we can easily clear the pipe by reading until EAGAIN or EWOULDBLOCK.
283 * Also, make both FDs close-on-exec, since we surely do not want any
284 * child processes messing with them.
286 if (pipe(pipefd
) < 0)
287 elog(FATAL
, "pipe() failed: %m");
288 if (fcntl(pipefd
[0], F_SETFL
, O_NONBLOCK
) == -1)
289 elog(FATAL
, "fcntl(F_SETFL) failed on read-end of self-pipe: %m");
290 if (fcntl(pipefd
[1], F_SETFL
, O_NONBLOCK
) == -1)
291 elog(FATAL
, "fcntl(F_SETFL) failed on write-end of self-pipe: %m");
292 if (fcntl(pipefd
[0], F_SETFD
, FD_CLOEXEC
) == -1)
293 elog(FATAL
, "fcntl(F_SETFD) failed on read-end of self-pipe: %m");
294 if (fcntl(pipefd
[1], F_SETFD
, FD_CLOEXEC
) == -1)
295 elog(FATAL
, "fcntl(F_SETFD) failed on write-end of self-pipe: %m");
297 selfpipe_readfd
= pipefd
[0];
298 selfpipe_writefd
= pipefd
[1];
299 selfpipe_owner_pid
= MyProcPid
;
301 /* Tell fd.c about these two long-lived FDs */
305 pqsignal(SIGURG
, latch_sigurg_handler
);
308 #ifdef WAIT_USE_SIGNALFD
309 sigset_t signalfd_mask
;
311 if (IsUnderPostmaster
)
314 * It would probably be safe to re-use the inherited signalfd since
315 * signalfds only see the current process's pending signals, but it
316 * seems less surprising to close it and create our own.
320 /* Release postmaster's signal FD; ignore any error */
321 (void) close(signal_fd
);
327 /* Block SIGURG, because we'll receive it through a signalfd. */
328 sigaddset(&UnBlockSig
, SIGURG
);
330 /* Set up the signalfd to receive SIGURG notifications. */
331 sigemptyset(&signalfd_mask
);
332 sigaddset(&signalfd_mask
, SIGURG
);
333 signal_fd
= signalfd(-1, &signalfd_mask
, SFD_NONBLOCK
| SFD_CLOEXEC
);
335 elog(FATAL
, "signalfd() failed");
339 #ifdef WAIT_USE_KQUEUE
340 /* Ignore SIGURG, because we'll receive it via kqueue. */
341 pqsignal(SIGURG
, SIG_IGN
);
346 InitializeLatchWaitSet(void)
348 int latch_pos PG_USED_FOR_ASSERTS_ONLY
;
350 Assert(LatchWaitSet
== NULL
);
352 /* Set up the WaitEventSet used by WaitLatch(). */
353 LatchWaitSet
= CreateWaitEventSet(NULL
, 2);
354 latch_pos
= AddWaitEventToSet(LatchWaitSet
, WL_LATCH_SET
, PGINVALID_SOCKET
,
356 if (IsUnderPostmaster
)
357 AddWaitEventToSet(LatchWaitSet
, WL_EXIT_ON_PM_DEATH
,
358 PGINVALID_SOCKET
, NULL
, NULL
);
360 Assert(latch_pos
== LatchWaitSetLatchPos
);
364 ShutdownLatchSupport(void)
366 #if defined(WAIT_USE_POLL)
367 pqsignal(SIGURG
, SIG_IGN
);
372 FreeWaitEventSet(LatchWaitSet
);
376 #if defined(WAIT_USE_SELF_PIPE)
377 close(selfpipe_readfd
);
378 close(selfpipe_writefd
);
379 selfpipe_readfd
= -1;
380 selfpipe_writefd
= -1;
381 selfpipe_owner_pid
= InvalidPid
;
384 #if defined(WAIT_USE_SIGNALFD)
391 * Initialize a process-local latch.
394 InitLatch(Latch
*latch
)
396 latch
->is_set
= false;
397 latch
->maybe_sleeping
= false;
398 latch
->owner_pid
= MyProcPid
;
399 latch
->is_shared
= false;
401 #if defined(WAIT_USE_SELF_PIPE)
402 /* Assert InitializeLatchSupport has been called in this process */
403 Assert(selfpipe_readfd
>= 0 && selfpipe_owner_pid
== MyProcPid
);
404 #elif defined(WAIT_USE_SIGNALFD)
405 /* Assert InitializeLatchSupport has been called in this process */
406 Assert(signal_fd
>= 0);
407 #elif defined(WAIT_USE_WIN32)
408 latch
->event
= CreateEvent(NULL
, TRUE
, FALSE
, NULL
);
409 if (latch
->event
== NULL
)
410 elog(ERROR
, "CreateEvent failed: error code %lu", GetLastError());
415 * Initialize a shared latch that can be set from other processes. The latch
416 * is initially owned by no-one; use OwnLatch to associate it with the
419 * InitSharedLatch needs to be called in postmaster before forking child
420 * processes, usually right after allocating the shared memory block
421 * containing the latch with ShmemInitStruct. (The Unix implementation
422 * doesn't actually require that, but the Windows one does.) Because of
423 * this restriction, we have no concurrency issues to worry about here.
425 * Note that other handles created in this module are never marked as
426 * inheritable. Thus we do not need to worry about cleaning up child
427 * process references to postmaster-private latches or WaitEventSets.
430 InitSharedLatch(Latch
*latch
)
433 SECURITY_ATTRIBUTES sa
;
436 * Set up security attributes to specify that the events are inherited.
438 ZeroMemory(&sa
, sizeof(sa
));
439 sa
.nLength
= sizeof(sa
);
440 sa
.bInheritHandle
= TRUE
;
442 latch
->event
= CreateEvent(&sa
, TRUE
, FALSE
, NULL
);
443 if (latch
->event
== NULL
)
444 elog(ERROR
, "CreateEvent failed: error code %lu", GetLastError());
447 latch
->is_set
= false;
448 latch
->maybe_sleeping
= false;
449 latch
->owner_pid
= 0;
450 latch
->is_shared
= true;
454 * Associate a shared latch with the current process, allowing it to
457 * Although there is a sanity check for latch-already-owned, we don't do
458 * any sort of locking here, meaning that we could fail to detect the error
459 * if two processes try to own the same latch at about the same time. If
460 * there is any risk of that, caller must provide an interlock to prevent it.
463 OwnLatch(Latch
*latch
)
468 Assert(latch
->is_shared
);
470 #if defined(WAIT_USE_SELF_PIPE)
471 /* Assert InitializeLatchSupport has been called in this process */
472 Assert(selfpipe_readfd
>= 0 && selfpipe_owner_pid
== MyProcPid
);
473 #elif defined(WAIT_USE_SIGNALFD)
474 /* Assert InitializeLatchSupport has been called in this process */
475 Assert(signal_fd
>= 0);
478 owner_pid
= latch
->owner_pid
;
480 elog(PANIC
, "latch already owned by PID %d", owner_pid
);
482 latch
->owner_pid
= MyProcPid
;
486 * Disown a shared latch currently owned by the current process.
489 DisownLatch(Latch
*latch
)
491 Assert(latch
->is_shared
);
492 Assert(latch
->owner_pid
== MyProcPid
);
494 latch
->owner_pid
= 0;
498 * Wait for a given latch to be set, or for postmaster death, or until timeout
499 * is exceeded. 'wakeEvents' is a bitmask that specifies which of those events
500 * to wait for. If the latch is already set (and WL_LATCH_SET is given), the
501 * function returns immediately.
503 * The "timeout" is given in milliseconds. It must be >= 0 if WL_TIMEOUT flag
504 * is given. Although it is declared as "long", we don't actually support
505 * timeouts longer than INT_MAX milliseconds. Note that some extra overhead
506 * is incurred when WL_TIMEOUT is given, so avoid using a timeout if possible.
508 * The latch must be owned by the current process, ie. it must be a
509 * process-local latch initialized with InitLatch, or a shared latch
510 * associated with the current process by calling OwnLatch.
512 * Returns bit mask indicating which condition(s) caused the wake-up. Note
513 * that if multiple wake-up conditions are true, there is no guarantee that
514 * we return all of them in one call, but we will return at least one.
517 WaitLatch(Latch
*latch
, int wakeEvents
, long timeout
,
518 uint32 wait_event_info
)
522 /* Postmaster-managed callers must handle postmaster death somehow. */
523 Assert(!IsUnderPostmaster
||
524 (wakeEvents
& WL_EXIT_ON_PM_DEATH
) ||
525 (wakeEvents
& WL_POSTMASTER_DEATH
));
528 * Some callers may have a latch other than MyLatch, or no latch at all,
529 * or want to handle postmaster death differently. It's cheap to assign
530 * those, so just do it every time.
532 if (!(wakeEvents
& WL_LATCH_SET
))
534 ModifyWaitEvent(LatchWaitSet
, LatchWaitSetLatchPos
, WL_LATCH_SET
, latch
);
535 LatchWaitSet
->exit_on_postmaster_death
=
536 ((wakeEvents
& WL_EXIT_ON_PM_DEATH
) != 0);
538 if (WaitEventSetWait(LatchWaitSet
,
539 (wakeEvents
& WL_TIMEOUT
) ? timeout
: -1,
541 wait_event_info
) == 0)
548 * Like WaitLatch, but with an extra socket argument for WL_SOCKET_*
551 * When waiting on a socket, EOF and error conditions always cause the socket
552 * to be reported as readable/writable/connected, so that the caller can deal
553 * with the condition.
555 * wakeEvents must include either WL_EXIT_ON_PM_DEATH for automatic exit
556 * if the postmaster dies or WL_POSTMASTER_DEATH for a flag set in the
557 * return value if the postmaster dies. The latter is useful for rare cases
558 * where some behavior other than immediate exit is needed.
560 * NB: These days this is just a wrapper around the WaitEventSet API. When
561 * using a latch very frequently, consider creating a longer living
562 * WaitEventSet instead; that's more efficient.
565 WaitLatchOrSocket(Latch
*latch
, int wakeEvents
, pgsocket sock
,
566 long timeout
, uint32 wait_event_info
)
571 WaitEventSet
*set
= CreateWaitEventSet(CurrentResourceOwner
, 3);
573 if (wakeEvents
& WL_TIMEOUT
)
574 Assert(timeout
>= 0);
578 if (wakeEvents
& WL_LATCH_SET
)
579 AddWaitEventToSet(set
, WL_LATCH_SET
, PGINVALID_SOCKET
,
582 /* Postmaster-managed callers must handle postmaster death somehow. */
583 Assert(!IsUnderPostmaster
||
584 (wakeEvents
& WL_EXIT_ON_PM_DEATH
) ||
585 (wakeEvents
& WL_POSTMASTER_DEATH
));
587 if ((wakeEvents
& WL_POSTMASTER_DEATH
) && IsUnderPostmaster
)
588 AddWaitEventToSet(set
, WL_POSTMASTER_DEATH
, PGINVALID_SOCKET
,
591 if ((wakeEvents
& WL_EXIT_ON_PM_DEATH
) && IsUnderPostmaster
)
592 AddWaitEventToSet(set
, WL_EXIT_ON_PM_DEATH
, PGINVALID_SOCKET
,
595 if (wakeEvents
& WL_SOCKET_MASK
)
599 ev
= wakeEvents
& WL_SOCKET_MASK
;
600 AddWaitEventToSet(set
, ev
, sock
, NULL
, NULL
);
603 rc
= WaitEventSetWait(set
, timeout
, &event
, 1, wait_event_info
);
609 ret
|= event
.events
& (WL_LATCH_SET
|
610 WL_POSTMASTER_DEATH
|
614 FreeWaitEventSet(set
);
620 * Sets a latch and wakes up anyone waiting on it.
622 * This is cheap if the latch is already set, otherwise not so much.
624 * NB: when calling this in a signal handler, be sure to save and restore
625 * errno around it. (That's standard practice in most signal handlers, of
626 * course, but we used to omit it in handlers that only set a flag.)
628 * NB: this function is called from critical sections and signal handlers so
629 * throwing an error is not a good idea.
632 SetLatch(Latch
*latch
)
641 * The memory barrier has to be placed here to ensure that any flag
642 * variables possibly changed by this process have been flushed to main
643 * memory, before we check/set is_set.
647 /* Quick exit if already set */
651 latch
->is_set
= true;
654 if (!latch
->maybe_sleeping
)
660 * See if anyone's waiting for the latch. It can be the current process if
661 * we're in a signal handler. We use the self-pipe or SIGURG to ourselves
662 * to wake up WaitEventSetWaitBlock() without races in that case. If it's
663 * another process, send a signal.
665 * Fetch owner_pid only once, in case the latch is concurrently getting
666 * owned or disowned. XXX: This assumes that pid_t is atomic, which isn't
667 * guaranteed to be true! In practice, the effective range of pid_t fits
668 * in a 32 bit integer, and so should be atomic. In the worst case, we
669 * might end up signaling the wrong process. Even then, you're very
670 * unlucky if a process with that bogus pid exists and belongs to
671 * Postgres; and PG database processes should handle excess SIGUSR1
672 * interrupts without a problem anyhow.
674 * Another sort of race condition that's possible here is for a new
675 * process to own the latch immediately after we look, so we don't signal
676 * it. This is okay so long as all callers of ResetLatch/WaitLatch follow
677 * the standard coding convention of waiting at the bottom of their loops,
678 * not the top, so that they'll correctly process latch-setting events
679 * that happen before they enter the loop.
681 owner_pid
= latch
->owner_pid
;
684 else if (owner_pid
== MyProcPid
)
686 #if defined(WAIT_USE_SELF_PIPE)
691 kill(MyProcPid
, SIGURG
);
695 kill(owner_pid
, SIGURG
);
700 * See if anyone's waiting for the latch. It can be the current process if
701 * we're in a signal handler.
703 * Use a local variable here just in case somebody changes the event field
704 * concurrently (which really should not happen).
706 handle
= latch
->event
;
712 * Note that we silently ignore any errors. We might be in a signal
713 * handler or other critical path where it's not safe to call elog().
720 * Clear the latch. Calling WaitLatch after this will sleep, unless
721 * the latch is set again before the WaitLatch call.
724 ResetLatch(Latch
*latch
)
726 /* Only the owner should reset the latch */
727 Assert(latch
->owner_pid
== MyProcPid
);
728 Assert(latch
->maybe_sleeping
== false);
730 latch
->is_set
= false;
733 * Ensure that the write to is_set gets flushed to main memory before we
734 * examine any flag variables. Otherwise a concurrent SetLatch might
735 * falsely conclude that it needn't signal us, even though we have missed
736 * seeing some flag updates that SetLatch was supposed to inform us of.
742 * Create a WaitEventSet with space for nevents different events to wait for.
744 * These events can then be efficiently waited upon together, using
745 * WaitEventSetWait().
747 * The WaitEventSet is tracked by the given 'resowner'. Use NULL for session
751 CreateWaitEventSet(ResourceOwner resowner
, int nevents
)
758 * Use MAXALIGN size/alignment to guarantee that later uses of memory are
759 * aligned correctly. E.g. epoll_event might need 8 byte alignment on some
760 * platforms, but earlier allocations like WaitEventSet and WaitEvent
761 * might not be sized to guarantee that when purely using sizeof().
763 sz
+= MAXALIGN(sizeof(WaitEventSet
));
764 sz
+= MAXALIGN(sizeof(WaitEvent
) * nevents
);
766 #if defined(WAIT_USE_EPOLL)
767 sz
+= MAXALIGN(sizeof(struct epoll_event
) * nevents
);
768 #elif defined(WAIT_USE_KQUEUE)
769 sz
+= MAXALIGN(sizeof(struct kevent
) * nevents
);
770 #elif defined(WAIT_USE_POLL)
771 sz
+= MAXALIGN(sizeof(struct pollfd
) * nevents
);
772 #elif defined(WAIT_USE_WIN32)
773 /* need space for the pgwin32_signal_event */
774 sz
+= MAXALIGN(sizeof(HANDLE
) * (nevents
+ 1));
777 if (resowner
!= NULL
)
778 ResourceOwnerEnlarge(resowner
);
780 data
= (char *) MemoryContextAllocZero(TopMemoryContext
, sz
);
782 set
= (WaitEventSet
*) data
;
783 data
+= MAXALIGN(sizeof(WaitEventSet
));
785 set
->events
= (WaitEvent
*) data
;
786 data
+= MAXALIGN(sizeof(WaitEvent
) * nevents
);
788 #if defined(WAIT_USE_EPOLL)
789 set
->epoll_ret_events
= (struct epoll_event
*) data
;
790 data
+= MAXALIGN(sizeof(struct epoll_event
) * nevents
);
791 #elif defined(WAIT_USE_KQUEUE)
792 set
->kqueue_ret_events
= (struct kevent
*) data
;
793 data
+= MAXALIGN(sizeof(struct kevent
) * nevents
);
794 #elif defined(WAIT_USE_POLL)
795 set
->pollfds
= (struct pollfd
*) data
;
796 data
+= MAXALIGN(sizeof(struct pollfd
) * nevents
);
797 #elif defined(WAIT_USE_WIN32)
798 set
->handles
= (HANDLE
) data
;
799 data
+= MAXALIGN(sizeof(HANDLE
) * nevents
);
803 set
->nevents_space
= nevents
;
804 set
->exit_on_postmaster_death
= false;
806 if (resowner
!= NULL
)
808 ResourceOwnerRememberWaitEventSet(resowner
, set
);
809 set
->owner
= resowner
;
812 #if defined(WAIT_USE_EPOLL)
813 if (!AcquireExternalFD())
814 elog(ERROR
, "AcquireExternalFD, for epoll_create1, failed: %m");
815 set
->epoll_fd
= epoll_create1(EPOLL_CLOEXEC
);
816 if (set
->epoll_fd
< 0)
819 elog(ERROR
, "epoll_create1 failed: %m");
821 #elif defined(WAIT_USE_KQUEUE)
822 if (!AcquireExternalFD())
823 elog(ERROR
, "AcquireExternalFD, for kqueue, failed: %m");
824 set
->kqueue_fd
= kqueue();
825 if (set
->kqueue_fd
< 0)
828 elog(ERROR
, "kqueue failed: %m");
830 if (fcntl(set
->kqueue_fd
, F_SETFD
, FD_CLOEXEC
) == -1)
832 int save_errno
= errno
;
834 close(set
->kqueue_fd
);
837 elog(ERROR
, "fcntl(F_SETFD) failed on kqueue descriptor: %m");
839 set
->report_postmaster_not_running
= false;
840 #elif defined(WAIT_USE_WIN32)
843 * To handle signals while waiting, we need to add a win32 specific event.
844 * We accounted for the additional event at the top of this routine. See
845 * port/win32/signal.c for more details.
847 * Note: pgwin32_signal_event should be first to ensure that it will be
848 * reported when multiple events are set. We want to guarantee that
849 * pending signals are serviced.
851 set
->handles
[0] = pgwin32_signal_event
;
852 StaticAssertStmt(WSA_INVALID_EVENT
== NULL
, "");
859 * Free a previously created WaitEventSet.
861 * Note: preferably, this shouldn't have to free any resources that could be
862 * inherited across an exec(). If it did, we'd likely leak those resources in
863 * many scenarios. For the epoll case, we ensure that by setting EPOLL_CLOEXEC
864 * when the FD is created. For the Windows case, we assume that the handles
865 * involved are non-inheritable.
868 FreeWaitEventSet(WaitEventSet
*set
)
872 ResourceOwnerForgetWaitEventSet(set
->owner
, set
);
876 #if defined(WAIT_USE_EPOLL)
877 close(set
->epoll_fd
);
879 #elif defined(WAIT_USE_KQUEUE)
880 close(set
->kqueue_fd
);
882 #elif defined(WAIT_USE_WIN32)
883 for (WaitEvent
*cur_event
= set
->events
;
884 cur_event
< (set
->events
+ set
->nevents
);
887 if (cur_event
->events
& WL_LATCH_SET
)
889 /* uses the latch's HANDLE */
891 else if (cur_event
->events
& WL_POSTMASTER_DEATH
)
893 /* uses PostmasterHandle */
897 /* Clean up the event object we created for the socket */
898 WSAEventSelect(cur_event
->fd
, NULL
, 0);
899 WSACloseEvent(set
->handles
[cur_event
->pos
+ 1]);
908 * Free a previously created WaitEventSet in a child process after a fork().
911 FreeWaitEventSetAfterFork(WaitEventSet
*set
)
913 #if defined(WAIT_USE_EPOLL)
914 close(set
->epoll_fd
);
916 #elif defined(WAIT_USE_KQUEUE)
917 /* kqueues are not normally inherited by child processes */
925 * Add an event to the set. Possible events are:
926 * - WL_LATCH_SET: Wait for the latch to be set
927 * - WL_POSTMASTER_DEATH: Wait for postmaster to die
928 * - WL_SOCKET_READABLE: Wait for socket to become readable,
929 * can be combined in one event with other WL_SOCKET_* events
930 * - WL_SOCKET_WRITEABLE: Wait for socket to become writeable,
931 * can be combined with other WL_SOCKET_* events
932 * - WL_SOCKET_CONNECTED: Wait for socket connection to be established,
933 * can be combined with other WL_SOCKET_* events (on non-Windows
934 * platforms, this is the same as WL_SOCKET_WRITEABLE)
935 * - WL_SOCKET_ACCEPT: Wait for new connection to a server socket,
936 * can be combined with other WL_SOCKET_* events (on non-Windows
937 * platforms, this is the same as WL_SOCKET_READABLE)
938 * - WL_SOCKET_CLOSED: Wait for socket to be closed by remote peer.
939 * - WL_EXIT_ON_PM_DEATH: Exit immediately if the postmaster dies
941 * Returns the offset in WaitEventSet->events (starting from 0), which can be
942 * used to modify previously added wait events using ModifyWaitEvent().
944 * In the WL_LATCH_SET case the latch must be owned by the current process,
945 * i.e. it must be a process-local latch initialized with InitLatch, or a
946 * shared latch associated with the current process by calling OwnLatch.
948 * In the WL_SOCKET_READABLE/WRITEABLE/CONNECTED/ACCEPT cases, EOF and error
949 * conditions cause the socket to be reported as readable/writable/connected,
950 * so that the caller can deal with the condition.
952 * The user_data pointer specified here will be set for the events returned
953 * by WaitEventSetWait(), allowing to easily associate additional data with
957 AddWaitEventToSet(WaitEventSet
*set
, uint32 events
, pgsocket fd
, Latch
*latch
,
962 /* not enough space */
963 Assert(set
->nevents
< set
->nevents_space
);
965 if (events
== WL_EXIT_ON_PM_DEATH
)
967 events
= WL_POSTMASTER_DEATH
;
968 set
->exit_on_postmaster_death
= true;
973 if (latch
->owner_pid
!= MyProcPid
)
974 elog(ERROR
, "cannot wait on a latch owned by another process");
976 elog(ERROR
, "cannot wait on more than one latch");
977 if ((events
& WL_LATCH_SET
) != WL_LATCH_SET
)
978 elog(ERROR
, "latch events only support being set");
982 if (events
& WL_LATCH_SET
)
983 elog(ERROR
, "cannot wait on latch without a specified latch");
986 /* waiting for socket readiness without a socket indicates a bug */
987 if (fd
== PGINVALID_SOCKET
&& (events
& WL_SOCKET_MASK
))
988 elog(ERROR
, "cannot wait on socket event without a socket");
990 event
= &set
->events
[set
->nevents
];
991 event
->pos
= set
->nevents
++;
993 event
->events
= events
;
994 event
->user_data
= user_data
;
996 event
->reset
= false;
999 if (events
== WL_LATCH_SET
)
1002 set
->latch_pos
= event
->pos
;
1003 #if defined(WAIT_USE_SELF_PIPE)
1004 event
->fd
= selfpipe_readfd
;
1005 #elif defined(WAIT_USE_SIGNALFD)
1006 event
->fd
= signal_fd
;
1008 event
->fd
= PGINVALID_SOCKET
;
1009 #ifdef WAIT_USE_EPOLL
1014 else if (events
== WL_POSTMASTER_DEATH
)
1017 event
->fd
= postmaster_alive_fds
[POSTMASTER_FD_WATCH
];
1021 /* perform wait primitive specific initialization, if needed */
1022 #if defined(WAIT_USE_EPOLL)
1023 WaitEventAdjustEpoll(set
, event
, EPOLL_CTL_ADD
);
1024 #elif defined(WAIT_USE_KQUEUE)
1025 WaitEventAdjustKqueue(set
, event
, 0);
1026 #elif defined(WAIT_USE_POLL)
1027 WaitEventAdjustPoll(set
, event
);
1028 #elif defined(WAIT_USE_WIN32)
1029 WaitEventAdjustWin32(set
, event
);
1036 * Change the event mask and, in the WL_LATCH_SET case, the latch associated
1037 * with the WaitEvent. The latch may be changed to NULL to disable the latch
1038 * temporarily, and then set back to a latch later.
1040 * 'pos' is the id returned by AddWaitEventToSet.
1043 ModifyWaitEvent(WaitEventSet
*set
, int pos
, uint32 events
, Latch
*latch
)
1046 #if defined(WAIT_USE_KQUEUE)
1050 Assert(pos
< set
->nevents
);
1052 event
= &set
->events
[pos
];
1053 #if defined(WAIT_USE_KQUEUE)
1054 old_events
= event
->events
;
1058 * If neither the event mask nor the associated latch changes, return
1059 * early. That's an important optimization for some sockets, where
1060 * ModifyWaitEvent is frequently used to switch from waiting for reads to
1061 * waiting on writes.
1063 if (events
== event
->events
&&
1064 (!(event
->events
& WL_LATCH_SET
) || set
->latch
== latch
))
1067 if (event
->events
& WL_LATCH_SET
&&
1068 events
!= event
->events
)
1070 elog(ERROR
, "cannot modify latch event");
1073 if (event
->events
& WL_POSTMASTER_DEATH
)
1075 elog(ERROR
, "cannot modify postmaster death event");
1078 /* FIXME: validate event mask */
1079 event
->events
= events
;
1081 if (events
== WL_LATCH_SET
)
1083 if (latch
&& latch
->owner_pid
!= MyProcPid
)
1084 elog(ERROR
, "cannot wait on a latch owned by another process");
1088 * On Unix, we don't need to modify the kernel object because the
1089 * underlying pipe (if there is one) is the same for all latches so we
1090 * can return immediately. On Windows, we need to update our array of
1091 * handles, but we leave the old one in place and tolerate spurious
1092 * wakeups if the latch is disabled.
1094 #if defined(WAIT_USE_WIN32)
1102 #if defined(WAIT_USE_EPOLL)
1103 WaitEventAdjustEpoll(set
, event
, EPOLL_CTL_MOD
);
1104 #elif defined(WAIT_USE_KQUEUE)
1105 WaitEventAdjustKqueue(set
, event
, old_events
);
1106 #elif defined(WAIT_USE_POLL)
1107 WaitEventAdjustPoll(set
, event
);
1108 #elif defined(WAIT_USE_WIN32)
1109 WaitEventAdjustWin32(set
, event
);
1113 #if defined(WAIT_USE_EPOLL)
1115 * action can be one of EPOLL_CTL_ADD | EPOLL_CTL_MOD | EPOLL_CTL_DEL
1118 WaitEventAdjustEpoll(WaitEventSet
*set
, WaitEvent
*event
, int action
)
1120 struct epoll_event epoll_ev
;
1123 /* pointer to our event, returned by epoll_wait */
1124 epoll_ev
.data
.ptr
= event
;
1125 /* always wait for errors */
1126 epoll_ev
.events
= EPOLLERR
| EPOLLHUP
;
1128 /* prepare pollfd entry once */
1129 if (event
->events
== WL_LATCH_SET
)
1131 Assert(set
->latch
!= NULL
);
1132 epoll_ev
.events
|= EPOLLIN
;
1134 else if (event
->events
== WL_POSTMASTER_DEATH
)
1136 epoll_ev
.events
|= EPOLLIN
;
1140 Assert(event
->fd
!= PGINVALID_SOCKET
);
1141 Assert(event
->events
& (WL_SOCKET_READABLE
|
1142 WL_SOCKET_WRITEABLE
|
1145 if (event
->events
& WL_SOCKET_READABLE
)
1146 epoll_ev
.events
|= EPOLLIN
;
1147 if (event
->events
& WL_SOCKET_WRITEABLE
)
1148 epoll_ev
.events
|= EPOLLOUT
;
1149 if (event
->events
& WL_SOCKET_CLOSED
)
1150 epoll_ev
.events
|= EPOLLRDHUP
;
1154 * Even though unused, we also pass epoll_ev as the data argument if
1155 * EPOLL_CTL_DEL is passed as action. There used to be an epoll bug
1156 * requiring that, and actually it makes the code simpler...
1158 rc
= epoll_ctl(set
->epoll_fd
, action
, event
->fd
, &epoll_ev
);
1162 (errcode_for_socket_access(),
1163 errmsg("%s() failed: %m",
1168 #if defined(WAIT_USE_POLL)
1170 WaitEventAdjustPoll(WaitEventSet
*set
, WaitEvent
*event
)
1172 struct pollfd
*pollfd
= &set
->pollfds
[event
->pos
];
1174 pollfd
->revents
= 0;
1175 pollfd
->fd
= event
->fd
;
1177 /* prepare pollfd entry once */
1178 if (event
->events
== WL_LATCH_SET
)
1180 Assert(set
->latch
!= NULL
);
1181 pollfd
->events
= POLLIN
;
1183 else if (event
->events
== WL_POSTMASTER_DEATH
)
1185 pollfd
->events
= POLLIN
;
1189 Assert(event
->events
& (WL_SOCKET_READABLE
|
1190 WL_SOCKET_WRITEABLE
|
1193 if (event
->events
& WL_SOCKET_READABLE
)
1194 pollfd
->events
|= POLLIN
;
1195 if (event
->events
& WL_SOCKET_WRITEABLE
)
1196 pollfd
->events
|= POLLOUT
;
1198 if (event
->events
& WL_SOCKET_CLOSED
)
1199 pollfd
->events
|= POLLRDHUP
;
1203 Assert(event
->fd
!= PGINVALID_SOCKET
);
1207 #if defined(WAIT_USE_KQUEUE)
1210 * On most BSD family systems, the udata member of struct kevent is of type
1211 * void *, so we could directly convert to/from WaitEvent *. Unfortunately,
1212 * NetBSD has it as intptr_t, so here we wallpaper over that difference with
1215 #define AccessWaitEvent(k_ev) (*((WaitEvent **)(&(k_ev)->udata)))
1218 WaitEventAdjustKqueueAdd(struct kevent
*k_ev
, int filter
, int action
,
1221 k_ev
->ident
= event
->fd
;
1222 k_ev
->filter
= filter
;
1223 k_ev
->flags
= action
;
1226 AccessWaitEvent(k_ev
) = event
;
1230 WaitEventAdjustKqueueAddPostmaster(struct kevent
*k_ev
, WaitEvent
*event
)
1232 /* For now postmaster death can only be added, not removed. */
1233 k_ev
->ident
= PostmasterPid
;
1234 k_ev
->filter
= EVFILT_PROC
;
1235 k_ev
->flags
= EV_ADD
;
1236 k_ev
->fflags
= NOTE_EXIT
;
1238 AccessWaitEvent(k_ev
) = event
;
1242 WaitEventAdjustKqueueAddLatch(struct kevent
*k_ev
, WaitEvent
*event
)
1244 /* For now latch can only be added, not removed. */
1245 k_ev
->ident
= SIGURG
;
1246 k_ev
->filter
= EVFILT_SIGNAL
;
1247 k_ev
->flags
= EV_ADD
;
1250 AccessWaitEvent(k_ev
) = event
;
1254 * old_events is the previous event mask, used to compute what has changed.
1257 WaitEventAdjustKqueue(WaitEventSet
*set
, WaitEvent
*event
, int old_events
)
1260 struct kevent k_ev
[2];
1262 bool new_filt_read
= false;
1263 bool old_filt_read
= false;
1264 bool new_filt_write
= false;
1265 bool old_filt_write
= false;
1267 if (old_events
== event
->events
)
1270 Assert(event
->events
!= WL_LATCH_SET
|| set
->latch
!= NULL
);
1271 Assert(event
->events
== WL_LATCH_SET
||
1272 event
->events
== WL_POSTMASTER_DEATH
||
1273 (event
->events
& (WL_SOCKET_READABLE
|
1274 WL_SOCKET_WRITEABLE
|
1275 WL_SOCKET_CLOSED
)));
1277 if (event
->events
== WL_POSTMASTER_DEATH
)
1280 * Unlike all the other implementations, we detect postmaster death
1281 * using process notification instead of waiting on the postmaster
1284 WaitEventAdjustKqueueAddPostmaster(&k_ev
[count
++], event
);
1286 else if (event
->events
== WL_LATCH_SET
)
1288 /* We detect latch wakeup using a signal event. */
1289 WaitEventAdjustKqueueAddLatch(&k_ev
[count
++], event
);
1294 * We need to compute the adds and deletes required to get from the
1295 * old event mask to the new event mask, since kevent treats readable
1296 * and writable as separate events.
1298 if (old_events
& (WL_SOCKET_READABLE
| WL_SOCKET_CLOSED
))
1299 old_filt_read
= true;
1300 if (event
->events
& (WL_SOCKET_READABLE
| WL_SOCKET_CLOSED
))
1301 new_filt_read
= true;
1302 if (old_events
& WL_SOCKET_WRITEABLE
)
1303 old_filt_write
= true;
1304 if (event
->events
& WL_SOCKET_WRITEABLE
)
1305 new_filt_write
= true;
1306 if (old_filt_read
&& !new_filt_read
)
1307 WaitEventAdjustKqueueAdd(&k_ev
[count
++], EVFILT_READ
, EV_DELETE
,
1309 else if (!old_filt_read
&& new_filt_read
)
1310 WaitEventAdjustKqueueAdd(&k_ev
[count
++], EVFILT_READ
, EV_ADD
,
1312 if (old_filt_write
&& !new_filt_write
)
1313 WaitEventAdjustKqueueAdd(&k_ev
[count
++], EVFILT_WRITE
, EV_DELETE
,
1315 else if (!old_filt_write
&& new_filt_write
)
1316 WaitEventAdjustKqueueAdd(&k_ev
[count
++], EVFILT_WRITE
, EV_ADD
,
1320 /* For WL_SOCKET_READ -> WL_SOCKET_CLOSED, no change needed. */
1326 rc
= kevent(set
->kqueue_fd
, &k_ev
[0], count
, NULL
, 0, NULL
);
1329 * When adding the postmaster's pid, we have to consider that it might
1330 * already have exited and perhaps even been replaced by another process
1331 * with the same pid. If so, we have to defer reporting this as an event
1332 * until the next call to WaitEventSetWaitBlock().
1337 if (event
->events
== WL_POSTMASTER_DEATH
&&
1338 (errno
== ESRCH
|| errno
== EACCES
))
1339 set
->report_postmaster_not_running
= true;
1342 (errcode_for_socket_access(),
1343 errmsg("%s() failed: %m",
1346 else if (event
->events
== WL_POSTMASTER_DEATH
&&
1347 PostmasterPid
!= getppid() &&
1348 !PostmasterIsAlive())
1351 * The extra PostmasterIsAliveInternal() check prevents false alarms
1352 * on systems that give a different value for getppid() while being
1353 * traced by a debugger.
1355 set
->report_postmaster_not_running
= true;
1361 #if defined(WAIT_USE_WIN32)
1363 WaitEventAdjustWin32(WaitEventSet
*set
, WaitEvent
*event
)
1365 HANDLE
*handle
= &set
->handles
[event
->pos
+ 1];
1367 if (event
->events
== WL_LATCH_SET
)
1369 Assert(set
->latch
!= NULL
);
1370 *handle
= set
->latch
->event
;
1372 else if (event
->events
== WL_POSTMASTER_DEATH
)
1374 *handle
= PostmasterHandle
;
1378 int flags
= FD_CLOSE
; /* always check for errors/EOF */
1380 if (event
->events
& WL_SOCKET_READABLE
)
1382 if (event
->events
& WL_SOCKET_WRITEABLE
)
1384 if (event
->events
& WL_SOCKET_CONNECTED
)
1385 flags
|= FD_CONNECT
;
1386 if (event
->events
& WL_SOCKET_ACCEPT
)
1389 if (*handle
== WSA_INVALID_EVENT
)
1391 *handle
= WSACreateEvent();
1392 if (*handle
== WSA_INVALID_EVENT
)
1393 elog(ERROR
, "failed to create event for socket: error code %d",
1396 if (WSAEventSelect(event
->fd
, *handle
, flags
) != 0)
1397 elog(ERROR
, "failed to set up event for socket: error code %d",
1400 Assert(event
->fd
!= PGINVALID_SOCKET
);
1406 * Wait for events added to the set to happen, or until the timeout is
1407 * reached. At most nevents occurred events are returned.
1409 * If timeout = -1, block until an event occurs; if 0, check sockets for
1410 * readiness, but don't block; if > 0, block for at most timeout milliseconds.
1412 * Returns the number of events occurred, or 0 if the timeout was reached.
1414 * Returned events will have the fd, pos, user_data fields set to the
1415 * values associated with the registered event.
1418 WaitEventSetWait(WaitEventSet
*set
, long timeout
,
1419 WaitEvent
*occurred_events
, int nevents
,
1420 uint32 wait_event_info
)
1422 int returned_events
= 0;
1423 instr_time start_time
;
1424 instr_time cur_time
;
1425 long cur_timeout
= -1;
1427 Assert(nevents
> 0);
1430 * Initialize timeout if requested. We must record the current time so
1431 * that we can determine the remaining timeout if interrupted.
1435 INSTR_TIME_SET_CURRENT(start_time
);
1436 Assert(timeout
>= 0 && timeout
<= INT_MAX
);
1437 cur_timeout
= timeout
;
1440 INSTR_TIME_SET_ZERO(start_time
);
1442 pgstat_report_wait_start(wait_event_info
);
1447 /* Ensure that signals are serviced even if latch is already set */
1448 pgwin32_dispatch_queued_signals();
1450 while (returned_events
== 0)
1455 * Check if the latch is set already. If so, leave the loop
1456 * immediately, avoid blocking again. We don't attempt to report any
1457 * other events that might also be satisfied.
1459 * If someone sets the latch between this and the
1460 * WaitEventSetWaitBlock() below, the setter will write a byte to the
1461 * pipe (or signal us and the signal handler will do that), and the
1462 * readiness routine will return immediately.
1464 * On unix, If there's a pending byte in the self pipe, we'll notice
1465 * whenever blocking. Only clearing the pipe in that case avoids
1466 * having to drain it every time WaitLatchOrSocket() is used. Should
1467 * the pipe-buffer fill up we're still ok, because the pipe is in
1468 * nonblocking mode. It's unlikely for that to happen, because the
1469 * self pipe isn't filled unless we're blocking (waiting = true), or
1470 * from inside a signal handler in latch_sigurg_handler().
1472 * On windows, we'll also notice if there's a pending event for the
1473 * latch when blocking, but there's no danger of anything filling up,
1474 * as "Setting an event that is already set has no effect.".
1476 * Note: we assume that the kernel calls involved in latch management
1477 * will provide adequate synchronization on machines with weak memory
1478 * ordering, so that we cannot miss seeing is_set if a notification
1479 * has already been queued.
1481 if (set
->latch
&& !set
->latch
->is_set
)
1483 /* about to sleep on a latch */
1484 set
->latch
->maybe_sleeping
= true;
1485 pg_memory_barrier();
1489 if (set
->latch
&& set
->latch
->is_set
)
1491 occurred_events
->fd
= PGINVALID_SOCKET
;
1492 occurred_events
->pos
= set
->latch_pos
;
1493 occurred_events
->user_data
=
1494 set
->events
[set
->latch_pos
].user_data
;
1495 occurred_events
->events
= WL_LATCH_SET
;
1499 /* could have been set above */
1500 set
->latch
->maybe_sleeping
= false;
1506 * Wait for events using the readiness primitive chosen at the top of
1507 * this file. If -1 is returned, a timeout has occurred, if 0 we have
1508 * to retry, everything >= 1 is the number of returned events.
1510 rc
= WaitEventSetWaitBlock(set
, cur_timeout
,
1511 occurred_events
, nevents
);
1515 Assert(set
->latch
->maybe_sleeping
);
1516 set
->latch
->maybe_sleeping
= false;
1520 break; /* timeout occurred */
1522 returned_events
= rc
;
1524 /* If we're not done, update cur_timeout for next iteration */
1525 if (returned_events
== 0 && timeout
>= 0)
1527 INSTR_TIME_SET_CURRENT(cur_time
);
1528 INSTR_TIME_SUBTRACT(cur_time
, start_time
);
1529 cur_timeout
= timeout
- (long) INSTR_TIME_GET_MILLISEC(cur_time
);
1530 if (cur_timeout
<= 0)
1538 pgstat_report_wait_end();
1540 return returned_events
;
1544 #if defined(WAIT_USE_EPOLL)
1547 * Wait using linux's epoll_wait(2).
1549 * This is the preferable wait method, as several readiness notifications are
1550 * delivered, without having to iterate through all of set->events. The return
1551 * epoll_event struct contain a pointer to our events, making association
1555 WaitEventSetWaitBlock(WaitEventSet
*set
, int cur_timeout
,
1556 WaitEvent
*occurred_events
, int nevents
)
1558 int returned_events
= 0;
1560 WaitEvent
*cur_event
;
1561 struct epoll_event
*cur_epoll_event
;
1564 rc
= epoll_wait(set
->epoll_fd
, set
->epoll_ret_events
,
1565 Min(nevents
, set
->nevents_space
), cur_timeout
);
1567 /* Check return code */
1570 /* EINTR is okay, otherwise complain */
1575 (errcode_for_socket_access(),
1576 errmsg("%s() failed: %m",
1583 /* timeout exceeded */
1588 * At least one event occurred, iterate over the returned epoll events
1589 * until they're either all processed, or we've returned all the events
1590 * the caller desired.
1592 for (cur_epoll_event
= set
->epoll_ret_events
;
1593 cur_epoll_event
< (set
->epoll_ret_events
+ rc
) &&
1594 returned_events
< nevents
;
1597 /* epoll's data pointer is set to the associated WaitEvent */
1598 cur_event
= (WaitEvent
*) cur_epoll_event
->data
.ptr
;
1600 occurred_events
->pos
= cur_event
->pos
;
1601 occurred_events
->user_data
= cur_event
->user_data
;
1602 occurred_events
->events
= 0;
1604 if (cur_event
->events
== WL_LATCH_SET
&&
1605 cur_epoll_event
->events
& (EPOLLIN
| EPOLLERR
| EPOLLHUP
))
1607 /* Drain the signalfd. */
1610 if (set
->latch
&& set
->latch
->is_set
)
1612 occurred_events
->fd
= PGINVALID_SOCKET
;
1613 occurred_events
->events
= WL_LATCH_SET
;
1618 else if (cur_event
->events
== WL_POSTMASTER_DEATH
&&
1619 cur_epoll_event
->events
& (EPOLLIN
| EPOLLERR
| EPOLLHUP
))
1622 * We expect an EPOLLHUP when the remote end is closed, but
1623 * because we don't expect the pipe to become readable or to have
1624 * any errors either, treat those cases as postmaster death, too.
1626 * Be paranoid about a spurious event signaling the postmaster as
1627 * being dead. There have been reports about that happening with
1628 * older primitives (select(2) to be specific), and a spurious
1629 * WL_POSTMASTER_DEATH event would be painful. Re-checking doesn't
1632 if (!PostmasterIsAliveInternal())
1634 if (set
->exit_on_postmaster_death
)
1636 occurred_events
->fd
= PGINVALID_SOCKET
;
1637 occurred_events
->events
= WL_POSTMASTER_DEATH
;
1642 else if (cur_event
->events
& (WL_SOCKET_READABLE
|
1643 WL_SOCKET_WRITEABLE
|
1646 Assert(cur_event
->fd
!= PGINVALID_SOCKET
);
1648 if ((cur_event
->events
& WL_SOCKET_READABLE
) &&
1649 (cur_epoll_event
->events
& (EPOLLIN
| EPOLLERR
| EPOLLHUP
)))
1651 /* data available in socket, or EOF */
1652 occurred_events
->events
|= WL_SOCKET_READABLE
;
1655 if ((cur_event
->events
& WL_SOCKET_WRITEABLE
) &&
1656 (cur_epoll_event
->events
& (EPOLLOUT
| EPOLLERR
| EPOLLHUP
)))
1658 /* writable, or EOF */
1659 occurred_events
->events
|= WL_SOCKET_WRITEABLE
;
1662 if ((cur_event
->events
& WL_SOCKET_CLOSED
) &&
1663 (cur_epoll_event
->events
& (EPOLLRDHUP
| EPOLLERR
| EPOLLHUP
)))
1665 /* remote peer shut down, or error */
1666 occurred_events
->events
|= WL_SOCKET_CLOSED
;
1669 if (occurred_events
->events
!= 0)
1671 occurred_events
->fd
= cur_event
->fd
;
1678 return returned_events
;
1681 #elif defined(WAIT_USE_KQUEUE)
1684 * Wait using kevent(2) on BSD-family systems and macOS.
1686 * For now this mirrors the epoll code, but in future it could modify the fd
1687 * set in the same call to kevent as it uses for waiting instead of doing that
1688 * with separate system calls.
1691 WaitEventSetWaitBlock(WaitEventSet
*set
, int cur_timeout
,
1692 WaitEvent
*occurred_events
, int nevents
)
1694 int returned_events
= 0;
1696 WaitEvent
*cur_event
;
1697 struct kevent
*cur_kqueue_event
;
1698 struct timespec timeout
;
1699 struct timespec
*timeout_p
;
1701 if (cur_timeout
< 0)
1705 timeout
.tv_sec
= cur_timeout
/ 1000;
1706 timeout
.tv_nsec
= (cur_timeout
% 1000) * 1000000;
1707 timeout_p
= &timeout
;
1711 * Report postmaster events discovered by WaitEventAdjustKqueue() or an
1712 * earlier call to WaitEventSetWait().
1714 if (unlikely(set
->report_postmaster_not_running
))
1716 if (set
->exit_on_postmaster_death
)
1718 occurred_events
->fd
= PGINVALID_SOCKET
;
1719 occurred_events
->events
= WL_POSTMASTER_DEATH
;
1724 rc
= kevent(set
->kqueue_fd
, NULL
, 0,
1725 set
->kqueue_ret_events
,
1726 Min(nevents
, set
->nevents_space
),
1729 /* Check return code */
1732 /* EINTR is okay, otherwise complain */
1737 (errcode_for_socket_access(),
1738 errmsg("%s() failed: %m",
1745 /* timeout exceeded */
1750 * At least one event occurred, iterate over the returned kqueue events
1751 * until they're either all processed, or we've returned all the events
1752 * the caller desired.
1754 for (cur_kqueue_event
= set
->kqueue_ret_events
;
1755 cur_kqueue_event
< (set
->kqueue_ret_events
+ rc
) &&
1756 returned_events
< nevents
;
1759 /* kevent's udata points to the associated WaitEvent */
1760 cur_event
= AccessWaitEvent(cur_kqueue_event
);
1762 occurred_events
->pos
= cur_event
->pos
;
1763 occurred_events
->user_data
= cur_event
->user_data
;
1764 occurred_events
->events
= 0;
1766 if (cur_event
->events
== WL_LATCH_SET
&&
1767 cur_kqueue_event
->filter
== EVFILT_SIGNAL
)
1769 if (set
->latch
&& set
->latch
->is_set
)
1771 occurred_events
->fd
= PGINVALID_SOCKET
;
1772 occurred_events
->events
= WL_LATCH_SET
;
1777 else if (cur_event
->events
== WL_POSTMASTER_DEATH
&&
1778 cur_kqueue_event
->filter
== EVFILT_PROC
&&
1779 (cur_kqueue_event
->fflags
& NOTE_EXIT
) != 0)
1782 * The kernel will tell this kqueue object only once about the
1783 * exit of the postmaster, so let's remember that for next time so
1784 * that we provide level-triggered semantics.
1786 set
->report_postmaster_not_running
= true;
1788 if (set
->exit_on_postmaster_death
)
1790 occurred_events
->fd
= PGINVALID_SOCKET
;
1791 occurred_events
->events
= WL_POSTMASTER_DEATH
;
1795 else if (cur_event
->events
& (WL_SOCKET_READABLE
|
1796 WL_SOCKET_WRITEABLE
|
1799 Assert(cur_event
->fd
>= 0);
1801 if ((cur_event
->events
& WL_SOCKET_READABLE
) &&
1802 (cur_kqueue_event
->filter
== EVFILT_READ
))
1804 /* readable, or EOF */
1805 occurred_events
->events
|= WL_SOCKET_READABLE
;
1808 if ((cur_event
->events
& WL_SOCKET_CLOSED
) &&
1809 (cur_kqueue_event
->filter
== EVFILT_READ
) &&
1810 (cur_kqueue_event
->flags
& EV_EOF
))
1812 /* the remote peer has shut down */
1813 occurred_events
->events
|= WL_SOCKET_CLOSED
;
1816 if ((cur_event
->events
& WL_SOCKET_WRITEABLE
) &&
1817 (cur_kqueue_event
->filter
== EVFILT_WRITE
))
1819 /* writable, or EOF */
1820 occurred_events
->events
|= WL_SOCKET_WRITEABLE
;
1823 if (occurred_events
->events
!= 0)
1825 occurred_events
->fd
= cur_event
->fd
;
1832 return returned_events
;
1835 #elif defined(WAIT_USE_POLL)
1838 * Wait using poll(2).
1840 * This allows to receive readiness notifications for several events at once,
1841 * but requires iterating through all of set->pollfds.
1844 WaitEventSetWaitBlock(WaitEventSet
*set
, int cur_timeout
,
1845 WaitEvent
*occurred_events
, int nevents
)
1847 int returned_events
= 0;
1849 WaitEvent
*cur_event
;
1850 struct pollfd
*cur_pollfd
;
1853 rc
= poll(set
->pollfds
, set
->nevents
, (int) cur_timeout
);
1855 /* Check return code */
1858 /* EINTR is okay, otherwise complain */
1863 (errcode_for_socket_access(),
1864 errmsg("%s() failed: %m",
1871 /* timeout exceeded */
1875 for (cur_event
= set
->events
, cur_pollfd
= set
->pollfds
;
1876 cur_event
< (set
->events
+ set
->nevents
) &&
1877 returned_events
< nevents
;
1878 cur_event
++, cur_pollfd
++)
1880 /* no activity on this FD, skip */
1881 if (cur_pollfd
->revents
== 0)
1884 occurred_events
->pos
= cur_event
->pos
;
1885 occurred_events
->user_data
= cur_event
->user_data
;
1886 occurred_events
->events
= 0;
1888 if (cur_event
->events
== WL_LATCH_SET
&&
1889 (cur_pollfd
->revents
& (POLLIN
| POLLHUP
| POLLERR
| POLLNVAL
)))
1891 /* There's data in the self-pipe, clear it. */
1894 if (set
->latch
&& set
->latch
->is_set
)
1896 occurred_events
->fd
= PGINVALID_SOCKET
;
1897 occurred_events
->events
= WL_LATCH_SET
;
1902 else if (cur_event
->events
== WL_POSTMASTER_DEATH
&&
1903 (cur_pollfd
->revents
& (POLLIN
| POLLHUP
| POLLERR
| POLLNVAL
)))
1906 * We expect an POLLHUP when the remote end is closed, but because
1907 * we don't expect the pipe to become readable or to have any
1908 * errors either, treat those cases as postmaster death, too.
1910 * Be paranoid about a spurious event signaling the postmaster as
1911 * being dead. There have been reports about that happening with
1912 * older primitives (select(2) to be specific), and a spurious
1913 * WL_POSTMASTER_DEATH event would be painful. Re-checking doesn't
1916 if (!PostmasterIsAliveInternal())
1918 if (set
->exit_on_postmaster_death
)
1920 occurred_events
->fd
= PGINVALID_SOCKET
;
1921 occurred_events
->events
= WL_POSTMASTER_DEATH
;
1926 else if (cur_event
->events
& (WL_SOCKET_READABLE
|
1927 WL_SOCKET_WRITEABLE
|
1930 int errflags
= POLLHUP
| POLLERR
| POLLNVAL
;
1932 Assert(cur_event
->fd
>= PGINVALID_SOCKET
);
1934 if ((cur_event
->events
& WL_SOCKET_READABLE
) &&
1935 (cur_pollfd
->revents
& (POLLIN
| errflags
)))
1937 /* data available in socket, or EOF */
1938 occurred_events
->events
|= WL_SOCKET_READABLE
;
1941 if ((cur_event
->events
& WL_SOCKET_WRITEABLE
) &&
1942 (cur_pollfd
->revents
& (POLLOUT
| errflags
)))
1944 /* writeable, or EOF */
1945 occurred_events
->events
|= WL_SOCKET_WRITEABLE
;
1949 if ((cur_event
->events
& WL_SOCKET_CLOSED
) &&
1950 (cur_pollfd
->revents
& (POLLRDHUP
| errflags
)))
1952 /* remote peer closed, or error */
1953 occurred_events
->events
|= WL_SOCKET_CLOSED
;
1957 if (occurred_events
->events
!= 0)
1959 occurred_events
->fd
= cur_event
->fd
;
1965 return returned_events
;
1968 #elif defined(WAIT_USE_WIN32)
1971 * Wait using Windows' WaitForMultipleObjects(). Each call only "consumes" one
1972 * event, so we keep calling until we've filled up our output buffer to match
1973 * the behavior of the other implementations.
1975 * https://blogs.msdn.microsoft.com/oldnewthing/20150409-00/?p=44273
1978 WaitEventSetWaitBlock(WaitEventSet
*set
, int cur_timeout
,
1979 WaitEvent
*occurred_events
, int nevents
)
1981 int returned_events
= 0;
1983 WaitEvent
*cur_event
;
1985 /* Reset any wait events that need it */
1986 for (cur_event
= set
->events
;
1987 cur_event
< (set
->events
+ set
->nevents
);
1990 if (cur_event
->reset
)
1992 WaitEventAdjustWin32(set
, cur_event
);
1993 cur_event
->reset
= false;
1997 * We associate the socket with a new event handle for each
1998 * WaitEventSet. FD_CLOSE is only generated once if the other end
1999 * closes gracefully. Therefore we might miss the FD_CLOSE
2000 * notification, if it was delivered to another event after we stopped
2001 * waiting for it. Close that race by peeking for EOF after setting
2002 * up this handle to receive notifications, and before entering the
2005 * XXX If we had one event handle for the lifetime of a socket, we
2006 * wouldn't need this.
2008 if (cur_event
->events
& WL_SOCKET_READABLE
)
2018 if (WSARecv(cur_event
->fd
, &buf
, 1, &received
, &flags
, NULL
, NULL
) == 0)
2020 occurred_events
->pos
= cur_event
->pos
;
2021 occurred_events
->user_data
= cur_event
->user_data
;
2022 occurred_events
->events
= WL_SOCKET_READABLE
;
2023 occurred_events
->fd
= cur_event
->fd
;
2029 * Windows does not guarantee to log an FD_WRITE network event
2030 * indicating that more data can be sent unless the previous send()
2031 * failed with WSAEWOULDBLOCK. While our caller might well have made
2032 * such a call, we cannot assume that here. Therefore, if waiting for
2033 * write-ready, force the issue by doing a dummy send(). If the dummy
2034 * send() succeeds, assume that the socket is in fact write-ready, and
2035 * return immediately. Also, if it fails with something other than
2036 * WSAEWOULDBLOCK, return a write-ready indication to let our caller
2037 * deal with the error condition.
2039 if (cur_event
->events
& WL_SOCKET_WRITEABLE
)
2049 r
= WSASend(cur_event
->fd
, &buf
, 1, &sent
, 0, NULL
, NULL
);
2050 if (r
== 0 || WSAGetLastError() != WSAEWOULDBLOCK
)
2052 occurred_events
->pos
= cur_event
->pos
;
2053 occurred_events
->user_data
= cur_event
->user_data
;
2054 occurred_events
->events
= WL_SOCKET_WRITEABLE
;
2055 occurred_events
->fd
= cur_event
->fd
;
2064 * Need to wait for ->nevents + 1, because signal handle is in [0].
2066 rc
= WaitForMultipleObjects(set
->nevents
+ 1, set
->handles
, FALSE
,
2069 /* Check return code */
2070 if (rc
== WAIT_FAILED
)
2071 elog(ERROR
, "WaitForMultipleObjects() failed: error code %lu",
2073 else if (rc
== WAIT_TIMEOUT
)
2075 /* timeout exceeded */
2079 if (rc
== WAIT_OBJECT_0
)
2081 /* Service newly-arrived signals */
2082 pgwin32_dispatch_queued_signals();
2083 return 0; /* retry */
2087 * With an offset of one, due to the always present pgwin32_signal_event,
2088 * the handle offset directly corresponds to a wait event.
2090 cur_event
= (WaitEvent
*) &set
->events
[rc
- WAIT_OBJECT_0
- 1];
2097 occurred_events
->pos
= cur_event
->pos
;
2098 occurred_events
->user_data
= cur_event
->user_data
;
2099 occurred_events
->events
= 0;
2101 if (cur_event
->events
== WL_LATCH_SET
)
2104 * We cannot use set->latch->event to reset the fired event if we
2105 * aren't waiting on this latch now.
2107 if (!ResetEvent(set
->handles
[cur_event
->pos
+ 1]))
2108 elog(ERROR
, "ResetEvent failed: error code %lu", GetLastError());
2110 if (set
->latch
&& set
->latch
->is_set
)
2112 occurred_events
->fd
= PGINVALID_SOCKET
;
2113 occurred_events
->events
= WL_LATCH_SET
;
2118 else if (cur_event
->events
== WL_POSTMASTER_DEATH
)
2121 * Postmaster apparently died. Since the consequences of falsely
2122 * returning WL_POSTMASTER_DEATH could be pretty unpleasant, we
2123 * take the trouble to positively verify this with
2124 * PostmasterIsAlive(), even though there is no known reason to
2125 * think that the event could be falsely set on Windows.
2127 if (!PostmasterIsAliveInternal())
2129 if (set
->exit_on_postmaster_death
)
2131 occurred_events
->fd
= PGINVALID_SOCKET
;
2132 occurred_events
->events
= WL_POSTMASTER_DEATH
;
2137 else if (cur_event
->events
& WL_SOCKET_MASK
)
2139 WSANETWORKEVENTS resEvents
;
2140 HANDLE handle
= set
->handles
[cur_event
->pos
+ 1];
2142 Assert(cur_event
->fd
);
2144 occurred_events
->fd
= cur_event
->fd
;
2146 ZeroMemory(&resEvents
, sizeof(resEvents
));
2147 if (WSAEnumNetworkEvents(cur_event
->fd
, handle
, &resEvents
) != 0)
2148 elog(ERROR
, "failed to enumerate network events: error code %d",
2150 if ((cur_event
->events
& WL_SOCKET_READABLE
) &&
2151 (resEvents
.lNetworkEvents
& FD_READ
))
2153 /* data available in socket */
2154 occurred_events
->events
|= WL_SOCKET_READABLE
;
2157 * WaitForMultipleObjects doesn't guarantee that a read event
2158 * will be returned if the latch is set at the same time. Even
2159 * if it did, the caller might drop that event expecting it to
2160 * reoccur on next call. So, we must force the event to be
2161 * reset if this WaitEventSet is used again in order to avoid
2162 * an indefinite hang.
2165 * https://msdn.microsoft.com/en-us/library/windows/desktop/ms741576(v=vs.85).aspx
2166 * for the behavior of socket events.
2169 cur_event
->reset
= true;
2171 if ((cur_event
->events
& WL_SOCKET_WRITEABLE
) &&
2172 (resEvents
.lNetworkEvents
& FD_WRITE
))
2175 occurred_events
->events
|= WL_SOCKET_WRITEABLE
;
2177 if ((cur_event
->events
& WL_SOCKET_CONNECTED
) &&
2178 (resEvents
.lNetworkEvents
& FD_CONNECT
))
2181 occurred_events
->events
|= WL_SOCKET_CONNECTED
;
2183 if ((cur_event
->events
& WL_SOCKET_ACCEPT
) &&
2184 (resEvents
.lNetworkEvents
& FD_ACCEPT
))
2186 /* incoming connection could be accepted */
2187 occurred_events
->events
|= WL_SOCKET_ACCEPT
;
2189 if (resEvents
.lNetworkEvents
& FD_CLOSE
)
2191 /* EOF/error, so signal all caller-requested socket flags */
2192 occurred_events
->events
|= (cur_event
->events
& WL_SOCKET_MASK
);
2195 if (occurred_events
->events
!= 0)
2202 /* Is the output buffer full? */
2203 if (returned_events
== nevents
)
2206 /* Have we run out of possible events? */
2207 next_pos
= cur_event
->pos
+ 1;
2208 if (next_pos
== set
->nevents
)
2212 * Poll the rest of the event handles in the array starting at
2213 * next_pos being careful to skip over the initial signal handle too.
2214 * This time we use a zero timeout.
2216 count
= set
->nevents
- next_pos
;
2217 rc
= WaitForMultipleObjects(count
,
2218 set
->handles
+ 1 + next_pos
,
2223 * We don't distinguish between errors and WAIT_TIMEOUT here because
2224 * we already have events to report.
2226 if (rc
< WAIT_OBJECT_0
|| rc
>= WAIT_OBJECT_0
+ count
)
2229 /* We have another event to decode. */
2230 cur_event
= &set
->events
[next_pos
+ (rc
- WAIT_OBJECT_0
)];
2233 return returned_events
;
2238 * Return whether the current build options can report WL_SOCKET_CLOSED.
2241 WaitEventSetCanReportClosed(void)
2243 #if (defined(WAIT_USE_POLL) && defined(POLLRDHUP)) || \
2244 defined(WAIT_USE_EPOLL) || \
2245 defined(WAIT_USE_KQUEUE)
2253 * Get the number of wait events registered in a given WaitEventSet.
2256 GetNumRegisteredWaitEvents(WaitEventSet
*set
)
2258 return set
->nevents
;
2261 #if defined(WAIT_USE_SELF_PIPE)
2264 * SetLatch uses SIGURG to wake up the process waiting on the latch.
2266 * Wake up WaitLatch, if we're waiting.
2269 latch_sigurg_handler(SIGNAL_ARGS
)
2275 /* Send one byte to the self-pipe, to wake up WaitLatch */
2277 sendSelfPipeByte(void)
2283 rc
= write(selfpipe_writefd
, &dummy
, 1);
2286 /* If interrupted by signal, just retry */
2291 * If the pipe is full, we don't need to retry, the data that's there
2292 * already is enough to wake up WaitLatch.
2294 if (errno
== EAGAIN
|| errno
== EWOULDBLOCK
)
2298 * Oops, the write() failed for some other reason. We might be in a
2299 * signal handler, so it's not safe to elog(). We have no choice but
2300 * silently ignore the error.
2308 #if defined(WAIT_USE_SELF_PIPE) || defined(WAIT_USE_SIGNALFD)
2311 * Read all available data from self-pipe or signalfd.
2313 * Note: this is only called when waiting = true. If it fails and doesn't
2314 * return, it must reset that flag first (though ideally, this will never
2324 #ifdef WAIT_USE_SELF_PIPE
2325 fd
= selfpipe_readfd
;
2332 rc
= read(fd
, buf
, sizeof(buf
));
2335 if (errno
== EAGAIN
|| errno
== EWOULDBLOCK
)
2336 break; /* the descriptor is empty */
2337 else if (errno
== EINTR
)
2338 continue; /* retry */
2342 #ifdef WAIT_USE_SELF_PIPE
2343 elog(ERROR
, "read() on self-pipe failed: %m");
2345 elog(ERROR
, "read() on signalfd failed: %m");
2352 #ifdef WAIT_USE_SELF_PIPE
2353 elog(ERROR
, "unexpected EOF on self-pipe");
2355 elog(ERROR
, "unexpected EOF on signalfd");
2358 else if (rc
< sizeof(buf
))
2360 /* we successfully drained the pipe; no need to read() again */
2363 /* else buffer wasn't big enough, so read again */
2370 ResOwnerReleaseWaitEventSet(Datum res
)
2372 WaitEventSet
*set
= (WaitEventSet
*) DatumGetPointer(res
);
2374 Assert(set
->owner
!= NULL
);
2376 FreeWaitEventSet(set
);