1 /*-------------------------------------------------------------------------
4 * Simple LRU buffering for wrap-around-able permanent metadata
6 * This module is used to maintain various pieces of transaction status
7 * indexed by TransactionId (such as commit status, parent transaction ID,
8 * commit timestamp), as well as storage for multixacts, serializable
9 * isolation locks and NOTIFY traffic. Extensions can define their own
12 * Under ordinary circumstances we expect that write traffic will occur
13 * mostly to the latest page (and to the just-prior page, soon after a
14 * page transition). Read traffic will probably touch a larger span of
15 * pages, but a relatively small number of buffers should be sufficient.
17 * We use a simple least-recently-used scheme to manage a pool of shared
18 * page buffers, split in banks by the lowest bits of the page number, and
19 * the management algorithm only processes the bank to which the desired
20 * page belongs, so a linear search is sufficient; there's no need for a
21 * hashtable or anything fancy. The algorithm is straight LRU except that
22 * we will never swap out the latest page (since we know it's going to be
23 * hit again eventually).
25 * We use per-bank control LWLocks to protect the shared data structures,
26 * plus per-buffer LWLocks that synchronize I/O for each buffer. The
27 * bank's control lock must be held to examine or modify any of the bank's
28 * shared state. A process that is reading in or writing out a page
29 * buffer does not hold the control lock, only the per-buffer lock for the
30 * buffer it is working on. One exception is latest_page_number, which is
31 * read and written using atomic ops.
33 * "Holding the bank control lock" means exclusive lock in all cases
34 * except for SimpleLruReadPage_ReadOnly(); see comments for
35 * SlruRecentlyUsed() for the implications of that.
37 * When initiating I/O on a buffer, we acquire the per-buffer lock exclusively
38 * before releasing the control lock. The per-buffer lock is released after
39 * completing the I/O, re-acquiring the control lock, and updating the shared
40 * state. (Deadlock is not possible here, because we never try to initiate
41 * I/O when someone else is already doing I/O on the same buffer.)
42 * To wait for I/O to complete, release the control lock, acquire the
43 * per-buffer lock in shared mode, immediately release the per-buffer lock,
44 * reacquire the control lock, and then recheck state (since arbitrary things
45 * could have happened while we didn't have the lock).
47 * As with the regular buffer manager, it is possible for another process
48 * to re-dirty a page that is currently being written out. This is handled
49 * by re-setting the page's page_dirty flag.
52 * Portions Copyright (c) 1996-2025, PostgreSQL Global Development Group
53 * Portions Copyright (c) 1994, Regents of the University of California
55 * src/backend/access/transam/slru.c
57 *-------------------------------------------------------------------------
65 #include "access/slru.h"
66 #include "access/transam.h"
67 #include "access/xlog.h"
68 #include "access/xlogutils.h"
69 #include "miscadmin.h"
71 #include "storage/fd.h"
72 #include "storage/shmem.h"
73 #include "utils/guc.h"
76 * Converts segment number to the filename of the segment.
78 * "path" should point to a buffer at least MAXPGPATH characters long.
80 * If ctl->long_segment_names is true, segno can be in the range [0, 2^60-1].
81 * The resulting file name is made of 15 characters, e.g. dir/123456789ABCDEF.
83 * If ctl->long_segment_names is false, segno can be in the range [0, 2^24-1].
84 * The resulting file name is made of 4 to 6 characters, as of:
86 * dir/1234 for [0, 2^16-1]
87 * dir/12345 for [2^16, 2^20-1]
88 * dir/123456 for [2^20, 2^24-1]
91 SlruFileName(SlruCtl ctl
, char *path
, int64 segno
)
93 if (ctl
->long_segment_names
)
96 * We could use 16 characters here but the disadvantage would be that
97 * the SLRU segments will be hard to distinguish from WAL segments.
99 * For this reason we use 15 characters. It is enough but also means
100 * that in the future we can't decrease SLRU_PAGES_PER_SEGMENT easily.
102 Assert(segno
>= 0 && segno
<= INT64CONST(0xFFFFFFFFFFFFFFF));
103 return snprintf(path
, MAXPGPATH
, "%s/%015llX", ctl
->Dir
,
109 * Despite the fact that %04X format string is used up to 24 bit
110 * integers are allowed. See SlruCorrectSegmentFilenameLength()
112 Assert(segno
>= 0 && segno
<= INT64CONST(0xFFFFFF));
113 return snprintf(path
, MAXPGPATH
, "%s/%04X", (ctl
)->Dir
,
114 (unsigned int) segno
);
119 * During SimpleLruWriteAll(), we will usually not need to write more than one
120 * or two physical files, but we may need to write several pages per file. We
121 * can consolidate the I/O requests by leaving files open until control returns
122 * to SimpleLruWriteAll(). This data structure remembers which files are open.
124 #define MAX_WRITEALL_BUFFERS 16
126 typedef struct SlruWriteAllData
128 int num_files
; /* # files actually open */
129 int fd
[MAX_WRITEALL_BUFFERS
]; /* their FD's */
130 int64 segno
[MAX_WRITEALL_BUFFERS
]; /* their log seg#s */
133 typedef struct SlruWriteAllData
*SlruWriteAll
;
137 * Bank size for the slot array. Pages are assigned a bank according to their
138 * page number, with each bank being this size. We want a power of 2 so that
139 * we can determine the bank number for a page with just bit shifting; we also
140 * want to keep the bank size small so that LRU victim search is fast. 16
141 * buffers per bank seems a good number.
143 #define SLRU_BANK_BITSHIFT 4
144 #define SLRU_BANK_SIZE (1 << SLRU_BANK_BITSHIFT)
147 * Macro to get the bank number to which the slot belongs.
149 #define SlotGetBankNumber(slotno) ((slotno) >> SLRU_BANK_BITSHIFT)
153 * Populate a file tag describing a segment file. We only use the segment
154 * number, since we can derive everything else we need by having separate
155 * sync handler functions for clog, multixact etc.
157 #define INIT_SLRUFILETAG(a,xx_handler,xx_segno) \
159 memset(&(a), 0, sizeof(FileTag)), \
160 (a).handler = (xx_handler), \
161 (a).segno = (xx_segno) \
164 /* Saved info for SlruReportIOError */
175 static SlruErrorCause slru_errcause
;
176 static int slru_errno
;
179 static void SimpleLruZeroLSNs(SlruCtl ctl
, int slotno
);
180 static void SimpleLruWaitIO(SlruCtl ctl
, int slotno
);
181 static void SlruInternalWritePage(SlruCtl ctl
, int slotno
, SlruWriteAll fdata
);
182 static bool SlruPhysicalReadPage(SlruCtl ctl
, int64 pageno
, int slotno
);
183 static bool SlruPhysicalWritePage(SlruCtl ctl
, int64 pageno
, int slotno
,
185 static void SlruReportIOError(SlruCtl ctl
, int64 pageno
, TransactionId xid
);
186 static int SlruSelectLRUPage(SlruCtl ctl
, int64 pageno
);
188 static bool SlruScanDirCbDeleteCutoff(SlruCtl ctl
, char *filename
,
189 int64 segpage
, void *data
);
190 static void SlruInternalDeleteSegment(SlruCtl ctl
, int64 segno
);
191 static inline void SlruRecentlyUsed(SlruShared shared
, int slotno
);
195 * Initialization of shared memory
199 SimpleLruShmemSize(int nslots
, int nlsns
)
201 int nbanks
= nslots
/ SLRU_BANK_SIZE
;
204 Assert(nslots
<= SLRU_MAX_ALLOWED_BUFFERS
);
205 Assert(nslots
% SLRU_BANK_SIZE
== 0);
207 /* we assume nslots isn't so large as to risk overflow */
208 sz
= MAXALIGN(sizeof(SlruSharedData
));
209 sz
+= MAXALIGN(nslots
* sizeof(char *)); /* page_buffer[] */
210 sz
+= MAXALIGN(nslots
* sizeof(SlruPageStatus
)); /* page_status[] */
211 sz
+= MAXALIGN(nslots
* sizeof(bool)); /* page_dirty[] */
212 sz
+= MAXALIGN(nslots
* sizeof(int64
)); /* page_number[] */
213 sz
+= MAXALIGN(nslots
* sizeof(int)); /* page_lru_count[] */
214 sz
+= MAXALIGN(nslots
* sizeof(LWLockPadded
)); /* buffer_locks[] */
215 sz
+= MAXALIGN(nbanks
* sizeof(LWLockPadded
)); /* bank_locks[] */
216 sz
+= MAXALIGN(nbanks
* sizeof(int)); /* bank_cur_lru_count[] */
219 sz
+= MAXALIGN(nslots
* nlsns
* sizeof(XLogRecPtr
)); /* group_lsn[] */
221 return BUFFERALIGN(sz
) + BLCKSZ
* nslots
;
225 * Determine a number of SLRU buffers to use.
227 * We simply divide shared_buffers by the divisor given and cap
228 * that at the maximum given; but always at least SLRU_BANK_SIZE.
229 * Round down to the nearest multiple of SLRU_BANK_SIZE.
232 SimpleLruAutotuneBuffers(int divisor
, int max
)
234 return Min(max
- (max
% SLRU_BANK_SIZE
),
236 NBuffers
/ divisor
- (NBuffers
/ divisor
) % SLRU_BANK_SIZE
));
240 * Initialize, or attach to, a simple LRU cache in shared memory.
242 * ctl: address of local (unshared) control structure.
243 * name: name of SLRU. (This is user-visible, pick with care!)
244 * nslots: number of page slots to use.
245 * nlsns: number of LSN groups per page (set to zero if not relevant).
246 * subdir: PGDATA-relative subdirectory that will contain the files.
247 * buffer_tranche_id: tranche ID to use for the SLRU's per-buffer LWLocks.
248 * bank_tranche_id: tranche ID to use for the bank LWLocks.
249 * sync_handler: which set of functions to use to handle sync requests
252 SimpleLruInit(SlruCtl ctl
, const char *name
, int nslots
, int nlsns
,
253 const char *subdir
, int buffer_tranche_id
, int bank_tranche_id
,
254 SyncRequestHandler sync_handler
, bool long_segment_names
)
258 int nbanks
= nslots
/ SLRU_BANK_SIZE
;
260 Assert(nslots
<= SLRU_MAX_ALLOWED_BUFFERS
);
262 shared
= (SlruShared
) ShmemInitStruct(name
,
263 SimpleLruShmemSize(nslots
, nlsns
),
266 if (!IsUnderPostmaster
)
268 /* Initialize locks and shared memory area */
274 memset(shared
, 0, sizeof(SlruSharedData
));
276 shared
->num_slots
= nslots
;
277 shared
->lsn_groups_per_page
= nlsns
;
279 pg_atomic_init_u64(&shared
->latest_page_number
, 0);
281 shared
->slru_stats_idx
= pgstat_get_slru_index(name
);
283 ptr
= (char *) shared
;
284 offset
= MAXALIGN(sizeof(SlruSharedData
));
285 shared
->page_buffer
= (char **) (ptr
+ offset
);
286 offset
+= MAXALIGN(nslots
* sizeof(char *));
287 shared
->page_status
= (SlruPageStatus
*) (ptr
+ offset
);
288 offset
+= MAXALIGN(nslots
* sizeof(SlruPageStatus
));
289 shared
->page_dirty
= (bool *) (ptr
+ offset
);
290 offset
+= MAXALIGN(nslots
* sizeof(bool));
291 shared
->page_number
= (int64
*) (ptr
+ offset
);
292 offset
+= MAXALIGN(nslots
* sizeof(int64
));
293 shared
->page_lru_count
= (int *) (ptr
+ offset
);
294 offset
+= MAXALIGN(nslots
* sizeof(int));
296 /* Initialize LWLocks */
297 shared
->buffer_locks
= (LWLockPadded
*) (ptr
+ offset
);
298 offset
+= MAXALIGN(nslots
* sizeof(LWLockPadded
));
299 shared
->bank_locks
= (LWLockPadded
*) (ptr
+ offset
);
300 offset
+= MAXALIGN(nbanks
* sizeof(LWLockPadded
));
301 shared
->bank_cur_lru_count
= (int *) (ptr
+ offset
);
302 offset
+= MAXALIGN(nbanks
* sizeof(int));
306 shared
->group_lsn
= (XLogRecPtr
*) (ptr
+ offset
);
307 offset
+= MAXALIGN(nslots
* nlsns
* sizeof(XLogRecPtr
));
310 ptr
+= BUFFERALIGN(offset
);
311 for (int slotno
= 0; slotno
< nslots
; slotno
++)
313 LWLockInitialize(&shared
->buffer_locks
[slotno
].lock
,
316 shared
->page_buffer
[slotno
] = ptr
;
317 shared
->page_status
[slotno
] = SLRU_PAGE_EMPTY
;
318 shared
->page_dirty
[slotno
] = false;
319 shared
->page_lru_count
[slotno
] = 0;
323 /* Initialize the slot banks. */
324 for (int bankno
= 0; bankno
< nbanks
; bankno
++)
326 LWLockInitialize(&shared
->bank_locks
[bankno
].lock
, bank_tranche_id
);
327 shared
->bank_cur_lru_count
[bankno
] = 0;
330 /* Should fit to estimated shmem size */
331 Assert(ptr
- (char *) shared
<= SimpleLruShmemSize(nslots
, nlsns
));
336 Assert(shared
->num_slots
== nslots
);
340 * Initialize the unshared control struct, including directory path. We
341 * assume caller set PagePrecedes.
343 ctl
->shared
= shared
;
344 ctl
->sync_handler
= sync_handler
;
345 ctl
->long_segment_names
= long_segment_names
;
346 ctl
->nbanks
= nbanks
;
347 strlcpy(ctl
->Dir
, subdir
, sizeof(ctl
->Dir
));
351 * Helper function for GUC check_hook to check whether slru buffers are in
352 * multiples of SLRU_BANK_SIZE.
355 check_slru_buffers(const char *name
, int *newval
)
357 /* Valid values are multiples of SLRU_BANK_SIZE */
358 if (*newval
% SLRU_BANK_SIZE
== 0)
361 GUC_check_errdetail("\"%s\" must be a multiple of %d.", name
,
367 * Initialize (or reinitialize) a page to zeroes.
369 * The page is not actually written, just set up in shared memory.
370 * The slot number of the new page is returned.
372 * Bank lock must be held at entry, and will be held at exit.
375 SimpleLruZeroPage(SlruCtl ctl
, int64 pageno
)
377 SlruShared shared
= ctl
->shared
;
380 Assert(LWLockHeldByMeInMode(SimpleLruGetBankLock(ctl
, pageno
), LW_EXCLUSIVE
));
382 /* Find a suitable buffer slot for the page */
383 slotno
= SlruSelectLRUPage(ctl
, pageno
);
384 Assert(shared
->page_status
[slotno
] == SLRU_PAGE_EMPTY
||
385 (shared
->page_status
[slotno
] == SLRU_PAGE_VALID
&&
386 !shared
->page_dirty
[slotno
]) ||
387 shared
->page_number
[slotno
] == pageno
);
389 /* Mark the slot as containing this page */
390 shared
->page_number
[slotno
] = pageno
;
391 shared
->page_status
[slotno
] = SLRU_PAGE_VALID
;
392 shared
->page_dirty
[slotno
] = true;
393 SlruRecentlyUsed(shared
, slotno
);
395 /* Set the buffer to zeroes */
396 MemSet(shared
->page_buffer
[slotno
], 0, BLCKSZ
);
398 /* Set the LSNs for this new page to zero */
399 SimpleLruZeroLSNs(ctl
, slotno
);
402 * Assume this page is now the latest active page.
404 * Note that because both this routine and SlruSelectLRUPage run with
405 * ControlLock held, it is not possible for this to be zeroing a page that
406 * SlruSelectLRUPage is going to evict simultaneously. Therefore, there's
407 * no memory barrier here.
409 pg_atomic_write_u64(&shared
->latest_page_number
, pageno
);
411 /* update the stats counter of zeroed pages */
412 pgstat_count_slru_page_zeroed(shared
->slru_stats_idx
);
418 * Zero all the LSNs we store for this slru page.
420 * This should be called each time we create a new page, and each time we read
421 * in a page from disk into an existing buffer. (Such an old page cannot
422 * have any interesting LSNs, since we'd have flushed them before writing
423 * the page in the first place.)
425 * This assumes that InvalidXLogRecPtr is bitwise-all-0.
428 SimpleLruZeroLSNs(SlruCtl ctl
, int slotno
)
430 SlruShared shared
= ctl
->shared
;
432 if (shared
->lsn_groups_per_page
> 0)
433 MemSet(&shared
->group_lsn
[slotno
* shared
->lsn_groups_per_page
], 0,
434 shared
->lsn_groups_per_page
* sizeof(XLogRecPtr
));
438 * Wait for any active I/O on a page slot to finish. (This does not
439 * guarantee that new I/O hasn't been started before we return, though.
440 * In fact the slot might not even contain the same page anymore.)
442 * Bank lock must be held at entry, and will be held at exit.
445 SimpleLruWaitIO(SlruCtl ctl
, int slotno
)
447 SlruShared shared
= ctl
->shared
;
448 int bankno
= SlotGetBankNumber(slotno
);
450 Assert(shared
->page_status
[slotno
] != SLRU_PAGE_EMPTY
);
452 /* See notes at top of file */
453 LWLockRelease(&shared
->bank_locks
[bankno
].lock
);
454 LWLockAcquire(&shared
->buffer_locks
[slotno
].lock
, LW_SHARED
);
455 LWLockRelease(&shared
->buffer_locks
[slotno
].lock
);
456 LWLockAcquire(&shared
->bank_locks
[bankno
].lock
, LW_EXCLUSIVE
);
459 * If the slot is still in an io-in-progress state, then either someone
460 * already started a new I/O on the slot, or a previous I/O failed and
461 * neglected to reset the page state. That shouldn't happen, really, but
462 * it seems worth a few extra cycles to check and recover from it. We can
463 * cheaply test for failure by seeing if the buffer lock is still held (we
464 * assume that transaction abort would release the lock).
466 if (shared
->page_status
[slotno
] == SLRU_PAGE_READ_IN_PROGRESS
||
467 shared
->page_status
[slotno
] == SLRU_PAGE_WRITE_IN_PROGRESS
)
469 if (LWLockConditionalAcquire(&shared
->buffer_locks
[slotno
].lock
, LW_SHARED
))
471 /* indeed, the I/O must have failed */
472 if (shared
->page_status
[slotno
] == SLRU_PAGE_READ_IN_PROGRESS
)
473 shared
->page_status
[slotno
] = SLRU_PAGE_EMPTY
;
474 else /* write_in_progress */
476 shared
->page_status
[slotno
] = SLRU_PAGE_VALID
;
477 shared
->page_dirty
[slotno
] = true;
479 LWLockRelease(&shared
->buffer_locks
[slotno
].lock
);
485 * Find a page in a shared buffer, reading it in if necessary.
486 * The page number must correspond to an already-initialized page.
488 * If write_ok is true then it is OK to return a page that is in
489 * WRITE_IN_PROGRESS state; it is the caller's responsibility to be sure
490 * that modification of the page is safe. If write_ok is false then we
491 * will not return the page until it is not undergoing active I/O.
493 * The passed-in xid is used only for error reporting, and may be
494 * InvalidTransactionId if no specific xid is associated with the action.
496 * Return value is the shared-buffer slot number now holding the page.
497 * The buffer's LRU access info is updated.
499 * The correct bank lock must be held at entry, and will be held at exit.
502 SimpleLruReadPage(SlruCtl ctl
, int64 pageno
, bool write_ok
,
505 SlruShared shared
= ctl
->shared
;
506 LWLock
*banklock
= SimpleLruGetBankLock(ctl
, pageno
);
508 Assert(LWLockHeldByMeInMode(banklock
, LW_EXCLUSIVE
));
510 /* Outer loop handles restart if we must wait for someone else's I/O */
516 /* See if page already is in memory; if not, pick victim slot */
517 slotno
= SlruSelectLRUPage(ctl
, pageno
);
519 /* Did we find the page in memory? */
520 if (shared
->page_status
[slotno
] != SLRU_PAGE_EMPTY
&&
521 shared
->page_number
[slotno
] == pageno
)
524 * If page is still being read in, we must wait for I/O. Likewise
525 * if the page is being written and the caller said that's not OK.
527 if (shared
->page_status
[slotno
] == SLRU_PAGE_READ_IN_PROGRESS
||
528 (shared
->page_status
[slotno
] == SLRU_PAGE_WRITE_IN_PROGRESS
&&
531 SimpleLruWaitIO(ctl
, slotno
);
532 /* Now we must recheck state from the top */
535 /* Otherwise, it's ready to use */
536 SlruRecentlyUsed(shared
, slotno
);
538 /* update the stats counter of pages found in the SLRU */
539 pgstat_count_slru_page_hit(shared
->slru_stats_idx
);
544 /* We found no match; assert we selected a freeable slot */
545 Assert(shared
->page_status
[slotno
] == SLRU_PAGE_EMPTY
||
546 (shared
->page_status
[slotno
] == SLRU_PAGE_VALID
&&
547 !shared
->page_dirty
[slotno
]));
549 /* Mark the slot read-busy */
550 shared
->page_number
[slotno
] = pageno
;
551 shared
->page_status
[slotno
] = SLRU_PAGE_READ_IN_PROGRESS
;
552 shared
->page_dirty
[slotno
] = false;
554 /* Acquire per-buffer lock (cannot deadlock, see notes at top) */
555 LWLockAcquire(&shared
->buffer_locks
[slotno
].lock
, LW_EXCLUSIVE
);
557 /* Release bank lock while doing I/O */
558 LWLockRelease(banklock
);
561 ok
= SlruPhysicalReadPage(ctl
, pageno
, slotno
);
563 /* Set the LSNs for this newly read-in page to zero */
564 SimpleLruZeroLSNs(ctl
, slotno
);
566 /* Re-acquire bank control lock and update page state */
567 LWLockAcquire(banklock
, LW_EXCLUSIVE
);
569 Assert(shared
->page_number
[slotno
] == pageno
&&
570 shared
->page_status
[slotno
] == SLRU_PAGE_READ_IN_PROGRESS
&&
571 !shared
->page_dirty
[slotno
]);
573 shared
->page_status
[slotno
] = ok
? SLRU_PAGE_VALID
: SLRU_PAGE_EMPTY
;
575 LWLockRelease(&shared
->buffer_locks
[slotno
].lock
);
577 /* Now it's okay to ereport if we failed */
579 SlruReportIOError(ctl
, pageno
, xid
);
581 SlruRecentlyUsed(shared
, slotno
);
583 /* update the stats counter of pages not found in SLRU */
584 pgstat_count_slru_page_read(shared
->slru_stats_idx
);
591 * Find a page in a shared buffer, reading it in if necessary.
592 * The page number must correspond to an already-initialized page.
593 * The caller must intend only read-only access to the page.
595 * The passed-in xid is used only for error reporting, and may be
596 * InvalidTransactionId if no specific xid is associated with the action.
598 * Return value is the shared-buffer slot number now holding the page.
599 * The buffer's LRU access info is updated.
601 * Bank control lock must NOT be held at entry, but will be held at exit.
602 * It is unspecified whether the lock will be shared or exclusive.
605 SimpleLruReadPage_ReadOnly(SlruCtl ctl
, int64 pageno
, TransactionId xid
)
607 SlruShared shared
= ctl
->shared
;
608 LWLock
*banklock
= SimpleLruGetBankLock(ctl
, pageno
);
609 int bankno
= pageno
% ctl
->nbanks
;
610 int bankstart
= bankno
* SLRU_BANK_SIZE
;
611 int bankend
= bankstart
+ SLRU_BANK_SIZE
;
613 /* Try to find the page while holding only shared lock */
614 LWLockAcquire(banklock
, LW_SHARED
);
616 /* See if page is already in a buffer */
617 for (int slotno
= bankstart
; slotno
< bankend
; slotno
++)
619 if (shared
->page_status
[slotno
] != SLRU_PAGE_EMPTY
&&
620 shared
->page_number
[slotno
] == pageno
&&
621 shared
->page_status
[slotno
] != SLRU_PAGE_READ_IN_PROGRESS
)
623 /* See comments for SlruRecentlyUsed macro */
624 SlruRecentlyUsed(shared
, slotno
);
626 /* update the stats counter of pages found in the SLRU */
627 pgstat_count_slru_page_hit(shared
->slru_stats_idx
);
633 /* No luck, so switch to normal exclusive lock and do regular read */
634 LWLockRelease(banklock
);
635 LWLockAcquire(banklock
, LW_EXCLUSIVE
);
637 return SimpleLruReadPage(ctl
, pageno
, true, xid
);
641 * Write a page from a shared buffer, if necessary.
642 * Does nothing if the specified slot is not dirty.
644 * NOTE: only one write attempt is made here. Hence, it is possible that
645 * the page is still dirty at exit (if someone else re-dirtied it during
646 * the write). However, we *do* attempt a fresh write even if the page
647 * is already being written; this is for checkpoints.
649 * Bank lock must be held at entry, and will be held at exit.
652 SlruInternalWritePage(SlruCtl ctl
, int slotno
, SlruWriteAll fdata
)
654 SlruShared shared
= ctl
->shared
;
655 int64 pageno
= shared
->page_number
[slotno
];
656 int bankno
= SlotGetBankNumber(slotno
);
659 Assert(shared
->page_status
[slotno
] != SLRU_PAGE_EMPTY
);
660 Assert(LWLockHeldByMeInMode(SimpleLruGetBankLock(ctl
, pageno
), LW_EXCLUSIVE
));
662 /* If a write is in progress, wait for it to finish */
663 while (shared
->page_status
[slotno
] == SLRU_PAGE_WRITE_IN_PROGRESS
&&
664 shared
->page_number
[slotno
] == pageno
)
666 SimpleLruWaitIO(ctl
, slotno
);
670 * Do nothing if page is not dirty, or if buffer no longer contains the
671 * same page we were called for.
673 if (!shared
->page_dirty
[slotno
] ||
674 shared
->page_status
[slotno
] != SLRU_PAGE_VALID
||
675 shared
->page_number
[slotno
] != pageno
)
679 * Mark the slot write-busy, and clear the dirtybit. After this point, a
680 * transaction status update on this page will mark it dirty again.
682 shared
->page_status
[slotno
] = SLRU_PAGE_WRITE_IN_PROGRESS
;
683 shared
->page_dirty
[slotno
] = false;
685 /* Acquire per-buffer lock (cannot deadlock, see notes at top) */
686 LWLockAcquire(&shared
->buffer_locks
[slotno
].lock
, LW_EXCLUSIVE
);
688 /* Release bank lock while doing I/O */
689 LWLockRelease(&shared
->bank_locks
[bankno
].lock
);
692 ok
= SlruPhysicalWritePage(ctl
, pageno
, slotno
, fdata
);
694 /* If we failed, and we're in a flush, better close the files */
697 for (int i
= 0; i
< fdata
->num_files
; i
++)
698 CloseTransientFile(fdata
->fd
[i
]);
701 /* Re-acquire bank lock and update page state */
702 LWLockAcquire(&shared
->bank_locks
[bankno
].lock
, LW_EXCLUSIVE
);
704 Assert(shared
->page_number
[slotno
] == pageno
&&
705 shared
->page_status
[slotno
] == SLRU_PAGE_WRITE_IN_PROGRESS
);
707 /* If we failed to write, mark the page dirty again */
709 shared
->page_dirty
[slotno
] = true;
711 shared
->page_status
[slotno
] = SLRU_PAGE_VALID
;
713 LWLockRelease(&shared
->buffer_locks
[slotno
].lock
);
715 /* Now it's okay to ereport if we failed */
717 SlruReportIOError(ctl
, pageno
, InvalidTransactionId
);
719 /* If part of a checkpoint, count this as a SLRU buffer written. */
722 CheckpointStats
.ckpt_slru_written
++;
723 PendingCheckpointerStats
.slru_written
++;
728 * Wrapper of SlruInternalWritePage, for external callers.
729 * fdata is always passed a NULL here.
732 SimpleLruWritePage(SlruCtl ctl
, int slotno
)
734 Assert(ctl
->shared
->page_status
[slotno
] != SLRU_PAGE_EMPTY
);
736 SlruInternalWritePage(ctl
, slotno
, NULL
);
740 * Return whether the given page exists on disk.
742 * A false return means that either the file does not exist, or that it's not
743 * large enough to contain the given page.
746 SimpleLruDoesPhysicalPageExist(SlruCtl ctl
, int64 pageno
)
748 int64 segno
= pageno
/ SLRU_PAGES_PER_SEGMENT
;
749 int rpageno
= pageno
% SLRU_PAGES_PER_SEGMENT
;
750 int offset
= rpageno
* BLCKSZ
;
751 char path
[MAXPGPATH
];
756 /* update the stats counter of checked pages */
757 pgstat_count_slru_page_exists(ctl
->shared
->slru_stats_idx
);
759 SlruFileName(ctl
, path
, segno
);
761 fd
= OpenTransientFile(path
, O_RDONLY
| PG_BINARY
);
764 /* expected: file doesn't exist */
768 /* report error normally */
769 slru_errcause
= SLRU_OPEN_FAILED
;
771 SlruReportIOError(ctl
, pageno
, 0);
774 if ((endpos
= lseek(fd
, 0, SEEK_END
)) < 0)
776 slru_errcause
= SLRU_SEEK_FAILED
;
778 SlruReportIOError(ctl
, pageno
, 0);
781 result
= endpos
>= (off_t
) (offset
+ BLCKSZ
);
783 if (CloseTransientFile(fd
) != 0)
785 slru_errcause
= SLRU_CLOSE_FAILED
;
794 * Physical read of a (previously existing) page into a buffer slot
796 * On failure, we cannot just ereport(ERROR) since caller has put state in
797 * shared memory that must be undone. So, we return false and save enough
798 * info in static variables to let SlruReportIOError make the report.
800 * For now, assume it's not worth keeping a file pointer open across
801 * read/write operations. We could cache one virtual file pointer ...
804 SlruPhysicalReadPage(SlruCtl ctl
, int64 pageno
, int slotno
)
806 SlruShared shared
= ctl
->shared
;
807 int64 segno
= pageno
/ SLRU_PAGES_PER_SEGMENT
;
808 int rpageno
= pageno
% SLRU_PAGES_PER_SEGMENT
;
809 off_t offset
= rpageno
* BLCKSZ
;
810 char path
[MAXPGPATH
];
813 SlruFileName(ctl
, path
, segno
);
816 * In a crash-and-restart situation, it's possible for us to receive
817 * commands to set the commit status of transactions whose bits are in
818 * already-truncated segments of the commit log (see notes in
819 * SlruPhysicalWritePage). Hence, if we are InRecovery, allow the case
820 * where the file doesn't exist, and return zeroes instead.
822 fd
= OpenTransientFile(path
, O_RDONLY
| PG_BINARY
);
825 if (errno
!= ENOENT
|| !InRecovery
)
827 slru_errcause
= SLRU_OPEN_FAILED
;
833 (errmsg("file \"%s\" doesn't exist, reading as zeroes",
835 MemSet(shared
->page_buffer
[slotno
], 0, BLCKSZ
);
840 pgstat_report_wait_start(WAIT_EVENT_SLRU_READ
);
841 if (pg_pread(fd
, shared
->page_buffer
[slotno
], BLCKSZ
, offset
) != BLCKSZ
)
843 pgstat_report_wait_end();
844 slru_errcause
= SLRU_READ_FAILED
;
846 CloseTransientFile(fd
);
849 pgstat_report_wait_end();
851 if (CloseTransientFile(fd
) != 0)
853 slru_errcause
= SLRU_CLOSE_FAILED
;
862 * Physical write of a page from a buffer slot
864 * On failure, we cannot just ereport(ERROR) since caller has put state in
865 * shared memory that must be undone. So, we return false and save enough
866 * info in static variables to let SlruReportIOError make the report.
868 * For now, assume it's not worth keeping a file pointer open across
869 * independent read/write operations. We do batch operations during
870 * SimpleLruWriteAll, though.
872 * fdata is NULL for a standalone write, pointer to open-file info during
876 SlruPhysicalWritePage(SlruCtl ctl
, int64 pageno
, int slotno
, SlruWriteAll fdata
)
878 SlruShared shared
= ctl
->shared
;
879 int64 segno
= pageno
/ SLRU_PAGES_PER_SEGMENT
;
880 int rpageno
= pageno
% SLRU_PAGES_PER_SEGMENT
;
881 off_t offset
= rpageno
* BLCKSZ
;
882 char path
[MAXPGPATH
];
885 /* update the stats counter of written pages */
886 pgstat_count_slru_page_written(shared
->slru_stats_idx
);
889 * Honor the write-WAL-before-data rule, if appropriate, so that we do not
890 * write out data before associated WAL records. This is the same action
891 * performed during FlushBuffer() in the main buffer manager.
893 if (shared
->group_lsn
!= NULL
)
896 * We must determine the largest async-commit LSN for the page. This
897 * is a bit tedious, but since this entire function is a slow path
898 * anyway, it seems better to do this here than to maintain a per-page
899 * LSN variable (which'd need an extra comparison in the
900 * transaction-commit path).
905 lsnindex
= slotno
* shared
->lsn_groups_per_page
;
906 max_lsn
= shared
->group_lsn
[lsnindex
++];
907 for (int lsnoff
= 1; lsnoff
< shared
->lsn_groups_per_page
; lsnoff
++)
909 XLogRecPtr this_lsn
= shared
->group_lsn
[lsnindex
++];
911 if (max_lsn
< this_lsn
)
915 if (!XLogRecPtrIsInvalid(max_lsn
))
918 * As noted above, elog(ERROR) is not acceptable here, so if
919 * XLogFlush were to fail, we must PANIC. This isn't much of a
920 * restriction because XLogFlush is just about all critical
921 * section anyway, but let's make sure.
923 START_CRIT_SECTION();
930 * During a SimpleLruWriteAll, we may already have the desired file open.
934 for (int i
= 0; i
< fdata
->num_files
; i
++)
936 if (fdata
->segno
[i
] == segno
)
947 * If the file doesn't already exist, we should create it. It is
948 * possible for this to need to happen when writing a page that's not
949 * first in its segment; we assume the OS can cope with that. (Note:
950 * it might seem that it'd be okay to create files only when
951 * SimpleLruZeroPage is called for the first page of a segment.
952 * However, if after a crash and restart the REDO logic elects to
953 * replay the log from a checkpoint before the latest one, then it's
954 * possible that we will get commands to set transaction status of
955 * transactions that have already been truncated from the commit log.
956 * Easiest way to deal with that is to accept references to
957 * nonexistent files here and in SlruPhysicalReadPage.)
959 * Note: it is possible for more than one backend to be executing this
960 * code simultaneously for different pages of the same file. Hence,
961 * don't use O_EXCL or O_TRUNC or anything like that.
963 SlruFileName(ctl
, path
, segno
);
964 fd
= OpenTransientFile(path
, O_RDWR
| O_CREAT
| PG_BINARY
);
967 slru_errcause
= SLRU_OPEN_FAILED
;
974 if (fdata
->num_files
< MAX_WRITEALL_BUFFERS
)
976 fdata
->fd
[fdata
->num_files
] = fd
;
977 fdata
->segno
[fdata
->num_files
] = segno
;
983 * In the unlikely event that we exceed MAX_WRITEALL_BUFFERS,
984 * fall back to treating it as a standalone write.
992 pgstat_report_wait_start(WAIT_EVENT_SLRU_WRITE
);
993 if (pg_pwrite(fd
, shared
->page_buffer
[slotno
], BLCKSZ
, offset
) != BLCKSZ
)
995 pgstat_report_wait_end();
996 /* if write didn't set errno, assume problem is no disk space */
999 slru_errcause
= SLRU_WRITE_FAILED
;
1002 CloseTransientFile(fd
);
1005 pgstat_report_wait_end();
1007 /* Queue up a sync request for the checkpointer. */
1008 if (ctl
->sync_handler
!= SYNC_HANDLER_NONE
)
1012 INIT_SLRUFILETAG(tag
, ctl
->sync_handler
, segno
);
1013 if (!RegisterSyncRequest(&tag
, SYNC_REQUEST
, false))
1015 /* No space to enqueue sync request. Do it synchronously. */
1016 pgstat_report_wait_start(WAIT_EVENT_SLRU_SYNC
);
1017 if (pg_fsync(fd
) != 0)
1019 pgstat_report_wait_end();
1020 slru_errcause
= SLRU_FSYNC_FAILED
;
1022 CloseTransientFile(fd
);
1025 pgstat_report_wait_end();
1029 /* Close file, unless part of flush request. */
1032 if (CloseTransientFile(fd
) != 0)
1034 slru_errcause
= SLRU_CLOSE_FAILED
;
1044 * Issue the error message after failure of SlruPhysicalReadPage or
1045 * SlruPhysicalWritePage. Call this after cleaning up shared-memory state.
1048 SlruReportIOError(SlruCtl ctl
, int64 pageno
, TransactionId xid
)
1050 int64 segno
= pageno
/ SLRU_PAGES_PER_SEGMENT
;
1051 int rpageno
= pageno
% SLRU_PAGES_PER_SEGMENT
;
1052 int offset
= rpageno
* BLCKSZ
;
1053 char path
[MAXPGPATH
];
1055 SlruFileName(ctl
, path
, segno
);
1057 switch (slru_errcause
)
1059 case SLRU_OPEN_FAILED
:
1061 (errcode_for_file_access(),
1062 errmsg("could not access status of transaction %u", xid
),
1063 errdetail("Could not open file \"%s\": %m.", path
)));
1065 case SLRU_SEEK_FAILED
:
1067 (errcode_for_file_access(),
1068 errmsg("could not access status of transaction %u", xid
),
1069 errdetail("Could not seek in file \"%s\" to offset %d: %m.",
1072 case SLRU_READ_FAILED
:
1075 (errcode_for_file_access(),
1076 errmsg("could not access status of transaction %u", xid
),
1077 errdetail("Could not read from file \"%s\" at offset %d: %m.",
1081 (errmsg("could not access status of transaction %u", xid
),
1082 errdetail("Could not read from file \"%s\" at offset %d: read too few bytes.", path
, offset
)));
1084 case SLRU_WRITE_FAILED
:
1087 (errcode_for_file_access(),
1088 errmsg("could not access status of transaction %u", xid
),
1089 errdetail("Could not write to file \"%s\" at offset %d: %m.",
1093 (errmsg("could not access status of transaction %u", xid
),
1094 errdetail("Could not write to file \"%s\" at offset %d: wrote too few bytes.",
1097 case SLRU_FSYNC_FAILED
:
1098 ereport(data_sync_elevel(ERROR
),
1099 (errcode_for_file_access(),
1100 errmsg("could not access status of transaction %u", xid
),
1101 errdetail("Could not fsync file \"%s\": %m.",
1104 case SLRU_CLOSE_FAILED
:
1106 (errcode_for_file_access(),
1107 errmsg("could not access status of transaction %u", xid
),
1108 errdetail("Could not close file \"%s\": %m.",
1112 /* can't get here, we trust */
1113 elog(ERROR
, "unrecognized SimpleLru error cause: %d",
1114 (int) slru_errcause
);
1120 * Mark a buffer slot "most recently used".
1123 SlruRecentlyUsed(SlruShared shared
, int slotno
)
1125 int bankno
= SlotGetBankNumber(slotno
);
1126 int new_lru_count
= shared
->bank_cur_lru_count
[bankno
];
1128 Assert(shared
->page_status
[slotno
] != SLRU_PAGE_EMPTY
);
1131 * The reason for the if-test is that there are often many consecutive
1132 * accesses to the same page (particularly the latest page). By
1133 * suppressing useless increments of bank_cur_lru_count, we reduce the
1134 * probability that old pages' counts will "wrap around" and make them
1135 * appear recently used.
1137 * We allow this code to be executed concurrently by multiple processes
1138 * within SimpleLruReadPage_ReadOnly(). As long as int reads and writes
1139 * are atomic, this should not cause any completely-bogus values to enter
1140 * the computation. However, it is possible for either bank_cur_lru_count
1141 * or individual page_lru_count entries to be "reset" to lower values than
1142 * they should have, in case a process is delayed while it executes this
1143 * function. With care in SlruSelectLRUPage(), this does little harm, and
1144 * in any case the absolute worst possible consequence is a nonoptimal
1145 * choice of page to evict. The gain from allowing concurrent reads of
1146 * SLRU pages seems worth it.
1148 if (new_lru_count
!= shared
->page_lru_count
[slotno
])
1150 shared
->bank_cur_lru_count
[bankno
] = ++new_lru_count
;
1151 shared
->page_lru_count
[slotno
] = new_lru_count
;
1156 * Select the slot to re-use when we need a free slot for the given page.
1158 * The target page number is passed not only because we need to know the
1159 * correct bank to use, but also because we need to consider the possibility
1160 * that some other process reads in the target page while we are doing I/O to
1161 * free a slot. Hence, check or recheck to see if any slot already holds the
1162 * target page, and return that slot if so. Thus, the returned slot is
1163 * *either* a slot already holding the pageno (could be any state except
1164 * EMPTY), *or* a freeable slot (state EMPTY or CLEAN).
1166 * The correct bank lock must be held at entry, and will be held at exit.
1169 SlruSelectLRUPage(SlruCtl ctl
, int64 pageno
)
1171 SlruShared shared
= ctl
->shared
;
1173 /* Outer loop handles restart after I/O */
1177 int bestvalidslot
= 0; /* keep compiler quiet */
1178 int best_valid_delta
= -1;
1179 int64 best_valid_page_number
= 0; /* keep compiler quiet */
1180 int bestinvalidslot
= 0; /* keep compiler quiet */
1181 int best_invalid_delta
= -1;
1182 int64 best_invalid_page_number
= 0; /* keep compiler quiet */
1183 int bankno
= pageno
% ctl
->nbanks
;
1184 int bankstart
= bankno
* SLRU_BANK_SIZE
;
1185 int bankend
= bankstart
+ SLRU_BANK_SIZE
;
1187 Assert(LWLockHeldByMe(SimpleLruGetBankLock(ctl
, pageno
)));
1189 /* See if page already has a buffer assigned */
1190 for (int slotno
= bankstart
; slotno
< bankend
; slotno
++)
1192 if (shared
->page_status
[slotno
] != SLRU_PAGE_EMPTY
&&
1193 shared
->page_number
[slotno
] == pageno
)
1198 * If we find any EMPTY slot, just select that one. Else choose a
1199 * victim page to replace. We normally take the least recently used
1200 * valid page, but we will never take the slot containing
1201 * latest_page_number, even if it appears least recently used. We
1202 * will select a slot that is already I/O busy only if there is no
1203 * other choice: a read-busy slot will not be least recently used once
1204 * the read finishes, and waiting for an I/O on a write-busy slot is
1205 * inferior to just picking some other slot. Testing shows the slot
1206 * we pick instead will often be clean, allowing us to begin a read at
1209 * Normally the page_lru_count values will all be different and so
1210 * there will be a well-defined LRU page. But since we allow
1211 * concurrent execution of SlruRecentlyUsed() within
1212 * SimpleLruReadPage_ReadOnly(), it is possible that multiple pages
1213 * acquire the same lru_count values. In that case we break ties by
1214 * choosing the furthest-back page.
1216 * Notice that this next line forcibly advances cur_lru_count to a
1217 * value that is certainly beyond any value that will be in the
1218 * page_lru_count array after the loop finishes. This ensures that
1219 * the next execution of SlruRecentlyUsed will mark the page newly
1220 * used, even if it's for a page that has the current counter value.
1221 * That gets us back on the path to having good data when there are
1222 * multiple pages with the same lru_count.
1224 cur_count
= (shared
->bank_cur_lru_count
[bankno
])++;
1225 for (int slotno
= bankstart
; slotno
< bankend
; slotno
++)
1228 int64 this_page_number
;
1230 if (shared
->page_status
[slotno
] == SLRU_PAGE_EMPTY
)
1233 this_delta
= cur_count
- shared
->page_lru_count
[slotno
];
1237 * Clean up in case shared updates have caused cur_count
1238 * increments to get "lost". We back off the page counts,
1239 * rather than trying to increase cur_count, to avoid any
1240 * question of infinite loops or failure in the presence of
1241 * wrapped-around counts.
1243 shared
->page_lru_count
[slotno
] = cur_count
;
1248 * If this page is the one most recently zeroed, don't consider it
1249 * an eviction candidate. See comments in SimpleLruZeroPage for an
1250 * explanation about the lack of a memory barrier here.
1252 this_page_number
= shared
->page_number
[slotno
];
1253 if (this_page_number
==
1254 pg_atomic_read_u64(&shared
->latest_page_number
))
1257 if (shared
->page_status
[slotno
] == SLRU_PAGE_VALID
)
1259 if (this_delta
> best_valid_delta
||
1260 (this_delta
== best_valid_delta
&&
1261 ctl
->PagePrecedes(this_page_number
,
1262 best_valid_page_number
)))
1264 bestvalidslot
= slotno
;
1265 best_valid_delta
= this_delta
;
1266 best_valid_page_number
= this_page_number
;
1271 if (this_delta
> best_invalid_delta
||
1272 (this_delta
== best_invalid_delta
&&
1273 ctl
->PagePrecedes(this_page_number
,
1274 best_invalid_page_number
)))
1276 bestinvalidslot
= slotno
;
1277 best_invalid_delta
= this_delta
;
1278 best_invalid_page_number
= this_page_number
;
1284 * If all pages (except possibly the latest one) are I/O busy, we'll
1285 * have to wait for an I/O to complete and then retry. In that
1286 * unhappy case, we choose to wait for the I/O on the least recently
1287 * used slot, on the assumption that it was likely initiated first of
1288 * all the I/Os in progress and may therefore finish first.
1290 if (best_valid_delta
< 0)
1292 SimpleLruWaitIO(ctl
, bestinvalidslot
);
1297 * If the selected page is clean, we're set.
1299 if (!shared
->page_dirty
[bestvalidslot
])
1300 return bestvalidslot
;
1305 SlruInternalWritePage(ctl
, bestvalidslot
, NULL
);
1308 * Now loop back and try again. This is the easiest way of dealing
1309 * with corner cases such as the victim page being re-dirtied while we
1316 * Write dirty pages to disk during checkpoint or database shutdown. Flushing
1317 * is deferred until the next call to ProcessSyncRequests(), though we do fsync
1318 * the containing directory here to make sure that newly created directory
1319 * entries are on disk.
1322 SimpleLruWriteAll(SlruCtl ctl
, bool allow_redirtied
)
1324 SlruShared shared
= ctl
->shared
;
1325 SlruWriteAllData fdata
;
1327 int prevbank
= SlotGetBankNumber(0);
1330 /* update the stats counter of flushes */
1331 pgstat_count_slru_flush(shared
->slru_stats_idx
);
1334 * Find and write dirty pages
1336 fdata
.num_files
= 0;
1338 LWLockAcquire(&shared
->bank_locks
[prevbank
].lock
, LW_EXCLUSIVE
);
1340 for (int slotno
= 0; slotno
< shared
->num_slots
; slotno
++)
1342 int curbank
= SlotGetBankNumber(slotno
);
1345 * If the current bank lock is not same as the previous bank lock then
1346 * release the previous lock and acquire the new lock.
1348 if (curbank
!= prevbank
)
1350 LWLockRelease(&shared
->bank_locks
[prevbank
].lock
);
1351 LWLockAcquire(&shared
->bank_locks
[curbank
].lock
, LW_EXCLUSIVE
);
1355 /* Do nothing if slot is unused */
1356 if (shared
->page_status
[slotno
] == SLRU_PAGE_EMPTY
)
1359 SlruInternalWritePage(ctl
, slotno
, &fdata
);
1362 * In some places (e.g. checkpoints), we cannot assert that the slot
1363 * is clean now, since another process might have re-dirtied it
1364 * already. That's okay.
1366 Assert(allow_redirtied
||
1367 shared
->page_status
[slotno
] == SLRU_PAGE_EMPTY
||
1368 (shared
->page_status
[slotno
] == SLRU_PAGE_VALID
&&
1369 !shared
->page_dirty
[slotno
]));
1372 LWLockRelease(&shared
->bank_locks
[prevbank
].lock
);
1375 * Now close any files that were open
1378 for (int i
= 0; i
< fdata
.num_files
; i
++)
1380 if (CloseTransientFile(fdata
.fd
[i
]) != 0)
1382 slru_errcause
= SLRU_CLOSE_FAILED
;
1384 pageno
= fdata
.segno
[i
] * SLRU_PAGES_PER_SEGMENT
;
1389 SlruReportIOError(ctl
, pageno
, InvalidTransactionId
);
1391 /* Ensure that directory entries for new files are on disk. */
1392 if (ctl
->sync_handler
!= SYNC_HANDLER_NONE
)
1393 fsync_fname(ctl
->Dir
, true);
1397 * Remove all segments before the one holding the passed page number
1399 * All SLRUs prevent concurrent calls to this function, either with an LWLock
1400 * or by calling it only as part of a checkpoint. Mutual exclusion must begin
1401 * before computing cutoffPage. Mutual exclusion must end after any limit
1402 * update that would permit other backends to write fresh data into the
1403 * segment immediately preceding the one containing cutoffPage. Otherwise,
1404 * when the SLRU is quite full, SimpleLruTruncate() might delete that segment
1405 * after it has accrued freshly-written data.
1408 SimpleLruTruncate(SlruCtl ctl
, int64 cutoffPage
)
1410 SlruShared shared
= ctl
->shared
;
1413 /* update the stats counter of truncates */
1414 pgstat_count_slru_truncate(shared
->slru_stats_idx
);
1417 * Scan shared memory and remove any pages preceding the cutoff page, to
1418 * ensure we won't rewrite them later. (Since this is normally called in
1419 * or just after a checkpoint, any dirty pages should have been flushed
1420 * already ... we're just being extra careful here.)
1425 * An important safety check: the current endpoint page must not be
1426 * eligible for removal. This check is just a backstop against wraparound
1427 * bugs elsewhere in SLRU handling, so we don't care if we read a slightly
1428 * outdated value; therefore we don't add a memory barrier.
1430 if (ctl
->PagePrecedes(pg_atomic_read_u64(&shared
->latest_page_number
),
1434 (errmsg("could not truncate directory \"%s\": apparent wraparound",
1439 prevbank
= SlotGetBankNumber(0);
1440 LWLockAcquire(&shared
->bank_locks
[prevbank
].lock
, LW_EXCLUSIVE
);
1441 for (int slotno
= 0; slotno
< shared
->num_slots
; slotno
++)
1443 int curbank
= SlotGetBankNumber(slotno
);
1446 * If the current bank lock is not same as the previous bank lock then
1447 * release the previous lock and acquire the new lock.
1449 if (curbank
!= prevbank
)
1451 LWLockRelease(&shared
->bank_locks
[prevbank
].lock
);
1452 LWLockAcquire(&shared
->bank_locks
[curbank
].lock
, LW_EXCLUSIVE
);
1456 if (shared
->page_status
[slotno
] == SLRU_PAGE_EMPTY
)
1458 if (!ctl
->PagePrecedes(shared
->page_number
[slotno
], cutoffPage
))
1462 * If page is clean, just change state to EMPTY (expected case).
1464 if (shared
->page_status
[slotno
] == SLRU_PAGE_VALID
&&
1465 !shared
->page_dirty
[slotno
])
1467 shared
->page_status
[slotno
] = SLRU_PAGE_EMPTY
;
1472 * Hmm, we have (or may have) I/O operations acting on the page, so
1473 * we've got to wait for them to finish and then start again. This is
1474 * the same logic as in SlruSelectLRUPage. (XXX if page is dirty,
1475 * wouldn't it be OK to just discard it without writing it?
1476 * SlruMayDeleteSegment() uses a stricter qualification, so we might
1477 * not delete this page in the end; even if we don't delete it, we
1478 * won't have cause to read its data again. For now, keep the logic
1479 * the same as it was.)
1481 if (shared
->page_status
[slotno
] == SLRU_PAGE_VALID
)
1482 SlruInternalWritePage(ctl
, slotno
, NULL
);
1484 SimpleLruWaitIO(ctl
, slotno
);
1486 LWLockRelease(&shared
->bank_locks
[prevbank
].lock
);
1490 LWLockRelease(&shared
->bank_locks
[prevbank
].lock
);
1492 /* Now we can remove the old segment(s) */
1493 (void) SlruScanDirectory(ctl
, SlruScanDirCbDeleteCutoff
, &cutoffPage
);
1497 * Delete an individual SLRU segment.
1499 * NB: This does not touch the SLRU buffers themselves, callers have to ensure
1500 * they either can't yet contain anything, or have already been cleaned out.
1503 SlruInternalDeleteSegment(SlruCtl ctl
, int64 segno
)
1505 char path
[MAXPGPATH
];
1507 /* Forget any fsync requests queued for this segment. */
1508 if (ctl
->sync_handler
!= SYNC_HANDLER_NONE
)
1512 INIT_SLRUFILETAG(tag
, ctl
->sync_handler
, segno
);
1513 RegisterSyncRequest(&tag
, SYNC_FORGET_REQUEST
, true);
1516 /* Unlink the file. */
1517 SlruFileName(ctl
, path
, segno
);
1518 ereport(DEBUG2
, (errmsg_internal("removing file \"%s\"", path
)));
1523 * Delete an individual SLRU segment, identified by the segment number.
1526 SlruDeleteSegment(SlruCtl ctl
, int64 segno
)
1528 SlruShared shared
= ctl
->shared
;
1529 int prevbank
= SlotGetBankNumber(0);
1532 /* Clean out any possibly existing references to the segment. */
1533 LWLockAcquire(&shared
->bank_locks
[prevbank
].lock
, LW_EXCLUSIVE
);
1536 for (int slotno
= 0; slotno
< shared
->num_slots
; slotno
++)
1539 int curbank
= SlotGetBankNumber(slotno
);
1542 * If the current bank lock is not same as the previous bank lock then
1543 * release the previous lock and acquire the new lock.
1545 if (curbank
!= prevbank
)
1547 LWLockRelease(&shared
->bank_locks
[prevbank
].lock
);
1548 LWLockAcquire(&shared
->bank_locks
[curbank
].lock
, LW_EXCLUSIVE
);
1552 if (shared
->page_status
[slotno
] == SLRU_PAGE_EMPTY
)
1555 pagesegno
= shared
->page_number
[slotno
] / SLRU_PAGES_PER_SEGMENT
;
1556 /* not the segment we're looking for */
1557 if (pagesegno
!= segno
)
1560 /* If page is clean, just change state to EMPTY (expected case). */
1561 if (shared
->page_status
[slotno
] == SLRU_PAGE_VALID
&&
1562 !shared
->page_dirty
[slotno
])
1564 shared
->page_status
[slotno
] = SLRU_PAGE_EMPTY
;
1568 /* Same logic as SimpleLruTruncate() */
1569 if (shared
->page_status
[slotno
] == SLRU_PAGE_VALID
)
1570 SlruInternalWritePage(ctl
, slotno
, NULL
);
1572 SimpleLruWaitIO(ctl
, slotno
);
1578 * Be extra careful and re-check. The IO functions release the control
1579 * lock, so new pages could have been read in.
1584 SlruInternalDeleteSegment(ctl
, segno
);
1586 LWLockRelease(&shared
->bank_locks
[prevbank
].lock
);
1590 * Determine whether a segment is okay to delete.
1592 * segpage is the first page of the segment, and cutoffPage is the oldest (in
1593 * PagePrecedes order) page in the SLRU containing still-useful data. Since
1594 * every core PagePrecedes callback implements "wrap around", check the
1595 * segment's first and last pages:
1597 * first<cutoff && last<cutoff: yes
1598 * first<cutoff && last>=cutoff: no; cutoff falls inside this segment
1599 * first>=cutoff && last<cutoff: no; wrap point falls inside this segment
1600 * first>=cutoff && last>=cutoff: no; every page of this segment is too young
1603 SlruMayDeleteSegment(SlruCtl ctl
, int64 segpage
, int64 cutoffPage
)
1605 int64 seg_last_page
= segpage
+ SLRU_PAGES_PER_SEGMENT
- 1;
1607 Assert(segpage
% SLRU_PAGES_PER_SEGMENT
== 0);
1609 return (ctl
->PagePrecedes(segpage
, cutoffPage
) &&
1610 ctl
->PagePrecedes(seg_last_page
, cutoffPage
));
1613 #ifdef USE_ASSERT_CHECKING
1615 SlruPagePrecedesTestOffset(SlruCtl ctl
, int per_page
, uint32 offset
)
1621 TransactionId newestXact
,
1625 * Compare an XID pair having undefined order (see RFC 1982), a pair at
1626 * "opposite ends" of the XID space. TransactionIdPrecedes() treats each
1627 * as preceding the other. If RHS is oldestXact, LHS is the first XID we
1630 lhs
= per_page
+ offset
; /* skip first page to avoid non-normal XIDs */
1631 rhs
= lhs
+ (1U << 31);
1632 Assert(TransactionIdPrecedes(lhs
, rhs
));
1633 Assert(TransactionIdPrecedes(rhs
, lhs
));
1634 Assert(!TransactionIdPrecedes(lhs
- 1, rhs
));
1635 Assert(TransactionIdPrecedes(rhs
, lhs
- 1));
1636 Assert(TransactionIdPrecedes(lhs
+ 1, rhs
));
1637 Assert(!TransactionIdPrecedes(rhs
, lhs
+ 1));
1638 Assert(!TransactionIdFollowsOrEquals(lhs
, rhs
));
1639 Assert(!TransactionIdFollowsOrEquals(rhs
, lhs
));
1640 Assert(!ctl
->PagePrecedes(lhs
/ per_page
, lhs
/ per_page
));
1641 Assert(!ctl
->PagePrecedes(lhs
/ per_page
, rhs
/ per_page
));
1642 Assert(!ctl
->PagePrecedes(rhs
/ per_page
, lhs
/ per_page
));
1643 Assert(!ctl
->PagePrecedes((lhs
- per_page
) / per_page
, rhs
/ per_page
));
1644 Assert(ctl
->PagePrecedes(rhs
/ per_page
, (lhs
- 3 * per_page
) / per_page
));
1645 Assert(ctl
->PagePrecedes(rhs
/ per_page
, (lhs
- 2 * per_page
) / per_page
));
1646 Assert(ctl
->PagePrecedes(rhs
/ per_page
, (lhs
- 1 * per_page
) / per_page
)
1647 || (1U << 31) % per_page
!= 0); /* See CommitTsPagePrecedes() */
1648 Assert(ctl
->PagePrecedes((lhs
+ 1 * per_page
) / per_page
, rhs
/ per_page
)
1649 || (1U << 31) % per_page
!= 0);
1650 Assert(ctl
->PagePrecedes((lhs
+ 2 * per_page
) / per_page
, rhs
/ per_page
));
1651 Assert(ctl
->PagePrecedes((lhs
+ 3 * per_page
) / per_page
, rhs
/ per_page
));
1652 Assert(!ctl
->PagePrecedes(rhs
/ per_page
, (lhs
+ per_page
) / per_page
));
1655 * GetNewTransactionId() has assigned the last XID it can safely use, and
1656 * that XID is in the *LAST* page of the second segment. We must not
1657 * delete that segment.
1659 newestPage
= 2 * SLRU_PAGES_PER_SEGMENT
- 1;
1660 newestXact
= newestPage
* per_page
+ offset
;
1661 Assert(newestXact
/ per_page
== newestPage
);
1662 oldestXact
= newestXact
+ 1;
1663 oldestXact
-= 1U << 31;
1664 oldestPage
= oldestXact
/ per_page
;
1665 Assert(!SlruMayDeleteSegment(ctl
,
1667 newestPage
% SLRU_PAGES_PER_SEGMENT
),
1671 * GetNewTransactionId() has assigned the last XID it can safely use, and
1672 * that XID is in the *FIRST* page of the second segment. We must not
1673 * delete that segment.
1675 newestPage
= SLRU_PAGES_PER_SEGMENT
;
1676 newestXact
= newestPage
* per_page
+ offset
;
1677 Assert(newestXact
/ per_page
== newestPage
);
1678 oldestXact
= newestXact
+ 1;
1679 oldestXact
-= 1U << 31;
1680 oldestPage
= oldestXact
/ per_page
;
1681 Assert(!SlruMayDeleteSegment(ctl
,
1683 newestPage
% SLRU_PAGES_PER_SEGMENT
),
1688 * Unit-test a PagePrecedes function.
1690 * This assumes every uint32 >= FirstNormalTransactionId is a valid key. It
1691 * assumes each value occupies a contiguous, fixed-size region of SLRU bytes.
1692 * (MultiXactMemberCtl separates flags from XIDs. NotifyCtl has
1693 * variable-length entries, no keys, and no random access. These unit tests
1694 * do not apply to them.)
1697 SlruPagePrecedesUnitTests(SlruCtl ctl
, int per_page
)
1699 /* Test first, middle and last entries of a page. */
1700 SlruPagePrecedesTestOffset(ctl
, per_page
, 0);
1701 SlruPagePrecedesTestOffset(ctl
, per_page
, per_page
/ 2);
1702 SlruPagePrecedesTestOffset(ctl
, per_page
, per_page
- 1);
1707 * SlruScanDirectory callback
1708 * This callback reports true if there's any segment wholly prior to the
1709 * one containing the page passed as "data".
1712 SlruScanDirCbReportPresence(SlruCtl ctl
, char *filename
, int64 segpage
,
1715 int64 cutoffPage
= *(int64
*) data
;
1717 if (SlruMayDeleteSegment(ctl
, segpage
, cutoffPage
))
1718 return true; /* found one; don't iterate any more */
1720 return false; /* keep going */
1724 * SlruScanDirectory callback.
1725 * This callback deletes segments prior to the one passed in as "data".
1728 SlruScanDirCbDeleteCutoff(SlruCtl ctl
, char *filename
, int64 segpage
,
1731 int64 cutoffPage
= *(int64
*) data
;
1733 if (SlruMayDeleteSegment(ctl
, segpage
, cutoffPage
))
1734 SlruInternalDeleteSegment(ctl
, segpage
/ SLRU_PAGES_PER_SEGMENT
);
1736 return false; /* keep going */
1740 * SlruScanDirectory callback.
1741 * This callback deletes all segments.
1744 SlruScanDirCbDeleteAll(SlruCtl ctl
, char *filename
, int64 segpage
, void *data
)
1746 SlruInternalDeleteSegment(ctl
, segpage
/ SLRU_PAGES_PER_SEGMENT
);
1748 return false; /* keep going */
1752 * An internal function used by SlruScanDirectory().
1754 * Returns true if a file with a name of a given length may be a correct
1758 SlruCorrectSegmentFilenameLength(SlruCtl ctl
, size_t len
)
1760 if (ctl
->long_segment_names
)
1761 return (len
== 15); /* see SlruFileName() */
1765 * Commit 638cf09e76d allowed 5-character lengths. Later commit
1766 * 73c986adde5 allowed 6-character length.
1768 * Note: There is an ongoing plan to migrate all SLRUs to 64-bit page
1769 * numbers, and the corresponding 15-character file names, which may
1770 * eventually deprecate the support for 4, 5, and 6-character names.
1772 return (len
== 4 || len
== 5 || len
== 6);
1776 * Scan the SimpleLru directory and apply a callback to each file found in it.
1778 * If the callback returns true, the scan is stopped. The last return value
1779 * from the callback is returned.
1781 * The callback receives the following arguments: 1. the SlruCtl struct for the
1782 * slru being truncated; 2. the filename being considered; 3. the page number
1783 * for the first page of that file; 4. a pointer to the opaque data given to us
1786 * Note that the ordering in which the directory is scanned is not guaranteed.
1788 * Note that no locking is applied.
1791 SlruScanDirectory(SlruCtl ctl
, SlruScanCallback callback
, void *data
)
1793 bool retval
= false;
1795 struct dirent
*clde
;
1799 cldir
= AllocateDir(ctl
->Dir
);
1800 while ((clde
= ReadDir(cldir
, ctl
->Dir
)) != NULL
)
1804 len
= strlen(clde
->d_name
);
1806 if (SlruCorrectSegmentFilenameLength(ctl
, len
) &&
1807 strspn(clde
->d_name
, "0123456789ABCDEF") == len
)
1809 segno
= strtoi64(clde
->d_name
, NULL
, 16);
1810 segpage
= segno
* SLRU_PAGES_PER_SEGMENT
;
1812 elog(DEBUG2
, "SlruScanDirectory invoking callback on %s/%s",
1813 ctl
->Dir
, clde
->d_name
);
1814 retval
= callback(ctl
, clde
->d_name
, segpage
, data
);
1825 * Individual SLRUs (clog, ...) have to provide a sync.c handler function so
1826 * that they can provide the correct "SlruCtl" (otherwise we don't know how to
1827 * build the path), but they just forward to this common implementation that
1828 * performs the fsync.
1831 SlruSyncFileTag(SlruCtl ctl
, const FileTag
*ftag
, char *path
)
1837 SlruFileName(ctl
, path
, ftag
->segno
);
1839 fd
= OpenTransientFile(path
, O_RDWR
| PG_BINARY
);
1843 pgstat_report_wait_start(WAIT_EVENT_SLRU_FLUSH_SYNC
);
1844 result
= pg_fsync(fd
);
1845 pgstat_report_wait_end();
1848 CloseTransientFile(fd
);