Repair ALTER EXTENSION ... SET SCHEMA.
[pgsql.git] / src / backend / commands / trigger.c
blob35eb7180f7eb9d2860d9287d33eb5e889b0d9d93
1 /*-------------------------------------------------------------------------
3 * trigger.c
4 * PostgreSQL TRIGGERs support code.
6 * Portions Copyright (c) 1996-2024, PostgreSQL Global Development Group
7 * Portions Copyright (c) 1994, Regents of the University of California
9 * IDENTIFICATION
10 * src/backend/commands/trigger.c
12 *-------------------------------------------------------------------------
14 #include "postgres.h"
16 #include "access/genam.h"
17 #include "access/htup_details.h"
18 #include "access/relation.h"
19 #include "access/sysattr.h"
20 #include "access/table.h"
21 #include "access/tableam.h"
22 #include "access/xact.h"
23 #include "catalog/catalog.h"
24 #include "catalog/dependency.h"
25 #include "catalog/indexing.h"
26 #include "catalog/objectaccess.h"
27 #include "catalog/partition.h"
28 #include "catalog/pg_constraint.h"
29 #include "catalog/pg_inherits.h"
30 #include "catalog/pg_proc.h"
31 #include "catalog/pg_trigger.h"
32 #include "catalog/pg_type.h"
33 #include "commands/dbcommands.h"
34 #include "commands/trigger.h"
35 #include "executor/executor.h"
36 #include "miscadmin.h"
37 #include "nodes/bitmapset.h"
38 #include "nodes/makefuncs.h"
39 #include "optimizer/optimizer.h"
40 #include "parser/parse_clause.h"
41 #include "parser/parse_collate.h"
42 #include "parser/parse_func.h"
43 #include "parser/parse_relation.h"
44 #include "partitioning/partdesc.h"
45 #include "pgstat.h"
46 #include "rewrite/rewriteManip.h"
47 #include "storage/lmgr.h"
48 #include "utils/acl.h"
49 #include "utils/builtins.h"
50 #include "utils/fmgroids.h"
51 #include "utils/guc_hooks.h"
52 #include "utils/inval.h"
53 #include "utils/lsyscache.h"
54 #include "utils/memutils.h"
55 #include "utils/plancache.h"
56 #include "utils/rel.h"
57 #include "utils/snapmgr.h"
58 #include "utils/syscache.h"
59 #include "utils/tuplestore.h"
62 /* GUC variables */
63 int SessionReplicationRole = SESSION_REPLICATION_ROLE_ORIGIN;
65 /* How many levels deep into trigger execution are we? */
66 static int MyTriggerDepth = 0;
68 /* Local function prototypes */
69 static void renametrig_internal(Relation tgrel, Relation targetrel,
70 HeapTuple trigtup, const char *newname,
71 const char *expected_name);
72 static void renametrig_partition(Relation tgrel, Oid partitionId,
73 Oid parentTriggerOid, const char *newname,
74 const char *expected_name);
75 static void SetTriggerFlags(TriggerDesc *trigdesc, Trigger *trigger);
76 static bool GetTupleForTrigger(EState *estate,
77 EPQState *epqstate,
78 ResultRelInfo *relinfo,
79 ItemPointer tid,
80 LockTupleMode lockmode,
81 TupleTableSlot *oldslot,
82 TupleTableSlot **epqslot,
83 TM_Result *tmresultp,
84 TM_FailureData *tmfdp);
85 static bool TriggerEnabled(EState *estate, ResultRelInfo *relinfo,
86 Trigger *trigger, TriggerEvent event,
87 Bitmapset *modifiedCols,
88 TupleTableSlot *oldslot, TupleTableSlot *newslot);
89 static HeapTuple ExecCallTriggerFunc(TriggerData *trigdata,
90 int tgindx,
91 FmgrInfo *finfo,
92 Instrumentation *instr,
93 MemoryContext per_tuple_context);
94 static void AfterTriggerSaveEvent(EState *estate, ResultRelInfo *relinfo,
95 ResultRelInfo *src_partinfo,
96 ResultRelInfo *dst_partinfo,
97 int event, bool row_trigger,
98 TupleTableSlot *oldslot, TupleTableSlot *newslot,
99 List *recheckIndexes, Bitmapset *modifiedCols,
100 TransitionCaptureState *transition_capture,
101 bool is_crosspart_update);
102 static void AfterTriggerEnlargeQueryState(void);
103 static bool before_stmt_triggers_fired(Oid relid, CmdType cmdType);
107 * Create a trigger. Returns the address of the created trigger.
109 * queryString is the source text of the CREATE TRIGGER command.
110 * This must be supplied if a whenClause is specified, else it can be NULL.
112 * relOid, if nonzero, is the relation on which the trigger should be
113 * created. If zero, the name provided in the statement will be looked up.
115 * refRelOid, if nonzero, is the relation to which the constraint trigger
116 * refers. If zero, the constraint relation name provided in the statement
117 * will be looked up as needed.
119 * constraintOid, if nonzero, says that this trigger is being created
120 * internally to implement that constraint. A suitable pg_depend entry will
121 * be made to link the trigger to that constraint. constraintOid is zero when
122 * executing a user-entered CREATE TRIGGER command. (For CREATE CONSTRAINT
123 * TRIGGER, we build a pg_constraint entry internally.)
125 * indexOid, if nonzero, is the OID of an index associated with the constraint.
126 * We do nothing with this except store it into pg_trigger.tgconstrindid;
127 * but when creating a trigger for a deferrable unique constraint on a
128 * partitioned table, its children are looked up. Note we don't cope with
129 * invalid indexes in that case.
131 * funcoid, if nonzero, is the OID of the function to invoke. When this is
132 * given, stmt->funcname is ignored.
134 * parentTriggerOid, if nonzero, is a trigger that begets this one; so that
135 * if that trigger is dropped, this one should be too. There are two cases
136 * when a nonzero value is passed for this: 1) when this function recurses to
137 * create the trigger on partitions, 2) when creating child foreign key
138 * triggers; see CreateFKCheckTrigger() and createForeignKeyActionTriggers().
140 * If whenClause is passed, it is an already-transformed expression for
141 * WHEN. In this case, we ignore any that may come in stmt->whenClause.
143 * If isInternal is true then this is an internally-generated trigger.
144 * This argument sets the tgisinternal field of the pg_trigger entry, and
145 * if true causes us to modify the given trigger name to ensure uniqueness.
147 * When isInternal is not true we require ACL_TRIGGER permissions on the
148 * relation, as well as ACL_EXECUTE on the trigger function. For internal
149 * triggers the caller must apply any required permission checks.
151 * When called on partitioned tables, this function recurses to create the
152 * trigger on all the partitions, except if isInternal is true, in which
153 * case caller is expected to execute recursion on its own. in_partition
154 * indicates such a recursive call; outside callers should pass "false"
155 * (but see CloneRowTriggersToPartition).
157 ObjectAddress
158 CreateTrigger(CreateTrigStmt *stmt, const char *queryString,
159 Oid relOid, Oid refRelOid, Oid constraintOid, Oid indexOid,
160 Oid funcoid, Oid parentTriggerOid, Node *whenClause,
161 bool isInternal, bool in_partition)
163 return
164 CreateTriggerFiringOn(stmt, queryString, relOid, refRelOid,
165 constraintOid, indexOid, funcoid,
166 parentTriggerOid, whenClause, isInternal,
167 in_partition, TRIGGER_FIRES_ON_ORIGIN);
171 * Like the above; additionally the firing condition
172 * (always/origin/replica/disabled) can be specified.
174 ObjectAddress
175 CreateTriggerFiringOn(CreateTrigStmt *stmt, const char *queryString,
176 Oid relOid, Oid refRelOid, Oid constraintOid,
177 Oid indexOid, Oid funcoid, Oid parentTriggerOid,
178 Node *whenClause, bool isInternal, bool in_partition,
179 char trigger_fires_when)
181 int16 tgtype;
182 int ncolumns;
183 int16 *columns;
184 int2vector *tgattr;
185 List *whenRtable;
186 char *qual;
187 Datum values[Natts_pg_trigger];
188 bool nulls[Natts_pg_trigger];
189 Relation rel;
190 AclResult aclresult;
191 Relation tgrel;
192 Relation pgrel;
193 HeapTuple tuple = NULL;
194 Oid funcrettype;
195 Oid trigoid = InvalidOid;
196 char internaltrigname[NAMEDATALEN];
197 char *trigname;
198 Oid constrrelid = InvalidOid;
199 ObjectAddress myself,
200 referenced;
201 char *oldtablename = NULL;
202 char *newtablename = NULL;
203 bool partition_recurse;
204 bool trigger_exists = false;
205 Oid existing_constraint_oid = InvalidOid;
206 bool existing_isInternal = false;
207 bool existing_isClone = false;
209 if (OidIsValid(relOid))
210 rel = table_open(relOid, ShareRowExclusiveLock);
211 else
212 rel = table_openrv(stmt->relation, ShareRowExclusiveLock);
215 * Triggers must be on tables or views, and there are additional
216 * relation-type-specific restrictions.
218 if (rel->rd_rel->relkind == RELKIND_RELATION)
220 /* Tables can't have INSTEAD OF triggers */
221 if (stmt->timing != TRIGGER_TYPE_BEFORE &&
222 stmt->timing != TRIGGER_TYPE_AFTER)
223 ereport(ERROR,
224 (errcode(ERRCODE_WRONG_OBJECT_TYPE),
225 errmsg("\"%s\" is a table",
226 RelationGetRelationName(rel)),
227 errdetail("Tables cannot have INSTEAD OF triggers.")));
229 else if (rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
231 /* Partitioned tables can't have INSTEAD OF triggers */
232 if (stmt->timing != TRIGGER_TYPE_BEFORE &&
233 stmt->timing != TRIGGER_TYPE_AFTER)
234 ereport(ERROR,
235 (errcode(ERRCODE_WRONG_OBJECT_TYPE),
236 errmsg("\"%s\" is a table",
237 RelationGetRelationName(rel)),
238 errdetail("Tables cannot have INSTEAD OF triggers.")));
241 * FOR EACH ROW triggers have further restrictions
243 if (stmt->row)
246 * Disallow use of transition tables.
248 * Note that we have another restriction about transition tables
249 * in partitions; search for 'has_superclass' below for an
250 * explanation. The check here is just to protect from the fact
251 * that if we allowed it here, the creation would succeed for a
252 * partitioned table with no partitions, but would be blocked by
253 * the other restriction when the first partition was created,
254 * which is very unfriendly behavior.
256 if (stmt->transitionRels != NIL)
257 ereport(ERROR,
258 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
259 errmsg("\"%s\" is a partitioned table",
260 RelationGetRelationName(rel)),
261 errdetail("ROW triggers with transition tables are not supported on partitioned tables.")));
264 else if (rel->rd_rel->relkind == RELKIND_VIEW)
267 * Views can have INSTEAD OF triggers (which we check below are
268 * row-level), or statement-level BEFORE/AFTER triggers.
270 if (stmt->timing != TRIGGER_TYPE_INSTEAD && stmt->row)
271 ereport(ERROR,
272 (errcode(ERRCODE_WRONG_OBJECT_TYPE),
273 errmsg("\"%s\" is a view",
274 RelationGetRelationName(rel)),
275 errdetail("Views cannot have row-level BEFORE or AFTER triggers.")));
276 /* Disallow TRUNCATE triggers on VIEWs */
277 if (TRIGGER_FOR_TRUNCATE(stmt->events))
278 ereport(ERROR,
279 (errcode(ERRCODE_WRONG_OBJECT_TYPE),
280 errmsg("\"%s\" is a view",
281 RelationGetRelationName(rel)),
282 errdetail("Views cannot have TRUNCATE triggers.")));
284 else if (rel->rd_rel->relkind == RELKIND_FOREIGN_TABLE)
286 if (stmt->timing != TRIGGER_TYPE_BEFORE &&
287 stmt->timing != TRIGGER_TYPE_AFTER)
288 ereport(ERROR,
289 (errcode(ERRCODE_WRONG_OBJECT_TYPE),
290 errmsg("\"%s\" is a foreign table",
291 RelationGetRelationName(rel)),
292 errdetail("Foreign tables cannot have INSTEAD OF triggers.")));
295 * We disallow constraint triggers to protect the assumption that
296 * triggers on FKs can't be deferred. See notes with AfterTriggers
297 * data structures, below.
299 if (stmt->isconstraint)
300 ereport(ERROR,
301 (errcode(ERRCODE_WRONG_OBJECT_TYPE),
302 errmsg("\"%s\" is a foreign table",
303 RelationGetRelationName(rel)),
304 errdetail("Foreign tables cannot have constraint triggers.")));
306 else
307 ereport(ERROR,
308 (errcode(ERRCODE_WRONG_OBJECT_TYPE),
309 errmsg("relation \"%s\" cannot have triggers",
310 RelationGetRelationName(rel)),
311 errdetail_relkind_not_supported(rel->rd_rel->relkind)));
313 if (!allowSystemTableMods && IsSystemRelation(rel))
314 ereport(ERROR,
315 (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
316 errmsg("permission denied: \"%s\" is a system catalog",
317 RelationGetRelationName(rel))));
319 if (stmt->isconstraint)
322 * We must take a lock on the target relation to protect against
323 * concurrent drop. It's not clear that AccessShareLock is strong
324 * enough, but we certainly need at least that much... otherwise, we
325 * might end up creating a pg_constraint entry referencing a
326 * nonexistent table.
328 if (OidIsValid(refRelOid))
330 LockRelationOid(refRelOid, AccessShareLock);
331 constrrelid = refRelOid;
333 else if (stmt->constrrel != NULL)
334 constrrelid = RangeVarGetRelid(stmt->constrrel, AccessShareLock,
335 false);
338 /* permission checks */
339 if (!isInternal)
341 aclresult = pg_class_aclcheck(RelationGetRelid(rel), GetUserId(),
342 ACL_TRIGGER);
343 if (aclresult != ACLCHECK_OK)
344 aclcheck_error(aclresult, get_relkind_objtype(rel->rd_rel->relkind),
345 RelationGetRelationName(rel));
347 if (OidIsValid(constrrelid))
349 aclresult = pg_class_aclcheck(constrrelid, GetUserId(),
350 ACL_TRIGGER);
351 if (aclresult != ACLCHECK_OK)
352 aclcheck_error(aclresult, get_relkind_objtype(get_rel_relkind(constrrelid)),
353 get_rel_name(constrrelid));
358 * When called on a partitioned table to create a FOR EACH ROW trigger
359 * that's not internal, we create one trigger for each partition, too.
361 * For that, we'd better hold lock on all of them ahead of time.
363 partition_recurse = !isInternal && stmt->row &&
364 rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE;
365 if (partition_recurse)
366 list_free(find_all_inheritors(RelationGetRelid(rel),
367 ShareRowExclusiveLock, NULL));
369 /* Compute tgtype */
370 TRIGGER_CLEAR_TYPE(tgtype);
371 if (stmt->row)
372 TRIGGER_SETT_ROW(tgtype);
373 tgtype |= stmt->timing;
374 tgtype |= stmt->events;
376 /* Disallow ROW-level TRUNCATE triggers */
377 if (TRIGGER_FOR_ROW(tgtype) && TRIGGER_FOR_TRUNCATE(tgtype))
378 ereport(ERROR,
379 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
380 errmsg("TRUNCATE FOR EACH ROW triggers are not supported")));
382 /* INSTEAD triggers must be row-level, and can't have WHEN or columns */
383 if (TRIGGER_FOR_INSTEAD(tgtype))
385 if (!TRIGGER_FOR_ROW(tgtype))
386 ereport(ERROR,
387 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
388 errmsg("INSTEAD OF triggers must be FOR EACH ROW")));
389 if (stmt->whenClause)
390 ereport(ERROR,
391 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
392 errmsg("INSTEAD OF triggers cannot have WHEN conditions")));
393 if (stmt->columns != NIL)
394 ereport(ERROR,
395 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
396 errmsg("INSTEAD OF triggers cannot have column lists")));
400 * We don't yet support naming ROW transition variables, but the parser
401 * recognizes the syntax so we can give a nicer message here.
403 * Per standard, REFERENCING TABLE names are only allowed on AFTER
404 * triggers. Per standard, REFERENCING ROW names are not allowed with FOR
405 * EACH STATEMENT. Per standard, each OLD/NEW, ROW/TABLE permutation is
406 * only allowed once. Per standard, OLD may not be specified when
407 * creating a trigger only for INSERT, and NEW may not be specified when
408 * creating a trigger only for DELETE.
410 * Notice that the standard allows an AFTER ... FOR EACH ROW trigger to
411 * reference both ROW and TABLE transition data.
413 if (stmt->transitionRels != NIL)
415 List *varList = stmt->transitionRels;
416 ListCell *lc;
418 foreach(lc, varList)
420 TriggerTransition *tt = lfirst_node(TriggerTransition, lc);
422 if (!(tt->isTable))
423 ereport(ERROR,
424 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
425 errmsg("ROW variable naming in the REFERENCING clause is not supported"),
426 errhint("Use OLD TABLE or NEW TABLE for naming transition tables.")));
429 * Because of the above test, we omit further ROW-related testing
430 * below. If we later allow naming OLD and NEW ROW variables,
431 * adjustments will be needed below.
434 if (rel->rd_rel->relkind == RELKIND_FOREIGN_TABLE)
435 ereport(ERROR,
436 (errcode(ERRCODE_WRONG_OBJECT_TYPE),
437 errmsg("\"%s\" is a foreign table",
438 RelationGetRelationName(rel)),
439 errdetail("Triggers on foreign tables cannot have transition tables.")));
441 if (rel->rd_rel->relkind == RELKIND_VIEW)
442 ereport(ERROR,
443 (errcode(ERRCODE_WRONG_OBJECT_TYPE),
444 errmsg("\"%s\" is a view",
445 RelationGetRelationName(rel)),
446 errdetail("Triggers on views cannot have transition tables.")));
449 * We currently don't allow row-level triggers with transition
450 * tables on partition or inheritance children. Such triggers
451 * would somehow need to see tuples converted to the format of the
452 * table they're attached to, and it's not clear which subset of
453 * tuples each child should see. See also the prohibitions in
454 * ATExecAttachPartition() and ATExecAddInherit().
456 if (TRIGGER_FOR_ROW(tgtype) && has_superclass(rel->rd_id))
458 /* Use appropriate error message. */
459 if (rel->rd_rel->relispartition)
460 ereport(ERROR,
461 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
462 errmsg("ROW triggers with transition tables are not supported on partitions")));
463 else
464 ereport(ERROR,
465 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
466 errmsg("ROW triggers with transition tables are not supported on inheritance children")));
469 if (stmt->timing != TRIGGER_TYPE_AFTER)
470 ereport(ERROR,
471 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
472 errmsg("transition table name can only be specified for an AFTER trigger")));
474 if (TRIGGER_FOR_TRUNCATE(tgtype))
475 ereport(ERROR,
476 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
477 errmsg("TRUNCATE triggers with transition tables are not supported")));
480 * We currently don't allow multi-event triggers ("INSERT OR
481 * UPDATE") with transition tables, because it's not clear how to
482 * handle INSERT ... ON CONFLICT statements which can fire both
483 * INSERT and UPDATE triggers. We show the inserted tuples to
484 * INSERT triggers and the updated tuples to UPDATE triggers, but
485 * it's not yet clear what INSERT OR UPDATE trigger should see.
486 * This restriction could be lifted if we can decide on the right
487 * semantics in a later release.
489 if (((TRIGGER_FOR_INSERT(tgtype) ? 1 : 0) +
490 (TRIGGER_FOR_UPDATE(tgtype) ? 1 : 0) +
491 (TRIGGER_FOR_DELETE(tgtype) ? 1 : 0)) != 1)
492 ereport(ERROR,
493 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
494 errmsg("transition tables cannot be specified for triggers with more than one event")));
497 * We currently don't allow column-specific triggers with
498 * transition tables. Per spec, that seems to require
499 * accumulating separate transition tables for each combination of
500 * columns, which is a lot of work for a rather marginal feature.
502 if (stmt->columns != NIL)
503 ereport(ERROR,
504 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
505 errmsg("transition tables cannot be specified for triggers with column lists")));
508 * We disallow constraint triggers with transition tables, to
509 * protect the assumption that such triggers can't be deferred.
510 * See notes with AfterTriggers data structures, below.
512 * Currently this is enforced by the grammar, so just Assert here.
514 Assert(!stmt->isconstraint);
516 if (tt->isNew)
518 if (!(TRIGGER_FOR_INSERT(tgtype) ||
519 TRIGGER_FOR_UPDATE(tgtype)))
520 ereport(ERROR,
521 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
522 errmsg("NEW TABLE can only be specified for an INSERT or UPDATE trigger")));
524 if (newtablename != NULL)
525 ereport(ERROR,
526 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
527 errmsg("NEW TABLE cannot be specified multiple times")));
529 newtablename = tt->name;
531 else
533 if (!(TRIGGER_FOR_DELETE(tgtype) ||
534 TRIGGER_FOR_UPDATE(tgtype)))
535 ereport(ERROR,
536 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
537 errmsg("OLD TABLE can only be specified for a DELETE or UPDATE trigger")));
539 if (oldtablename != NULL)
540 ereport(ERROR,
541 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
542 errmsg("OLD TABLE cannot be specified multiple times")));
544 oldtablename = tt->name;
548 if (newtablename != NULL && oldtablename != NULL &&
549 strcmp(newtablename, oldtablename) == 0)
550 ereport(ERROR,
551 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
552 errmsg("OLD TABLE name and NEW TABLE name cannot be the same")));
556 * Parse the WHEN clause, if any and we weren't passed an already
557 * transformed one.
559 * Note that as a side effect, we fill whenRtable when parsing. If we got
560 * an already parsed clause, this does not occur, which is what we want --
561 * no point in adding redundant dependencies below.
563 if (!whenClause && stmt->whenClause)
565 ParseState *pstate;
566 ParseNamespaceItem *nsitem;
567 List *varList;
568 ListCell *lc;
570 /* Set up a pstate to parse with */
571 pstate = make_parsestate(NULL);
572 pstate->p_sourcetext = queryString;
575 * Set up nsitems for OLD and NEW references.
577 * 'OLD' must always have varno equal to 1 and 'NEW' equal to 2.
579 nsitem = addRangeTableEntryForRelation(pstate, rel,
580 AccessShareLock,
581 makeAlias("old", NIL),
582 false, false);
583 addNSItemToQuery(pstate, nsitem, false, true, true);
584 nsitem = addRangeTableEntryForRelation(pstate, rel,
585 AccessShareLock,
586 makeAlias("new", NIL),
587 false, false);
588 addNSItemToQuery(pstate, nsitem, false, true, true);
590 /* Transform expression. Copy to be sure we don't modify original */
591 whenClause = transformWhereClause(pstate,
592 copyObject(stmt->whenClause),
593 EXPR_KIND_TRIGGER_WHEN,
594 "WHEN");
595 /* we have to fix its collations too */
596 assign_expr_collations(pstate, whenClause);
599 * Check for disallowed references to OLD/NEW.
601 * NB: pull_var_clause is okay here only because we don't allow
602 * subselects in WHEN clauses; it would fail to examine the contents
603 * of subselects.
605 varList = pull_var_clause(whenClause, 0);
606 foreach(lc, varList)
608 Var *var = (Var *) lfirst(lc);
610 switch (var->varno)
612 case PRS2_OLD_VARNO:
613 if (!TRIGGER_FOR_ROW(tgtype))
614 ereport(ERROR,
615 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
616 errmsg("statement trigger's WHEN condition cannot reference column values"),
617 parser_errposition(pstate, var->location)));
618 if (TRIGGER_FOR_INSERT(tgtype))
619 ereport(ERROR,
620 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
621 errmsg("INSERT trigger's WHEN condition cannot reference OLD values"),
622 parser_errposition(pstate, var->location)));
623 /* system columns are okay here */
624 break;
625 case PRS2_NEW_VARNO:
626 if (!TRIGGER_FOR_ROW(tgtype))
627 ereport(ERROR,
628 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
629 errmsg("statement trigger's WHEN condition cannot reference column values"),
630 parser_errposition(pstate, var->location)));
631 if (TRIGGER_FOR_DELETE(tgtype))
632 ereport(ERROR,
633 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
634 errmsg("DELETE trigger's WHEN condition cannot reference NEW values"),
635 parser_errposition(pstate, var->location)));
636 if (var->varattno < 0 && TRIGGER_FOR_BEFORE(tgtype))
637 ereport(ERROR,
638 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
639 errmsg("BEFORE trigger's WHEN condition cannot reference NEW system columns"),
640 parser_errposition(pstate, var->location)));
641 if (TRIGGER_FOR_BEFORE(tgtype) &&
642 var->varattno == 0 &&
643 RelationGetDescr(rel)->constr &&
644 RelationGetDescr(rel)->constr->has_generated_stored)
645 ereport(ERROR,
646 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
647 errmsg("BEFORE trigger's WHEN condition cannot reference NEW generated columns"),
648 errdetail("A whole-row reference is used and the table contains generated columns."),
649 parser_errposition(pstate, var->location)));
650 if (TRIGGER_FOR_BEFORE(tgtype) &&
651 var->varattno > 0 &&
652 TupleDescAttr(RelationGetDescr(rel), var->varattno - 1)->attgenerated)
653 ereport(ERROR,
654 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
655 errmsg("BEFORE trigger's WHEN condition cannot reference NEW generated columns"),
656 errdetail("Column \"%s\" is a generated column.",
657 NameStr(TupleDescAttr(RelationGetDescr(rel), var->varattno - 1)->attname)),
658 parser_errposition(pstate, var->location)));
659 break;
660 default:
661 /* can't happen without add_missing_from, so just elog */
662 elog(ERROR, "trigger WHEN condition cannot contain references to other relations");
663 break;
667 /* we'll need the rtable for recordDependencyOnExpr */
668 whenRtable = pstate->p_rtable;
670 qual = nodeToString(whenClause);
672 free_parsestate(pstate);
674 else if (!whenClause)
676 whenClause = NULL;
677 whenRtable = NIL;
678 qual = NULL;
680 else
682 qual = nodeToString(whenClause);
683 whenRtable = NIL;
687 * Find and validate the trigger function.
689 if (!OidIsValid(funcoid))
690 funcoid = LookupFuncName(stmt->funcname, 0, NULL, false);
691 if (!isInternal)
693 aclresult = object_aclcheck(ProcedureRelationId, funcoid, GetUserId(), ACL_EXECUTE);
694 if (aclresult != ACLCHECK_OK)
695 aclcheck_error(aclresult, OBJECT_FUNCTION,
696 NameListToString(stmt->funcname));
698 funcrettype = get_func_rettype(funcoid);
699 if (funcrettype != TRIGGEROID)
700 ereport(ERROR,
701 (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
702 errmsg("function %s must return type %s",
703 NameListToString(stmt->funcname), "trigger")));
706 * Scan pg_trigger to see if there is already a trigger of the same name.
707 * Skip this for internally generated triggers, since we'll modify the
708 * name to be unique below.
710 * NOTE that this is cool only because we have ShareRowExclusiveLock on
711 * the relation, so the trigger set won't be changing underneath us.
713 tgrel = table_open(TriggerRelationId, RowExclusiveLock);
714 if (!isInternal)
716 ScanKeyData skeys[2];
717 SysScanDesc tgscan;
719 ScanKeyInit(&skeys[0],
720 Anum_pg_trigger_tgrelid,
721 BTEqualStrategyNumber, F_OIDEQ,
722 ObjectIdGetDatum(RelationGetRelid(rel)));
724 ScanKeyInit(&skeys[1],
725 Anum_pg_trigger_tgname,
726 BTEqualStrategyNumber, F_NAMEEQ,
727 CStringGetDatum(stmt->trigname));
729 tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
730 NULL, 2, skeys);
732 /* There should be at most one matching tuple */
733 if (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
735 Form_pg_trigger oldtrigger = (Form_pg_trigger) GETSTRUCT(tuple);
737 trigoid = oldtrigger->oid;
738 existing_constraint_oid = oldtrigger->tgconstraint;
739 existing_isInternal = oldtrigger->tgisinternal;
740 existing_isClone = OidIsValid(oldtrigger->tgparentid);
741 trigger_exists = true;
742 /* copy the tuple to use in CatalogTupleUpdate() */
743 tuple = heap_copytuple(tuple);
745 systable_endscan(tgscan);
748 if (!trigger_exists)
750 /* Generate the OID for the new trigger. */
751 trigoid = GetNewOidWithIndex(tgrel, TriggerOidIndexId,
752 Anum_pg_trigger_oid);
754 else
757 * If OR REPLACE was specified, we'll replace the old trigger;
758 * otherwise complain about the duplicate name.
760 if (!stmt->replace)
761 ereport(ERROR,
762 (errcode(ERRCODE_DUPLICATE_OBJECT),
763 errmsg("trigger \"%s\" for relation \"%s\" already exists",
764 stmt->trigname, RelationGetRelationName(rel))));
767 * An internal trigger or a child trigger (isClone) cannot be replaced
768 * by a user-defined trigger. However, skip this test when
769 * in_partition, because then we're recursing from a partitioned table
770 * and the check was made at the parent level.
772 if ((existing_isInternal || existing_isClone) &&
773 !isInternal && !in_partition)
774 ereport(ERROR,
775 (errcode(ERRCODE_DUPLICATE_OBJECT),
776 errmsg("trigger \"%s\" for relation \"%s\" is an internal or a child trigger",
777 stmt->trigname, RelationGetRelationName(rel))));
780 * It is not allowed to replace with a constraint trigger; gram.y
781 * should have enforced this already.
783 Assert(!stmt->isconstraint);
786 * It is not allowed to replace an existing constraint trigger,
787 * either. (The reason for these restrictions is partly that it seems
788 * difficult to deal with pending trigger events in such cases, and
789 * partly that the command might imply changing the constraint's
790 * properties as well, which doesn't seem nice.)
792 if (OidIsValid(existing_constraint_oid))
793 ereport(ERROR,
794 (errcode(ERRCODE_DUPLICATE_OBJECT),
795 errmsg("trigger \"%s\" for relation \"%s\" is a constraint trigger",
796 stmt->trigname, RelationGetRelationName(rel))));
800 * If it's a user-entered CREATE CONSTRAINT TRIGGER command, make a
801 * corresponding pg_constraint entry.
803 if (stmt->isconstraint && !OidIsValid(constraintOid))
805 /* Internal callers should have made their own constraints */
806 Assert(!isInternal);
807 constraintOid = CreateConstraintEntry(stmt->trigname,
808 RelationGetNamespace(rel),
809 CONSTRAINT_TRIGGER,
810 stmt->deferrable,
811 stmt->initdeferred,
812 true,
813 InvalidOid, /* no parent */
814 RelationGetRelid(rel),
815 NULL, /* no conkey */
818 InvalidOid, /* no domain */
819 InvalidOid, /* no index */
820 InvalidOid, /* no foreign key */
821 NULL,
822 NULL,
823 NULL,
824 NULL,
826 ' ',
827 ' ',
828 NULL,
830 ' ',
831 NULL, /* no exclusion */
832 NULL, /* no check constraint */
833 NULL,
834 true, /* islocal */
835 0, /* inhcount */
836 true, /* noinherit */
837 false, /* conperiod */
838 isInternal); /* is_internal */
842 * If trigger is internally generated, modify the provided trigger name to
843 * ensure uniqueness by appending the trigger OID. (Callers will usually
844 * supply a simple constant trigger name in these cases.)
846 if (isInternal)
848 snprintf(internaltrigname, sizeof(internaltrigname),
849 "%s_%u", stmt->trigname, trigoid);
850 trigname = internaltrigname;
852 else
854 /* user-defined trigger; use the specified trigger name as-is */
855 trigname = stmt->trigname;
859 * Build the new pg_trigger tuple.
861 memset(nulls, false, sizeof(nulls));
863 values[Anum_pg_trigger_oid - 1] = ObjectIdGetDatum(trigoid);
864 values[Anum_pg_trigger_tgrelid - 1] = ObjectIdGetDatum(RelationGetRelid(rel));
865 values[Anum_pg_trigger_tgparentid - 1] = ObjectIdGetDatum(parentTriggerOid);
866 values[Anum_pg_trigger_tgname - 1] = DirectFunctionCall1(namein,
867 CStringGetDatum(trigname));
868 values[Anum_pg_trigger_tgfoid - 1] = ObjectIdGetDatum(funcoid);
869 values[Anum_pg_trigger_tgtype - 1] = Int16GetDatum(tgtype);
870 values[Anum_pg_trigger_tgenabled - 1] = trigger_fires_when;
871 values[Anum_pg_trigger_tgisinternal - 1] = BoolGetDatum(isInternal);
872 values[Anum_pg_trigger_tgconstrrelid - 1] = ObjectIdGetDatum(constrrelid);
873 values[Anum_pg_trigger_tgconstrindid - 1] = ObjectIdGetDatum(indexOid);
874 values[Anum_pg_trigger_tgconstraint - 1] = ObjectIdGetDatum(constraintOid);
875 values[Anum_pg_trigger_tgdeferrable - 1] = BoolGetDatum(stmt->deferrable);
876 values[Anum_pg_trigger_tginitdeferred - 1] = BoolGetDatum(stmt->initdeferred);
878 if (stmt->args)
880 ListCell *le;
881 char *args;
882 int16 nargs = list_length(stmt->args);
883 int len = 0;
885 foreach(le, stmt->args)
887 char *ar = strVal(lfirst(le));
889 len += strlen(ar) + 4;
890 for (; *ar; ar++)
892 if (*ar == '\\')
893 len++;
896 args = (char *) palloc(len + 1);
897 args[0] = '\0';
898 foreach(le, stmt->args)
900 char *s = strVal(lfirst(le));
901 char *d = args + strlen(args);
903 while (*s)
905 if (*s == '\\')
906 *d++ = '\\';
907 *d++ = *s++;
909 strcpy(d, "\\000");
911 values[Anum_pg_trigger_tgnargs - 1] = Int16GetDatum(nargs);
912 values[Anum_pg_trigger_tgargs - 1] = DirectFunctionCall1(byteain,
913 CStringGetDatum(args));
915 else
917 values[Anum_pg_trigger_tgnargs - 1] = Int16GetDatum(0);
918 values[Anum_pg_trigger_tgargs - 1] = DirectFunctionCall1(byteain,
919 CStringGetDatum(""));
922 /* build column number array if it's a column-specific trigger */
923 ncolumns = list_length(stmt->columns);
924 if (ncolumns == 0)
925 columns = NULL;
926 else
928 ListCell *cell;
929 int i = 0;
931 columns = (int16 *) palloc(ncolumns * sizeof(int16));
932 foreach(cell, stmt->columns)
934 char *name = strVal(lfirst(cell));
935 int16 attnum;
936 int j;
938 /* Lookup column name. System columns are not allowed */
939 attnum = attnameAttNum(rel, name, false);
940 if (attnum == InvalidAttrNumber)
941 ereport(ERROR,
942 (errcode(ERRCODE_UNDEFINED_COLUMN),
943 errmsg("column \"%s\" of relation \"%s\" does not exist",
944 name, RelationGetRelationName(rel))));
946 /* Check for duplicates */
947 for (j = i - 1; j >= 0; j--)
949 if (columns[j] == attnum)
950 ereport(ERROR,
951 (errcode(ERRCODE_DUPLICATE_COLUMN),
952 errmsg("column \"%s\" specified more than once",
953 name)));
956 columns[i++] = attnum;
959 tgattr = buildint2vector(columns, ncolumns);
960 values[Anum_pg_trigger_tgattr - 1] = PointerGetDatum(tgattr);
962 /* set tgqual if trigger has WHEN clause */
963 if (qual)
964 values[Anum_pg_trigger_tgqual - 1] = CStringGetTextDatum(qual);
965 else
966 nulls[Anum_pg_trigger_tgqual - 1] = true;
968 if (oldtablename)
969 values[Anum_pg_trigger_tgoldtable - 1] = DirectFunctionCall1(namein,
970 CStringGetDatum(oldtablename));
971 else
972 nulls[Anum_pg_trigger_tgoldtable - 1] = true;
973 if (newtablename)
974 values[Anum_pg_trigger_tgnewtable - 1] = DirectFunctionCall1(namein,
975 CStringGetDatum(newtablename));
976 else
977 nulls[Anum_pg_trigger_tgnewtable - 1] = true;
980 * Insert or replace tuple in pg_trigger.
982 if (!trigger_exists)
984 tuple = heap_form_tuple(tgrel->rd_att, values, nulls);
985 CatalogTupleInsert(tgrel, tuple);
987 else
989 HeapTuple newtup;
991 newtup = heap_form_tuple(tgrel->rd_att, values, nulls);
992 CatalogTupleUpdate(tgrel, &tuple->t_self, newtup);
993 heap_freetuple(newtup);
996 heap_freetuple(tuple); /* free either original or new tuple */
997 table_close(tgrel, RowExclusiveLock);
999 pfree(DatumGetPointer(values[Anum_pg_trigger_tgname - 1]));
1000 pfree(DatumGetPointer(values[Anum_pg_trigger_tgargs - 1]));
1001 pfree(DatumGetPointer(values[Anum_pg_trigger_tgattr - 1]));
1002 if (oldtablename)
1003 pfree(DatumGetPointer(values[Anum_pg_trigger_tgoldtable - 1]));
1004 if (newtablename)
1005 pfree(DatumGetPointer(values[Anum_pg_trigger_tgnewtable - 1]));
1008 * Update relation's pg_class entry; if necessary; and if not, send an SI
1009 * message to make other backends (and this one) rebuild relcache entries.
1011 pgrel = table_open(RelationRelationId, RowExclusiveLock);
1012 tuple = SearchSysCacheCopy1(RELOID,
1013 ObjectIdGetDatum(RelationGetRelid(rel)));
1014 if (!HeapTupleIsValid(tuple))
1015 elog(ERROR, "cache lookup failed for relation %u",
1016 RelationGetRelid(rel));
1017 if (!((Form_pg_class) GETSTRUCT(tuple))->relhastriggers)
1019 ((Form_pg_class) GETSTRUCT(tuple))->relhastriggers = true;
1021 CatalogTupleUpdate(pgrel, &tuple->t_self, tuple);
1023 CommandCounterIncrement();
1025 else
1026 CacheInvalidateRelcacheByTuple(tuple);
1028 heap_freetuple(tuple);
1029 table_close(pgrel, RowExclusiveLock);
1032 * If we're replacing a trigger, flush all the old dependencies before
1033 * recording new ones.
1035 if (trigger_exists)
1036 deleteDependencyRecordsFor(TriggerRelationId, trigoid, true);
1039 * Record dependencies for trigger. Always place a normal dependency on
1040 * the function.
1042 myself.classId = TriggerRelationId;
1043 myself.objectId = trigoid;
1044 myself.objectSubId = 0;
1046 referenced.classId = ProcedureRelationId;
1047 referenced.objectId = funcoid;
1048 referenced.objectSubId = 0;
1049 recordDependencyOn(&myself, &referenced, DEPENDENCY_NORMAL);
1051 if (isInternal && OidIsValid(constraintOid))
1054 * Internally-generated trigger for a constraint, so make it an
1055 * internal dependency of the constraint. We can skip depending on
1056 * the relation(s), as there'll be an indirect dependency via the
1057 * constraint.
1059 referenced.classId = ConstraintRelationId;
1060 referenced.objectId = constraintOid;
1061 referenced.objectSubId = 0;
1062 recordDependencyOn(&myself, &referenced, DEPENDENCY_INTERNAL);
1064 else
1067 * User CREATE TRIGGER, so place dependencies. We make trigger be
1068 * auto-dropped if its relation is dropped or if the FK relation is
1069 * dropped. (Auto drop is compatible with our pre-7.3 behavior.)
1071 referenced.classId = RelationRelationId;
1072 referenced.objectId = RelationGetRelid(rel);
1073 referenced.objectSubId = 0;
1074 recordDependencyOn(&myself, &referenced, DEPENDENCY_AUTO);
1076 if (OidIsValid(constrrelid))
1078 referenced.classId = RelationRelationId;
1079 referenced.objectId = constrrelid;
1080 referenced.objectSubId = 0;
1081 recordDependencyOn(&myself, &referenced, DEPENDENCY_AUTO);
1083 /* Not possible to have an index dependency in this case */
1084 Assert(!OidIsValid(indexOid));
1087 * If it's a user-specified constraint trigger, make the constraint
1088 * internally dependent on the trigger instead of vice versa.
1090 if (OidIsValid(constraintOid))
1092 referenced.classId = ConstraintRelationId;
1093 referenced.objectId = constraintOid;
1094 referenced.objectSubId = 0;
1095 recordDependencyOn(&referenced, &myself, DEPENDENCY_INTERNAL);
1099 * If it's a partition trigger, create the partition dependencies.
1101 if (OidIsValid(parentTriggerOid))
1103 ObjectAddressSet(referenced, TriggerRelationId, parentTriggerOid);
1104 recordDependencyOn(&myself, &referenced, DEPENDENCY_PARTITION_PRI);
1105 ObjectAddressSet(referenced, RelationRelationId, RelationGetRelid(rel));
1106 recordDependencyOn(&myself, &referenced, DEPENDENCY_PARTITION_SEC);
1110 /* If column-specific trigger, add normal dependencies on columns */
1111 if (columns != NULL)
1113 int i;
1115 referenced.classId = RelationRelationId;
1116 referenced.objectId = RelationGetRelid(rel);
1117 for (i = 0; i < ncolumns; i++)
1119 referenced.objectSubId = columns[i];
1120 recordDependencyOn(&myself, &referenced, DEPENDENCY_NORMAL);
1125 * If it has a WHEN clause, add dependencies on objects mentioned in the
1126 * expression (eg, functions, as well as any columns used).
1128 if (whenRtable != NIL)
1129 recordDependencyOnExpr(&myself, whenClause, whenRtable,
1130 DEPENDENCY_NORMAL);
1132 /* Post creation hook for new trigger */
1133 InvokeObjectPostCreateHookArg(TriggerRelationId, trigoid, 0,
1134 isInternal);
1137 * Lastly, create the trigger on child relations, if needed.
1139 if (partition_recurse)
1141 PartitionDesc partdesc = RelationGetPartitionDesc(rel, true);
1142 int i;
1143 MemoryContext oldcxt,
1144 perChildCxt;
1146 perChildCxt = AllocSetContextCreate(CurrentMemoryContext,
1147 "part trig clone",
1148 ALLOCSET_SMALL_SIZES);
1151 * We don't currently expect to be called with a valid indexOid. If
1152 * that ever changes then we'll need to write code here to find the
1153 * corresponding child index.
1155 Assert(!OidIsValid(indexOid));
1157 oldcxt = MemoryContextSwitchTo(perChildCxt);
1159 /* Iterate to create the trigger on each existing partition */
1160 for (i = 0; i < partdesc->nparts; i++)
1162 CreateTrigStmt *childStmt;
1163 Relation childTbl;
1164 Node *qual;
1166 childTbl = table_open(partdesc->oids[i], ShareRowExclusiveLock);
1169 * Initialize our fabricated parse node by copying the original
1170 * one, then resetting fields that we pass separately.
1172 childStmt = (CreateTrigStmt *) copyObject(stmt);
1173 childStmt->funcname = NIL;
1174 childStmt->whenClause = NULL;
1176 /* If there is a WHEN clause, create a modified copy of it */
1177 qual = copyObject(whenClause);
1178 qual = (Node *)
1179 map_partition_varattnos((List *) qual, PRS2_OLD_VARNO,
1180 childTbl, rel);
1181 qual = (Node *)
1182 map_partition_varattnos((List *) qual, PRS2_NEW_VARNO,
1183 childTbl, rel);
1185 CreateTriggerFiringOn(childStmt, queryString,
1186 partdesc->oids[i], refRelOid,
1187 InvalidOid, InvalidOid,
1188 funcoid, trigoid, qual,
1189 isInternal, true, trigger_fires_when);
1191 table_close(childTbl, NoLock);
1193 MemoryContextReset(perChildCxt);
1196 MemoryContextSwitchTo(oldcxt);
1197 MemoryContextDelete(perChildCxt);
1200 /* Keep lock on target rel until end of xact */
1201 table_close(rel, NoLock);
1203 return myself;
1207 * TriggerSetParentTrigger
1208 * Set a partition's trigger as child of its parent trigger,
1209 * or remove the linkage if parentTrigId is InvalidOid.
1211 * This updates the constraint's pg_trigger row to show it as inherited, and
1212 * adds PARTITION dependencies to prevent the trigger from being deleted
1213 * on its own. Alternatively, reverse that.
1215 void
1216 TriggerSetParentTrigger(Relation trigRel,
1217 Oid childTrigId,
1218 Oid parentTrigId,
1219 Oid childTableId)
1221 SysScanDesc tgscan;
1222 ScanKeyData skey[1];
1223 Form_pg_trigger trigForm;
1224 HeapTuple tuple,
1225 newtup;
1226 ObjectAddress depender;
1227 ObjectAddress referenced;
1230 * Find the trigger to delete.
1232 ScanKeyInit(&skey[0],
1233 Anum_pg_trigger_oid,
1234 BTEqualStrategyNumber, F_OIDEQ,
1235 ObjectIdGetDatum(childTrigId));
1237 tgscan = systable_beginscan(trigRel, TriggerOidIndexId, true,
1238 NULL, 1, skey);
1240 tuple = systable_getnext(tgscan);
1241 if (!HeapTupleIsValid(tuple))
1242 elog(ERROR, "could not find tuple for trigger %u", childTrigId);
1243 newtup = heap_copytuple(tuple);
1244 trigForm = (Form_pg_trigger) GETSTRUCT(newtup);
1245 if (OidIsValid(parentTrigId))
1247 /* don't allow setting parent for a constraint that already has one */
1248 if (OidIsValid(trigForm->tgparentid))
1249 elog(ERROR, "trigger %u already has a parent trigger",
1250 childTrigId);
1252 trigForm->tgparentid = parentTrigId;
1254 CatalogTupleUpdate(trigRel, &tuple->t_self, newtup);
1256 ObjectAddressSet(depender, TriggerRelationId, childTrigId);
1258 ObjectAddressSet(referenced, TriggerRelationId, parentTrigId);
1259 recordDependencyOn(&depender, &referenced, DEPENDENCY_PARTITION_PRI);
1261 ObjectAddressSet(referenced, RelationRelationId, childTableId);
1262 recordDependencyOn(&depender, &referenced, DEPENDENCY_PARTITION_SEC);
1264 else
1266 trigForm->tgparentid = InvalidOid;
1268 CatalogTupleUpdate(trigRel, &tuple->t_self, newtup);
1270 deleteDependencyRecordsForClass(TriggerRelationId, childTrigId,
1271 TriggerRelationId,
1272 DEPENDENCY_PARTITION_PRI);
1273 deleteDependencyRecordsForClass(TriggerRelationId, childTrigId,
1274 RelationRelationId,
1275 DEPENDENCY_PARTITION_SEC);
1278 heap_freetuple(newtup);
1279 systable_endscan(tgscan);
1284 * Guts of trigger deletion.
1286 void
1287 RemoveTriggerById(Oid trigOid)
1289 Relation tgrel;
1290 SysScanDesc tgscan;
1291 ScanKeyData skey[1];
1292 HeapTuple tup;
1293 Oid relid;
1294 Relation rel;
1296 tgrel = table_open(TriggerRelationId, RowExclusiveLock);
1299 * Find the trigger to delete.
1301 ScanKeyInit(&skey[0],
1302 Anum_pg_trigger_oid,
1303 BTEqualStrategyNumber, F_OIDEQ,
1304 ObjectIdGetDatum(trigOid));
1306 tgscan = systable_beginscan(tgrel, TriggerOidIndexId, true,
1307 NULL, 1, skey);
1309 tup = systable_getnext(tgscan);
1310 if (!HeapTupleIsValid(tup))
1311 elog(ERROR, "could not find tuple for trigger %u", trigOid);
1314 * Open and exclusive-lock the relation the trigger belongs to.
1316 relid = ((Form_pg_trigger) GETSTRUCT(tup))->tgrelid;
1318 rel = table_open(relid, AccessExclusiveLock);
1320 if (rel->rd_rel->relkind != RELKIND_RELATION &&
1321 rel->rd_rel->relkind != RELKIND_VIEW &&
1322 rel->rd_rel->relkind != RELKIND_FOREIGN_TABLE &&
1323 rel->rd_rel->relkind != RELKIND_PARTITIONED_TABLE)
1324 ereport(ERROR,
1325 (errcode(ERRCODE_WRONG_OBJECT_TYPE),
1326 errmsg("relation \"%s\" cannot have triggers",
1327 RelationGetRelationName(rel)),
1328 errdetail_relkind_not_supported(rel->rd_rel->relkind)));
1330 if (!allowSystemTableMods && IsSystemRelation(rel))
1331 ereport(ERROR,
1332 (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
1333 errmsg("permission denied: \"%s\" is a system catalog",
1334 RelationGetRelationName(rel))));
1337 * Delete the pg_trigger tuple.
1339 CatalogTupleDelete(tgrel, &tup->t_self);
1341 systable_endscan(tgscan);
1342 table_close(tgrel, RowExclusiveLock);
1345 * We do not bother to try to determine whether any other triggers remain,
1346 * which would be needed in order to decide whether it's safe to clear the
1347 * relation's relhastriggers. (In any case, there might be a concurrent
1348 * process adding new triggers.) Instead, just force a relcache inval to
1349 * make other backends (and this one too!) rebuild their relcache entries.
1350 * There's no great harm in leaving relhastriggers true even if there are
1351 * no triggers left.
1353 CacheInvalidateRelcache(rel);
1355 /* Keep lock on trigger's rel until end of xact */
1356 table_close(rel, NoLock);
1360 * get_trigger_oid - Look up a trigger by name to find its OID.
1362 * If missing_ok is false, throw an error if trigger not found. If
1363 * true, just return InvalidOid.
1366 get_trigger_oid(Oid relid, const char *trigname, bool missing_ok)
1368 Relation tgrel;
1369 ScanKeyData skey[2];
1370 SysScanDesc tgscan;
1371 HeapTuple tup;
1372 Oid oid;
1375 * Find the trigger, verify permissions, set up object address
1377 tgrel = table_open(TriggerRelationId, AccessShareLock);
1379 ScanKeyInit(&skey[0],
1380 Anum_pg_trigger_tgrelid,
1381 BTEqualStrategyNumber, F_OIDEQ,
1382 ObjectIdGetDatum(relid));
1383 ScanKeyInit(&skey[1],
1384 Anum_pg_trigger_tgname,
1385 BTEqualStrategyNumber, F_NAMEEQ,
1386 CStringGetDatum(trigname));
1388 tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1389 NULL, 2, skey);
1391 tup = systable_getnext(tgscan);
1393 if (!HeapTupleIsValid(tup))
1395 if (!missing_ok)
1396 ereport(ERROR,
1397 (errcode(ERRCODE_UNDEFINED_OBJECT),
1398 errmsg("trigger \"%s\" for table \"%s\" does not exist",
1399 trigname, get_rel_name(relid))));
1400 oid = InvalidOid;
1402 else
1404 oid = ((Form_pg_trigger) GETSTRUCT(tup))->oid;
1407 systable_endscan(tgscan);
1408 table_close(tgrel, AccessShareLock);
1409 return oid;
1413 * Perform permissions and integrity checks before acquiring a relation lock.
1415 static void
1416 RangeVarCallbackForRenameTrigger(const RangeVar *rv, Oid relid, Oid oldrelid,
1417 void *arg)
1419 HeapTuple tuple;
1420 Form_pg_class form;
1422 tuple = SearchSysCache1(RELOID, ObjectIdGetDatum(relid));
1423 if (!HeapTupleIsValid(tuple))
1424 return; /* concurrently dropped */
1425 form = (Form_pg_class) GETSTRUCT(tuple);
1427 /* only tables and views can have triggers */
1428 if (form->relkind != RELKIND_RELATION && form->relkind != RELKIND_VIEW &&
1429 form->relkind != RELKIND_FOREIGN_TABLE &&
1430 form->relkind != RELKIND_PARTITIONED_TABLE)
1431 ereport(ERROR,
1432 (errcode(ERRCODE_WRONG_OBJECT_TYPE),
1433 errmsg("relation \"%s\" cannot have triggers",
1434 rv->relname),
1435 errdetail_relkind_not_supported(form->relkind)));
1437 /* you must own the table to rename one of its triggers */
1438 if (!object_ownercheck(RelationRelationId, relid, GetUserId()))
1439 aclcheck_error(ACLCHECK_NOT_OWNER, get_relkind_objtype(get_rel_relkind(relid)), rv->relname);
1440 if (!allowSystemTableMods && IsSystemClass(relid, form))
1441 ereport(ERROR,
1442 (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
1443 errmsg("permission denied: \"%s\" is a system catalog",
1444 rv->relname)));
1446 ReleaseSysCache(tuple);
1450 * renametrig - changes the name of a trigger on a relation
1452 * trigger name is changed in trigger catalog.
1453 * No record of the previous name is kept.
1455 * get proper relrelation from relation catalog (if not arg)
1456 * scan trigger catalog
1457 * for name conflict (within rel)
1458 * for original trigger (if not arg)
1459 * modify tgname in trigger tuple
1460 * update row in catalog
1462 ObjectAddress
1463 renametrig(RenameStmt *stmt)
1465 Oid tgoid;
1466 Relation targetrel;
1467 Relation tgrel;
1468 HeapTuple tuple;
1469 SysScanDesc tgscan;
1470 ScanKeyData key[2];
1471 Oid relid;
1472 ObjectAddress address;
1475 * Look up name, check permissions, and acquire lock (which we will NOT
1476 * release until end of transaction).
1478 relid = RangeVarGetRelidExtended(stmt->relation, AccessExclusiveLock,
1480 RangeVarCallbackForRenameTrigger,
1481 NULL);
1483 /* Have lock already, so just need to build relcache entry. */
1484 targetrel = relation_open(relid, NoLock);
1487 * On partitioned tables, this operation recurses to partitions. Lock all
1488 * tables upfront.
1490 if (targetrel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
1491 (void) find_all_inheritors(relid, AccessExclusiveLock, NULL);
1493 tgrel = table_open(TriggerRelationId, RowExclusiveLock);
1496 * Search for the trigger to modify.
1498 ScanKeyInit(&key[0],
1499 Anum_pg_trigger_tgrelid,
1500 BTEqualStrategyNumber, F_OIDEQ,
1501 ObjectIdGetDatum(relid));
1502 ScanKeyInit(&key[1],
1503 Anum_pg_trigger_tgname,
1504 BTEqualStrategyNumber, F_NAMEEQ,
1505 PointerGetDatum(stmt->subname));
1506 tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1507 NULL, 2, key);
1508 if (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
1510 Form_pg_trigger trigform;
1512 trigform = (Form_pg_trigger) GETSTRUCT(tuple);
1513 tgoid = trigform->oid;
1516 * If the trigger descends from a trigger on a parent partitioned
1517 * table, reject the rename. We don't allow a trigger in a partition
1518 * to differ in name from that of its parent: that would lead to an
1519 * inconsistency that pg_dump would not reproduce.
1521 if (OidIsValid(trigform->tgparentid))
1522 ereport(ERROR,
1523 errmsg("cannot rename trigger \"%s\" on table \"%s\"",
1524 stmt->subname, RelationGetRelationName(targetrel)),
1525 errhint("Rename the trigger on the partitioned table \"%s\" instead.",
1526 get_rel_name(get_partition_parent(relid, false))));
1529 /* Rename the trigger on this relation ... */
1530 renametrig_internal(tgrel, targetrel, tuple, stmt->newname,
1531 stmt->subname);
1533 /* ... and if it is partitioned, recurse to its partitions */
1534 if (targetrel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
1536 PartitionDesc partdesc = RelationGetPartitionDesc(targetrel, true);
1538 for (int i = 0; i < partdesc->nparts; i++)
1540 Oid partitionId = partdesc->oids[i];
1542 renametrig_partition(tgrel, partitionId, trigform->oid,
1543 stmt->newname, stmt->subname);
1547 else
1549 ereport(ERROR,
1550 (errcode(ERRCODE_UNDEFINED_OBJECT),
1551 errmsg("trigger \"%s\" for table \"%s\" does not exist",
1552 stmt->subname, RelationGetRelationName(targetrel))));
1555 ObjectAddressSet(address, TriggerRelationId, tgoid);
1557 systable_endscan(tgscan);
1559 table_close(tgrel, RowExclusiveLock);
1562 * Close rel, but keep exclusive lock!
1564 relation_close(targetrel, NoLock);
1566 return address;
1570 * Subroutine for renametrig -- perform the actual work of renaming one
1571 * trigger on one table.
1573 * If the trigger has a name different from the expected one, raise a
1574 * NOTICE about it.
1576 static void
1577 renametrig_internal(Relation tgrel, Relation targetrel, HeapTuple trigtup,
1578 const char *newname, const char *expected_name)
1580 HeapTuple tuple;
1581 Form_pg_trigger tgform;
1582 ScanKeyData key[2];
1583 SysScanDesc tgscan;
1585 /* If the trigger already has the new name, nothing to do. */
1586 tgform = (Form_pg_trigger) GETSTRUCT(trigtup);
1587 if (strcmp(NameStr(tgform->tgname), newname) == 0)
1588 return;
1591 * Before actually trying the rename, search for triggers with the same
1592 * name. The update would fail with an ugly message in that case, and it
1593 * is better to throw a nicer error.
1595 ScanKeyInit(&key[0],
1596 Anum_pg_trigger_tgrelid,
1597 BTEqualStrategyNumber, F_OIDEQ,
1598 ObjectIdGetDatum(RelationGetRelid(targetrel)));
1599 ScanKeyInit(&key[1],
1600 Anum_pg_trigger_tgname,
1601 BTEqualStrategyNumber, F_NAMEEQ,
1602 PointerGetDatum(newname));
1603 tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1604 NULL, 2, key);
1605 if (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
1606 ereport(ERROR,
1607 (errcode(ERRCODE_DUPLICATE_OBJECT),
1608 errmsg("trigger \"%s\" for relation \"%s\" already exists",
1609 newname, RelationGetRelationName(targetrel))));
1610 systable_endscan(tgscan);
1613 * The target name is free; update the existing pg_trigger tuple with it.
1615 tuple = heap_copytuple(trigtup); /* need a modifiable copy */
1616 tgform = (Form_pg_trigger) GETSTRUCT(tuple);
1619 * If the trigger has a name different from what we expected, let the user
1620 * know. (We can proceed anyway, since we must have reached here following
1621 * a tgparentid link.)
1623 if (strcmp(NameStr(tgform->tgname), expected_name) != 0)
1624 ereport(NOTICE,
1625 errmsg("renamed trigger \"%s\" on relation \"%s\"",
1626 NameStr(tgform->tgname),
1627 RelationGetRelationName(targetrel)));
1629 namestrcpy(&tgform->tgname, newname);
1631 CatalogTupleUpdate(tgrel, &tuple->t_self, tuple);
1633 InvokeObjectPostAlterHook(TriggerRelationId, tgform->oid, 0);
1636 * Invalidate relation's relcache entry so that other backends (and this
1637 * one too!) are sent SI message to make them rebuild relcache entries.
1638 * (Ideally this should happen automatically...)
1640 CacheInvalidateRelcache(targetrel);
1644 * Subroutine for renametrig -- Helper for recursing to partitions when
1645 * renaming triggers on a partitioned table.
1647 static void
1648 renametrig_partition(Relation tgrel, Oid partitionId, Oid parentTriggerOid,
1649 const char *newname, const char *expected_name)
1651 SysScanDesc tgscan;
1652 ScanKeyData key;
1653 HeapTuple tuple;
1656 * Given a relation and the OID of a trigger on parent relation, find the
1657 * corresponding trigger in the child and rename that trigger to the given
1658 * name.
1660 ScanKeyInit(&key,
1661 Anum_pg_trigger_tgrelid,
1662 BTEqualStrategyNumber, F_OIDEQ,
1663 ObjectIdGetDatum(partitionId));
1664 tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1665 NULL, 1, &key);
1666 while (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
1668 Form_pg_trigger tgform = (Form_pg_trigger) GETSTRUCT(tuple);
1669 Relation partitionRel;
1671 if (tgform->tgparentid != parentTriggerOid)
1672 continue; /* not our trigger */
1674 partitionRel = table_open(partitionId, NoLock);
1676 /* Rename the trigger on this partition */
1677 renametrig_internal(tgrel, partitionRel, tuple, newname, expected_name);
1679 /* And if this relation is partitioned, recurse to its partitions */
1680 if (partitionRel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
1682 PartitionDesc partdesc = RelationGetPartitionDesc(partitionRel,
1683 true);
1685 for (int i = 0; i < partdesc->nparts; i++)
1687 Oid partoid = partdesc->oids[i];
1689 renametrig_partition(tgrel, partoid, tgform->oid, newname,
1690 NameStr(tgform->tgname));
1693 table_close(partitionRel, NoLock);
1695 /* There should be at most one matching tuple */
1696 break;
1698 systable_endscan(tgscan);
1702 * EnableDisableTrigger()
1704 * Called by ALTER TABLE ENABLE/DISABLE [ REPLICA | ALWAYS ] TRIGGER
1705 * to change 'tgenabled' field for the specified trigger(s)
1707 * rel: relation to process (caller must hold suitable lock on it)
1708 * tgname: name of trigger to process, or NULL to scan all triggers
1709 * tgparent: if not zero, process only triggers with this tgparentid
1710 * fires_when: new value for tgenabled field. In addition to generic
1711 * enablement/disablement, this also defines when the trigger
1712 * should be fired in session replication roles.
1713 * skip_system: if true, skip "system" triggers (constraint triggers)
1714 * recurse: if true, recurse to partitions
1716 * Caller should have checked permissions for the table; here we also
1717 * enforce that superuser privilege is required to alter the state of
1718 * system triggers
1720 void
1721 EnableDisableTrigger(Relation rel, const char *tgname, Oid tgparent,
1722 char fires_when, bool skip_system, bool recurse,
1723 LOCKMODE lockmode)
1725 Relation tgrel;
1726 int nkeys;
1727 ScanKeyData keys[2];
1728 SysScanDesc tgscan;
1729 HeapTuple tuple;
1730 bool found;
1731 bool changed;
1733 /* Scan the relevant entries in pg_triggers */
1734 tgrel = table_open(TriggerRelationId, RowExclusiveLock);
1736 ScanKeyInit(&keys[0],
1737 Anum_pg_trigger_tgrelid,
1738 BTEqualStrategyNumber, F_OIDEQ,
1739 ObjectIdGetDatum(RelationGetRelid(rel)));
1740 if (tgname)
1742 ScanKeyInit(&keys[1],
1743 Anum_pg_trigger_tgname,
1744 BTEqualStrategyNumber, F_NAMEEQ,
1745 CStringGetDatum(tgname));
1746 nkeys = 2;
1748 else
1749 nkeys = 1;
1751 tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1752 NULL, nkeys, keys);
1754 found = changed = false;
1756 while (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
1758 Form_pg_trigger oldtrig = (Form_pg_trigger) GETSTRUCT(tuple);
1760 if (OidIsValid(tgparent) && tgparent != oldtrig->tgparentid)
1761 continue;
1763 if (oldtrig->tgisinternal)
1765 /* system trigger ... ok to process? */
1766 if (skip_system)
1767 continue;
1768 if (!superuser())
1769 ereport(ERROR,
1770 (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
1771 errmsg("permission denied: \"%s\" is a system trigger",
1772 NameStr(oldtrig->tgname))));
1775 found = true;
1777 if (oldtrig->tgenabled != fires_when)
1779 /* need to change this one ... make a copy to scribble on */
1780 HeapTuple newtup = heap_copytuple(tuple);
1781 Form_pg_trigger newtrig = (Form_pg_trigger) GETSTRUCT(newtup);
1783 newtrig->tgenabled = fires_when;
1785 CatalogTupleUpdate(tgrel, &newtup->t_self, newtup);
1787 heap_freetuple(newtup);
1789 changed = true;
1793 * When altering FOR EACH ROW triggers on a partitioned table, do the
1794 * same on the partitions as well, unless ONLY is specified.
1796 * Note that we recurse even if we didn't change the trigger above,
1797 * because the partitions' copy of the trigger may have a different
1798 * value of tgenabled than the parent's trigger and thus might need to
1799 * be changed.
1801 if (recurse &&
1802 rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE &&
1803 (TRIGGER_FOR_ROW(oldtrig->tgtype)))
1805 PartitionDesc partdesc = RelationGetPartitionDesc(rel, true);
1806 int i;
1808 for (i = 0; i < partdesc->nparts; i++)
1810 Relation part;
1812 part = relation_open(partdesc->oids[i], lockmode);
1813 /* Match on child triggers' tgparentid, not their name */
1814 EnableDisableTrigger(part, NULL, oldtrig->oid,
1815 fires_when, skip_system, recurse,
1816 lockmode);
1817 table_close(part, NoLock); /* keep lock till commit */
1821 InvokeObjectPostAlterHook(TriggerRelationId,
1822 oldtrig->oid, 0);
1825 systable_endscan(tgscan);
1827 table_close(tgrel, RowExclusiveLock);
1829 if (tgname && !found)
1830 ereport(ERROR,
1831 (errcode(ERRCODE_UNDEFINED_OBJECT),
1832 errmsg("trigger \"%s\" for table \"%s\" does not exist",
1833 tgname, RelationGetRelationName(rel))));
1836 * If we changed anything, broadcast a SI inval message to force each
1837 * backend (including our own!) to rebuild relation's relcache entry.
1838 * Otherwise they will fail to apply the change promptly.
1840 if (changed)
1841 CacheInvalidateRelcache(rel);
1846 * Build trigger data to attach to the given relcache entry.
1848 * Note that trigger data attached to a relcache entry must be stored in
1849 * CacheMemoryContext to ensure it survives as long as the relcache entry.
1850 * But we should be running in a less long-lived working context. To avoid
1851 * leaking cache memory if this routine fails partway through, we build a
1852 * temporary TriggerDesc in working memory and then copy the completed
1853 * structure into cache memory.
1855 void
1856 RelationBuildTriggers(Relation relation)
1858 TriggerDesc *trigdesc;
1859 int numtrigs;
1860 int maxtrigs;
1861 Trigger *triggers;
1862 Relation tgrel;
1863 ScanKeyData skey;
1864 SysScanDesc tgscan;
1865 HeapTuple htup;
1866 MemoryContext oldContext;
1867 int i;
1870 * Allocate a working array to hold the triggers (the array is extended if
1871 * necessary)
1873 maxtrigs = 16;
1874 triggers = (Trigger *) palloc(maxtrigs * sizeof(Trigger));
1875 numtrigs = 0;
1878 * Note: since we scan the triggers using TriggerRelidNameIndexId, we will
1879 * be reading the triggers in name order, except possibly during
1880 * emergency-recovery operations (ie, IgnoreSystemIndexes). This in turn
1881 * ensures that triggers will be fired in name order.
1883 ScanKeyInit(&skey,
1884 Anum_pg_trigger_tgrelid,
1885 BTEqualStrategyNumber, F_OIDEQ,
1886 ObjectIdGetDatum(RelationGetRelid(relation)));
1888 tgrel = table_open(TriggerRelationId, AccessShareLock);
1889 tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1890 NULL, 1, &skey);
1892 while (HeapTupleIsValid(htup = systable_getnext(tgscan)))
1894 Form_pg_trigger pg_trigger = (Form_pg_trigger) GETSTRUCT(htup);
1895 Trigger *build;
1896 Datum datum;
1897 bool isnull;
1899 if (numtrigs >= maxtrigs)
1901 maxtrigs *= 2;
1902 triggers = (Trigger *) repalloc(triggers, maxtrigs * sizeof(Trigger));
1904 build = &(triggers[numtrigs]);
1906 build->tgoid = pg_trigger->oid;
1907 build->tgname = DatumGetCString(DirectFunctionCall1(nameout,
1908 NameGetDatum(&pg_trigger->tgname)));
1909 build->tgfoid = pg_trigger->tgfoid;
1910 build->tgtype = pg_trigger->tgtype;
1911 build->tgenabled = pg_trigger->tgenabled;
1912 build->tgisinternal = pg_trigger->tgisinternal;
1913 build->tgisclone = OidIsValid(pg_trigger->tgparentid);
1914 build->tgconstrrelid = pg_trigger->tgconstrrelid;
1915 build->tgconstrindid = pg_trigger->tgconstrindid;
1916 build->tgconstraint = pg_trigger->tgconstraint;
1917 build->tgdeferrable = pg_trigger->tgdeferrable;
1918 build->tginitdeferred = pg_trigger->tginitdeferred;
1919 build->tgnargs = pg_trigger->tgnargs;
1920 /* tgattr is first var-width field, so OK to access directly */
1921 build->tgnattr = pg_trigger->tgattr.dim1;
1922 if (build->tgnattr > 0)
1924 build->tgattr = (int16 *) palloc(build->tgnattr * sizeof(int16));
1925 memcpy(build->tgattr, &(pg_trigger->tgattr.values),
1926 build->tgnattr * sizeof(int16));
1928 else
1929 build->tgattr = NULL;
1930 if (build->tgnargs > 0)
1932 bytea *val;
1933 char *p;
1935 val = DatumGetByteaPP(fastgetattr(htup,
1936 Anum_pg_trigger_tgargs,
1937 tgrel->rd_att, &isnull));
1938 if (isnull)
1939 elog(ERROR, "tgargs is null in trigger for relation \"%s\"",
1940 RelationGetRelationName(relation));
1941 p = (char *) VARDATA_ANY(val);
1942 build->tgargs = (char **) palloc(build->tgnargs * sizeof(char *));
1943 for (i = 0; i < build->tgnargs; i++)
1945 build->tgargs[i] = pstrdup(p);
1946 p += strlen(p) + 1;
1949 else
1950 build->tgargs = NULL;
1952 datum = fastgetattr(htup, Anum_pg_trigger_tgoldtable,
1953 tgrel->rd_att, &isnull);
1954 if (!isnull)
1955 build->tgoldtable =
1956 DatumGetCString(DirectFunctionCall1(nameout, datum));
1957 else
1958 build->tgoldtable = NULL;
1960 datum = fastgetattr(htup, Anum_pg_trigger_tgnewtable,
1961 tgrel->rd_att, &isnull);
1962 if (!isnull)
1963 build->tgnewtable =
1964 DatumGetCString(DirectFunctionCall1(nameout, datum));
1965 else
1966 build->tgnewtable = NULL;
1968 datum = fastgetattr(htup, Anum_pg_trigger_tgqual,
1969 tgrel->rd_att, &isnull);
1970 if (!isnull)
1971 build->tgqual = TextDatumGetCString(datum);
1972 else
1973 build->tgqual = NULL;
1975 numtrigs++;
1978 systable_endscan(tgscan);
1979 table_close(tgrel, AccessShareLock);
1981 /* There might not be any triggers */
1982 if (numtrigs == 0)
1984 pfree(triggers);
1985 return;
1988 /* Build trigdesc */
1989 trigdesc = (TriggerDesc *) palloc0(sizeof(TriggerDesc));
1990 trigdesc->triggers = triggers;
1991 trigdesc->numtriggers = numtrigs;
1992 for (i = 0; i < numtrigs; i++)
1993 SetTriggerFlags(trigdesc, &(triggers[i]));
1995 /* Copy completed trigdesc into cache storage */
1996 oldContext = MemoryContextSwitchTo(CacheMemoryContext);
1997 relation->trigdesc = CopyTriggerDesc(trigdesc);
1998 MemoryContextSwitchTo(oldContext);
2000 /* Release working memory */
2001 FreeTriggerDesc(trigdesc);
2005 * Update the TriggerDesc's hint flags to include the specified trigger
2007 static void
2008 SetTriggerFlags(TriggerDesc *trigdesc, Trigger *trigger)
2010 int16 tgtype = trigger->tgtype;
2012 trigdesc->trig_insert_before_row |=
2013 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2014 TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_INSERT);
2015 trigdesc->trig_insert_after_row |=
2016 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2017 TRIGGER_TYPE_AFTER, TRIGGER_TYPE_INSERT);
2018 trigdesc->trig_insert_instead_row |=
2019 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2020 TRIGGER_TYPE_INSTEAD, TRIGGER_TYPE_INSERT);
2021 trigdesc->trig_insert_before_statement |=
2022 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2023 TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_INSERT);
2024 trigdesc->trig_insert_after_statement |=
2025 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2026 TRIGGER_TYPE_AFTER, TRIGGER_TYPE_INSERT);
2027 trigdesc->trig_update_before_row |=
2028 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2029 TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_UPDATE);
2030 trigdesc->trig_update_after_row |=
2031 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2032 TRIGGER_TYPE_AFTER, TRIGGER_TYPE_UPDATE);
2033 trigdesc->trig_update_instead_row |=
2034 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2035 TRIGGER_TYPE_INSTEAD, TRIGGER_TYPE_UPDATE);
2036 trigdesc->trig_update_before_statement |=
2037 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2038 TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_UPDATE);
2039 trigdesc->trig_update_after_statement |=
2040 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2041 TRIGGER_TYPE_AFTER, TRIGGER_TYPE_UPDATE);
2042 trigdesc->trig_delete_before_row |=
2043 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2044 TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_DELETE);
2045 trigdesc->trig_delete_after_row |=
2046 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2047 TRIGGER_TYPE_AFTER, TRIGGER_TYPE_DELETE);
2048 trigdesc->trig_delete_instead_row |=
2049 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2050 TRIGGER_TYPE_INSTEAD, TRIGGER_TYPE_DELETE);
2051 trigdesc->trig_delete_before_statement |=
2052 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2053 TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_DELETE);
2054 trigdesc->trig_delete_after_statement |=
2055 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2056 TRIGGER_TYPE_AFTER, TRIGGER_TYPE_DELETE);
2057 /* there are no row-level truncate triggers */
2058 trigdesc->trig_truncate_before_statement |=
2059 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2060 TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_TRUNCATE);
2061 trigdesc->trig_truncate_after_statement |=
2062 TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2063 TRIGGER_TYPE_AFTER, TRIGGER_TYPE_TRUNCATE);
2065 trigdesc->trig_insert_new_table |=
2066 (TRIGGER_FOR_INSERT(tgtype) &&
2067 TRIGGER_USES_TRANSITION_TABLE(trigger->tgnewtable));
2068 trigdesc->trig_update_old_table |=
2069 (TRIGGER_FOR_UPDATE(tgtype) &&
2070 TRIGGER_USES_TRANSITION_TABLE(trigger->tgoldtable));
2071 trigdesc->trig_update_new_table |=
2072 (TRIGGER_FOR_UPDATE(tgtype) &&
2073 TRIGGER_USES_TRANSITION_TABLE(trigger->tgnewtable));
2074 trigdesc->trig_delete_old_table |=
2075 (TRIGGER_FOR_DELETE(tgtype) &&
2076 TRIGGER_USES_TRANSITION_TABLE(trigger->tgoldtable));
2080 * Copy a TriggerDesc data structure.
2082 * The copy is allocated in the current memory context.
2084 TriggerDesc *
2085 CopyTriggerDesc(TriggerDesc *trigdesc)
2087 TriggerDesc *newdesc;
2088 Trigger *trigger;
2089 int i;
2091 if (trigdesc == NULL || trigdesc->numtriggers <= 0)
2092 return NULL;
2094 newdesc = (TriggerDesc *) palloc(sizeof(TriggerDesc));
2095 memcpy(newdesc, trigdesc, sizeof(TriggerDesc));
2097 trigger = (Trigger *) palloc(trigdesc->numtriggers * sizeof(Trigger));
2098 memcpy(trigger, trigdesc->triggers,
2099 trigdesc->numtriggers * sizeof(Trigger));
2100 newdesc->triggers = trigger;
2102 for (i = 0; i < trigdesc->numtriggers; i++)
2104 trigger->tgname = pstrdup(trigger->tgname);
2105 if (trigger->tgnattr > 0)
2107 int16 *newattr;
2109 newattr = (int16 *) palloc(trigger->tgnattr * sizeof(int16));
2110 memcpy(newattr, trigger->tgattr,
2111 trigger->tgnattr * sizeof(int16));
2112 trigger->tgattr = newattr;
2114 if (trigger->tgnargs > 0)
2116 char **newargs;
2117 int16 j;
2119 newargs = (char **) palloc(trigger->tgnargs * sizeof(char *));
2120 for (j = 0; j < trigger->tgnargs; j++)
2121 newargs[j] = pstrdup(trigger->tgargs[j]);
2122 trigger->tgargs = newargs;
2124 if (trigger->tgqual)
2125 trigger->tgqual = pstrdup(trigger->tgqual);
2126 if (trigger->tgoldtable)
2127 trigger->tgoldtable = pstrdup(trigger->tgoldtable);
2128 if (trigger->tgnewtable)
2129 trigger->tgnewtable = pstrdup(trigger->tgnewtable);
2130 trigger++;
2133 return newdesc;
2137 * Free a TriggerDesc data structure.
2139 void
2140 FreeTriggerDesc(TriggerDesc *trigdesc)
2142 Trigger *trigger;
2143 int i;
2145 if (trigdesc == NULL)
2146 return;
2148 trigger = trigdesc->triggers;
2149 for (i = 0; i < trigdesc->numtriggers; i++)
2151 pfree(trigger->tgname);
2152 if (trigger->tgnattr > 0)
2153 pfree(trigger->tgattr);
2154 if (trigger->tgnargs > 0)
2156 while (--(trigger->tgnargs) >= 0)
2157 pfree(trigger->tgargs[trigger->tgnargs]);
2158 pfree(trigger->tgargs);
2160 if (trigger->tgqual)
2161 pfree(trigger->tgqual);
2162 if (trigger->tgoldtable)
2163 pfree(trigger->tgoldtable);
2164 if (trigger->tgnewtable)
2165 pfree(trigger->tgnewtable);
2166 trigger++;
2168 pfree(trigdesc->triggers);
2169 pfree(trigdesc);
2173 * Compare two TriggerDesc structures for logical equality.
2175 #ifdef NOT_USED
2176 bool
2177 equalTriggerDescs(TriggerDesc *trigdesc1, TriggerDesc *trigdesc2)
2179 int i,
2183 * We need not examine the hint flags, just the trigger array itself; if
2184 * we have the same triggers with the same types, the flags should match.
2186 * As of 7.3 we assume trigger set ordering is significant in the
2187 * comparison; so we just compare corresponding slots of the two sets.
2189 * Note: comparing the stringToNode forms of the WHEN clauses means that
2190 * parse column locations will affect the result. This is okay as long as
2191 * this function is only used for detecting exact equality, as for example
2192 * in checking for staleness of a cache entry.
2194 if (trigdesc1 != NULL)
2196 if (trigdesc2 == NULL)
2197 return false;
2198 if (trigdesc1->numtriggers != trigdesc2->numtriggers)
2199 return false;
2200 for (i = 0; i < trigdesc1->numtriggers; i++)
2202 Trigger *trig1 = trigdesc1->triggers + i;
2203 Trigger *trig2 = trigdesc2->triggers + i;
2205 if (trig1->tgoid != trig2->tgoid)
2206 return false;
2207 if (strcmp(trig1->tgname, trig2->tgname) != 0)
2208 return false;
2209 if (trig1->tgfoid != trig2->tgfoid)
2210 return false;
2211 if (trig1->tgtype != trig2->tgtype)
2212 return false;
2213 if (trig1->tgenabled != trig2->tgenabled)
2214 return false;
2215 if (trig1->tgisinternal != trig2->tgisinternal)
2216 return false;
2217 if (trig1->tgisclone != trig2->tgisclone)
2218 return false;
2219 if (trig1->tgconstrrelid != trig2->tgconstrrelid)
2220 return false;
2221 if (trig1->tgconstrindid != trig2->tgconstrindid)
2222 return false;
2223 if (trig1->tgconstraint != trig2->tgconstraint)
2224 return false;
2225 if (trig1->tgdeferrable != trig2->tgdeferrable)
2226 return false;
2227 if (trig1->tginitdeferred != trig2->tginitdeferred)
2228 return false;
2229 if (trig1->tgnargs != trig2->tgnargs)
2230 return false;
2231 if (trig1->tgnattr != trig2->tgnattr)
2232 return false;
2233 if (trig1->tgnattr > 0 &&
2234 memcmp(trig1->tgattr, trig2->tgattr,
2235 trig1->tgnattr * sizeof(int16)) != 0)
2236 return false;
2237 for (j = 0; j < trig1->tgnargs; j++)
2238 if (strcmp(trig1->tgargs[j], trig2->tgargs[j]) != 0)
2239 return false;
2240 if (trig1->tgqual == NULL && trig2->tgqual == NULL)
2241 /* ok */ ;
2242 else if (trig1->tgqual == NULL || trig2->tgqual == NULL)
2243 return false;
2244 else if (strcmp(trig1->tgqual, trig2->tgqual) != 0)
2245 return false;
2246 if (trig1->tgoldtable == NULL && trig2->tgoldtable == NULL)
2247 /* ok */ ;
2248 else if (trig1->tgoldtable == NULL || trig2->tgoldtable == NULL)
2249 return false;
2250 else if (strcmp(trig1->tgoldtable, trig2->tgoldtable) != 0)
2251 return false;
2252 if (trig1->tgnewtable == NULL && trig2->tgnewtable == NULL)
2253 /* ok */ ;
2254 else if (trig1->tgnewtable == NULL || trig2->tgnewtable == NULL)
2255 return false;
2256 else if (strcmp(trig1->tgnewtable, trig2->tgnewtable) != 0)
2257 return false;
2260 else if (trigdesc2 != NULL)
2261 return false;
2262 return true;
2264 #endif /* NOT_USED */
2267 * Check if there is a row-level trigger with transition tables that prevents
2268 * a table from becoming an inheritance child or partition. Return the name
2269 * of the first such incompatible trigger, or NULL if there is none.
2271 const char *
2272 FindTriggerIncompatibleWithInheritance(TriggerDesc *trigdesc)
2274 if (trigdesc != NULL)
2276 int i;
2278 for (i = 0; i < trigdesc->numtriggers; ++i)
2280 Trigger *trigger = &trigdesc->triggers[i];
2282 if (trigger->tgoldtable != NULL || trigger->tgnewtable != NULL)
2283 return trigger->tgname;
2287 return NULL;
2291 * Call a trigger function.
2293 * trigdata: trigger descriptor.
2294 * tgindx: trigger's index in finfo and instr arrays.
2295 * finfo: array of cached trigger function call information.
2296 * instr: optional array of EXPLAIN ANALYZE instrumentation state.
2297 * per_tuple_context: memory context to execute the function in.
2299 * Returns the tuple (or NULL) as returned by the function.
2301 static HeapTuple
2302 ExecCallTriggerFunc(TriggerData *trigdata,
2303 int tgindx,
2304 FmgrInfo *finfo,
2305 Instrumentation *instr,
2306 MemoryContext per_tuple_context)
2308 LOCAL_FCINFO(fcinfo, 0);
2309 PgStat_FunctionCallUsage fcusage;
2310 Datum result;
2311 MemoryContext oldContext;
2314 * Protect against code paths that may fail to initialize transition table
2315 * info.
2317 Assert(((TRIGGER_FIRED_BY_INSERT(trigdata->tg_event) ||
2318 TRIGGER_FIRED_BY_UPDATE(trigdata->tg_event) ||
2319 TRIGGER_FIRED_BY_DELETE(trigdata->tg_event)) &&
2320 TRIGGER_FIRED_AFTER(trigdata->tg_event) &&
2321 !(trigdata->tg_event & AFTER_TRIGGER_DEFERRABLE) &&
2322 !(trigdata->tg_event & AFTER_TRIGGER_INITDEFERRED)) ||
2323 (trigdata->tg_oldtable == NULL && trigdata->tg_newtable == NULL));
2325 finfo += tgindx;
2328 * We cache fmgr lookup info, to avoid making the lookup again on each
2329 * call.
2331 if (finfo->fn_oid == InvalidOid)
2332 fmgr_info(trigdata->tg_trigger->tgfoid, finfo);
2334 Assert(finfo->fn_oid == trigdata->tg_trigger->tgfoid);
2337 * If doing EXPLAIN ANALYZE, start charging time to this trigger.
2339 if (instr)
2340 InstrStartNode(instr + tgindx);
2343 * Do the function evaluation in the per-tuple memory context, so that
2344 * leaked memory will be reclaimed once per tuple. Note in particular that
2345 * any new tuple created by the trigger function will live till the end of
2346 * the tuple cycle.
2348 oldContext = MemoryContextSwitchTo(per_tuple_context);
2351 * Call the function, passing no arguments but setting a context.
2353 InitFunctionCallInfoData(*fcinfo, finfo, 0,
2354 InvalidOid, (Node *) trigdata, NULL);
2356 pgstat_init_function_usage(fcinfo, &fcusage);
2358 MyTriggerDepth++;
2359 PG_TRY();
2361 result = FunctionCallInvoke(fcinfo);
2363 PG_FINALLY();
2365 MyTriggerDepth--;
2367 PG_END_TRY();
2369 pgstat_end_function_usage(&fcusage, true);
2371 MemoryContextSwitchTo(oldContext);
2374 * Trigger protocol allows function to return a null pointer, but NOT to
2375 * set the isnull result flag.
2377 if (fcinfo->isnull)
2378 ereport(ERROR,
2379 (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
2380 errmsg("trigger function %u returned null value",
2381 fcinfo->flinfo->fn_oid)));
2384 * If doing EXPLAIN ANALYZE, stop charging time to this trigger, and count
2385 * one "tuple returned" (really the number of firings).
2387 if (instr)
2388 InstrStopNode(instr + tgindx, 1);
2390 return (HeapTuple) DatumGetPointer(result);
2393 void
2394 ExecBSInsertTriggers(EState *estate, ResultRelInfo *relinfo)
2396 TriggerDesc *trigdesc;
2397 int i;
2398 TriggerData LocTriggerData = {0};
2400 trigdesc = relinfo->ri_TrigDesc;
2402 if (trigdesc == NULL)
2403 return;
2404 if (!trigdesc->trig_insert_before_statement)
2405 return;
2407 /* no-op if we already fired BS triggers in this context */
2408 if (before_stmt_triggers_fired(RelationGetRelid(relinfo->ri_RelationDesc),
2409 CMD_INSERT))
2410 return;
2412 LocTriggerData.type = T_TriggerData;
2413 LocTriggerData.tg_event = TRIGGER_EVENT_INSERT |
2414 TRIGGER_EVENT_BEFORE;
2415 LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2416 for (i = 0; i < trigdesc->numtriggers; i++)
2418 Trigger *trigger = &trigdesc->triggers[i];
2419 HeapTuple newtuple;
2421 if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2422 TRIGGER_TYPE_STATEMENT,
2423 TRIGGER_TYPE_BEFORE,
2424 TRIGGER_TYPE_INSERT))
2425 continue;
2426 if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2427 NULL, NULL, NULL))
2428 continue;
2430 LocTriggerData.tg_trigger = trigger;
2431 newtuple = ExecCallTriggerFunc(&LocTriggerData,
2433 relinfo->ri_TrigFunctions,
2434 relinfo->ri_TrigInstrument,
2435 GetPerTupleMemoryContext(estate));
2437 if (newtuple)
2438 ereport(ERROR,
2439 (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
2440 errmsg("BEFORE STATEMENT trigger cannot return a value")));
2444 void
2445 ExecASInsertTriggers(EState *estate, ResultRelInfo *relinfo,
2446 TransitionCaptureState *transition_capture)
2448 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2450 if (trigdesc && trigdesc->trig_insert_after_statement)
2451 AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2452 TRIGGER_EVENT_INSERT,
2453 false, NULL, NULL, NIL, NULL, transition_capture,
2454 false);
2457 bool
2458 ExecBRInsertTriggers(EState *estate, ResultRelInfo *relinfo,
2459 TupleTableSlot *slot)
2461 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2462 HeapTuple newtuple = NULL;
2463 bool should_free;
2464 TriggerData LocTriggerData = {0};
2465 int i;
2467 LocTriggerData.type = T_TriggerData;
2468 LocTriggerData.tg_event = TRIGGER_EVENT_INSERT |
2469 TRIGGER_EVENT_ROW |
2470 TRIGGER_EVENT_BEFORE;
2471 LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2472 for (i = 0; i < trigdesc->numtriggers; i++)
2474 Trigger *trigger = &trigdesc->triggers[i];
2475 HeapTuple oldtuple;
2477 if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2478 TRIGGER_TYPE_ROW,
2479 TRIGGER_TYPE_BEFORE,
2480 TRIGGER_TYPE_INSERT))
2481 continue;
2482 if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2483 NULL, NULL, slot))
2484 continue;
2486 if (!newtuple)
2487 newtuple = ExecFetchSlotHeapTuple(slot, true, &should_free);
2489 LocTriggerData.tg_trigslot = slot;
2490 LocTriggerData.tg_trigtuple = oldtuple = newtuple;
2491 LocTriggerData.tg_trigger = trigger;
2492 newtuple = ExecCallTriggerFunc(&LocTriggerData,
2494 relinfo->ri_TrigFunctions,
2495 relinfo->ri_TrigInstrument,
2496 GetPerTupleMemoryContext(estate));
2497 if (newtuple == NULL)
2499 if (should_free)
2500 heap_freetuple(oldtuple);
2501 return false; /* "do nothing" */
2503 else if (newtuple != oldtuple)
2505 ExecForceStoreHeapTuple(newtuple, slot, false);
2508 * After a tuple in a partition goes through a trigger, the user
2509 * could have changed the partition key enough that the tuple no
2510 * longer fits the partition. Verify that.
2512 if (trigger->tgisclone &&
2513 !ExecPartitionCheck(relinfo, slot, estate, false))
2514 ereport(ERROR,
2515 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
2516 errmsg("moving row to another partition during a BEFORE FOR EACH ROW trigger is not supported"),
2517 errdetail("Before executing trigger \"%s\", the row was to be in partition \"%s.%s\".",
2518 trigger->tgname,
2519 get_namespace_name(RelationGetNamespace(relinfo->ri_RelationDesc)),
2520 RelationGetRelationName(relinfo->ri_RelationDesc))));
2522 if (should_free)
2523 heap_freetuple(oldtuple);
2525 /* signal tuple should be re-fetched if used */
2526 newtuple = NULL;
2530 return true;
2533 void
2534 ExecARInsertTriggers(EState *estate, ResultRelInfo *relinfo,
2535 TupleTableSlot *slot, List *recheckIndexes,
2536 TransitionCaptureState *transition_capture)
2538 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2540 if ((trigdesc && trigdesc->trig_insert_after_row) ||
2541 (transition_capture && transition_capture->tcs_insert_new_table))
2542 AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2543 TRIGGER_EVENT_INSERT,
2544 true, NULL, slot,
2545 recheckIndexes, NULL,
2546 transition_capture,
2547 false);
2550 bool
2551 ExecIRInsertTriggers(EState *estate, ResultRelInfo *relinfo,
2552 TupleTableSlot *slot)
2554 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2555 HeapTuple newtuple = NULL;
2556 bool should_free;
2557 TriggerData LocTriggerData = {0};
2558 int i;
2560 LocTriggerData.type = T_TriggerData;
2561 LocTriggerData.tg_event = TRIGGER_EVENT_INSERT |
2562 TRIGGER_EVENT_ROW |
2563 TRIGGER_EVENT_INSTEAD;
2564 LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2565 for (i = 0; i < trigdesc->numtriggers; i++)
2567 Trigger *trigger = &trigdesc->triggers[i];
2568 HeapTuple oldtuple;
2570 if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2571 TRIGGER_TYPE_ROW,
2572 TRIGGER_TYPE_INSTEAD,
2573 TRIGGER_TYPE_INSERT))
2574 continue;
2575 if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2576 NULL, NULL, slot))
2577 continue;
2579 if (!newtuple)
2580 newtuple = ExecFetchSlotHeapTuple(slot, true, &should_free);
2582 LocTriggerData.tg_trigslot = slot;
2583 LocTriggerData.tg_trigtuple = oldtuple = newtuple;
2584 LocTriggerData.tg_trigger = trigger;
2585 newtuple = ExecCallTriggerFunc(&LocTriggerData,
2587 relinfo->ri_TrigFunctions,
2588 relinfo->ri_TrigInstrument,
2589 GetPerTupleMemoryContext(estate));
2590 if (newtuple == NULL)
2592 if (should_free)
2593 heap_freetuple(oldtuple);
2594 return false; /* "do nothing" */
2596 else if (newtuple != oldtuple)
2598 ExecForceStoreHeapTuple(newtuple, slot, false);
2600 if (should_free)
2601 heap_freetuple(oldtuple);
2603 /* signal tuple should be re-fetched if used */
2604 newtuple = NULL;
2608 return true;
2611 void
2612 ExecBSDeleteTriggers(EState *estate, ResultRelInfo *relinfo)
2614 TriggerDesc *trigdesc;
2615 int i;
2616 TriggerData LocTriggerData = {0};
2618 trigdesc = relinfo->ri_TrigDesc;
2620 if (trigdesc == NULL)
2621 return;
2622 if (!trigdesc->trig_delete_before_statement)
2623 return;
2625 /* no-op if we already fired BS triggers in this context */
2626 if (before_stmt_triggers_fired(RelationGetRelid(relinfo->ri_RelationDesc),
2627 CMD_DELETE))
2628 return;
2630 LocTriggerData.type = T_TriggerData;
2631 LocTriggerData.tg_event = TRIGGER_EVENT_DELETE |
2632 TRIGGER_EVENT_BEFORE;
2633 LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2634 for (i = 0; i < trigdesc->numtriggers; i++)
2636 Trigger *trigger = &trigdesc->triggers[i];
2637 HeapTuple newtuple;
2639 if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2640 TRIGGER_TYPE_STATEMENT,
2641 TRIGGER_TYPE_BEFORE,
2642 TRIGGER_TYPE_DELETE))
2643 continue;
2644 if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2645 NULL, NULL, NULL))
2646 continue;
2648 LocTriggerData.tg_trigger = trigger;
2649 newtuple = ExecCallTriggerFunc(&LocTriggerData,
2651 relinfo->ri_TrigFunctions,
2652 relinfo->ri_TrigInstrument,
2653 GetPerTupleMemoryContext(estate));
2655 if (newtuple)
2656 ereport(ERROR,
2657 (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
2658 errmsg("BEFORE STATEMENT trigger cannot return a value")));
2662 void
2663 ExecASDeleteTriggers(EState *estate, ResultRelInfo *relinfo,
2664 TransitionCaptureState *transition_capture)
2666 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2668 if (trigdesc && trigdesc->trig_delete_after_statement)
2669 AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2670 TRIGGER_EVENT_DELETE,
2671 false, NULL, NULL, NIL, NULL, transition_capture,
2672 false);
2676 * Execute BEFORE ROW DELETE triggers.
2678 * True indicates caller can proceed with the delete. False indicates caller
2679 * need to suppress the delete and additionally if requested, we need to pass
2680 * back the concurrently updated tuple if any.
2682 bool
2683 ExecBRDeleteTriggers(EState *estate, EPQState *epqstate,
2684 ResultRelInfo *relinfo,
2685 ItemPointer tupleid,
2686 HeapTuple fdw_trigtuple,
2687 TupleTableSlot **epqslot,
2688 TM_Result *tmresult,
2689 TM_FailureData *tmfd)
2691 TupleTableSlot *slot = ExecGetTriggerOldSlot(estate, relinfo);
2692 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2693 bool result = true;
2694 TriggerData LocTriggerData = {0};
2695 HeapTuple trigtuple;
2696 bool should_free = false;
2697 int i;
2699 Assert(HeapTupleIsValid(fdw_trigtuple) ^ ItemPointerIsValid(tupleid));
2700 if (fdw_trigtuple == NULL)
2702 TupleTableSlot *epqslot_candidate = NULL;
2704 if (!GetTupleForTrigger(estate, epqstate, relinfo, tupleid,
2705 LockTupleExclusive, slot, &epqslot_candidate,
2706 tmresult, tmfd))
2707 return false;
2710 * If the tuple was concurrently updated and the caller of this
2711 * function requested for the updated tuple, skip the trigger
2712 * execution.
2714 if (epqslot_candidate != NULL && epqslot != NULL)
2716 *epqslot = epqslot_candidate;
2717 return false;
2720 trigtuple = ExecFetchSlotHeapTuple(slot, true, &should_free);
2722 else
2724 trigtuple = fdw_trigtuple;
2725 ExecForceStoreHeapTuple(trigtuple, slot, false);
2728 LocTriggerData.type = T_TriggerData;
2729 LocTriggerData.tg_event = TRIGGER_EVENT_DELETE |
2730 TRIGGER_EVENT_ROW |
2731 TRIGGER_EVENT_BEFORE;
2732 LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2733 for (i = 0; i < trigdesc->numtriggers; i++)
2735 HeapTuple newtuple;
2736 Trigger *trigger = &trigdesc->triggers[i];
2738 if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2739 TRIGGER_TYPE_ROW,
2740 TRIGGER_TYPE_BEFORE,
2741 TRIGGER_TYPE_DELETE))
2742 continue;
2743 if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2744 NULL, slot, NULL))
2745 continue;
2747 LocTriggerData.tg_trigslot = slot;
2748 LocTriggerData.tg_trigtuple = trigtuple;
2749 LocTriggerData.tg_trigger = trigger;
2750 newtuple = ExecCallTriggerFunc(&LocTriggerData,
2752 relinfo->ri_TrigFunctions,
2753 relinfo->ri_TrigInstrument,
2754 GetPerTupleMemoryContext(estate));
2755 if (newtuple == NULL)
2757 result = false; /* tell caller to suppress delete */
2758 break;
2760 if (newtuple != trigtuple)
2761 heap_freetuple(newtuple);
2763 if (should_free)
2764 heap_freetuple(trigtuple);
2766 return result;
2770 * Note: is_crosspart_update must be true if the DELETE is being performed
2771 * as part of a cross-partition update.
2773 void
2774 ExecARDeleteTriggers(EState *estate,
2775 ResultRelInfo *relinfo,
2776 ItemPointer tupleid,
2777 HeapTuple fdw_trigtuple,
2778 TransitionCaptureState *transition_capture,
2779 bool is_crosspart_update)
2781 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2783 if ((trigdesc && trigdesc->trig_delete_after_row) ||
2784 (transition_capture && transition_capture->tcs_delete_old_table))
2786 TupleTableSlot *slot = ExecGetTriggerOldSlot(estate, relinfo);
2788 Assert(HeapTupleIsValid(fdw_trigtuple) ^ ItemPointerIsValid(tupleid));
2789 if (fdw_trigtuple == NULL)
2790 GetTupleForTrigger(estate,
2791 NULL,
2792 relinfo,
2793 tupleid,
2794 LockTupleExclusive,
2795 slot,
2796 NULL,
2797 NULL,
2798 NULL);
2799 else
2800 ExecForceStoreHeapTuple(fdw_trigtuple, slot, false);
2802 AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2803 TRIGGER_EVENT_DELETE,
2804 true, slot, NULL, NIL, NULL,
2805 transition_capture,
2806 is_crosspart_update);
2810 bool
2811 ExecIRDeleteTriggers(EState *estate, ResultRelInfo *relinfo,
2812 HeapTuple trigtuple)
2814 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2815 TupleTableSlot *slot = ExecGetTriggerOldSlot(estate, relinfo);
2816 TriggerData LocTriggerData = {0};
2817 int i;
2819 LocTriggerData.type = T_TriggerData;
2820 LocTriggerData.tg_event = TRIGGER_EVENT_DELETE |
2821 TRIGGER_EVENT_ROW |
2822 TRIGGER_EVENT_INSTEAD;
2823 LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2825 ExecForceStoreHeapTuple(trigtuple, slot, false);
2827 for (i = 0; i < trigdesc->numtriggers; i++)
2829 HeapTuple rettuple;
2830 Trigger *trigger = &trigdesc->triggers[i];
2832 if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2833 TRIGGER_TYPE_ROW,
2834 TRIGGER_TYPE_INSTEAD,
2835 TRIGGER_TYPE_DELETE))
2836 continue;
2837 if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2838 NULL, slot, NULL))
2839 continue;
2841 LocTriggerData.tg_trigslot = slot;
2842 LocTriggerData.tg_trigtuple = trigtuple;
2843 LocTriggerData.tg_trigger = trigger;
2844 rettuple = ExecCallTriggerFunc(&LocTriggerData,
2846 relinfo->ri_TrigFunctions,
2847 relinfo->ri_TrigInstrument,
2848 GetPerTupleMemoryContext(estate));
2849 if (rettuple == NULL)
2850 return false; /* Delete was suppressed */
2851 if (rettuple != trigtuple)
2852 heap_freetuple(rettuple);
2854 return true;
2857 void
2858 ExecBSUpdateTriggers(EState *estate, ResultRelInfo *relinfo)
2860 TriggerDesc *trigdesc;
2861 int i;
2862 TriggerData LocTriggerData = {0};
2863 Bitmapset *updatedCols;
2865 trigdesc = relinfo->ri_TrigDesc;
2867 if (trigdesc == NULL)
2868 return;
2869 if (!trigdesc->trig_update_before_statement)
2870 return;
2872 /* no-op if we already fired BS triggers in this context */
2873 if (before_stmt_triggers_fired(RelationGetRelid(relinfo->ri_RelationDesc),
2874 CMD_UPDATE))
2875 return;
2877 /* statement-level triggers operate on the parent table */
2878 Assert(relinfo->ri_RootResultRelInfo == NULL);
2880 updatedCols = ExecGetAllUpdatedCols(relinfo, estate);
2882 LocTriggerData.type = T_TriggerData;
2883 LocTriggerData.tg_event = TRIGGER_EVENT_UPDATE |
2884 TRIGGER_EVENT_BEFORE;
2885 LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2886 LocTriggerData.tg_updatedcols = updatedCols;
2887 for (i = 0; i < trigdesc->numtriggers; i++)
2889 Trigger *trigger = &trigdesc->triggers[i];
2890 HeapTuple newtuple;
2892 if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2893 TRIGGER_TYPE_STATEMENT,
2894 TRIGGER_TYPE_BEFORE,
2895 TRIGGER_TYPE_UPDATE))
2896 continue;
2897 if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2898 updatedCols, NULL, NULL))
2899 continue;
2901 LocTriggerData.tg_trigger = trigger;
2902 newtuple = ExecCallTriggerFunc(&LocTriggerData,
2904 relinfo->ri_TrigFunctions,
2905 relinfo->ri_TrigInstrument,
2906 GetPerTupleMemoryContext(estate));
2908 if (newtuple)
2909 ereport(ERROR,
2910 (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
2911 errmsg("BEFORE STATEMENT trigger cannot return a value")));
2915 void
2916 ExecASUpdateTriggers(EState *estate, ResultRelInfo *relinfo,
2917 TransitionCaptureState *transition_capture)
2919 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2921 /* statement-level triggers operate on the parent table */
2922 Assert(relinfo->ri_RootResultRelInfo == NULL);
2924 if (trigdesc && trigdesc->trig_update_after_statement)
2925 AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2926 TRIGGER_EVENT_UPDATE,
2927 false, NULL, NULL, NIL,
2928 ExecGetAllUpdatedCols(relinfo, estate),
2929 transition_capture,
2930 false);
2933 bool
2934 ExecBRUpdateTriggers(EState *estate, EPQState *epqstate,
2935 ResultRelInfo *relinfo,
2936 ItemPointer tupleid,
2937 HeapTuple fdw_trigtuple,
2938 TupleTableSlot *newslot,
2939 TM_Result *tmresult,
2940 TM_FailureData *tmfd)
2942 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2943 TupleTableSlot *oldslot = ExecGetTriggerOldSlot(estate, relinfo);
2944 HeapTuple newtuple = NULL;
2945 HeapTuple trigtuple;
2946 bool should_free_trig = false;
2947 bool should_free_new = false;
2948 TriggerData LocTriggerData = {0};
2949 int i;
2950 Bitmapset *updatedCols;
2951 LockTupleMode lockmode;
2953 /* Determine lock mode to use */
2954 lockmode = ExecUpdateLockMode(estate, relinfo);
2956 Assert(HeapTupleIsValid(fdw_trigtuple) ^ ItemPointerIsValid(tupleid));
2957 if (fdw_trigtuple == NULL)
2959 TupleTableSlot *epqslot_candidate = NULL;
2961 /* get a copy of the on-disk tuple we are planning to update */
2962 if (!GetTupleForTrigger(estate, epqstate, relinfo, tupleid,
2963 lockmode, oldslot, &epqslot_candidate,
2964 tmresult, tmfd))
2965 return false; /* cancel the update action */
2968 * In READ COMMITTED isolation level it's possible that target tuple
2969 * was changed due to concurrent update. In that case we have a raw
2970 * subplan output tuple in epqslot_candidate, and need to form a new
2971 * insertable tuple using ExecGetUpdateNewTuple to replace the one we
2972 * received in newslot. Neither we nor our callers have any further
2973 * interest in the passed-in tuple, so it's okay to overwrite newslot
2974 * with the newer data.
2976 if (epqslot_candidate != NULL)
2978 TupleTableSlot *epqslot_clean;
2980 epqslot_clean = ExecGetUpdateNewTuple(relinfo, epqslot_candidate,
2981 oldslot);
2984 * Typically, the caller's newslot was also generated by
2985 * ExecGetUpdateNewTuple, so that epqslot_clean will be the same
2986 * slot and copying is not needed. But do the right thing if it
2987 * isn't.
2989 if (unlikely(newslot != epqslot_clean))
2990 ExecCopySlot(newslot, epqslot_clean);
2993 * At this point newslot contains a virtual tuple that may
2994 * reference some fields of oldslot's tuple in some disk buffer.
2995 * If that tuple is in a different page than the original target
2996 * tuple, then our only pin on that buffer is oldslot's, and we're
2997 * about to release it. Hence we'd better materialize newslot to
2998 * ensure it doesn't contain references into an unpinned buffer.
2999 * (We'd materialize it below anyway, but too late for safety.)
3001 ExecMaterializeSlot(newslot);
3005 * Here we convert oldslot to a materialized slot holding trigtuple.
3006 * Neither slot passed to the triggers will hold any buffer pin.
3008 trigtuple = ExecFetchSlotHeapTuple(oldslot, true, &should_free_trig);
3010 else
3012 /* Put the FDW-supplied tuple into oldslot to unify the cases */
3013 ExecForceStoreHeapTuple(fdw_trigtuple, oldslot, false);
3014 trigtuple = fdw_trigtuple;
3017 LocTriggerData.type = T_TriggerData;
3018 LocTriggerData.tg_event = TRIGGER_EVENT_UPDATE |
3019 TRIGGER_EVENT_ROW |
3020 TRIGGER_EVENT_BEFORE;
3021 LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
3022 updatedCols = ExecGetAllUpdatedCols(relinfo, estate);
3023 LocTriggerData.tg_updatedcols = updatedCols;
3024 for (i = 0; i < trigdesc->numtriggers; i++)
3026 Trigger *trigger = &trigdesc->triggers[i];
3027 HeapTuple oldtuple;
3029 if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
3030 TRIGGER_TYPE_ROW,
3031 TRIGGER_TYPE_BEFORE,
3032 TRIGGER_TYPE_UPDATE))
3033 continue;
3034 if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
3035 updatedCols, oldslot, newslot))
3036 continue;
3038 if (!newtuple)
3039 newtuple = ExecFetchSlotHeapTuple(newslot, true, &should_free_new);
3041 LocTriggerData.tg_trigslot = oldslot;
3042 LocTriggerData.tg_trigtuple = trigtuple;
3043 LocTriggerData.tg_newtuple = oldtuple = newtuple;
3044 LocTriggerData.tg_newslot = newslot;
3045 LocTriggerData.tg_trigger = trigger;
3046 newtuple = ExecCallTriggerFunc(&LocTriggerData,
3048 relinfo->ri_TrigFunctions,
3049 relinfo->ri_TrigInstrument,
3050 GetPerTupleMemoryContext(estate));
3052 if (newtuple == NULL)
3054 if (should_free_trig)
3055 heap_freetuple(trigtuple);
3056 if (should_free_new)
3057 heap_freetuple(oldtuple);
3058 return false; /* "do nothing" */
3060 else if (newtuple != oldtuple)
3062 ExecForceStoreHeapTuple(newtuple, newslot, false);
3065 * If the tuple returned by the trigger / being stored, is the old
3066 * row version, and the heap tuple passed to the trigger was
3067 * allocated locally, materialize the slot. Otherwise we might
3068 * free it while still referenced by the slot.
3070 if (should_free_trig && newtuple == trigtuple)
3071 ExecMaterializeSlot(newslot);
3073 if (should_free_new)
3074 heap_freetuple(oldtuple);
3076 /* signal tuple should be re-fetched if used */
3077 newtuple = NULL;
3080 if (should_free_trig)
3081 heap_freetuple(trigtuple);
3083 return true;
3087 * Note: 'src_partinfo' and 'dst_partinfo', when non-NULL, refer to the source
3088 * and destination partitions, respectively, of a cross-partition update of
3089 * the root partitioned table mentioned in the query, given by 'relinfo'.
3090 * 'tupleid' in that case refers to the ctid of the "old" tuple in the source
3091 * partition, and 'newslot' contains the "new" tuple in the destination
3092 * partition. This interface allows to support the requirements of
3093 * ExecCrossPartitionUpdateForeignKey(); is_crosspart_update must be true in
3094 * that case.
3096 void
3097 ExecARUpdateTriggers(EState *estate, ResultRelInfo *relinfo,
3098 ResultRelInfo *src_partinfo,
3099 ResultRelInfo *dst_partinfo,
3100 ItemPointer tupleid,
3101 HeapTuple fdw_trigtuple,
3102 TupleTableSlot *newslot,
3103 List *recheckIndexes,
3104 TransitionCaptureState *transition_capture,
3105 bool is_crosspart_update)
3107 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
3109 if ((trigdesc && trigdesc->trig_update_after_row) ||
3110 (transition_capture &&
3111 (transition_capture->tcs_update_old_table ||
3112 transition_capture->tcs_update_new_table)))
3115 * Note: if the UPDATE is converted into a DELETE+INSERT as part of
3116 * update-partition-key operation, then this function is also called
3117 * separately for DELETE and INSERT to capture transition table rows.
3118 * In such case, either old tuple or new tuple can be NULL.
3120 TupleTableSlot *oldslot;
3121 ResultRelInfo *tupsrc;
3123 Assert((src_partinfo != NULL && dst_partinfo != NULL) ||
3124 !is_crosspart_update);
3126 tupsrc = src_partinfo ? src_partinfo : relinfo;
3127 oldslot = ExecGetTriggerOldSlot(estate, tupsrc);
3129 if (fdw_trigtuple == NULL && ItemPointerIsValid(tupleid))
3130 GetTupleForTrigger(estate,
3131 NULL,
3132 tupsrc,
3133 tupleid,
3134 LockTupleExclusive,
3135 oldslot,
3136 NULL,
3137 NULL,
3138 NULL);
3139 else if (fdw_trigtuple != NULL)
3140 ExecForceStoreHeapTuple(fdw_trigtuple, oldslot, false);
3141 else
3142 ExecClearTuple(oldslot);
3144 AfterTriggerSaveEvent(estate, relinfo,
3145 src_partinfo, dst_partinfo,
3146 TRIGGER_EVENT_UPDATE,
3147 true,
3148 oldslot, newslot, recheckIndexes,
3149 ExecGetAllUpdatedCols(relinfo, estate),
3150 transition_capture,
3151 is_crosspart_update);
3155 bool
3156 ExecIRUpdateTriggers(EState *estate, ResultRelInfo *relinfo,
3157 HeapTuple trigtuple, TupleTableSlot *newslot)
3159 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
3160 TupleTableSlot *oldslot = ExecGetTriggerOldSlot(estate, relinfo);
3161 HeapTuple newtuple = NULL;
3162 bool should_free;
3163 TriggerData LocTriggerData = {0};
3164 int i;
3166 LocTriggerData.type = T_TriggerData;
3167 LocTriggerData.tg_event = TRIGGER_EVENT_UPDATE |
3168 TRIGGER_EVENT_ROW |
3169 TRIGGER_EVENT_INSTEAD;
3170 LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
3172 ExecForceStoreHeapTuple(trigtuple, oldslot, false);
3174 for (i = 0; i < trigdesc->numtriggers; i++)
3176 Trigger *trigger = &trigdesc->triggers[i];
3177 HeapTuple oldtuple;
3179 if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
3180 TRIGGER_TYPE_ROW,
3181 TRIGGER_TYPE_INSTEAD,
3182 TRIGGER_TYPE_UPDATE))
3183 continue;
3184 if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
3185 NULL, oldslot, newslot))
3186 continue;
3188 if (!newtuple)
3189 newtuple = ExecFetchSlotHeapTuple(newslot, true, &should_free);
3191 LocTriggerData.tg_trigslot = oldslot;
3192 LocTriggerData.tg_trigtuple = trigtuple;
3193 LocTriggerData.tg_newslot = newslot;
3194 LocTriggerData.tg_newtuple = oldtuple = newtuple;
3196 LocTriggerData.tg_trigger = trigger;
3197 newtuple = ExecCallTriggerFunc(&LocTriggerData,
3199 relinfo->ri_TrigFunctions,
3200 relinfo->ri_TrigInstrument,
3201 GetPerTupleMemoryContext(estate));
3202 if (newtuple == NULL)
3204 return false; /* "do nothing" */
3206 else if (newtuple != oldtuple)
3208 ExecForceStoreHeapTuple(newtuple, newslot, false);
3210 if (should_free)
3211 heap_freetuple(oldtuple);
3213 /* signal tuple should be re-fetched if used */
3214 newtuple = NULL;
3218 return true;
3221 void
3222 ExecBSTruncateTriggers(EState *estate, ResultRelInfo *relinfo)
3224 TriggerDesc *trigdesc;
3225 int i;
3226 TriggerData LocTriggerData = {0};
3228 trigdesc = relinfo->ri_TrigDesc;
3230 if (trigdesc == NULL)
3231 return;
3232 if (!trigdesc->trig_truncate_before_statement)
3233 return;
3235 LocTriggerData.type = T_TriggerData;
3236 LocTriggerData.tg_event = TRIGGER_EVENT_TRUNCATE |
3237 TRIGGER_EVENT_BEFORE;
3238 LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
3240 for (i = 0; i < trigdesc->numtriggers; i++)
3242 Trigger *trigger = &trigdesc->triggers[i];
3243 HeapTuple newtuple;
3245 if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
3246 TRIGGER_TYPE_STATEMENT,
3247 TRIGGER_TYPE_BEFORE,
3248 TRIGGER_TYPE_TRUNCATE))
3249 continue;
3250 if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
3251 NULL, NULL, NULL))
3252 continue;
3254 LocTriggerData.tg_trigger = trigger;
3255 newtuple = ExecCallTriggerFunc(&LocTriggerData,
3257 relinfo->ri_TrigFunctions,
3258 relinfo->ri_TrigInstrument,
3259 GetPerTupleMemoryContext(estate));
3261 if (newtuple)
3262 ereport(ERROR,
3263 (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
3264 errmsg("BEFORE STATEMENT trigger cannot return a value")));
3268 void
3269 ExecASTruncateTriggers(EState *estate, ResultRelInfo *relinfo)
3271 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
3273 if (trigdesc && trigdesc->trig_truncate_after_statement)
3274 AfterTriggerSaveEvent(estate, relinfo,
3275 NULL, NULL,
3276 TRIGGER_EVENT_TRUNCATE,
3277 false, NULL, NULL, NIL, NULL, NULL,
3278 false);
3283 * Fetch tuple into "oldslot", dealing with locking and EPQ if necessary
3285 static bool
3286 GetTupleForTrigger(EState *estate,
3287 EPQState *epqstate,
3288 ResultRelInfo *relinfo,
3289 ItemPointer tid,
3290 LockTupleMode lockmode,
3291 TupleTableSlot *oldslot,
3292 TupleTableSlot **epqslot,
3293 TM_Result *tmresultp,
3294 TM_FailureData *tmfdp)
3296 Relation relation = relinfo->ri_RelationDesc;
3298 if (epqslot != NULL)
3300 TM_Result test;
3301 TM_FailureData tmfd;
3302 int lockflags = 0;
3304 *epqslot = NULL;
3306 /* caller must pass an epqstate if EvalPlanQual is possible */
3307 Assert(epqstate != NULL);
3310 * lock tuple for update
3312 if (!IsolationUsesXactSnapshot())
3313 lockflags |= TUPLE_LOCK_FLAG_FIND_LAST_VERSION;
3314 test = table_tuple_lock(relation, tid, estate->es_snapshot, oldslot,
3315 estate->es_output_cid,
3316 lockmode, LockWaitBlock,
3317 lockflags,
3318 &tmfd);
3320 /* Let the caller know about the status of this operation */
3321 if (tmresultp)
3322 *tmresultp = test;
3323 if (tmfdp)
3324 *tmfdp = tmfd;
3326 switch (test)
3328 case TM_SelfModified:
3331 * The target tuple was already updated or deleted by the
3332 * current command, or by a later command in the current
3333 * transaction. We ignore the tuple in the former case, and
3334 * throw error in the latter case, for the same reasons
3335 * enumerated in ExecUpdate and ExecDelete in
3336 * nodeModifyTable.c.
3338 if (tmfd.cmax != estate->es_output_cid)
3339 ereport(ERROR,
3340 (errcode(ERRCODE_TRIGGERED_DATA_CHANGE_VIOLATION),
3341 errmsg("tuple to be updated was already modified by an operation triggered by the current command"),
3342 errhint("Consider using an AFTER trigger instead of a BEFORE trigger to propagate changes to other rows.")));
3344 /* treat it as deleted; do not process */
3345 return false;
3347 case TM_Ok:
3348 if (tmfd.traversed)
3351 * Recheck the tuple using EPQ. For MERGE, we leave this
3352 * to the caller (it must do additional rechecking, and
3353 * might end up executing a different action entirely).
3355 if (estate->es_plannedstmt->commandType == CMD_MERGE)
3357 if (tmresultp)
3358 *tmresultp = TM_Updated;
3359 return false;
3362 *epqslot = EvalPlanQual(epqstate,
3363 relation,
3364 relinfo->ri_RangeTableIndex,
3365 oldslot);
3368 * If PlanQual failed for updated tuple - we must not
3369 * process this tuple!
3371 if (TupIsNull(*epqslot))
3373 *epqslot = NULL;
3374 return false;
3377 break;
3379 case TM_Updated:
3380 if (IsolationUsesXactSnapshot())
3381 ereport(ERROR,
3382 (errcode(ERRCODE_T_R_SERIALIZATION_FAILURE),
3383 errmsg("could not serialize access due to concurrent update")));
3384 elog(ERROR, "unexpected table_tuple_lock status: %u", test);
3385 break;
3387 case TM_Deleted:
3388 if (IsolationUsesXactSnapshot())
3389 ereport(ERROR,
3390 (errcode(ERRCODE_T_R_SERIALIZATION_FAILURE),
3391 errmsg("could not serialize access due to concurrent delete")));
3392 /* tuple was deleted */
3393 return false;
3395 case TM_Invisible:
3396 elog(ERROR, "attempted to lock invisible tuple");
3397 break;
3399 default:
3400 elog(ERROR, "unrecognized table_tuple_lock status: %u", test);
3401 return false; /* keep compiler quiet */
3404 else
3407 * We expect the tuple to be present, thus very simple error handling
3408 * suffices.
3410 if (!table_tuple_fetch_row_version(relation, tid, SnapshotAny,
3411 oldslot))
3412 elog(ERROR, "failed to fetch tuple for trigger");
3415 return true;
3419 * Is trigger enabled to fire?
3421 static bool
3422 TriggerEnabled(EState *estate, ResultRelInfo *relinfo,
3423 Trigger *trigger, TriggerEvent event,
3424 Bitmapset *modifiedCols,
3425 TupleTableSlot *oldslot, TupleTableSlot *newslot)
3427 /* Check replication-role-dependent enable state */
3428 if (SessionReplicationRole == SESSION_REPLICATION_ROLE_REPLICA)
3430 if (trigger->tgenabled == TRIGGER_FIRES_ON_ORIGIN ||
3431 trigger->tgenabled == TRIGGER_DISABLED)
3432 return false;
3434 else /* ORIGIN or LOCAL role */
3436 if (trigger->tgenabled == TRIGGER_FIRES_ON_REPLICA ||
3437 trigger->tgenabled == TRIGGER_DISABLED)
3438 return false;
3442 * Check for column-specific trigger (only possible for UPDATE, and in
3443 * fact we *must* ignore tgattr for other event types)
3445 if (trigger->tgnattr > 0 && TRIGGER_FIRED_BY_UPDATE(event))
3447 int i;
3448 bool modified;
3450 modified = false;
3451 for (i = 0; i < trigger->tgnattr; i++)
3453 if (bms_is_member(trigger->tgattr[i] - FirstLowInvalidHeapAttributeNumber,
3454 modifiedCols))
3456 modified = true;
3457 break;
3460 if (!modified)
3461 return false;
3464 /* Check for WHEN clause */
3465 if (trigger->tgqual)
3467 ExprState **predicate;
3468 ExprContext *econtext;
3469 MemoryContext oldContext;
3470 int i;
3472 Assert(estate != NULL);
3475 * trigger is an element of relinfo->ri_TrigDesc->triggers[]; find the
3476 * matching element of relinfo->ri_TrigWhenExprs[]
3478 i = trigger - relinfo->ri_TrigDesc->triggers;
3479 predicate = &relinfo->ri_TrigWhenExprs[i];
3482 * If first time through for this WHEN expression, build expression
3483 * nodetrees for it. Keep them in the per-query memory context so
3484 * they'll survive throughout the query.
3486 if (*predicate == NULL)
3488 Node *tgqual;
3490 oldContext = MemoryContextSwitchTo(estate->es_query_cxt);
3491 tgqual = stringToNode(trigger->tgqual);
3492 /* Change references to OLD and NEW to INNER_VAR and OUTER_VAR */
3493 ChangeVarNodes(tgqual, PRS2_OLD_VARNO, INNER_VAR, 0);
3494 ChangeVarNodes(tgqual, PRS2_NEW_VARNO, OUTER_VAR, 0);
3495 /* ExecPrepareQual wants implicit-AND form */
3496 tgqual = (Node *) make_ands_implicit((Expr *) tgqual);
3497 *predicate = ExecPrepareQual((List *) tgqual, estate);
3498 MemoryContextSwitchTo(oldContext);
3502 * We will use the EState's per-tuple context for evaluating WHEN
3503 * expressions (creating it if it's not already there).
3505 econtext = GetPerTupleExprContext(estate);
3508 * Finally evaluate the expression, making the old and/or new tuples
3509 * available as INNER_VAR/OUTER_VAR respectively.
3511 econtext->ecxt_innertuple = oldslot;
3512 econtext->ecxt_outertuple = newslot;
3513 if (!ExecQual(*predicate, econtext))
3514 return false;
3517 return true;
3521 /* ----------
3522 * After-trigger stuff
3524 * The AfterTriggersData struct holds data about pending AFTER trigger events
3525 * during the current transaction tree. (BEFORE triggers are fired
3526 * immediately so we don't need any persistent state about them.) The struct
3527 * and most of its subsidiary data are kept in TopTransactionContext; however
3528 * some data that can be discarded sooner appears in the CurTransactionContext
3529 * of the relevant subtransaction. Also, the individual event records are
3530 * kept in a separate sub-context of TopTransactionContext. This is done
3531 * mainly so that it's easy to tell from a memory context dump how much space
3532 * is being eaten by trigger events.
3534 * Because the list of pending events can grow large, we go to some
3535 * considerable effort to minimize per-event memory consumption. The event
3536 * records are grouped into chunks and common data for similar events in the
3537 * same chunk is only stored once.
3539 * XXX We need to be able to save the per-event data in a file if it grows too
3540 * large.
3541 * ----------
3544 /* Per-trigger SET CONSTRAINT status */
3545 typedef struct SetConstraintTriggerData
3547 Oid sct_tgoid;
3548 bool sct_tgisdeferred;
3549 } SetConstraintTriggerData;
3551 typedef struct SetConstraintTriggerData *SetConstraintTrigger;
3554 * SET CONSTRAINT intra-transaction status.
3556 * We make this a single palloc'd object so it can be copied and freed easily.
3558 * all_isset and all_isdeferred are used to keep track
3559 * of SET CONSTRAINTS ALL {DEFERRED, IMMEDIATE}.
3561 * trigstates[] stores per-trigger tgisdeferred settings.
3563 typedef struct SetConstraintStateData
3565 bool all_isset;
3566 bool all_isdeferred;
3567 int numstates; /* number of trigstates[] entries in use */
3568 int numalloc; /* allocated size of trigstates[] */
3569 SetConstraintTriggerData trigstates[FLEXIBLE_ARRAY_MEMBER];
3570 } SetConstraintStateData;
3572 typedef SetConstraintStateData *SetConstraintState;
3576 * Per-trigger-event data
3578 * The actual per-event data, AfterTriggerEventData, includes DONE/IN_PROGRESS
3579 * status bits, up to two tuple CTIDs, and optionally two OIDs of partitions.
3580 * Each event record also has an associated AfterTriggerSharedData that is
3581 * shared across all instances of similar events within a "chunk".
3583 * For row-level triggers, we arrange not to waste storage on unneeded ctid
3584 * fields. Updates of regular tables use two; inserts and deletes of regular
3585 * tables use one; foreign tables always use zero and save the tuple(s) to a
3586 * tuplestore. AFTER_TRIGGER_FDW_FETCH directs AfterTriggerExecute() to
3587 * retrieve a fresh tuple or pair of tuples from that tuplestore, while
3588 * AFTER_TRIGGER_FDW_REUSE directs it to use the most-recently-retrieved
3589 * tuple(s). This permits storing tuples once regardless of the number of
3590 * row-level triggers on a foreign table.
3592 * When updates on partitioned tables cause rows to move between partitions,
3593 * the OIDs of both partitions are stored too, so that the tuples can be
3594 * fetched; such entries are marked AFTER_TRIGGER_CP_UPDATE (for "cross-
3595 * partition update").
3597 * Note that we need triggers on foreign tables to be fired in exactly the
3598 * order they were queued, so that the tuples come out of the tuplestore in
3599 * the right order. To ensure that, we forbid deferrable (constraint)
3600 * triggers on foreign tables. This also ensures that such triggers do not
3601 * get deferred into outer trigger query levels, meaning that it's okay to
3602 * destroy the tuplestore at the end of the query level.
3604 * Statement-level triggers always bear AFTER_TRIGGER_1CTID, though they
3605 * require no ctid field. We lack the flag bit space to neatly represent that
3606 * distinct case, and it seems unlikely to be worth much trouble.
3608 * Note: ats_firing_id is initially zero and is set to something else when
3609 * AFTER_TRIGGER_IN_PROGRESS is set. It indicates which trigger firing
3610 * cycle the trigger will be fired in (or was fired in, if DONE is set).
3611 * Although this is mutable state, we can keep it in AfterTriggerSharedData
3612 * because all instances of the same type of event in a given event list will
3613 * be fired at the same time, if they were queued between the same firing
3614 * cycles. So we need only ensure that ats_firing_id is zero when attaching
3615 * a new event to an existing AfterTriggerSharedData record.
3617 typedef uint32 TriggerFlags;
3619 #define AFTER_TRIGGER_OFFSET 0x07FFFFFF /* must be low-order bits */
3620 #define AFTER_TRIGGER_DONE 0x80000000
3621 #define AFTER_TRIGGER_IN_PROGRESS 0x40000000
3622 /* bits describing the size and tuple sources of this event */
3623 #define AFTER_TRIGGER_FDW_REUSE 0x00000000
3624 #define AFTER_TRIGGER_FDW_FETCH 0x20000000
3625 #define AFTER_TRIGGER_1CTID 0x10000000
3626 #define AFTER_TRIGGER_2CTID 0x30000000
3627 #define AFTER_TRIGGER_CP_UPDATE 0x08000000
3628 #define AFTER_TRIGGER_TUP_BITS 0x38000000
3629 typedef struct AfterTriggerSharedData *AfterTriggerShared;
3631 typedef struct AfterTriggerSharedData
3633 TriggerEvent ats_event; /* event type indicator, see trigger.h */
3634 Oid ats_tgoid; /* the trigger's ID */
3635 Oid ats_relid; /* the relation it's on */
3636 CommandId ats_firing_id; /* ID for firing cycle */
3637 struct AfterTriggersTableData *ats_table; /* transition table access */
3638 Bitmapset *ats_modifiedcols; /* modified columns */
3639 } AfterTriggerSharedData;
3641 typedef struct AfterTriggerEventData *AfterTriggerEvent;
3643 typedef struct AfterTriggerEventData
3645 TriggerFlags ate_flags; /* status bits and offset to shared data */
3646 ItemPointerData ate_ctid1; /* inserted, deleted, or old updated tuple */
3647 ItemPointerData ate_ctid2; /* new updated tuple */
3650 * During a cross-partition update of a partitioned table, we also store
3651 * the OIDs of source and destination partitions that are needed to fetch
3652 * the old (ctid1) and the new tuple (ctid2) from, respectively.
3654 Oid ate_src_part;
3655 Oid ate_dst_part;
3656 } AfterTriggerEventData;
3658 /* AfterTriggerEventData, minus ate_src_part, ate_dst_part */
3659 typedef struct AfterTriggerEventDataNoOids
3661 TriggerFlags ate_flags;
3662 ItemPointerData ate_ctid1;
3663 ItemPointerData ate_ctid2;
3664 } AfterTriggerEventDataNoOids;
3666 /* AfterTriggerEventData, minus ate_*_part and ate_ctid2 */
3667 typedef struct AfterTriggerEventDataOneCtid
3669 TriggerFlags ate_flags; /* status bits and offset to shared data */
3670 ItemPointerData ate_ctid1; /* inserted, deleted, or old updated tuple */
3671 } AfterTriggerEventDataOneCtid;
3673 /* AfterTriggerEventData, minus ate_*_part, ate_ctid1 and ate_ctid2 */
3674 typedef struct AfterTriggerEventDataZeroCtids
3676 TriggerFlags ate_flags; /* status bits and offset to shared data */
3677 } AfterTriggerEventDataZeroCtids;
3679 #define SizeofTriggerEvent(evt) \
3680 (((evt)->ate_flags & AFTER_TRIGGER_TUP_BITS) == AFTER_TRIGGER_CP_UPDATE ? \
3681 sizeof(AfterTriggerEventData) : \
3682 (((evt)->ate_flags & AFTER_TRIGGER_TUP_BITS) == AFTER_TRIGGER_2CTID ? \
3683 sizeof(AfterTriggerEventDataNoOids) : \
3684 (((evt)->ate_flags & AFTER_TRIGGER_TUP_BITS) == AFTER_TRIGGER_1CTID ? \
3685 sizeof(AfterTriggerEventDataOneCtid) : \
3686 sizeof(AfterTriggerEventDataZeroCtids))))
3688 #define GetTriggerSharedData(evt) \
3689 ((AfterTriggerShared) ((char *) (evt) + ((evt)->ate_flags & AFTER_TRIGGER_OFFSET)))
3692 * To avoid palloc overhead, we keep trigger events in arrays in successively-
3693 * larger chunks (a slightly more sophisticated version of an expansible
3694 * array). The space between CHUNK_DATA_START and freeptr is occupied by
3695 * AfterTriggerEventData records; the space between endfree and endptr is
3696 * occupied by AfterTriggerSharedData records.
3698 typedef struct AfterTriggerEventChunk
3700 struct AfterTriggerEventChunk *next; /* list link */
3701 char *freeptr; /* start of free space in chunk */
3702 char *endfree; /* end of free space in chunk */
3703 char *endptr; /* end of chunk */
3704 /* event data follows here */
3705 } AfterTriggerEventChunk;
3707 #define CHUNK_DATA_START(cptr) ((char *) (cptr) + MAXALIGN(sizeof(AfterTriggerEventChunk)))
3709 /* A list of events */
3710 typedef struct AfterTriggerEventList
3712 AfterTriggerEventChunk *head;
3713 AfterTriggerEventChunk *tail;
3714 char *tailfree; /* freeptr of tail chunk */
3715 } AfterTriggerEventList;
3717 /* Macros to help in iterating over a list of events */
3718 #define for_each_chunk(cptr, evtlist) \
3719 for (cptr = (evtlist).head; cptr != NULL; cptr = cptr->next)
3720 #define for_each_event(eptr, cptr) \
3721 for (eptr = (AfterTriggerEvent) CHUNK_DATA_START(cptr); \
3722 (char *) eptr < (cptr)->freeptr; \
3723 eptr = (AfterTriggerEvent) (((char *) eptr) + SizeofTriggerEvent(eptr)))
3724 /* Use this if no special per-chunk processing is needed */
3725 #define for_each_event_chunk(eptr, cptr, evtlist) \
3726 for_each_chunk(cptr, evtlist) for_each_event(eptr, cptr)
3728 /* Macros for iterating from a start point that might not be list start */
3729 #define for_each_chunk_from(cptr) \
3730 for (; cptr != NULL; cptr = cptr->next)
3731 #define for_each_event_from(eptr, cptr) \
3732 for (; \
3733 (char *) eptr < (cptr)->freeptr; \
3734 eptr = (AfterTriggerEvent) (((char *) eptr) + SizeofTriggerEvent(eptr)))
3738 * All per-transaction data for the AFTER TRIGGERS module.
3740 * AfterTriggersData has the following fields:
3742 * firing_counter is incremented for each call of afterTriggerInvokeEvents.
3743 * We mark firable events with the current firing cycle's ID so that we can
3744 * tell which ones to work on. This ensures sane behavior if a trigger
3745 * function chooses to do SET CONSTRAINTS: the inner SET CONSTRAINTS will
3746 * only fire those events that weren't already scheduled for firing.
3748 * state keeps track of the transaction-local effects of SET CONSTRAINTS.
3749 * This is saved and restored across failed subtransactions.
3751 * events is the current list of deferred events. This is global across
3752 * all subtransactions of the current transaction. In a subtransaction
3753 * abort, we know that the events added by the subtransaction are at the
3754 * end of the list, so it is relatively easy to discard them. The event
3755 * list chunks themselves are stored in event_cxt.
3757 * query_depth is the current depth of nested AfterTriggerBeginQuery calls
3758 * (-1 when the stack is empty).
3760 * query_stack[query_depth] is the per-query-level data, including these fields:
3762 * events is a list of AFTER trigger events queued by the current query.
3763 * None of these are valid until the matching AfterTriggerEndQuery call
3764 * occurs. At that point we fire immediate-mode triggers, and append any
3765 * deferred events to the main events list.
3767 * fdw_tuplestore is a tuplestore containing the foreign-table tuples
3768 * needed by events queued by the current query. (Note: we use just one
3769 * tuplestore even though more than one foreign table might be involved.
3770 * This is okay because tuplestores don't really care what's in the tuples
3771 * they store; but it's possible that someday it'd break.)
3773 * tables is a List of AfterTriggersTableData structs for target tables
3774 * of the current query (see below).
3776 * maxquerydepth is just the allocated length of query_stack.
3778 * trans_stack holds per-subtransaction data, including these fields:
3780 * state is NULL or a pointer to a saved copy of the SET CONSTRAINTS
3781 * state data. Each subtransaction level that modifies that state first
3782 * saves a copy, which we use to restore the state if we abort.
3784 * events is a copy of the events head/tail pointers,
3785 * which we use to restore those values during subtransaction abort.
3787 * query_depth is the subtransaction-start-time value of query_depth,
3788 * which we similarly use to clean up at subtransaction abort.
3790 * firing_counter is the subtransaction-start-time value of firing_counter.
3791 * We use this to recognize which deferred triggers were fired (or marked
3792 * for firing) within an aborted subtransaction.
3794 * We use GetCurrentTransactionNestLevel() to determine the correct array
3795 * index in trans_stack. maxtransdepth is the number of allocated entries in
3796 * trans_stack. (By not keeping our own stack pointer, we can avoid trouble
3797 * in cases where errors during subxact abort cause multiple invocations
3798 * of AfterTriggerEndSubXact() at the same nesting depth.)
3800 * We create an AfterTriggersTableData struct for each target table of the
3801 * current query, and each operation mode (INSERT/UPDATE/DELETE), that has
3802 * either transition tables or statement-level triggers. This is used to
3803 * hold the relevant transition tables, as well as info tracking whether
3804 * we already queued the statement triggers. (We use that info to prevent
3805 * firing the same statement triggers more than once per statement, or really
3806 * once per transition table set.) These structs, along with the transition
3807 * table tuplestores, live in the (sub)transaction's CurTransactionContext.
3808 * That's sufficient lifespan because we don't allow transition tables to be
3809 * used by deferrable triggers, so they only need to survive until
3810 * AfterTriggerEndQuery.
3812 typedef struct AfterTriggersQueryData AfterTriggersQueryData;
3813 typedef struct AfterTriggersTransData AfterTriggersTransData;
3814 typedef struct AfterTriggersTableData AfterTriggersTableData;
3816 typedef struct AfterTriggersData
3818 CommandId firing_counter; /* next firing ID to assign */
3819 SetConstraintState state; /* the active S C state */
3820 AfterTriggerEventList events; /* deferred-event list */
3821 MemoryContext event_cxt; /* memory context for events, if any */
3823 /* per-query-level data: */
3824 AfterTriggersQueryData *query_stack; /* array of structs shown below */
3825 int query_depth; /* current index in above array */
3826 int maxquerydepth; /* allocated len of above array */
3828 /* per-subtransaction-level data: */
3829 AfterTriggersTransData *trans_stack; /* array of structs shown below */
3830 int maxtransdepth; /* allocated len of above array */
3831 } AfterTriggersData;
3833 struct AfterTriggersQueryData
3835 AfterTriggerEventList events; /* events pending from this query */
3836 Tuplestorestate *fdw_tuplestore; /* foreign tuples for said events */
3837 List *tables; /* list of AfterTriggersTableData, see below */
3840 struct AfterTriggersTransData
3842 /* these fields are just for resetting at subtrans abort: */
3843 SetConstraintState state; /* saved S C state, or NULL if not yet saved */
3844 AfterTriggerEventList events; /* saved list pointer */
3845 int query_depth; /* saved query_depth */
3846 CommandId firing_counter; /* saved firing_counter */
3849 struct AfterTriggersTableData
3851 /* relid + cmdType form the lookup key for these structs: */
3852 Oid relid; /* target table's OID */
3853 CmdType cmdType; /* event type, CMD_INSERT/UPDATE/DELETE */
3854 bool closed; /* true when no longer OK to add tuples */
3855 bool before_trig_done; /* did we already queue BS triggers? */
3856 bool after_trig_done; /* did we already queue AS triggers? */
3857 AfterTriggerEventList after_trig_events; /* if so, saved list pointer */
3860 * We maintain separate transition tables for UPDATE/INSERT/DELETE since
3861 * MERGE can run all three actions in a single statement. Note that UPDATE
3862 * needs both old and new transition tables whereas INSERT needs only new,
3863 * and DELETE needs only old.
3866 /* "old" transition table for UPDATE, if any */
3867 Tuplestorestate *old_upd_tuplestore;
3868 /* "new" transition table for UPDATE, if any */
3869 Tuplestorestate *new_upd_tuplestore;
3870 /* "old" transition table for DELETE, if any */
3871 Tuplestorestate *old_del_tuplestore;
3872 /* "new" transition table for INSERT, if any */
3873 Tuplestorestate *new_ins_tuplestore;
3875 TupleTableSlot *storeslot; /* for converting to tuplestore's format */
3878 static AfterTriggersData afterTriggers;
3880 static void AfterTriggerExecute(EState *estate,
3881 AfterTriggerEvent event,
3882 ResultRelInfo *relInfo,
3883 ResultRelInfo *src_relInfo,
3884 ResultRelInfo *dst_relInfo,
3885 TriggerDesc *trigdesc,
3886 FmgrInfo *finfo,
3887 Instrumentation *instr,
3888 MemoryContext per_tuple_context,
3889 TupleTableSlot *trig_tuple_slot1,
3890 TupleTableSlot *trig_tuple_slot2);
3891 static AfterTriggersTableData *GetAfterTriggersTableData(Oid relid,
3892 CmdType cmdType);
3893 static TupleTableSlot *GetAfterTriggersStoreSlot(AfterTriggersTableData *table,
3894 TupleDesc tupdesc);
3895 static Tuplestorestate *GetAfterTriggersTransitionTable(int event,
3896 TupleTableSlot *oldslot,
3897 TupleTableSlot *newslot,
3898 TransitionCaptureState *transition_capture);
3899 static void TransitionTableAddTuple(EState *estate,
3900 TransitionCaptureState *transition_capture,
3901 ResultRelInfo *relinfo,
3902 TupleTableSlot *slot,
3903 TupleTableSlot *original_insert_tuple,
3904 Tuplestorestate *tuplestore);
3905 static void AfterTriggerFreeQuery(AfterTriggersQueryData *qs);
3906 static SetConstraintState SetConstraintStateCreate(int numalloc);
3907 static SetConstraintState SetConstraintStateCopy(SetConstraintState origstate);
3908 static SetConstraintState SetConstraintStateAddItem(SetConstraintState state,
3909 Oid tgoid, bool tgisdeferred);
3910 static void cancel_prior_stmt_triggers(Oid relid, CmdType cmdType, int tgevent);
3914 * Get the FDW tuplestore for the current trigger query level, creating it
3915 * if necessary.
3917 static Tuplestorestate *
3918 GetCurrentFDWTuplestore(void)
3920 Tuplestorestate *ret;
3922 ret = afterTriggers.query_stack[afterTriggers.query_depth].fdw_tuplestore;
3923 if (ret == NULL)
3925 MemoryContext oldcxt;
3926 ResourceOwner saveResourceOwner;
3929 * Make the tuplestore valid until end of subtransaction. We really
3930 * only need it until AfterTriggerEndQuery().
3932 oldcxt = MemoryContextSwitchTo(CurTransactionContext);
3933 saveResourceOwner = CurrentResourceOwner;
3934 CurrentResourceOwner = CurTransactionResourceOwner;
3936 ret = tuplestore_begin_heap(false, false, work_mem);
3938 CurrentResourceOwner = saveResourceOwner;
3939 MemoryContextSwitchTo(oldcxt);
3941 afterTriggers.query_stack[afterTriggers.query_depth].fdw_tuplestore = ret;
3944 return ret;
3947 /* ----------
3948 * afterTriggerCheckState()
3950 * Returns true if the trigger event is actually in state DEFERRED.
3951 * ----------
3953 static bool
3954 afterTriggerCheckState(AfterTriggerShared evtshared)
3956 Oid tgoid = evtshared->ats_tgoid;
3957 SetConstraintState state = afterTriggers.state;
3958 int i;
3961 * For not-deferrable triggers (i.e. normal AFTER ROW triggers and
3962 * constraints declared NOT DEFERRABLE), the state is always false.
3964 if ((evtshared->ats_event & AFTER_TRIGGER_DEFERRABLE) == 0)
3965 return false;
3968 * If constraint state exists, SET CONSTRAINTS might have been executed
3969 * either for this trigger or for all triggers.
3971 if (state != NULL)
3973 /* Check for SET CONSTRAINTS for this specific trigger. */
3974 for (i = 0; i < state->numstates; i++)
3976 if (state->trigstates[i].sct_tgoid == tgoid)
3977 return state->trigstates[i].sct_tgisdeferred;
3980 /* Check for SET CONSTRAINTS ALL. */
3981 if (state->all_isset)
3982 return state->all_isdeferred;
3986 * Otherwise return the default state for the trigger.
3988 return ((evtshared->ats_event & AFTER_TRIGGER_INITDEFERRED) != 0);
3991 /* ----------
3992 * afterTriggerCopyBitmap()
3994 * Copy bitmap into AfterTriggerEvents memory context, which is where the after
3995 * trigger events are kept.
3996 * ----------
3998 static Bitmapset *
3999 afterTriggerCopyBitmap(Bitmapset *src)
4001 Bitmapset *dst;
4002 MemoryContext oldcxt;
4004 if (src == NULL)
4005 return NULL;
4007 /* Create event context if we didn't already */
4008 if (afterTriggers.event_cxt == NULL)
4009 afterTriggers.event_cxt =
4010 AllocSetContextCreate(TopTransactionContext,
4011 "AfterTriggerEvents",
4012 ALLOCSET_DEFAULT_SIZES);
4014 oldcxt = MemoryContextSwitchTo(afterTriggers.event_cxt);
4016 dst = bms_copy(src);
4018 MemoryContextSwitchTo(oldcxt);
4020 return dst;
4023 /* ----------
4024 * afterTriggerAddEvent()
4026 * Add a new trigger event to the specified queue.
4027 * The passed-in event data is copied.
4028 * ----------
4030 static void
4031 afterTriggerAddEvent(AfterTriggerEventList *events,
4032 AfterTriggerEvent event, AfterTriggerShared evtshared)
4034 Size eventsize = SizeofTriggerEvent(event);
4035 Size needed = eventsize + sizeof(AfterTriggerSharedData);
4036 AfterTriggerEventChunk *chunk;
4037 AfterTriggerShared newshared;
4038 AfterTriggerEvent newevent;
4041 * If empty list or not enough room in the tail chunk, make a new chunk.
4042 * We assume here that a new shared record will always be needed.
4044 chunk = events->tail;
4045 if (chunk == NULL ||
4046 chunk->endfree - chunk->freeptr < needed)
4048 Size chunksize;
4050 /* Create event context if we didn't already */
4051 if (afterTriggers.event_cxt == NULL)
4052 afterTriggers.event_cxt =
4053 AllocSetContextCreate(TopTransactionContext,
4054 "AfterTriggerEvents",
4055 ALLOCSET_DEFAULT_SIZES);
4058 * Chunk size starts at 1KB and is allowed to increase up to 1MB.
4059 * These numbers are fairly arbitrary, though there is a hard limit at
4060 * AFTER_TRIGGER_OFFSET; else we couldn't link event records to their
4061 * shared records using the available space in ate_flags. Another
4062 * constraint is that if the chunk size gets too huge, the search loop
4063 * below would get slow given a (not too common) usage pattern with
4064 * many distinct event types in a chunk. Therefore, we double the
4065 * preceding chunk size only if there weren't too many shared records
4066 * in the preceding chunk; otherwise we halve it. This gives us some
4067 * ability to adapt to the actual usage pattern of the current query
4068 * while still having large chunk sizes in typical usage. All chunk
4069 * sizes used should be MAXALIGN multiples, to ensure that the shared
4070 * records will be aligned safely.
4072 #define MIN_CHUNK_SIZE 1024
4073 #define MAX_CHUNK_SIZE (1024*1024)
4075 #if MAX_CHUNK_SIZE > (AFTER_TRIGGER_OFFSET+1)
4076 #error MAX_CHUNK_SIZE must not exceed AFTER_TRIGGER_OFFSET
4077 #endif
4079 if (chunk == NULL)
4080 chunksize = MIN_CHUNK_SIZE;
4081 else
4083 /* preceding chunk size... */
4084 chunksize = chunk->endptr - (char *) chunk;
4085 /* check number of shared records in preceding chunk */
4086 if ((chunk->endptr - chunk->endfree) <=
4087 (100 * sizeof(AfterTriggerSharedData)))
4088 chunksize *= 2; /* okay, double it */
4089 else
4090 chunksize /= 2; /* too many shared records */
4091 chunksize = Min(chunksize, MAX_CHUNK_SIZE);
4093 chunk = MemoryContextAlloc(afterTriggers.event_cxt, chunksize);
4094 chunk->next = NULL;
4095 chunk->freeptr = CHUNK_DATA_START(chunk);
4096 chunk->endptr = chunk->endfree = (char *) chunk + chunksize;
4097 Assert(chunk->endfree - chunk->freeptr >= needed);
4099 if (events->head == NULL)
4100 events->head = chunk;
4101 else
4102 events->tail->next = chunk;
4103 events->tail = chunk;
4104 /* events->tailfree is now out of sync, but we'll fix it below */
4108 * Try to locate a matching shared-data record already in the chunk. If
4109 * none, make a new one.
4111 for (newshared = ((AfterTriggerShared) chunk->endptr) - 1;
4112 (char *) newshared >= chunk->endfree;
4113 newshared--)
4115 if (newshared->ats_tgoid == evtshared->ats_tgoid &&
4116 newshared->ats_relid == evtshared->ats_relid &&
4117 newshared->ats_event == evtshared->ats_event &&
4118 newshared->ats_table == evtshared->ats_table &&
4119 newshared->ats_firing_id == 0)
4120 break;
4122 if ((char *) newshared < chunk->endfree)
4124 *newshared = *evtshared;
4125 newshared->ats_firing_id = 0; /* just to be sure */
4126 chunk->endfree = (char *) newshared;
4129 /* Insert the data */
4130 newevent = (AfterTriggerEvent) chunk->freeptr;
4131 memcpy(newevent, event, eventsize);
4132 /* ... and link the new event to its shared record */
4133 newevent->ate_flags &= ~AFTER_TRIGGER_OFFSET;
4134 newevent->ate_flags |= (char *) newshared - (char *) newevent;
4136 chunk->freeptr += eventsize;
4137 events->tailfree = chunk->freeptr;
4140 /* ----------
4141 * afterTriggerFreeEventList()
4143 * Free all the event storage in the given list.
4144 * ----------
4146 static void
4147 afterTriggerFreeEventList(AfterTriggerEventList *events)
4149 AfterTriggerEventChunk *chunk;
4151 while ((chunk = events->head) != NULL)
4153 events->head = chunk->next;
4154 pfree(chunk);
4156 events->tail = NULL;
4157 events->tailfree = NULL;
4160 /* ----------
4161 * afterTriggerRestoreEventList()
4163 * Restore an event list to its prior length, removing all the events
4164 * added since it had the value old_events.
4165 * ----------
4167 static void
4168 afterTriggerRestoreEventList(AfterTriggerEventList *events,
4169 const AfterTriggerEventList *old_events)
4171 AfterTriggerEventChunk *chunk;
4172 AfterTriggerEventChunk *next_chunk;
4174 if (old_events->tail == NULL)
4176 /* restoring to a completely empty state, so free everything */
4177 afterTriggerFreeEventList(events);
4179 else
4181 *events = *old_events;
4182 /* free any chunks after the last one we want to keep */
4183 for (chunk = events->tail->next; chunk != NULL; chunk = next_chunk)
4185 next_chunk = chunk->next;
4186 pfree(chunk);
4188 /* and clean up the tail chunk to be the right length */
4189 events->tail->next = NULL;
4190 events->tail->freeptr = events->tailfree;
4193 * We don't make any effort to remove now-unused shared data records.
4194 * They might still be useful, anyway.
4199 /* ----------
4200 * afterTriggerDeleteHeadEventChunk()
4202 * Remove the first chunk of events from the query level's event list.
4203 * Keep any event list pointers elsewhere in the query level's data
4204 * structures in sync.
4205 * ----------
4207 static void
4208 afterTriggerDeleteHeadEventChunk(AfterTriggersQueryData *qs)
4210 AfterTriggerEventChunk *target = qs->events.head;
4211 ListCell *lc;
4213 Assert(target && target->next);
4216 * First, update any pointers in the per-table data, so that they won't be
4217 * dangling. Resetting obsoleted pointers to NULL will make
4218 * cancel_prior_stmt_triggers start from the list head, which is fine.
4220 foreach(lc, qs->tables)
4222 AfterTriggersTableData *table = (AfterTriggersTableData *) lfirst(lc);
4224 if (table->after_trig_done &&
4225 table->after_trig_events.tail == target)
4227 table->after_trig_events.head = NULL;
4228 table->after_trig_events.tail = NULL;
4229 table->after_trig_events.tailfree = NULL;
4233 /* Now we can flush the head chunk */
4234 qs->events.head = target->next;
4235 pfree(target);
4239 /* ----------
4240 * AfterTriggerExecute()
4242 * Fetch the required tuples back from the heap and fire one
4243 * single trigger function.
4245 * Frequently, this will be fired many times in a row for triggers of
4246 * a single relation. Therefore, we cache the open relation and provide
4247 * fmgr lookup cache space at the caller level. (For triggers fired at
4248 * the end of a query, we can even piggyback on the executor's state.)
4250 * When fired for a cross-partition update of a partitioned table, the old
4251 * tuple is fetched using 'src_relInfo' (the source leaf partition) and
4252 * the new tuple using 'dst_relInfo' (the destination leaf partition), though
4253 * both are converted into the root partitioned table's format before passing
4254 * to the trigger function.
4256 * event: event currently being fired.
4257 * relInfo: result relation for event.
4258 * src_relInfo: source partition of a cross-partition update
4259 * dst_relInfo: its destination partition
4260 * trigdesc: working copy of rel's trigger info.
4261 * finfo: array of fmgr lookup cache entries (one per trigger in trigdesc).
4262 * instr: array of EXPLAIN ANALYZE instrumentation nodes (one per trigger),
4263 * or NULL if no instrumentation is wanted.
4264 * per_tuple_context: memory context to call trigger function in.
4265 * trig_tuple_slot1: scratch slot for tg_trigtuple (foreign tables only)
4266 * trig_tuple_slot2: scratch slot for tg_newtuple (foreign tables only)
4267 * ----------
4269 static void
4270 AfterTriggerExecute(EState *estate,
4271 AfterTriggerEvent event,
4272 ResultRelInfo *relInfo,
4273 ResultRelInfo *src_relInfo,
4274 ResultRelInfo *dst_relInfo,
4275 TriggerDesc *trigdesc,
4276 FmgrInfo *finfo, Instrumentation *instr,
4277 MemoryContext per_tuple_context,
4278 TupleTableSlot *trig_tuple_slot1,
4279 TupleTableSlot *trig_tuple_slot2)
4281 Relation rel = relInfo->ri_RelationDesc;
4282 Relation src_rel = src_relInfo->ri_RelationDesc;
4283 Relation dst_rel = dst_relInfo->ri_RelationDesc;
4284 AfterTriggerShared evtshared = GetTriggerSharedData(event);
4285 Oid tgoid = evtshared->ats_tgoid;
4286 TriggerData LocTriggerData = {0};
4287 HeapTuple rettuple;
4288 int tgindx;
4289 bool should_free_trig = false;
4290 bool should_free_new = false;
4293 * Locate trigger in trigdesc.
4295 for (tgindx = 0; tgindx < trigdesc->numtriggers; tgindx++)
4297 if (trigdesc->triggers[tgindx].tgoid == tgoid)
4299 LocTriggerData.tg_trigger = &(trigdesc->triggers[tgindx]);
4300 break;
4303 if (LocTriggerData.tg_trigger == NULL)
4304 elog(ERROR, "could not find trigger %u", tgoid);
4307 * If doing EXPLAIN ANALYZE, start charging time to this trigger. We want
4308 * to include time spent re-fetching tuples in the trigger cost.
4310 if (instr)
4311 InstrStartNode(instr + tgindx);
4314 * Fetch the required tuple(s).
4316 switch (event->ate_flags & AFTER_TRIGGER_TUP_BITS)
4318 case AFTER_TRIGGER_FDW_FETCH:
4320 Tuplestorestate *fdw_tuplestore = GetCurrentFDWTuplestore();
4322 if (!tuplestore_gettupleslot(fdw_tuplestore, true, false,
4323 trig_tuple_slot1))
4324 elog(ERROR, "failed to fetch tuple1 for AFTER trigger");
4326 if ((evtshared->ats_event & TRIGGER_EVENT_OPMASK) ==
4327 TRIGGER_EVENT_UPDATE &&
4328 !tuplestore_gettupleslot(fdw_tuplestore, true, false,
4329 trig_tuple_slot2))
4330 elog(ERROR, "failed to fetch tuple2 for AFTER trigger");
4332 /* fall through */
4333 case AFTER_TRIGGER_FDW_REUSE:
4336 * Store tuple in the slot so that tg_trigtuple does not reference
4337 * tuplestore memory. (It is formally possible for the trigger
4338 * function to queue trigger events that add to the same
4339 * tuplestore, which can push other tuples out of memory.) The
4340 * distinction is academic, because we start with a minimal tuple
4341 * that is stored as a heap tuple, constructed in different memory
4342 * context, in the slot anyway.
4344 LocTriggerData.tg_trigslot = trig_tuple_slot1;
4345 LocTriggerData.tg_trigtuple =
4346 ExecFetchSlotHeapTuple(trig_tuple_slot1, true, &should_free_trig);
4348 if ((evtshared->ats_event & TRIGGER_EVENT_OPMASK) ==
4349 TRIGGER_EVENT_UPDATE)
4351 LocTriggerData.tg_newslot = trig_tuple_slot2;
4352 LocTriggerData.tg_newtuple =
4353 ExecFetchSlotHeapTuple(trig_tuple_slot2, true, &should_free_new);
4355 else
4357 LocTriggerData.tg_newtuple = NULL;
4359 break;
4361 default:
4362 if (ItemPointerIsValid(&(event->ate_ctid1)))
4364 TupleTableSlot *src_slot = ExecGetTriggerOldSlot(estate,
4365 src_relInfo);
4367 if (!table_tuple_fetch_row_version(src_rel,
4368 &(event->ate_ctid1),
4369 SnapshotAny,
4370 src_slot))
4371 elog(ERROR, "failed to fetch tuple1 for AFTER trigger");
4374 * Store the tuple fetched from the source partition into the
4375 * target (root partitioned) table slot, converting if needed.
4377 if (src_relInfo != relInfo)
4379 TupleConversionMap *map = ExecGetChildToRootMap(src_relInfo);
4381 LocTriggerData.tg_trigslot = ExecGetTriggerOldSlot(estate, relInfo);
4382 if (map)
4384 execute_attr_map_slot(map->attrMap,
4385 src_slot,
4386 LocTriggerData.tg_trigslot);
4388 else
4389 ExecCopySlot(LocTriggerData.tg_trigslot, src_slot);
4391 else
4392 LocTriggerData.tg_trigslot = src_slot;
4393 LocTriggerData.tg_trigtuple =
4394 ExecFetchSlotHeapTuple(LocTriggerData.tg_trigslot, false, &should_free_trig);
4396 else
4398 LocTriggerData.tg_trigtuple = NULL;
4401 /* don't touch ctid2 if not there */
4402 if (((event->ate_flags & AFTER_TRIGGER_TUP_BITS) == AFTER_TRIGGER_2CTID ||
4403 (event->ate_flags & AFTER_TRIGGER_CP_UPDATE)) &&
4404 ItemPointerIsValid(&(event->ate_ctid2)))
4406 TupleTableSlot *dst_slot = ExecGetTriggerNewSlot(estate,
4407 dst_relInfo);
4409 if (!table_tuple_fetch_row_version(dst_rel,
4410 &(event->ate_ctid2),
4411 SnapshotAny,
4412 dst_slot))
4413 elog(ERROR, "failed to fetch tuple2 for AFTER trigger");
4416 * Store the tuple fetched from the destination partition into
4417 * the target (root partitioned) table slot, converting if
4418 * needed.
4420 if (dst_relInfo != relInfo)
4422 TupleConversionMap *map = ExecGetChildToRootMap(dst_relInfo);
4424 LocTriggerData.tg_newslot = ExecGetTriggerNewSlot(estate, relInfo);
4425 if (map)
4427 execute_attr_map_slot(map->attrMap,
4428 dst_slot,
4429 LocTriggerData.tg_newslot);
4431 else
4432 ExecCopySlot(LocTriggerData.tg_newslot, dst_slot);
4434 else
4435 LocTriggerData.tg_newslot = dst_slot;
4436 LocTriggerData.tg_newtuple =
4437 ExecFetchSlotHeapTuple(LocTriggerData.tg_newslot, false, &should_free_new);
4439 else
4441 LocTriggerData.tg_newtuple = NULL;
4446 * Set up the tuplestore information to let the trigger have access to
4447 * transition tables. When we first make a transition table available to
4448 * a trigger, mark it "closed" so that it cannot change anymore. If any
4449 * additional events of the same type get queued in the current trigger
4450 * query level, they'll go into new transition tables.
4452 LocTriggerData.tg_oldtable = LocTriggerData.tg_newtable = NULL;
4453 if (evtshared->ats_table)
4455 if (LocTriggerData.tg_trigger->tgoldtable)
4457 if (TRIGGER_FIRED_BY_UPDATE(evtshared->ats_event))
4458 LocTriggerData.tg_oldtable = evtshared->ats_table->old_upd_tuplestore;
4459 else
4460 LocTriggerData.tg_oldtable = evtshared->ats_table->old_del_tuplestore;
4461 evtshared->ats_table->closed = true;
4464 if (LocTriggerData.tg_trigger->tgnewtable)
4466 if (TRIGGER_FIRED_BY_INSERT(evtshared->ats_event))
4467 LocTriggerData.tg_newtable = evtshared->ats_table->new_ins_tuplestore;
4468 else
4469 LocTriggerData.tg_newtable = evtshared->ats_table->new_upd_tuplestore;
4470 evtshared->ats_table->closed = true;
4475 * Setup the remaining trigger information
4477 LocTriggerData.type = T_TriggerData;
4478 LocTriggerData.tg_event =
4479 evtshared->ats_event & (TRIGGER_EVENT_OPMASK | TRIGGER_EVENT_ROW);
4480 LocTriggerData.tg_relation = rel;
4481 if (TRIGGER_FOR_UPDATE(LocTriggerData.tg_trigger->tgtype))
4482 LocTriggerData.tg_updatedcols = evtshared->ats_modifiedcols;
4484 MemoryContextReset(per_tuple_context);
4487 * Call the trigger and throw away any possibly returned updated tuple.
4488 * (Don't let ExecCallTriggerFunc measure EXPLAIN time.)
4490 rettuple = ExecCallTriggerFunc(&LocTriggerData,
4491 tgindx,
4492 finfo,
4493 NULL,
4494 per_tuple_context);
4495 if (rettuple != NULL &&
4496 rettuple != LocTriggerData.tg_trigtuple &&
4497 rettuple != LocTriggerData.tg_newtuple)
4498 heap_freetuple(rettuple);
4501 * Release resources
4503 if (should_free_trig)
4504 heap_freetuple(LocTriggerData.tg_trigtuple);
4505 if (should_free_new)
4506 heap_freetuple(LocTriggerData.tg_newtuple);
4508 /* don't clear slots' contents if foreign table */
4509 if (trig_tuple_slot1 == NULL)
4511 if (LocTriggerData.tg_trigslot)
4512 ExecClearTuple(LocTriggerData.tg_trigslot);
4513 if (LocTriggerData.tg_newslot)
4514 ExecClearTuple(LocTriggerData.tg_newslot);
4518 * If doing EXPLAIN ANALYZE, stop charging time to this trigger, and count
4519 * one "tuple returned" (really the number of firings).
4521 if (instr)
4522 InstrStopNode(instr + tgindx, 1);
4527 * afterTriggerMarkEvents()
4529 * Scan the given event list for not yet invoked events. Mark the ones
4530 * that can be invoked now with the current firing ID.
4532 * If move_list isn't NULL, events that are not to be invoked now are
4533 * transferred to move_list.
4535 * When immediate_only is true, do not invoke currently-deferred triggers.
4536 * (This will be false only at main transaction exit.)
4538 * Returns true if any invokable events were found.
4540 static bool
4541 afterTriggerMarkEvents(AfterTriggerEventList *events,
4542 AfterTriggerEventList *move_list,
4543 bool immediate_only)
4545 bool found = false;
4546 bool deferred_found = false;
4547 AfterTriggerEvent event;
4548 AfterTriggerEventChunk *chunk;
4550 for_each_event_chunk(event, chunk, *events)
4552 AfterTriggerShared evtshared = GetTriggerSharedData(event);
4553 bool defer_it = false;
4555 if (!(event->ate_flags &
4556 (AFTER_TRIGGER_DONE | AFTER_TRIGGER_IN_PROGRESS)))
4559 * This trigger hasn't been called or scheduled yet. Check if we
4560 * should call it now.
4562 if (immediate_only && afterTriggerCheckState(evtshared))
4564 defer_it = true;
4566 else
4569 * Mark it as to be fired in this firing cycle.
4571 evtshared->ats_firing_id = afterTriggers.firing_counter;
4572 event->ate_flags |= AFTER_TRIGGER_IN_PROGRESS;
4573 found = true;
4578 * If it's deferred, move it to move_list, if requested.
4580 if (defer_it && move_list != NULL)
4582 deferred_found = true;
4583 /* add it to move_list */
4584 afterTriggerAddEvent(move_list, event, evtshared);
4585 /* mark original copy "done" so we don't do it again */
4586 event->ate_flags |= AFTER_TRIGGER_DONE;
4591 * We could allow deferred triggers if, before the end of the
4592 * security-restricted operation, we were to verify that a SET CONSTRAINTS
4593 * ... IMMEDIATE has fired all such triggers. For now, don't bother.
4595 if (deferred_found && InSecurityRestrictedOperation())
4596 ereport(ERROR,
4597 (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
4598 errmsg("cannot fire deferred trigger within security-restricted operation")));
4600 return found;
4604 * afterTriggerInvokeEvents()
4606 * Scan the given event list for events that are marked as to be fired
4607 * in the current firing cycle, and fire them.
4609 * If estate isn't NULL, we use its result relation info to avoid repeated
4610 * openings and closing of trigger target relations. If it is NULL, we
4611 * make one locally to cache the info in case there are multiple trigger
4612 * events per rel.
4614 * When delete_ok is true, it's safe to delete fully-processed events.
4615 * (We are not very tense about that: we simply reset a chunk to be empty
4616 * if all its events got fired. The objective here is just to avoid useless
4617 * rescanning of events when a trigger queues new events during transaction
4618 * end, so it's not necessary to worry much about the case where only
4619 * some events are fired.)
4621 * Returns true if no unfired events remain in the list (this allows us
4622 * to avoid repeating afterTriggerMarkEvents).
4624 static bool
4625 afterTriggerInvokeEvents(AfterTriggerEventList *events,
4626 CommandId firing_id,
4627 EState *estate,
4628 bool delete_ok)
4630 bool all_fired = true;
4631 AfterTriggerEventChunk *chunk;
4632 MemoryContext per_tuple_context;
4633 bool local_estate = false;
4634 ResultRelInfo *rInfo = NULL;
4635 Relation rel = NULL;
4636 TriggerDesc *trigdesc = NULL;
4637 FmgrInfo *finfo = NULL;
4638 Instrumentation *instr = NULL;
4639 TupleTableSlot *slot1 = NULL,
4640 *slot2 = NULL;
4642 /* Make a local EState if need be */
4643 if (estate == NULL)
4645 estate = CreateExecutorState();
4646 local_estate = true;
4649 /* Make a per-tuple memory context for trigger function calls */
4650 per_tuple_context =
4651 AllocSetContextCreate(CurrentMemoryContext,
4652 "AfterTriggerTupleContext",
4653 ALLOCSET_DEFAULT_SIZES);
4655 for_each_chunk(chunk, *events)
4657 AfterTriggerEvent event;
4658 bool all_fired_in_chunk = true;
4660 for_each_event(event, chunk)
4662 AfterTriggerShared evtshared = GetTriggerSharedData(event);
4665 * Is it one for me to fire?
4667 if ((event->ate_flags & AFTER_TRIGGER_IN_PROGRESS) &&
4668 evtshared->ats_firing_id == firing_id)
4670 ResultRelInfo *src_rInfo,
4671 *dst_rInfo;
4674 * So let's fire it... but first, find the correct relation if
4675 * this is not the same relation as before.
4677 if (rel == NULL || RelationGetRelid(rel) != evtshared->ats_relid)
4679 rInfo = ExecGetTriggerResultRel(estate, evtshared->ats_relid,
4680 NULL);
4681 rel = rInfo->ri_RelationDesc;
4682 /* Catch calls with insufficient relcache refcounting */
4683 Assert(!RelationHasReferenceCountZero(rel));
4684 trigdesc = rInfo->ri_TrigDesc;
4685 finfo = rInfo->ri_TrigFunctions;
4686 instr = rInfo->ri_TrigInstrument;
4687 if (slot1 != NULL)
4689 ExecDropSingleTupleTableSlot(slot1);
4690 ExecDropSingleTupleTableSlot(slot2);
4691 slot1 = slot2 = NULL;
4693 if (rel->rd_rel->relkind == RELKIND_FOREIGN_TABLE)
4695 slot1 = MakeSingleTupleTableSlot(rel->rd_att,
4696 &TTSOpsMinimalTuple);
4697 slot2 = MakeSingleTupleTableSlot(rel->rd_att,
4698 &TTSOpsMinimalTuple);
4700 if (trigdesc == NULL) /* should not happen */
4701 elog(ERROR, "relation %u has no triggers",
4702 evtshared->ats_relid);
4706 * Look up source and destination partition result rels of a
4707 * cross-partition update event.
4709 if ((event->ate_flags & AFTER_TRIGGER_TUP_BITS) ==
4710 AFTER_TRIGGER_CP_UPDATE)
4712 Assert(OidIsValid(event->ate_src_part) &&
4713 OidIsValid(event->ate_dst_part));
4714 src_rInfo = ExecGetTriggerResultRel(estate,
4715 event->ate_src_part,
4716 rInfo);
4717 dst_rInfo = ExecGetTriggerResultRel(estate,
4718 event->ate_dst_part,
4719 rInfo);
4721 else
4722 src_rInfo = dst_rInfo = rInfo;
4725 * Fire it. Note that the AFTER_TRIGGER_IN_PROGRESS flag is
4726 * still set, so recursive examinations of the event list
4727 * won't try to re-fire it.
4729 AfterTriggerExecute(estate, event, rInfo,
4730 src_rInfo, dst_rInfo,
4731 trigdesc, finfo, instr,
4732 per_tuple_context, slot1, slot2);
4735 * Mark the event as done.
4737 event->ate_flags &= ~AFTER_TRIGGER_IN_PROGRESS;
4738 event->ate_flags |= AFTER_TRIGGER_DONE;
4740 else if (!(event->ate_flags & AFTER_TRIGGER_DONE))
4742 /* something remains to be done */
4743 all_fired = all_fired_in_chunk = false;
4747 /* Clear the chunk if delete_ok and nothing left of interest */
4748 if (delete_ok && all_fired_in_chunk)
4750 chunk->freeptr = CHUNK_DATA_START(chunk);
4751 chunk->endfree = chunk->endptr;
4754 * If it's last chunk, must sync event list's tailfree too. Note
4755 * that delete_ok must NOT be passed as true if there could be
4756 * additional AfterTriggerEventList values pointing at this event
4757 * list, since we'd fail to fix their copies of tailfree.
4759 if (chunk == events->tail)
4760 events->tailfree = chunk->freeptr;
4763 if (slot1 != NULL)
4765 ExecDropSingleTupleTableSlot(slot1);
4766 ExecDropSingleTupleTableSlot(slot2);
4769 /* Release working resources */
4770 MemoryContextDelete(per_tuple_context);
4772 if (local_estate)
4774 ExecCloseResultRelations(estate);
4775 ExecResetTupleTable(estate->es_tupleTable, false);
4776 FreeExecutorState(estate);
4779 return all_fired;
4784 * GetAfterTriggersTableData
4786 * Find or create an AfterTriggersTableData struct for the specified
4787 * trigger event (relation + operation type). Ignore existing structs
4788 * marked "closed"; we don't want to put any additional tuples into them,
4789 * nor change their stmt-triggers-fired state.
4791 * Note: the AfterTriggersTableData list is allocated in the current
4792 * (sub)transaction's CurTransactionContext. This is OK because
4793 * we don't need it to live past AfterTriggerEndQuery.
4795 static AfterTriggersTableData *
4796 GetAfterTriggersTableData(Oid relid, CmdType cmdType)
4798 AfterTriggersTableData *table;
4799 AfterTriggersQueryData *qs;
4800 MemoryContext oldcxt;
4801 ListCell *lc;
4803 /* Caller should have ensured query_depth is OK. */
4804 Assert(afterTriggers.query_depth >= 0 &&
4805 afterTriggers.query_depth < afterTriggers.maxquerydepth);
4806 qs = &afterTriggers.query_stack[afterTriggers.query_depth];
4808 foreach(lc, qs->tables)
4810 table = (AfterTriggersTableData *) lfirst(lc);
4811 if (table->relid == relid && table->cmdType == cmdType &&
4812 !table->closed)
4813 return table;
4816 oldcxt = MemoryContextSwitchTo(CurTransactionContext);
4818 table = (AfterTriggersTableData *) palloc0(sizeof(AfterTriggersTableData));
4819 table->relid = relid;
4820 table->cmdType = cmdType;
4821 qs->tables = lappend(qs->tables, table);
4823 MemoryContextSwitchTo(oldcxt);
4825 return table;
4829 * Returns a TupleTableSlot suitable for holding the tuples to be put
4830 * into AfterTriggersTableData's transition table tuplestores.
4832 static TupleTableSlot *
4833 GetAfterTriggersStoreSlot(AfterTriggersTableData *table,
4834 TupleDesc tupdesc)
4836 /* Create it if not already done. */
4837 if (!table->storeslot)
4839 MemoryContext oldcxt;
4842 * We need this slot only until AfterTriggerEndQuery, but making it
4843 * last till end-of-subxact is good enough. It'll be freed by
4844 * AfterTriggerFreeQuery(). However, the passed-in tupdesc might have
4845 * a different lifespan, so we'd better make a copy of that.
4847 oldcxt = MemoryContextSwitchTo(CurTransactionContext);
4848 tupdesc = CreateTupleDescCopy(tupdesc);
4849 table->storeslot = MakeSingleTupleTableSlot(tupdesc, &TTSOpsVirtual);
4850 MemoryContextSwitchTo(oldcxt);
4853 return table->storeslot;
4857 * MakeTransitionCaptureState
4859 * Make a TransitionCaptureState object for the given TriggerDesc, target
4860 * relation, and operation type. The TCS object holds all the state needed
4861 * to decide whether to capture tuples in transition tables.
4863 * If there are no triggers in 'trigdesc' that request relevant transition
4864 * tables, then return NULL.
4866 * The resulting object can be passed to the ExecAR* functions. When
4867 * dealing with child tables, the caller can set tcs_original_insert_tuple
4868 * to avoid having to reconstruct the original tuple in the root table's
4869 * format.
4871 * Note that we copy the flags from a parent table into this struct (rather
4872 * than subsequently using the relation's TriggerDesc directly) so that we can
4873 * use it to control collection of transition tuples from child tables.
4875 * Per SQL spec, all operations of the same kind (INSERT/UPDATE/DELETE)
4876 * on the same table during one query should share one transition table.
4877 * Therefore, the Tuplestores are owned by an AfterTriggersTableData struct
4878 * looked up using the table OID + CmdType, and are merely referenced by
4879 * the TransitionCaptureState objects we hand out to callers.
4881 TransitionCaptureState *
4882 MakeTransitionCaptureState(TriggerDesc *trigdesc, Oid relid, CmdType cmdType)
4884 TransitionCaptureState *state;
4885 bool need_old_upd,
4886 need_new_upd,
4887 need_old_del,
4888 need_new_ins;
4889 AfterTriggersTableData *table;
4890 MemoryContext oldcxt;
4891 ResourceOwner saveResourceOwner;
4893 if (trigdesc == NULL)
4894 return NULL;
4896 /* Detect which table(s) we need. */
4897 switch (cmdType)
4899 case CMD_INSERT:
4900 need_old_upd = need_old_del = need_new_upd = false;
4901 need_new_ins = trigdesc->trig_insert_new_table;
4902 break;
4903 case CMD_UPDATE:
4904 need_old_upd = trigdesc->trig_update_old_table;
4905 need_new_upd = trigdesc->trig_update_new_table;
4906 need_old_del = need_new_ins = false;
4907 break;
4908 case CMD_DELETE:
4909 need_old_del = trigdesc->trig_delete_old_table;
4910 need_old_upd = need_new_upd = need_new_ins = false;
4911 break;
4912 case CMD_MERGE:
4913 need_old_upd = trigdesc->trig_update_old_table;
4914 need_new_upd = trigdesc->trig_update_new_table;
4915 need_old_del = trigdesc->trig_delete_old_table;
4916 need_new_ins = trigdesc->trig_insert_new_table;
4917 break;
4918 default:
4919 elog(ERROR, "unexpected CmdType: %d", (int) cmdType);
4920 /* keep compiler quiet */
4921 need_old_upd = need_new_upd = need_old_del = need_new_ins = false;
4922 break;
4924 if (!need_old_upd && !need_new_upd && !need_new_ins && !need_old_del)
4925 return NULL;
4927 /* Check state, like AfterTriggerSaveEvent. */
4928 if (afterTriggers.query_depth < 0)
4929 elog(ERROR, "MakeTransitionCaptureState() called outside of query");
4931 /* Be sure we have enough space to record events at this query depth. */
4932 if (afterTriggers.query_depth >= afterTriggers.maxquerydepth)
4933 AfterTriggerEnlargeQueryState();
4936 * Find or create an AfterTriggersTableData struct to hold the
4937 * tuplestore(s). If there's a matching struct but it's marked closed,
4938 * ignore it; we need a newer one.
4940 * Note: the AfterTriggersTableData list, as well as the tuplestores, are
4941 * allocated in the current (sub)transaction's CurTransactionContext, and
4942 * the tuplestores are managed by the (sub)transaction's resource owner.
4943 * This is sufficient lifespan because we do not allow triggers using
4944 * transition tables to be deferrable; they will be fired during
4945 * AfterTriggerEndQuery, after which it's okay to delete the data.
4947 table = GetAfterTriggersTableData(relid, cmdType);
4949 /* Now create required tuplestore(s), if we don't have them already. */
4950 oldcxt = MemoryContextSwitchTo(CurTransactionContext);
4951 saveResourceOwner = CurrentResourceOwner;
4952 CurrentResourceOwner = CurTransactionResourceOwner;
4954 if (need_old_upd && table->old_upd_tuplestore == NULL)
4955 table->old_upd_tuplestore = tuplestore_begin_heap(false, false, work_mem);
4956 if (need_new_upd && table->new_upd_tuplestore == NULL)
4957 table->new_upd_tuplestore = tuplestore_begin_heap(false, false, work_mem);
4958 if (need_old_del && table->old_del_tuplestore == NULL)
4959 table->old_del_tuplestore = tuplestore_begin_heap(false, false, work_mem);
4960 if (need_new_ins && table->new_ins_tuplestore == NULL)
4961 table->new_ins_tuplestore = tuplestore_begin_heap(false, false, work_mem);
4963 CurrentResourceOwner = saveResourceOwner;
4964 MemoryContextSwitchTo(oldcxt);
4966 /* Now build the TransitionCaptureState struct, in caller's context */
4967 state = (TransitionCaptureState *) palloc0(sizeof(TransitionCaptureState));
4968 state->tcs_delete_old_table = trigdesc->trig_delete_old_table;
4969 state->tcs_update_old_table = trigdesc->trig_update_old_table;
4970 state->tcs_update_new_table = trigdesc->trig_update_new_table;
4971 state->tcs_insert_new_table = trigdesc->trig_insert_new_table;
4972 state->tcs_private = table;
4974 return state;
4978 /* ----------
4979 * AfterTriggerBeginXact()
4981 * Called at transaction start (either BEGIN or implicit for single
4982 * statement outside of transaction block).
4983 * ----------
4985 void
4986 AfterTriggerBeginXact(void)
4989 * Initialize after-trigger state structure to empty
4991 afterTriggers.firing_counter = (CommandId) 1; /* mustn't be 0 */
4992 afterTriggers.query_depth = -1;
4995 * Verify that there is no leftover state remaining. If these assertions
4996 * trip, it means that AfterTriggerEndXact wasn't called or didn't clean
4997 * up properly.
4999 Assert(afterTriggers.state == NULL);
5000 Assert(afterTriggers.query_stack == NULL);
5001 Assert(afterTriggers.maxquerydepth == 0);
5002 Assert(afterTriggers.event_cxt == NULL);
5003 Assert(afterTriggers.events.head == NULL);
5004 Assert(afterTriggers.trans_stack == NULL);
5005 Assert(afterTriggers.maxtransdepth == 0);
5009 /* ----------
5010 * AfterTriggerBeginQuery()
5012 * Called just before we start processing a single query within a
5013 * transaction (or subtransaction). Most of the real work gets deferred
5014 * until somebody actually tries to queue a trigger event.
5015 * ----------
5017 void
5018 AfterTriggerBeginQuery(void)
5020 /* Increase the query stack depth */
5021 afterTriggers.query_depth++;
5025 /* ----------
5026 * AfterTriggerEndQuery()
5028 * Called after one query has been completely processed. At this time
5029 * we invoke all AFTER IMMEDIATE trigger events queued by the query, and
5030 * transfer deferred trigger events to the global deferred-trigger list.
5032 * Note that this must be called BEFORE closing down the executor
5033 * with ExecutorEnd, because we make use of the EState's info about
5034 * target relations. Normally it is called from ExecutorFinish.
5035 * ----------
5037 void
5038 AfterTriggerEndQuery(EState *estate)
5040 AfterTriggersQueryData *qs;
5042 /* Must be inside a query, too */
5043 Assert(afterTriggers.query_depth >= 0);
5046 * If we never even got as far as initializing the event stack, there
5047 * certainly won't be any events, so exit quickly.
5049 if (afterTriggers.query_depth >= afterTriggers.maxquerydepth)
5051 afterTriggers.query_depth--;
5052 return;
5056 * Process all immediate-mode triggers queued by the query, and move the
5057 * deferred ones to the main list of deferred events.
5059 * Notice that we decide which ones will be fired, and put the deferred
5060 * ones on the main list, before anything is actually fired. This ensures
5061 * reasonably sane behavior if a trigger function does SET CONSTRAINTS ...
5062 * IMMEDIATE: all events we have decided to defer will be available for it
5063 * to fire.
5065 * We loop in case a trigger queues more events at the same query level.
5066 * Ordinary trigger functions, including all PL/pgSQL trigger functions,
5067 * will instead fire any triggers in a dedicated query level. Foreign key
5068 * enforcement triggers do add to the current query level, thanks to their
5069 * passing fire_triggers = false to SPI_execute_snapshot(). Other
5070 * C-language triggers might do likewise.
5072 * If we find no firable events, we don't have to increment
5073 * firing_counter.
5075 qs = &afterTriggers.query_stack[afterTriggers.query_depth];
5077 for (;;)
5079 if (afterTriggerMarkEvents(&qs->events, &afterTriggers.events, true))
5081 CommandId firing_id = afterTriggers.firing_counter++;
5082 AfterTriggerEventChunk *oldtail = qs->events.tail;
5084 if (afterTriggerInvokeEvents(&qs->events, firing_id, estate, false))
5085 break; /* all fired */
5088 * Firing a trigger could result in query_stack being repalloc'd,
5089 * so we must recalculate qs after each afterTriggerInvokeEvents
5090 * call. Furthermore, it's unsafe to pass delete_ok = true here,
5091 * because that could cause afterTriggerInvokeEvents to try to
5092 * access qs->events after the stack has been repalloc'd.
5094 qs = &afterTriggers.query_stack[afterTriggers.query_depth];
5097 * We'll need to scan the events list again. To reduce the cost
5098 * of doing so, get rid of completely-fired chunks. We know that
5099 * all events were marked IN_PROGRESS or DONE at the conclusion of
5100 * afterTriggerMarkEvents, so any still-interesting events must
5101 * have been added after that, and so must be in the chunk that
5102 * was then the tail chunk, or in later chunks. So, zap all
5103 * chunks before oldtail. This is approximately the same set of
5104 * events we would have gotten rid of by passing delete_ok = true.
5106 Assert(oldtail != NULL);
5107 while (qs->events.head != oldtail)
5108 afterTriggerDeleteHeadEventChunk(qs);
5110 else
5111 break;
5114 /* Release query-level-local storage, including tuplestores if any */
5115 AfterTriggerFreeQuery(&afterTriggers.query_stack[afterTriggers.query_depth]);
5117 afterTriggers.query_depth--;
5122 * AfterTriggerFreeQuery
5123 * Release subsidiary storage for a trigger query level.
5124 * This includes closing down tuplestores.
5125 * Note: it's important for this to be safe if interrupted by an error
5126 * and then called again for the same query level.
5128 static void
5129 AfterTriggerFreeQuery(AfterTriggersQueryData *qs)
5131 Tuplestorestate *ts;
5132 List *tables;
5133 ListCell *lc;
5135 /* Drop the trigger events */
5136 afterTriggerFreeEventList(&qs->events);
5138 /* Drop FDW tuplestore if any */
5139 ts = qs->fdw_tuplestore;
5140 qs->fdw_tuplestore = NULL;
5141 if (ts)
5142 tuplestore_end(ts);
5144 /* Release per-table subsidiary storage */
5145 tables = qs->tables;
5146 foreach(lc, tables)
5148 AfterTriggersTableData *table = (AfterTriggersTableData *) lfirst(lc);
5150 ts = table->old_upd_tuplestore;
5151 table->old_upd_tuplestore = NULL;
5152 if (ts)
5153 tuplestore_end(ts);
5154 ts = table->new_upd_tuplestore;
5155 table->new_upd_tuplestore = NULL;
5156 if (ts)
5157 tuplestore_end(ts);
5158 ts = table->old_del_tuplestore;
5159 table->old_del_tuplestore = NULL;
5160 if (ts)
5161 tuplestore_end(ts);
5162 ts = table->new_ins_tuplestore;
5163 table->new_ins_tuplestore = NULL;
5164 if (ts)
5165 tuplestore_end(ts);
5166 if (table->storeslot)
5168 TupleTableSlot *slot = table->storeslot;
5170 table->storeslot = NULL;
5171 ExecDropSingleTupleTableSlot(slot);
5176 * Now free the AfterTriggersTableData structs and list cells. Reset list
5177 * pointer first; if list_free_deep somehow gets an error, better to leak
5178 * that storage than have an infinite loop.
5180 qs->tables = NIL;
5181 list_free_deep(tables);
5185 /* ----------
5186 * AfterTriggerFireDeferred()
5188 * Called just before the current transaction is committed. At this
5189 * time we invoke all pending DEFERRED triggers.
5191 * It is possible for other modules to queue additional deferred triggers
5192 * during pre-commit processing; therefore xact.c may have to call this
5193 * multiple times.
5194 * ----------
5196 void
5197 AfterTriggerFireDeferred(void)
5199 AfterTriggerEventList *events;
5200 bool snap_pushed = false;
5202 /* Must not be inside a query */
5203 Assert(afterTriggers.query_depth == -1);
5206 * If there are any triggers to fire, make sure we have set a snapshot for
5207 * them to use. (Since PortalRunUtility doesn't set a snap for COMMIT, we
5208 * can't assume ActiveSnapshot is valid on entry.)
5210 events = &afterTriggers.events;
5211 if (events->head != NULL)
5213 PushActiveSnapshot(GetTransactionSnapshot());
5214 snap_pushed = true;
5218 * Run all the remaining triggers. Loop until they are all gone, in case
5219 * some trigger queues more for us to do.
5221 while (afterTriggerMarkEvents(events, NULL, false))
5223 CommandId firing_id = afterTriggers.firing_counter++;
5225 if (afterTriggerInvokeEvents(events, firing_id, NULL, true))
5226 break; /* all fired */
5230 * We don't bother freeing the event list, since it will go away anyway
5231 * (and more efficiently than via pfree) in AfterTriggerEndXact.
5234 if (snap_pushed)
5235 PopActiveSnapshot();
5239 /* ----------
5240 * AfterTriggerEndXact()
5242 * The current transaction is finishing.
5244 * Any unfired triggers are canceled so we simply throw
5245 * away anything we know.
5247 * Note: it is possible for this to be called repeatedly in case of
5248 * error during transaction abort; therefore, do not complain if
5249 * already closed down.
5250 * ----------
5252 void
5253 AfterTriggerEndXact(bool isCommit)
5256 * Forget the pending-events list.
5258 * Since all the info is in TopTransactionContext or children thereof, we
5259 * don't really need to do anything to reclaim memory. However, the
5260 * pending-events list could be large, and so it's useful to discard it as
5261 * soon as possible --- especially if we are aborting because we ran out
5262 * of memory for the list!
5264 if (afterTriggers.event_cxt)
5266 MemoryContextDelete(afterTriggers.event_cxt);
5267 afterTriggers.event_cxt = NULL;
5268 afterTriggers.events.head = NULL;
5269 afterTriggers.events.tail = NULL;
5270 afterTriggers.events.tailfree = NULL;
5274 * Forget any subtransaction state as well. Since this can't be very
5275 * large, we let the eventual reset of TopTransactionContext free the
5276 * memory instead of doing it here.
5278 afterTriggers.trans_stack = NULL;
5279 afterTriggers.maxtransdepth = 0;
5283 * Forget the query stack and constraint-related state information. As
5284 * with the subtransaction state information, we don't bother freeing the
5285 * memory here.
5287 afterTriggers.query_stack = NULL;
5288 afterTriggers.maxquerydepth = 0;
5289 afterTriggers.state = NULL;
5291 /* No more afterTriggers manipulation until next transaction starts. */
5292 afterTriggers.query_depth = -1;
5296 * AfterTriggerBeginSubXact()
5298 * Start a subtransaction.
5300 void
5301 AfterTriggerBeginSubXact(void)
5303 int my_level = GetCurrentTransactionNestLevel();
5306 * Allocate more space in the trans_stack if needed. (Note: because the
5307 * minimum nest level of a subtransaction is 2, we waste the first couple
5308 * entries of the array; not worth the notational effort to avoid it.)
5310 while (my_level >= afterTriggers.maxtransdepth)
5312 if (afterTriggers.maxtransdepth == 0)
5314 /* Arbitrarily initialize for max of 8 subtransaction levels */
5315 afterTriggers.trans_stack = (AfterTriggersTransData *)
5316 MemoryContextAlloc(TopTransactionContext,
5317 8 * sizeof(AfterTriggersTransData));
5318 afterTriggers.maxtransdepth = 8;
5320 else
5322 /* repalloc will keep the stack in the same context */
5323 int new_alloc = afterTriggers.maxtransdepth * 2;
5325 afterTriggers.trans_stack = (AfterTriggersTransData *)
5326 repalloc(afterTriggers.trans_stack,
5327 new_alloc * sizeof(AfterTriggersTransData));
5328 afterTriggers.maxtransdepth = new_alloc;
5333 * Push the current information into the stack. The SET CONSTRAINTS state
5334 * is not saved until/unless changed. Likewise, we don't make a
5335 * per-subtransaction event context until needed.
5337 afterTriggers.trans_stack[my_level].state = NULL;
5338 afterTriggers.trans_stack[my_level].events = afterTriggers.events;
5339 afterTriggers.trans_stack[my_level].query_depth = afterTriggers.query_depth;
5340 afterTriggers.trans_stack[my_level].firing_counter = afterTriggers.firing_counter;
5344 * AfterTriggerEndSubXact()
5346 * The current subtransaction is ending.
5348 void
5349 AfterTriggerEndSubXact(bool isCommit)
5351 int my_level = GetCurrentTransactionNestLevel();
5352 SetConstraintState state;
5353 AfterTriggerEvent event;
5354 AfterTriggerEventChunk *chunk;
5355 CommandId subxact_firing_id;
5358 * Pop the prior state if needed.
5360 if (isCommit)
5362 Assert(my_level < afterTriggers.maxtransdepth);
5363 /* If we saved a prior state, we don't need it anymore */
5364 state = afterTriggers.trans_stack[my_level].state;
5365 if (state != NULL)
5366 pfree(state);
5367 /* this avoids double pfree if error later: */
5368 afterTriggers.trans_stack[my_level].state = NULL;
5369 Assert(afterTriggers.query_depth ==
5370 afterTriggers.trans_stack[my_level].query_depth);
5372 else
5375 * Aborting. It is possible subxact start failed before calling
5376 * AfterTriggerBeginSubXact, in which case we mustn't risk touching
5377 * trans_stack levels that aren't there.
5379 if (my_level >= afterTriggers.maxtransdepth)
5380 return;
5383 * Release query-level storage for queries being aborted, and restore
5384 * query_depth to its pre-subxact value. This assumes that a
5385 * subtransaction will not add events to query levels started in a
5386 * earlier transaction state.
5388 while (afterTriggers.query_depth > afterTriggers.trans_stack[my_level].query_depth)
5390 if (afterTriggers.query_depth < afterTriggers.maxquerydepth)
5391 AfterTriggerFreeQuery(&afterTriggers.query_stack[afterTriggers.query_depth]);
5392 afterTriggers.query_depth--;
5394 Assert(afterTriggers.query_depth ==
5395 afterTriggers.trans_stack[my_level].query_depth);
5398 * Restore the global deferred-event list to its former length,
5399 * discarding any events queued by the subxact.
5401 afterTriggerRestoreEventList(&afterTriggers.events,
5402 &afterTriggers.trans_stack[my_level].events);
5405 * Restore the trigger state. If the saved state is NULL, then this
5406 * subxact didn't save it, so it doesn't need restoring.
5408 state = afterTriggers.trans_stack[my_level].state;
5409 if (state != NULL)
5411 pfree(afterTriggers.state);
5412 afterTriggers.state = state;
5414 /* this avoids double pfree if error later: */
5415 afterTriggers.trans_stack[my_level].state = NULL;
5418 * Scan for any remaining deferred events that were marked DONE or IN
5419 * PROGRESS by this subxact or a child, and un-mark them. We can
5420 * recognize such events because they have a firing ID greater than or
5421 * equal to the firing_counter value we saved at subtransaction start.
5422 * (This essentially assumes that the current subxact includes all
5423 * subxacts started after it.)
5425 subxact_firing_id = afterTriggers.trans_stack[my_level].firing_counter;
5426 for_each_event_chunk(event, chunk, afterTriggers.events)
5428 AfterTriggerShared evtshared = GetTriggerSharedData(event);
5430 if (event->ate_flags &
5431 (AFTER_TRIGGER_DONE | AFTER_TRIGGER_IN_PROGRESS))
5433 if (evtshared->ats_firing_id >= subxact_firing_id)
5434 event->ate_flags &=
5435 ~(AFTER_TRIGGER_DONE | AFTER_TRIGGER_IN_PROGRESS);
5442 * Get the transition table for the given event and depending on whether we are
5443 * processing the old or the new tuple.
5445 static Tuplestorestate *
5446 GetAfterTriggersTransitionTable(int event,
5447 TupleTableSlot *oldslot,
5448 TupleTableSlot *newslot,
5449 TransitionCaptureState *transition_capture)
5451 Tuplestorestate *tuplestore = NULL;
5452 bool delete_old_table = transition_capture->tcs_delete_old_table;
5453 bool update_old_table = transition_capture->tcs_update_old_table;
5454 bool update_new_table = transition_capture->tcs_update_new_table;
5455 bool insert_new_table = transition_capture->tcs_insert_new_table;
5458 * For INSERT events NEW should be non-NULL, for DELETE events OLD should
5459 * be non-NULL, whereas for UPDATE events normally both OLD and NEW are
5460 * non-NULL. But for UPDATE events fired for capturing transition tuples
5461 * during UPDATE partition-key row movement, OLD is NULL when the event is
5462 * for a row being inserted, whereas NEW is NULL when the event is for a
5463 * row being deleted.
5465 Assert(!(event == TRIGGER_EVENT_DELETE && delete_old_table &&
5466 TupIsNull(oldslot)));
5467 Assert(!(event == TRIGGER_EVENT_INSERT && insert_new_table &&
5468 TupIsNull(newslot)));
5470 if (!TupIsNull(oldslot))
5472 Assert(TupIsNull(newslot));
5473 if (event == TRIGGER_EVENT_DELETE && delete_old_table)
5474 tuplestore = transition_capture->tcs_private->old_del_tuplestore;
5475 else if (event == TRIGGER_EVENT_UPDATE && update_old_table)
5476 tuplestore = transition_capture->tcs_private->old_upd_tuplestore;
5478 else if (!TupIsNull(newslot))
5480 Assert(TupIsNull(oldslot));
5481 if (event == TRIGGER_EVENT_INSERT && insert_new_table)
5482 tuplestore = transition_capture->tcs_private->new_ins_tuplestore;
5483 else if (event == TRIGGER_EVENT_UPDATE && update_new_table)
5484 tuplestore = transition_capture->tcs_private->new_upd_tuplestore;
5487 return tuplestore;
5491 * Add the given heap tuple to the given tuplestore, applying the conversion
5492 * map if necessary.
5494 * If original_insert_tuple is given, we can add that tuple without conversion.
5496 static void
5497 TransitionTableAddTuple(EState *estate,
5498 TransitionCaptureState *transition_capture,
5499 ResultRelInfo *relinfo,
5500 TupleTableSlot *slot,
5501 TupleTableSlot *original_insert_tuple,
5502 Tuplestorestate *tuplestore)
5504 TupleConversionMap *map;
5507 * Nothing needs to be done if we don't have a tuplestore.
5509 if (tuplestore == NULL)
5510 return;
5512 if (original_insert_tuple)
5513 tuplestore_puttupleslot(tuplestore, original_insert_tuple);
5514 else if ((map = ExecGetChildToRootMap(relinfo)) != NULL)
5516 AfterTriggersTableData *table = transition_capture->tcs_private;
5517 TupleTableSlot *storeslot;
5519 storeslot = GetAfterTriggersStoreSlot(table, map->outdesc);
5520 execute_attr_map_slot(map->attrMap, slot, storeslot);
5521 tuplestore_puttupleslot(tuplestore, storeslot);
5523 else
5524 tuplestore_puttupleslot(tuplestore, slot);
5527 /* ----------
5528 * AfterTriggerEnlargeQueryState()
5530 * Prepare the necessary state so that we can record AFTER trigger events
5531 * queued by a query. It is allowed to have nested queries within a
5532 * (sub)transaction, so we need to have separate state for each query
5533 * nesting level.
5534 * ----------
5536 static void
5537 AfterTriggerEnlargeQueryState(void)
5539 int init_depth = afterTriggers.maxquerydepth;
5541 Assert(afterTriggers.query_depth >= afterTriggers.maxquerydepth);
5543 if (afterTriggers.maxquerydepth == 0)
5545 int new_alloc = Max(afterTriggers.query_depth + 1, 8);
5547 afterTriggers.query_stack = (AfterTriggersQueryData *)
5548 MemoryContextAlloc(TopTransactionContext,
5549 new_alloc * sizeof(AfterTriggersQueryData));
5550 afterTriggers.maxquerydepth = new_alloc;
5552 else
5554 /* repalloc will keep the stack in the same context */
5555 int old_alloc = afterTriggers.maxquerydepth;
5556 int new_alloc = Max(afterTriggers.query_depth + 1,
5557 old_alloc * 2);
5559 afterTriggers.query_stack = (AfterTriggersQueryData *)
5560 repalloc(afterTriggers.query_stack,
5561 new_alloc * sizeof(AfterTriggersQueryData));
5562 afterTriggers.maxquerydepth = new_alloc;
5565 /* Initialize new array entries to empty */
5566 while (init_depth < afterTriggers.maxquerydepth)
5568 AfterTriggersQueryData *qs = &afterTriggers.query_stack[init_depth];
5570 qs->events.head = NULL;
5571 qs->events.tail = NULL;
5572 qs->events.tailfree = NULL;
5573 qs->fdw_tuplestore = NULL;
5574 qs->tables = NIL;
5576 ++init_depth;
5581 * Create an empty SetConstraintState with room for numalloc trigstates
5583 static SetConstraintState
5584 SetConstraintStateCreate(int numalloc)
5586 SetConstraintState state;
5588 /* Behave sanely with numalloc == 0 */
5589 if (numalloc <= 0)
5590 numalloc = 1;
5593 * We assume that zeroing will correctly initialize the state values.
5595 state = (SetConstraintState)
5596 MemoryContextAllocZero(TopTransactionContext,
5597 offsetof(SetConstraintStateData, trigstates) +
5598 numalloc * sizeof(SetConstraintTriggerData));
5600 state->numalloc = numalloc;
5602 return state;
5606 * Copy a SetConstraintState
5608 static SetConstraintState
5609 SetConstraintStateCopy(SetConstraintState origstate)
5611 SetConstraintState state;
5613 state = SetConstraintStateCreate(origstate->numstates);
5615 state->all_isset = origstate->all_isset;
5616 state->all_isdeferred = origstate->all_isdeferred;
5617 state->numstates = origstate->numstates;
5618 memcpy(state->trigstates, origstate->trigstates,
5619 origstate->numstates * sizeof(SetConstraintTriggerData));
5621 return state;
5625 * Add a per-trigger item to a SetConstraintState. Returns possibly-changed
5626 * pointer to the state object (it will change if we have to repalloc).
5628 static SetConstraintState
5629 SetConstraintStateAddItem(SetConstraintState state,
5630 Oid tgoid, bool tgisdeferred)
5632 if (state->numstates >= state->numalloc)
5634 int newalloc = state->numalloc * 2;
5636 newalloc = Max(newalloc, 8); /* in case original has size 0 */
5637 state = (SetConstraintState)
5638 repalloc(state,
5639 offsetof(SetConstraintStateData, trigstates) +
5640 newalloc * sizeof(SetConstraintTriggerData));
5641 state->numalloc = newalloc;
5642 Assert(state->numstates < state->numalloc);
5645 state->trigstates[state->numstates].sct_tgoid = tgoid;
5646 state->trigstates[state->numstates].sct_tgisdeferred = tgisdeferred;
5647 state->numstates++;
5649 return state;
5652 /* ----------
5653 * AfterTriggerSetState()
5655 * Execute the SET CONSTRAINTS ... utility command.
5656 * ----------
5658 void
5659 AfterTriggerSetState(ConstraintsSetStmt *stmt)
5661 int my_level = GetCurrentTransactionNestLevel();
5663 /* If we haven't already done so, initialize our state. */
5664 if (afterTriggers.state == NULL)
5665 afterTriggers.state = SetConstraintStateCreate(8);
5668 * If in a subtransaction, and we didn't save the current state already,
5669 * save it so it can be restored if the subtransaction aborts.
5671 if (my_level > 1 &&
5672 afterTriggers.trans_stack[my_level].state == NULL)
5674 afterTriggers.trans_stack[my_level].state =
5675 SetConstraintStateCopy(afterTriggers.state);
5679 * Handle SET CONSTRAINTS ALL ...
5681 if (stmt->constraints == NIL)
5684 * Forget any previous SET CONSTRAINTS commands in this transaction.
5686 afterTriggers.state->numstates = 0;
5689 * Set the per-transaction ALL state to known.
5691 afterTriggers.state->all_isset = true;
5692 afterTriggers.state->all_isdeferred = stmt->deferred;
5694 else
5696 Relation conrel;
5697 Relation tgrel;
5698 List *conoidlist = NIL;
5699 List *tgoidlist = NIL;
5700 ListCell *lc;
5703 * Handle SET CONSTRAINTS constraint-name [, ...]
5705 * First, identify all the named constraints and make a list of their
5706 * OIDs. Since, unlike the SQL spec, we allow multiple constraints of
5707 * the same name within a schema, the specifications are not
5708 * necessarily unique. Our strategy is to target all matching
5709 * constraints within the first search-path schema that has any
5710 * matches, but disregard matches in schemas beyond the first match.
5711 * (This is a bit odd but it's the historical behavior.)
5713 * A constraint in a partitioned table may have corresponding
5714 * constraints in the partitions. Grab those too.
5716 conrel = table_open(ConstraintRelationId, AccessShareLock);
5718 foreach(lc, stmt->constraints)
5720 RangeVar *constraint = lfirst(lc);
5721 bool found;
5722 List *namespacelist;
5723 ListCell *nslc;
5725 if (constraint->catalogname)
5727 if (strcmp(constraint->catalogname, get_database_name(MyDatabaseId)) != 0)
5728 ereport(ERROR,
5729 (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
5730 errmsg("cross-database references are not implemented: \"%s.%s.%s\"",
5731 constraint->catalogname, constraint->schemaname,
5732 constraint->relname)));
5736 * If we're given the schema name with the constraint, look only
5737 * in that schema. If given a bare constraint name, use the
5738 * search path to find the first matching constraint.
5740 if (constraint->schemaname)
5742 Oid namespaceId = LookupExplicitNamespace(constraint->schemaname,
5743 false);
5745 namespacelist = list_make1_oid(namespaceId);
5747 else
5749 namespacelist = fetch_search_path(true);
5752 found = false;
5753 foreach(nslc, namespacelist)
5755 Oid namespaceId = lfirst_oid(nslc);
5756 SysScanDesc conscan;
5757 ScanKeyData skey[2];
5758 HeapTuple tup;
5760 ScanKeyInit(&skey[0],
5761 Anum_pg_constraint_conname,
5762 BTEqualStrategyNumber, F_NAMEEQ,
5763 CStringGetDatum(constraint->relname));
5764 ScanKeyInit(&skey[1],
5765 Anum_pg_constraint_connamespace,
5766 BTEqualStrategyNumber, F_OIDEQ,
5767 ObjectIdGetDatum(namespaceId));
5769 conscan = systable_beginscan(conrel, ConstraintNameNspIndexId,
5770 true, NULL, 2, skey);
5772 while (HeapTupleIsValid(tup = systable_getnext(conscan)))
5774 Form_pg_constraint con = (Form_pg_constraint) GETSTRUCT(tup);
5776 if (con->condeferrable)
5777 conoidlist = lappend_oid(conoidlist, con->oid);
5778 else if (stmt->deferred)
5779 ereport(ERROR,
5780 (errcode(ERRCODE_WRONG_OBJECT_TYPE),
5781 errmsg("constraint \"%s\" is not deferrable",
5782 constraint->relname)));
5783 found = true;
5786 systable_endscan(conscan);
5789 * Once we've found a matching constraint we do not search
5790 * later parts of the search path.
5792 if (found)
5793 break;
5796 list_free(namespacelist);
5799 * Not found ?
5801 if (!found)
5802 ereport(ERROR,
5803 (errcode(ERRCODE_UNDEFINED_OBJECT),
5804 errmsg("constraint \"%s\" does not exist",
5805 constraint->relname)));
5809 * Scan for any possible descendants of the constraints. We append
5810 * whatever we find to the same list that we're scanning; this has the
5811 * effect that we create new scans for those, too, so if there are
5812 * further descendents, we'll also catch them.
5814 foreach(lc, conoidlist)
5816 Oid parent = lfirst_oid(lc);
5817 ScanKeyData key;
5818 SysScanDesc scan;
5819 HeapTuple tuple;
5821 ScanKeyInit(&key,
5822 Anum_pg_constraint_conparentid,
5823 BTEqualStrategyNumber, F_OIDEQ,
5824 ObjectIdGetDatum(parent));
5826 scan = systable_beginscan(conrel, ConstraintParentIndexId, true, NULL, 1, &key);
5828 while (HeapTupleIsValid(tuple = systable_getnext(scan)))
5830 Form_pg_constraint con = (Form_pg_constraint) GETSTRUCT(tuple);
5832 conoidlist = lappend_oid(conoidlist, con->oid);
5835 systable_endscan(scan);
5838 table_close(conrel, AccessShareLock);
5841 * Now, locate the trigger(s) implementing each of these constraints,
5842 * and make a list of their OIDs.
5844 tgrel = table_open(TriggerRelationId, AccessShareLock);
5846 foreach(lc, conoidlist)
5848 Oid conoid = lfirst_oid(lc);
5849 ScanKeyData skey;
5850 SysScanDesc tgscan;
5851 HeapTuple htup;
5853 ScanKeyInit(&skey,
5854 Anum_pg_trigger_tgconstraint,
5855 BTEqualStrategyNumber, F_OIDEQ,
5856 ObjectIdGetDatum(conoid));
5858 tgscan = systable_beginscan(tgrel, TriggerConstraintIndexId, true,
5859 NULL, 1, &skey);
5861 while (HeapTupleIsValid(htup = systable_getnext(tgscan)))
5863 Form_pg_trigger pg_trigger = (Form_pg_trigger) GETSTRUCT(htup);
5866 * Silently skip triggers that are marked as non-deferrable in
5867 * pg_trigger. This is not an error condition, since a
5868 * deferrable RI constraint may have some non-deferrable
5869 * actions.
5871 if (pg_trigger->tgdeferrable)
5872 tgoidlist = lappend_oid(tgoidlist, pg_trigger->oid);
5875 systable_endscan(tgscan);
5878 table_close(tgrel, AccessShareLock);
5881 * Now we can set the trigger states of individual triggers for this
5882 * xact.
5884 foreach(lc, tgoidlist)
5886 Oid tgoid = lfirst_oid(lc);
5887 SetConstraintState state = afterTriggers.state;
5888 bool found = false;
5889 int i;
5891 for (i = 0; i < state->numstates; i++)
5893 if (state->trigstates[i].sct_tgoid == tgoid)
5895 state->trigstates[i].sct_tgisdeferred = stmt->deferred;
5896 found = true;
5897 break;
5900 if (!found)
5902 afterTriggers.state =
5903 SetConstraintStateAddItem(state, tgoid, stmt->deferred);
5909 * SQL99 requires that when a constraint is set to IMMEDIATE, any deferred
5910 * checks against that constraint must be made when the SET CONSTRAINTS
5911 * command is executed -- i.e. the effects of the SET CONSTRAINTS command
5912 * apply retroactively. We've updated the constraints state, so scan the
5913 * list of previously deferred events to fire any that have now become
5914 * immediate.
5916 * Obviously, if this was SET ... DEFERRED then it can't have converted
5917 * any unfired events to immediate, so we need do nothing in that case.
5919 if (!stmt->deferred)
5921 AfterTriggerEventList *events = &afterTriggers.events;
5922 bool snapshot_set = false;
5924 while (afterTriggerMarkEvents(events, NULL, true))
5926 CommandId firing_id = afterTriggers.firing_counter++;
5929 * Make sure a snapshot has been established in case trigger
5930 * functions need one. Note that we avoid setting a snapshot if
5931 * we don't find at least one trigger that has to be fired now.
5932 * This is so that BEGIN; SET CONSTRAINTS ...; SET TRANSACTION
5933 * ISOLATION LEVEL SERIALIZABLE; ... works properly. (If we are
5934 * at the start of a transaction it's not possible for any trigger
5935 * events to be queued yet.)
5937 if (!snapshot_set)
5939 PushActiveSnapshot(GetTransactionSnapshot());
5940 snapshot_set = true;
5944 * We can delete fired events if we are at top transaction level,
5945 * but we'd better not if inside a subtransaction, since the
5946 * subtransaction could later get rolled back.
5948 if (afterTriggerInvokeEvents(events, firing_id, NULL,
5949 !IsSubTransaction()))
5950 break; /* all fired */
5953 if (snapshot_set)
5954 PopActiveSnapshot();
5958 /* ----------
5959 * AfterTriggerPendingOnRel()
5960 * Test to see if there are any pending after-trigger events for rel.
5962 * This is used by TRUNCATE, CLUSTER, ALTER TABLE, etc to detect whether
5963 * it is unsafe to perform major surgery on a relation. Note that only
5964 * local pending events are examined. We assume that having exclusive lock
5965 * on a rel guarantees there are no unserviced events in other backends ---
5966 * but having a lock does not prevent there being such events in our own.
5968 * In some scenarios it'd be reasonable to remove pending events (more
5969 * specifically, mark them DONE by the current subxact) but without a lot
5970 * of knowledge of the trigger semantics we can't do this in general.
5971 * ----------
5973 bool
5974 AfterTriggerPendingOnRel(Oid relid)
5976 AfterTriggerEvent event;
5977 AfterTriggerEventChunk *chunk;
5978 int depth;
5980 /* Scan queued events */
5981 for_each_event_chunk(event, chunk, afterTriggers.events)
5983 AfterTriggerShared evtshared = GetTriggerSharedData(event);
5986 * We can ignore completed events. (Even if a DONE flag is rolled
5987 * back by subxact abort, it's OK because the effects of the TRUNCATE
5988 * or whatever must get rolled back too.)
5990 if (event->ate_flags & AFTER_TRIGGER_DONE)
5991 continue;
5993 if (evtshared->ats_relid == relid)
5994 return true;
5998 * Also scan events queued by incomplete queries. This could only matter
5999 * if TRUNCATE/etc is executed by a function or trigger within an updating
6000 * query on the same relation, which is pretty perverse, but let's check.
6002 for (depth = 0; depth <= afterTriggers.query_depth && depth < afterTriggers.maxquerydepth; depth++)
6004 for_each_event_chunk(event, chunk, afterTriggers.query_stack[depth].events)
6006 AfterTriggerShared evtshared = GetTriggerSharedData(event);
6008 if (event->ate_flags & AFTER_TRIGGER_DONE)
6009 continue;
6011 if (evtshared->ats_relid == relid)
6012 return true;
6016 return false;
6019 /* ----------
6020 * AfterTriggerSaveEvent()
6022 * Called by ExecA[RS]...Triggers() to queue up the triggers that should
6023 * be fired for an event.
6025 * NOTE: this is called whenever there are any triggers associated with
6026 * the event (even if they are disabled). This function decides which
6027 * triggers actually need to be queued. It is also called after each row,
6028 * even if there are no triggers for that event, if there are any AFTER
6029 * STATEMENT triggers for the statement which use transition tables, so that
6030 * the transition tuplestores can be built. Furthermore, if the transition
6031 * capture is happening for UPDATEd rows being moved to another partition due
6032 * to the partition-key being changed, then this function is called once when
6033 * the row is deleted (to capture OLD row), and once when the row is inserted
6034 * into another partition (to capture NEW row). This is done separately because
6035 * DELETE and INSERT happen on different tables.
6037 * Transition tuplestores are built now, rather than when events are pulled
6038 * off of the queue because AFTER ROW triggers are allowed to select from the
6039 * transition tables for the statement.
6041 * This contains special support to queue the update events for the case where
6042 * a partitioned table undergoing a cross-partition update may have foreign
6043 * keys pointing into it. Normally, a partitioned table's row triggers are
6044 * not fired because the leaf partition(s) which are modified as a result of
6045 * the operation on the partitioned table contain the same triggers which are
6046 * fired instead. But that general scheme can cause problematic behavior with
6047 * foreign key triggers during cross-partition updates, which are implemented
6048 * as DELETE on the source partition followed by INSERT into the destination
6049 * partition. Specifically, firing DELETE triggers would lead to the wrong
6050 * foreign key action to be enforced considering that the original command is
6051 * UPDATE; in this case, this function is called with relinfo as the
6052 * partitioned table, and src_partinfo and dst_partinfo referring to the
6053 * source and target leaf partitions, respectively.
6055 * is_crosspart_update is true either when a DELETE event is fired on the
6056 * source partition (which is to be ignored) or an UPDATE event is fired on
6057 * the root partitioned table.
6058 * ----------
6060 static void
6061 AfterTriggerSaveEvent(EState *estate, ResultRelInfo *relinfo,
6062 ResultRelInfo *src_partinfo,
6063 ResultRelInfo *dst_partinfo,
6064 int event, bool row_trigger,
6065 TupleTableSlot *oldslot, TupleTableSlot *newslot,
6066 List *recheckIndexes, Bitmapset *modifiedCols,
6067 TransitionCaptureState *transition_capture,
6068 bool is_crosspart_update)
6070 Relation rel = relinfo->ri_RelationDesc;
6071 TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
6072 AfterTriggerEventData new_event;
6073 AfterTriggerSharedData new_shared;
6074 char relkind = rel->rd_rel->relkind;
6075 int tgtype_event;
6076 int tgtype_level;
6077 int i;
6078 Tuplestorestate *fdw_tuplestore = NULL;
6081 * Check state. We use a normal test not Assert because it is possible to
6082 * reach here in the wrong state given misconfigured RI triggers, in
6083 * particular deferring a cascade action trigger.
6085 if (afterTriggers.query_depth < 0)
6086 elog(ERROR, "AfterTriggerSaveEvent() called outside of query");
6088 /* Be sure we have enough space to record events at this query depth. */
6089 if (afterTriggers.query_depth >= afterTriggers.maxquerydepth)
6090 AfterTriggerEnlargeQueryState();
6093 * If the directly named relation has any triggers with transition tables,
6094 * then we need to capture transition tuples.
6096 if (row_trigger && transition_capture != NULL)
6098 TupleTableSlot *original_insert_tuple = transition_capture->tcs_original_insert_tuple;
6101 * Capture the old tuple in the appropriate transition table based on
6102 * the event.
6104 if (!TupIsNull(oldslot))
6106 Tuplestorestate *old_tuplestore;
6108 old_tuplestore = GetAfterTriggersTransitionTable(event,
6109 oldslot,
6110 NULL,
6111 transition_capture);
6112 TransitionTableAddTuple(estate, transition_capture, relinfo,
6113 oldslot, NULL, old_tuplestore);
6117 * Capture the new tuple in the appropriate transition table based on
6118 * the event.
6120 if (!TupIsNull(newslot))
6122 Tuplestorestate *new_tuplestore;
6124 new_tuplestore = GetAfterTriggersTransitionTable(event,
6125 NULL,
6126 newslot,
6127 transition_capture);
6128 TransitionTableAddTuple(estate, transition_capture, relinfo,
6129 newslot, original_insert_tuple, new_tuplestore);
6133 * If transition tables are the only reason we're here, return. As
6134 * mentioned above, we can also be here during update tuple routing in
6135 * presence of transition tables, in which case this function is
6136 * called separately for OLD and NEW, so we expect exactly one of them
6137 * to be NULL.
6139 if (trigdesc == NULL ||
6140 (event == TRIGGER_EVENT_DELETE && !trigdesc->trig_delete_after_row) ||
6141 (event == TRIGGER_EVENT_INSERT && !trigdesc->trig_insert_after_row) ||
6142 (event == TRIGGER_EVENT_UPDATE && !trigdesc->trig_update_after_row) ||
6143 (event == TRIGGER_EVENT_UPDATE && (TupIsNull(oldslot) ^ TupIsNull(newslot))))
6144 return;
6148 * We normally don't see partitioned tables here for row level triggers
6149 * except in the special case of a cross-partition update. In that case,
6150 * nodeModifyTable.c:ExecCrossPartitionUpdateForeignKey() calls here to
6151 * queue an update event on the root target partitioned table, also
6152 * passing the source and destination partitions and their tuples.
6154 Assert(!row_trigger ||
6155 rel->rd_rel->relkind != RELKIND_PARTITIONED_TABLE ||
6156 (is_crosspart_update &&
6157 TRIGGER_FIRED_BY_UPDATE(event) &&
6158 src_partinfo != NULL && dst_partinfo != NULL));
6161 * Validate the event code and collect the associated tuple CTIDs.
6163 * The event code will be used both as a bitmask and an array offset, so
6164 * validation is important to make sure we don't walk off the edge of our
6165 * arrays.
6167 * Also, if we're considering statement-level triggers, check whether we
6168 * already queued a set of them for this event, and cancel the prior set
6169 * if so. This preserves the behavior that statement-level triggers fire
6170 * just once per statement and fire after row-level triggers.
6172 switch (event)
6174 case TRIGGER_EVENT_INSERT:
6175 tgtype_event = TRIGGER_TYPE_INSERT;
6176 if (row_trigger)
6178 Assert(oldslot == NULL);
6179 Assert(newslot != NULL);
6180 ItemPointerCopy(&(newslot->tts_tid), &(new_event.ate_ctid1));
6181 ItemPointerSetInvalid(&(new_event.ate_ctid2));
6183 else
6185 Assert(oldslot == NULL);
6186 Assert(newslot == NULL);
6187 ItemPointerSetInvalid(&(new_event.ate_ctid1));
6188 ItemPointerSetInvalid(&(new_event.ate_ctid2));
6189 cancel_prior_stmt_triggers(RelationGetRelid(rel),
6190 CMD_INSERT, event);
6192 break;
6193 case TRIGGER_EVENT_DELETE:
6194 tgtype_event = TRIGGER_TYPE_DELETE;
6195 if (row_trigger)
6197 Assert(oldslot != NULL);
6198 Assert(newslot == NULL);
6199 ItemPointerCopy(&(oldslot->tts_tid), &(new_event.ate_ctid1));
6200 ItemPointerSetInvalid(&(new_event.ate_ctid2));
6202 else
6204 Assert(oldslot == NULL);
6205 Assert(newslot == NULL);
6206 ItemPointerSetInvalid(&(new_event.ate_ctid1));
6207 ItemPointerSetInvalid(&(new_event.ate_ctid2));
6208 cancel_prior_stmt_triggers(RelationGetRelid(rel),
6209 CMD_DELETE, event);
6211 break;
6212 case TRIGGER_EVENT_UPDATE:
6213 tgtype_event = TRIGGER_TYPE_UPDATE;
6214 if (row_trigger)
6216 Assert(oldslot != NULL);
6217 Assert(newslot != NULL);
6218 ItemPointerCopy(&(oldslot->tts_tid), &(new_event.ate_ctid1));
6219 ItemPointerCopy(&(newslot->tts_tid), &(new_event.ate_ctid2));
6222 * Also remember the OIDs of partitions to fetch these tuples
6223 * out of later in AfterTriggerExecute().
6225 if (rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
6227 Assert(src_partinfo != NULL && dst_partinfo != NULL);
6228 new_event.ate_src_part =
6229 RelationGetRelid(src_partinfo->ri_RelationDesc);
6230 new_event.ate_dst_part =
6231 RelationGetRelid(dst_partinfo->ri_RelationDesc);
6234 else
6236 Assert(oldslot == NULL);
6237 Assert(newslot == NULL);
6238 ItemPointerSetInvalid(&(new_event.ate_ctid1));
6239 ItemPointerSetInvalid(&(new_event.ate_ctid2));
6240 cancel_prior_stmt_triggers(RelationGetRelid(rel),
6241 CMD_UPDATE, event);
6243 break;
6244 case TRIGGER_EVENT_TRUNCATE:
6245 tgtype_event = TRIGGER_TYPE_TRUNCATE;
6246 Assert(oldslot == NULL);
6247 Assert(newslot == NULL);
6248 ItemPointerSetInvalid(&(new_event.ate_ctid1));
6249 ItemPointerSetInvalid(&(new_event.ate_ctid2));
6250 break;
6251 default:
6252 elog(ERROR, "invalid after-trigger event code: %d", event);
6253 tgtype_event = 0; /* keep compiler quiet */
6254 break;
6257 /* Determine flags */
6258 if (!(relkind == RELKIND_FOREIGN_TABLE && row_trigger))
6260 if (row_trigger && event == TRIGGER_EVENT_UPDATE)
6262 if (relkind == RELKIND_PARTITIONED_TABLE)
6263 new_event.ate_flags = AFTER_TRIGGER_CP_UPDATE;
6264 else
6265 new_event.ate_flags = AFTER_TRIGGER_2CTID;
6267 else
6268 new_event.ate_flags = AFTER_TRIGGER_1CTID;
6271 /* else, we'll initialize ate_flags for each trigger */
6273 tgtype_level = (row_trigger ? TRIGGER_TYPE_ROW : TRIGGER_TYPE_STATEMENT);
6276 * Must convert/copy the source and destination partition tuples into the
6277 * root partitioned table's format/slot, because the processing in the
6278 * loop below expects both oldslot and newslot tuples to be in that form.
6280 if (row_trigger && rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
6282 TupleTableSlot *rootslot;
6283 TupleConversionMap *map;
6285 rootslot = ExecGetTriggerOldSlot(estate, relinfo);
6286 map = ExecGetChildToRootMap(src_partinfo);
6287 if (map)
6288 oldslot = execute_attr_map_slot(map->attrMap,
6289 oldslot,
6290 rootslot);
6291 else
6292 oldslot = ExecCopySlot(rootslot, oldslot);
6294 rootslot = ExecGetTriggerNewSlot(estate, relinfo);
6295 map = ExecGetChildToRootMap(dst_partinfo);
6296 if (map)
6297 newslot = execute_attr_map_slot(map->attrMap,
6298 newslot,
6299 rootslot);
6300 else
6301 newslot = ExecCopySlot(rootslot, newslot);
6304 for (i = 0; i < trigdesc->numtriggers; i++)
6306 Trigger *trigger = &trigdesc->triggers[i];
6308 if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
6309 tgtype_level,
6310 TRIGGER_TYPE_AFTER,
6311 tgtype_event))
6312 continue;
6313 if (!TriggerEnabled(estate, relinfo, trigger, event,
6314 modifiedCols, oldslot, newslot))
6315 continue;
6317 if (relkind == RELKIND_FOREIGN_TABLE && row_trigger)
6319 if (fdw_tuplestore == NULL)
6321 fdw_tuplestore = GetCurrentFDWTuplestore();
6322 new_event.ate_flags = AFTER_TRIGGER_FDW_FETCH;
6324 else
6325 /* subsequent event for the same tuple */
6326 new_event.ate_flags = AFTER_TRIGGER_FDW_REUSE;
6330 * If the trigger is a foreign key enforcement trigger, there are
6331 * certain cases where we can skip queueing the event because we can
6332 * tell by inspection that the FK constraint will still pass. There
6333 * are also some cases during cross-partition updates of a partitioned
6334 * table where queuing the event can be skipped.
6336 if (TRIGGER_FIRED_BY_UPDATE(event) || TRIGGER_FIRED_BY_DELETE(event))
6338 switch (RI_FKey_trigger_type(trigger->tgfoid))
6340 case RI_TRIGGER_PK:
6343 * For cross-partitioned updates of partitioned PK table,
6344 * skip the event fired by the component delete on the
6345 * source leaf partition unless the constraint originates
6346 * in the partition itself (!tgisclone), because the
6347 * update event that will be fired on the root
6348 * (partitioned) target table will be used to perform the
6349 * necessary foreign key enforcement action.
6351 if (is_crosspart_update &&
6352 TRIGGER_FIRED_BY_DELETE(event) &&
6353 trigger->tgisclone)
6354 continue;
6356 /* Update or delete on trigger's PK table */
6357 if (!RI_FKey_pk_upd_check_required(trigger, rel,
6358 oldslot, newslot))
6360 /* skip queuing this event */
6361 continue;
6363 break;
6365 case RI_TRIGGER_FK:
6368 * Update on trigger's FK table. We can skip the update
6369 * event fired on a partitioned table during a
6370 * cross-partition of that table, because the insert event
6371 * that is fired on the destination leaf partition would
6372 * suffice to perform the necessary foreign key check.
6373 * Moreover, RI_FKey_fk_upd_check_required() expects to be
6374 * passed a tuple that contains system attributes, most of
6375 * which are not present in the virtual slot belonging to
6376 * a partitioned table.
6378 if (rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE ||
6379 !RI_FKey_fk_upd_check_required(trigger, rel,
6380 oldslot, newslot))
6382 /* skip queuing this event */
6383 continue;
6385 break;
6387 case RI_TRIGGER_NONE:
6390 * Not an FK trigger. No need to queue the update event
6391 * fired during a cross-partitioned update of a
6392 * partitioned table, because the same row trigger must be
6393 * present in the leaf partition(s) that are affected as
6394 * part of this update and the events fired on them are
6395 * queued instead.
6397 if (row_trigger &&
6398 rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
6399 continue;
6400 break;
6405 * If the trigger is a deferred unique constraint check trigger, only
6406 * queue it if the unique constraint was potentially violated, which
6407 * we know from index insertion time.
6409 if (trigger->tgfoid == F_UNIQUE_KEY_RECHECK)
6411 if (!list_member_oid(recheckIndexes, trigger->tgconstrindid))
6412 continue; /* Uniqueness definitely not violated */
6416 * Fill in event structure and add it to the current query's queue.
6417 * Note we set ats_table to NULL whenever this trigger doesn't use
6418 * transition tables, to improve sharability of the shared event data.
6420 new_shared.ats_event =
6421 (event & TRIGGER_EVENT_OPMASK) |
6422 (row_trigger ? TRIGGER_EVENT_ROW : 0) |
6423 (trigger->tgdeferrable ? AFTER_TRIGGER_DEFERRABLE : 0) |
6424 (trigger->tginitdeferred ? AFTER_TRIGGER_INITDEFERRED : 0);
6425 new_shared.ats_tgoid = trigger->tgoid;
6426 new_shared.ats_relid = RelationGetRelid(rel);
6427 new_shared.ats_firing_id = 0;
6428 if ((trigger->tgoldtable || trigger->tgnewtable) &&
6429 transition_capture != NULL)
6430 new_shared.ats_table = transition_capture->tcs_private;
6431 else
6432 new_shared.ats_table = NULL;
6433 new_shared.ats_modifiedcols = afterTriggerCopyBitmap(modifiedCols);
6435 afterTriggerAddEvent(&afterTriggers.query_stack[afterTriggers.query_depth].events,
6436 &new_event, &new_shared);
6440 * Finally, spool any foreign tuple(s). The tuplestore squashes them to
6441 * minimal tuples, so this loses any system columns. The executor lost
6442 * those columns before us, for an unrelated reason, so this is fine.
6444 if (fdw_tuplestore)
6446 if (oldslot != NULL)
6447 tuplestore_puttupleslot(fdw_tuplestore, oldslot);
6448 if (newslot != NULL)
6449 tuplestore_puttupleslot(fdw_tuplestore, newslot);
6454 * Detect whether we already queued BEFORE STATEMENT triggers for the given
6455 * relation + operation, and set the flag so the next call will report "true".
6457 static bool
6458 before_stmt_triggers_fired(Oid relid, CmdType cmdType)
6460 bool result;
6461 AfterTriggersTableData *table;
6463 /* Check state, like AfterTriggerSaveEvent. */
6464 if (afterTriggers.query_depth < 0)
6465 elog(ERROR, "before_stmt_triggers_fired() called outside of query");
6467 /* Be sure we have enough space to record events at this query depth. */
6468 if (afterTriggers.query_depth >= afterTriggers.maxquerydepth)
6469 AfterTriggerEnlargeQueryState();
6472 * We keep this state in the AfterTriggersTableData that also holds
6473 * transition tables for the relation + operation. In this way, if we are
6474 * forced to make a new set of transition tables because more tuples get
6475 * entered after we've already fired triggers, we will allow a new set of
6476 * statement triggers to get queued.
6478 table = GetAfterTriggersTableData(relid, cmdType);
6479 result = table->before_trig_done;
6480 table->before_trig_done = true;
6481 return result;
6485 * If we previously queued a set of AFTER STATEMENT triggers for the given
6486 * relation + operation, and they've not been fired yet, cancel them. The
6487 * caller will queue a fresh set that's after any row-level triggers that may
6488 * have been queued by the current sub-statement, preserving (as much as
6489 * possible) the property that AFTER ROW triggers fire before AFTER STATEMENT
6490 * triggers, and that the latter only fire once. This deals with the
6491 * situation where several FK enforcement triggers sequentially queue triggers
6492 * for the same table into the same trigger query level. We can't fully
6493 * prevent odd behavior though: if there are AFTER ROW triggers taking
6494 * transition tables, we don't want to change the transition tables once the
6495 * first such trigger has seen them. In such a case, any additional events
6496 * will result in creating new transition tables and allowing new firings of
6497 * statement triggers.
6499 * This also saves the current event list location so that a later invocation
6500 * of this function can cheaply find the triggers we're about to queue and
6501 * cancel them.
6503 static void
6504 cancel_prior_stmt_triggers(Oid relid, CmdType cmdType, int tgevent)
6506 AfterTriggersTableData *table;
6507 AfterTriggersQueryData *qs = &afterTriggers.query_stack[afterTriggers.query_depth];
6510 * We keep this state in the AfterTriggersTableData that also holds
6511 * transition tables for the relation + operation. In this way, if we are
6512 * forced to make a new set of transition tables because more tuples get
6513 * entered after we've already fired triggers, we will allow a new set of
6514 * statement triggers to get queued without canceling the old ones.
6516 table = GetAfterTriggersTableData(relid, cmdType);
6518 if (table->after_trig_done)
6521 * We want to start scanning from the tail location that existed just
6522 * before we inserted any statement triggers. But the events list
6523 * might've been entirely empty then, in which case scan from the
6524 * current head.
6526 AfterTriggerEvent event;
6527 AfterTriggerEventChunk *chunk;
6529 if (table->after_trig_events.tail)
6531 chunk = table->after_trig_events.tail;
6532 event = (AfterTriggerEvent) table->after_trig_events.tailfree;
6534 else
6536 chunk = qs->events.head;
6537 event = NULL;
6540 for_each_chunk_from(chunk)
6542 if (event == NULL)
6543 event = (AfterTriggerEvent) CHUNK_DATA_START(chunk);
6544 for_each_event_from(event, chunk)
6546 AfterTriggerShared evtshared = GetTriggerSharedData(event);
6549 * Exit loop when we reach events that aren't AS triggers for
6550 * the target relation.
6552 if (evtshared->ats_relid != relid)
6553 goto done;
6554 if ((evtshared->ats_event & TRIGGER_EVENT_OPMASK) != tgevent)
6555 goto done;
6556 if (!TRIGGER_FIRED_FOR_STATEMENT(evtshared->ats_event))
6557 goto done;
6558 if (!TRIGGER_FIRED_AFTER(evtshared->ats_event))
6559 goto done;
6560 /* OK, mark it DONE */
6561 event->ate_flags &= ~AFTER_TRIGGER_IN_PROGRESS;
6562 event->ate_flags |= AFTER_TRIGGER_DONE;
6564 /* signal we must reinitialize event ptr for next chunk */
6565 event = NULL;
6568 done:
6570 /* In any case, save current insertion point for next time */
6571 table->after_trig_done = true;
6572 table->after_trig_events = qs->events;
6576 * GUC assign_hook for session_replication_role
6578 void
6579 assign_session_replication_role(int newval, void *extra)
6582 * Must flush the plan cache when changing replication role; but don't
6583 * flush unnecessarily.
6585 if (SessionReplicationRole != newval)
6586 ResetPlanCache();
6590 * SQL function pg_trigger_depth()
6592 Datum
6593 pg_trigger_depth(PG_FUNCTION_ARGS)
6595 PG_RETURN_INT32(MyTriggerDepth);