diff --git a/contrib/basebackup_to_shell/t/001_basic.pl b/contrib/basebackup_to_shell/t/001_basic.pl index a278df3f91..84ad93f614 100644 --- a/contrib/basebackup_to_shell/t/001_basic.pl +++ b/contrib/basebackup_to_shell/t/001_basic.pl @@ -49,7 +49,7 @@ $node->command_fails_like( qr/shell command for backup is not configured/, 'fails if basebackup_to_shell.command is not set'); -# Configure basebackup_to_shell.command and reload the configuation file. +# Configure basebackup_to_shell.command and reload the configuration file. my $backup_path = PostgreSQL::Test::Utils::tempdir; my $escaped_backup_path = $backup_path; $escaped_backup_path =~ s{\\}{\\\\}g diff --git a/contrib/pg_walinspect/pg_walinspect.c b/contrib/pg_walinspect/pg_walinspect.c index e6f18e1c64..1cd3744d5d 100644 --- a/contrib/pg_walinspect/pg_walinspect.c +++ b/contrib/pg_walinspect/pg_walinspect.c @@ -85,7 +85,7 @@ GetCurrentLSN(void) } /* - * Intialize WAL reader and identify first valid LSN. + * Initialize WAL reader and identify first valid LSN. */ static XLogReaderState * InitXLogReaderState(XLogRecPtr lsn) diff --git a/src/backend/access/transam/parallel.c b/src/backend/access/transam/parallel.c index b26f2a64fb..7133ec0b22 100644 --- a/src/backend/access/transam/parallel.c +++ b/src/backend/access/transam/parallel.c @@ -376,7 +376,7 @@ InitializeParallelDSM(ParallelContext *pcxt) /* * Serialize the transaction snapshot if the transaction - * isolation-level uses a transaction snapshot. + * isolation level uses a transaction snapshot. */ if (IsolationUsesXactSnapshot()) { diff --git a/src/backend/access/transam/xlog.c b/src/backend/access/transam/xlog.c index b393f1b6c1..63481d826f 100644 --- a/src/backend/access/transam/xlog.c +++ b/src/backend/access/transam/xlog.c @@ -5835,7 +5835,7 @@ ReachedEndOfBackup(XLogRecPtr EndRecPtr, TimeLineID tli) { /* * We have reached the end of base backup, as indicated by pg_control. The - * data on disk is now consistent (unless minRecovery point is further + * data on disk is now consistent (unless minRecoveryPoint is further * ahead, which can happen if we crashed during previous recovery). Reset * backupStartPoint and backupEndPoint, and update minRecoveryPoint to * make sure we don't allow starting up at an earlier point even if diff --git a/src/backend/access/transam/xlogprefetcher.c b/src/backend/access/transam/xlogprefetcher.c index 7ba18f2a76..906e3d9469 100644 --- a/src/backend/access/transam/xlogprefetcher.c +++ b/src/backend/access/transam/xlogprefetcher.c @@ -457,9 +457,9 @@ XLogPrefetcherComputeStats(XLogPrefetcher *prefetcher) * *lsn, and the I/O will be considered to have completed once that LSN is * replayed. * - * Returns LRQ_NO_IO if we examined the next block reference and found that it - * was already in the buffer pool, or we decided for various reasons not to - * prefetch. + * Returns LRQ_NEXT_NO_IO if we examined the next block reference and found + * that it was already in the buffer pool, or we decided for various reasons + * not to prefetch. */ static LsnReadQueueNextStatus XLogPrefetcherNextBlock(uintptr_t pgsr_private, XLogRecPtr *lsn) diff --git a/src/backend/access/transam/xlogrecovery.c b/src/backend/access/transam/xlogrecovery.c index 02d1b2cd6d..188f6d6f85 100644 --- a/src/backend/access/transam/xlogrecovery.c +++ b/src/backend/access/transam/xlogrecovery.c @@ -8,7 +8,7 @@ * or standby mode, depending on configuration options and the state of * the control file and possible backup label file. PerformWalRecovery() * performs the actual WAL replay, calling the rmgr-specific redo routines. - * EndWalRecovery() performs end-of-recovery checks and cleanup actions, + * FinishWalRecovery() performs end-of-recovery checks and cleanup actions, * and prepares information needed to initialize the WAL for writes. In * addition to these three main functions, there are a bunch of functions * for interrogating recovery state and controlling the recovery process. @@ -505,7 +505,7 @@ EnableStandbyMode(void) * disk does after initializing other subsystems, but before calling * PerformWalRecovery(). * - * This initializes some global variables like ArchiveModeRequested, and + * This initializes some global variables like ArchiveRecoveryRequested, and * StandbyModeRequested and InRecovery. */ void @@ -1396,11 +1396,11 @@ read_tablespace_map(List **tablespaces) * * This does not close the 'xlogreader' yet, because in some cases the caller * still wants to re-read the last checkpoint record by calling - * ReadCheckPointRecord(). + * ReadCheckpointRecord(). * * Returns the position of the last valid or applied record, after which new * WAL should be appended, information about why recovery was ended, and some - * other things. See the WalRecoveryResult struct for details. + * other things. See the EndOfWalRecoveryInfo struct for details. */ EndOfWalRecoveryInfo * FinishWalRecovery(void) diff --git a/src/backend/backup/basebackup_copy.c b/src/backend/backup/basebackup_copy.c index 2bb6c89f8c..73a3f4a970 100644 --- a/src/backend/backup/basebackup_copy.c +++ b/src/backend/backup/basebackup_copy.c @@ -3,7 +3,7 @@ * basebackup_copy.c * send basebackup archives using COPY OUT * - * We send a result set with information about the tabelspaces to be included + * We send a result set with information about the tablespaces to be included * in the backup before starting COPY OUT. Then, we start a single COPY OUT * operation and transmits all the archives and the manifest if present during * the course of that single COPY OUT. Each CopyData message begins with a diff --git a/src/backend/catalog/pg_publication.c b/src/backend/catalog/pg_publication.c index 47637f28ab..c488b6370b 100644 --- a/src/backend/catalog/pg_publication.c +++ b/src/backend/catalog/pg_publication.c @@ -497,7 +497,7 @@ compare_int16(const void *a, const void *b) * and a Bitmapset with them; verify that each attribute is appropriate * to have in a publication column list (no system or generated attributes, * no duplicates). Additional checks with replica identity are done later; - * see check_publication_columns. + * see pub_collist_contains_invalid_column. * * Note that the attribute numbers are *not* offset by * FirstLowInvalidHeapAttributeNumber; system columns are forbidden so this diff --git a/src/backend/commands/vacuumparallel.c b/src/backend/commands/vacuumparallel.c index 87ea5c5242..a79067fd46 100644 --- a/src/backend/commands/vacuumparallel.c +++ b/src/backend/commands/vacuumparallel.c @@ -8,7 +8,7 @@ * * In a parallel vacuum, we perform both index bulk deletion and index cleanup * with parallel worker processes. Individual indexes are processed by one - * vacuum process. ParalleVacuumState contains shared information as well as + * vacuum process. ParallelVacuumState contains shared information as well as * the memory space for storing dead items allocated in the DSM segment. We * launch parallel worker processes at the start of parallel index * bulk-deletion and index cleanup and once all indexes are processed, the diff --git a/src/backend/executor/nodeAgg.c b/src/backend/executor/nodeAgg.c index 28205f7433..3aab5a0e80 100644 --- a/src/backend/executor/nodeAgg.c +++ b/src/backend/executor/nodeAgg.c @@ -4081,7 +4081,7 @@ build_pertrans_for_aggref(AggStatePerTrans pertrans, numTransArgs = pertrans->numTransInputs + 1; /* - * Set up infrastructure for calling the transfn. Note that invtrans is + * Set up infrastructure for calling the transfn. Note that invtransfn is * not needed here. */ build_aggregate_transfn_expr(inputTypes, diff --git a/src/backend/executor/nodeModifyTable.c b/src/backend/executor/nodeModifyTable.c index 5b596f1e7a..6aa8c03def 100644 --- a/src/backend/executor/nodeModifyTable.c +++ b/src/backend/executor/nodeModifyTable.c @@ -740,7 +740,7 @@ ExecGetUpdateNewTuple(ResultRelInfo *relinfo, * * Returns RETURNING result if any, otherwise NULL. * *inserted_tuple is the tuple that's effectively inserted; - * *inserted_destrel is the relation where it was inserted. + * *insert_destrel is the relation where it was inserted. * These are only set on success. * * This may change the currently active tuple conversion map in diff --git a/src/backend/executor/spi.c b/src/backend/executor/spi.c index e3a170c38b..256632c985 100644 --- a/src/backend/executor/spi.c +++ b/src/backend/executor/spi.c @@ -432,7 +432,7 @@ AtEOXact_SPI(bool isCommit) /* * Pop stack entries, stopping if we find one marked internal_xact (that - * one belongs to the caller of SPI_commit or SPI_abort). + * one belongs to the caller of SPI_commit or SPI_rollback). */ while (_SPI_connected >= 0) { diff --git a/src/backend/optimizer/plan/initsplan.c b/src/backend/optimizer/plan/initsplan.c index c6c21a7338..06f90882c4 100644 --- a/src/backend/optimizer/plan/initsplan.c +++ b/src/backend/optimizer/plan/initsplan.c @@ -3152,7 +3152,7 @@ check_hashjoinable(RestrictInfo *restrictinfo) /* * check_memoizable * If the restrictinfo's clause is suitable to be used for a Memoize node, - * set the lefthasheqoperator and righthasheqoperator to the hash equality + * set the left_hasheqoperator and right_hasheqoperator to the hash equality * operator that will be needed during caching. */ static void diff --git a/src/backend/replication/logical/tablesync.c b/src/backend/replication/logical/tablesync.c index 6dce355633..0c71ae9ba7 100644 --- a/src/backend/replication/logical/tablesync.c +++ b/src/backend/replication/logical/tablesync.c @@ -1556,7 +1556,7 @@ FetchTableStates(bool *started_tx) * Does the subscription have tables? * * If there were not-READY relations found then we know it does. But - * if table_state_not_ready was empty we still need to check again to + * if table_states_not_ready was empty we still need to check again to * see if there are 0 tables. */ has_subrels = (table_states_not_ready != NIL) || diff --git a/src/backend/storage/ipc/procarray.c b/src/backend/storage/ipc/procarray.c index ea91ce355f..fc05e2e293 100644 --- a/src/backend/storage/ipc/procarray.c +++ b/src/backend/storage/ipc/procarray.c @@ -2000,7 +2000,7 @@ static inline GlobalVisHorizonKind GlobalVisHorizonKindForRel(Relation rel) { /* - * Other relkkinds currently don't contain xids, nor always the necessary + * Other relkinds currently don't contain xids, nor always the necessary * logical decoding markers. */ Assert(!rel || diff --git a/src/backend/utils/mmgr/generation.c b/src/backend/utils/mmgr/generation.c index ae48cdbaa9..4fb8663cd6 100644 --- a/src/backend/utils/mmgr/generation.c +++ b/src/backend/utils/mmgr/generation.c @@ -544,7 +544,7 @@ GenerationBlockInit(GenerationContext *context, GenerationBlock *block, /* * GenerationBlockIsEmpty - * Returns true iif 'block' contains no chunks + * Returns true iff 'block' contains no chunks */ static inline bool GenerationBlockIsEmpty(GenerationBlock *block) diff --git a/src/backend/utils/sort/tuplesort.c b/src/backend/utils/sort/tuplesort.c index 9ca9835aab..95c3970437 100644 --- a/src/backend/utils/sort/tuplesort.c +++ b/src/backend/utils/sort/tuplesort.c @@ -494,7 +494,7 @@ static void tuplesort_updatemax(Tuplesortstate *state); * abbreviations of text or multi-key sorts. There could be! Is it worth it? */ -/* Used if first key's comparator is ssup_datum_unsigned_compare */ +/* Used if first key's comparator is ssup_datum_unsigned_cmp */ static pg_attribute_always_inline int qsort_tuple_unsigned_compare(SortTuple *a, SortTuple *b, Tuplesortstate *state) { @@ -517,7 +517,7 @@ qsort_tuple_unsigned_compare(SortTuple *a, SortTuple *b, Tuplesortstate *state) } #if SIZEOF_DATUM >= 8 -/* Used if first key's comparator is ssup_datum_signed_compare */ +/* Used if first key's comparator is ssup_datum_signed_cmp */ static pg_attribute_always_inline int qsort_tuple_signed_compare(SortTuple *a, SortTuple *b, Tuplesortstate *state) { @@ -541,7 +541,7 @@ qsort_tuple_signed_compare(SortTuple *a, SortTuple *b, Tuplesortstate *state) } #endif -/* Used if first key's comparator is ssup_datum_int32_compare */ +/* Used if first key's comparator is ssup_datum_int32_cmp */ static pg_attribute_always_inline int qsort_tuple_int32_compare(SortTuple *a, SortTuple *b, Tuplesortstate *state) { diff --git a/src/backend/utils/time/snapmgr.c b/src/backend/utils/time/snapmgr.c index 7d11ae3478..c9ca44d8b7 100644 --- a/src/backend/utils/time/snapmgr.c +++ b/src/backend/utils/time/snapmgr.c @@ -1630,7 +1630,7 @@ ThereAreNoPriorRegisteredSnapshots(void) } /* - * HaveRegisteredOrActiveSnapshots + * HaveRegisteredOrActiveSnapshot * Is there any registered or active snapshot? * * NB: Unless pushed or active, the cached catalog snapshot will not cause diff --git a/src/bin/pg_basebackup/bbstreamer_gzip.c b/src/bin/pg_basebackup/bbstreamer_gzip.c index 83bf5972fa..3bdbfa0bc4 100644 --- a/src/bin/pg_basebackup/bbstreamer_gzip.c +++ b/src/bin/pg_basebackup/bbstreamer_gzip.c @@ -150,7 +150,7 @@ bbstreamer_gzip_writer_content(bbstreamer *streamer, * calling gzclose. * * It makes no difference whether we opened the file or the caller did it, - * because libz provides no way of avoiding a close on the underling file + * because libz provides no way of avoiding a close on the underlying file * handle. Notice, however, that bbstreamer_gzip_writer_new() uses dup() to * work around this issue, so that the behavior from the caller's viewpoint * is the same as for bbstreamer_plain_writer. diff --git a/src/bin/pg_dump/pg_dump.c b/src/bin/pg_dump/pg_dump.c index 73a6c964ba..058244cd17 100644 --- a/src/bin/pg_dump/pg_dump.c +++ b/src/bin/pg_dump/pg_dump.c @@ -9636,7 +9636,7 @@ getAdditionalACLs(Archive *fout) { if (dobj->objType == DO_TABLE) { - /* For a column initpriv, set the table's ACL flags */ + /* For a column initprivs, set the table's ACL flags */ dobj->components |= DUMP_COMPONENT_ACL; ((TableInfo *) dobj)->hascolumnACLs = true; } diff --git a/src/include/utils/pgstat_internal.h b/src/include/utils/pgstat_internal.h index 6badb2fde4..60fbf9394b 100644 --- a/src/include/utils/pgstat_internal.h +++ b/src/include/utils/pgstat_internal.h @@ -102,7 +102,7 @@ typedef struct PgStatShared_HashEntry } PgStatShared_HashEntry; /* - * Common header struct for PgStatShm_Stat*Entry. + * Common header struct for PgStatShared_*. */ typedef struct PgStatShared_Common { diff --git a/src/include/utils/sortsupport.h b/src/include/utils/sortsupport.h index aaca6634cb..475ed1d5b8 100644 --- a/src/include/utils/sortsupport.h +++ b/src/include/utils/sortsupport.h @@ -372,7 +372,7 @@ ApplySortAbbrevFullComparator(Datum datum1, bool isNull1, /* * Datum comparison functions that we have specialized sort routines for. - * Datatypes that install these as their comparator or abbrevated comparator + * Datatypes that install these as their comparator or abbreviated comparator * are eligible for faster sorting. */ extern int ssup_datum_unsigned_cmp(Datum x, Datum y, SortSupport ssup); diff --git a/src/interfaces/libpq/fe-trace.c b/src/interfaces/libpq/fe-trace.c index abaab6a073..402784f40e 100644 --- a/src/interfaces/libpq/fe-trace.c +++ b/src/interfaces/libpq/fe-trace.c @@ -74,7 +74,7 @@ PQsetTraceFlags(PGconn *conn, int flags) /* * Print the current time, with microseconds, into a caller-supplied * buffer. - * Cribbed from setup_formatted_log_time, but much simpler. + * Cribbed from get_formatted_log_time, but much simpler. */ static void pqTraceFormatTimestamp(char *timestr, size_t ts_len) diff --git a/src/pl/plpython/plpy_main.c b/src/pl/plpython/plpy_main.c index a4d66f3057..010a97378c 100644 --- a/src/pl/plpython/plpy_main.c +++ b/src/pl/plpython/plpy_main.c @@ -306,7 +306,7 @@ plpython3_inline_handler(PG_FUNCTION_ARGS) /* * Setup error traceback support for ereport(). * plpython_inline_error_callback doesn't currently need exec_ctx, but - * for consistency with plpython_call_handler we do it the same way. + * for consistency with plpython3_call_handler we do it the same way. */ plerrcontext.callback = plpython_inline_error_callback; plerrcontext.arg = exec_ctx; diff --git a/src/test/subscription/t/013_partition.pl b/src/test/subscription/t/013_partition.pl index f03cd24246..275fb3b525 100644 --- a/src/test/subscription/t/013_partition.pl +++ b/src/test/subscription/t/013_partition.pl @@ -564,7 +564,7 @@ is($result, qq(), 'inserts into tab4_1 replicated'); # now switch the order of publications in the list, try again, the result -# should be the same (no dependence on order of pulications) +# should be the same (no dependence on order of publications) $node_subscriber2->safe_psql('postgres', "ALTER SUBSCRIPTION sub2 SET PUBLICATION pub_all, pub_lower_level"); diff --git a/src/test/subscription/t/023_twophase_stream.pl b/src/test/subscription/t/023_twophase_stream.pl index 1cc871fddb..f4af44414b 100644 --- a/src/test/subscription/t/023_twophase_stream.pl +++ b/src/test/subscription/t/023_twophase_stream.pl @@ -264,7 +264,7 @@ sub test_streaming is($result, qq(1), 'transaction is prepared on subscriber'); # Insert a different record (now we are outside of the 2PC transaction) - # Note: the 2PC transaction still holds row locks so make sure this insert is for a separare primary key + # Note: the 2PC transaction still holds row locks so make sure this insert is for a separate primary key $node_publisher->safe_psql('postgres', "INSERT INTO test_tab VALUES (99999, 'foobar')"); diff --git a/src/test/subscription/t/029_on_error.pl b/src/test/subscription/t/029_on_error.pl index 7797a949c2..7d6fb66985 100644 --- a/src/test/subscription/t/029_on_error.pl +++ b/src/test/subscription/t/029_on_error.pl @@ -13,7 +13,7 @@ my $offset = 0; # Test skipping the transaction. This function must be called after the caller # has inserted data that conflicts with the subscriber. The finish LSN of the # error transaction that is used to specify to ALTER SUBSCRIPTION ... SKIP is -# fetched from the server logs. After executing ALTER SUBSCRITPION ... SKIP, we +# fetched from the server logs. After executing ALTER SUBSCRIPTION ... SKIP, we # check if logical replication can continue working by inserting $nonconflict_data # on the publisher. sub test_skip_lsn diff --git a/src/tools/mark_pgdllimport.pl b/src/tools/mark_pgdllimport.pl index 69fa8e4ae9..0cf71dbc25 100755 --- a/src/tools/mark_pgdllimport.pl +++ b/src/tools/mark_pgdllimport.pl @@ -6,7 +6,7 @@ # Perl script that tries to add PGDLLIMPORT markings to PostgreSQL # header files. # -# This relies on a few idiosyncracies of the PostgreSQL coding style, +# This relies on a few idiosyncrasies of the PostgreSQL coding style, # such as the fact that we always use "extern" in function # declarations, and that we don't use // comments. It's not very # smart and may not catch all cases.