diff --git a/contrib/oid2name/oid2name.c b/contrib/oid2name/oid2name.c index c2785848f5..5180290713 100644 --- a/contrib/oid2name/oid2name.c +++ b/contrib/oid2name/oid2name.c @@ -353,7 +353,7 @@ sql_conn(struct options *my_opts) res = PQexec(conn, ALWAYS_SECURE_SEARCH_PATH_SQL); if (PQresultStatus(res) != PGRES_TUPLES_OK) { - pg_log_error("could not clear search_path: %s", + pg_log_error("could not clear \"search_path\": %s", PQerrorMessage(conn)); PQclear(res); PQfinish(conn); diff --git a/contrib/vacuumlo/vacuumlo.c b/contrib/vacuumlo/vacuumlo.c index 0d99428dec..ce1bbab12e 100644 --- a/contrib/vacuumlo/vacuumlo.c +++ b/contrib/vacuumlo/vacuumlo.c @@ -139,7 +139,7 @@ vacuumlo(const char *database, const struct _param *param) res = PQexec(conn, ALWAYS_SECURE_SEARCH_PATH_SQL); if (PQresultStatus(res) != PGRES_TUPLES_OK) { - pg_log_error("failed to set search_path: %s", PQerrorMessage(conn)); + pg_log_error("failed to set \"search_path\": %s", PQerrorMessage(conn)); PQclear(res); PQfinish(conn); return -1; diff --git a/src/backend/access/transam/xlog.c b/src/backend/access/transam/xlog.c index 5211e4e108..c1def6fdf5 100644 --- a/src/backend/access/transam/xlog.c +++ b/src/backend/access/transam/xlog.c @@ -9111,7 +9111,7 @@ do_pg_backup_stop(BackupState *state, bool waitforarchive) ereport(ERROR, (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), errmsg("WAL level not sufficient for making an online backup"), - errhint("wal_level must be set to \"replica\" or \"logical\" at server start."))); + errhint("\"wal_level\" must be set to \"replica\" or \"logical\" at server start."))); /* * OK to update backup counter and session-level lock. diff --git a/src/backend/archive/shell_archive.c b/src/backend/archive/shell_archive.c index 506c5a30ad..4de7e66f6e 100644 --- a/src/backend/archive/shell_archive.c +++ b/src/backend/archive/shell_archive.c @@ -48,7 +48,7 @@ shell_archive_configured(ArchiveModuleState *state) if (XLogArchiveCommand[0] != '\0') return true; - arch_module_check_errdetail("%s is not set.", + arch_module_check_errdetail("\"%s\" is not set.", "archive_command"); return false; } diff --git a/src/backend/postmaster/bgworker.c b/src/backend/postmaster/bgworker.c index b83967cda3..07bc5517fc 100644 --- a/src/backend/postmaster/bgworker.c +++ b/src/backend/postmaster/bgworker.c @@ -257,7 +257,7 @@ BackgroundWorkerStateChange(bool allow_new_workers) if (max_worker_processes != BackgroundWorkerData->total_slots) { ereport(LOG, - (errmsg("inconsistent background worker state (max_worker_processes=%d, total_slots=%d)", + (errmsg("inconsistent background worker state (\"max_worker_processes\"=%d, total slots=%d)", max_worker_processes, BackgroundWorkerData->total_slots))); return; diff --git a/src/backend/postmaster/checkpointer.c b/src/backend/postmaster/checkpointer.c index 199f008bcd..eeb73c8572 100644 --- a/src/backend/postmaster/checkpointer.c +++ b/src/backend/postmaster/checkpointer.c @@ -663,7 +663,7 @@ CheckArchiveTimeout(void) * assume nothing happened. */ if (XLogSegmentOffset(switchpoint, wal_segment_size) != 0) - elog(DEBUG1, "write-ahead log switch forced (archive_timeout=%d)", + elog(DEBUG1, "write-ahead log switch forced (\"archive_timeout\"=%d)", XLogArchiveTimeout); } diff --git a/src/backend/replication/logical/launcher.c b/src/backend/replication/logical/launcher.c index c566d50a07..e5fdca8bbf 100644 --- a/src/backend/replication/logical/launcher.c +++ b/src/backend/replication/logical/launcher.c @@ -337,7 +337,7 @@ logicalrep_worker_launch(LogicalRepWorkerType wtype, if (max_replication_slots == 0) ereport(ERROR, (errcode(ERRCODE_CONFIGURATION_LIMIT_EXCEEDED), - errmsg("cannot start logical replication workers when max_replication_slots = 0"))); + errmsg("cannot start logical replication workers when \"max_replication_slots\"=0"))); /* * We need to do the modification of the shared memory under lock so that diff --git a/src/backend/replication/logical/slotsync.c b/src/backend/replication/logical/slotsync.c index 33378bacea..f9649eec1a 100644 --- a/src/backend/replication/logical/slotsync.c +++ b/src/backend/replication/logical/slotsync.c @@ -962,14 +962,14 @@ validate_remote_info(WalReceiverConn *wrconn) if (res->status != WALRCV_OK_TUPLES) ereport(ERROR, - errmsg("could not fetch primary_slot_name \"%s\" info from the primary server: %s", + errmsg("could not fetch primary slot name \"%s\" info from the primary server: %s", PrimarySlotName, res->err), - errhint("Check if primary_slot_name is configured correctly.")); + errhint("Check if \"primary_slot_name\" is configured correctly.")); tupslot = MakeSingleTupleTableSlot(res->tupledesc, &TTSOpsMinimalTuple); if (!tuplestore_gettupleslot(res->tuplestore, true, false, tupslot)) elog(ERROR, - "failed to fetch tuple for the primary server slot specified by primary_slot_name"); + "failed to fetch tuple for the primary server slot specified by \"primary_slot_name\""); remote_in_recovery = DatumGetBool(slot_getattr(tupslot, 1, &isnull)); Assert(!isnull); diff --git a/src/backend/storage/file/fd.c b/src/backend/storage/file/fd.c index d5204b1eb9..8df0e2f81a 100644 --- a/src/backend/storage/file/fd.c +++ b/src/backend/storage/file/fd.c @@ -2254,7 +2254,7 @@ FileWriteV(File file, const struct iovec *iov, int iovcnt, off_t offset, if (newTotal > (uint64) temp_file_limit * (uint64) 1024) ereport(ERROR, (errcode(ERRCODE_CONFIGURATION_LIMIT_EXCEEDED), - errmsg("temporary file size exceeds temp_file_limit (%dkB)", + errmsg("temporary file size exceeds \"temp_file_limit\" (%dkB)", temp_file_limit))); } } diff --git a/src/backend/utils/misc/guc_tables.c b/src/backend/utils/misc/guc_tables.c index 521ec5591c..686309db58 100644 --- a/src/backend/utils/misc/guc_tables.c +++ b/src/backend/utils/misc/guc_tables.c @@ -2356,7 +2356,7 @@ struct config_int ConfigureNamesInt[] = { {"commit_timestamp_buffers", PGC_POSTMASTER, RESOURCES_MEM, gettext_noop("Sets the size of the dedicated buffer pool used for the commit timestamp cache."), - gettext_noop("Specify 0 to have this value determined as a fraction of shared_buffers."), + gettext_noop("Specify 0 to have this value determined as a fraction of \"shared_buffers\"."), GUC_UNIT_BLOCKS }, &commit_timestamp_buffers, @@ -2411,7 +2411,7 @@ struct config_int ConfigureNamesInt[] = { {"subtransaction_buffers", PGC_POSTMASTER, RESOURCES_MEM, gettext_noop("Sets the size of the dedicated buffer pool used for the subtransaction cache."), - gettext_noop("Specify 0 to have this value determined as a fraction of shared_buffers."), + gettext_noop("Specify 0 to have this value determined as a fraction of \"shared_buffers\"."), GUC_UNIT_BLOCKS }, &subtransaction_buffers, @@ -2422,7 +2422,7 @@ struct config_int ConfigureNamesInt[] = { {"transaction_buffers", PGC_POSTMASTER, RESOURCES_MEM, gettext_noop("Sets the size of the dedicated buffer pool used for the transaction status cache."), - gettext_noop("Specify 0 to have this value determined as a fraction of shared_buffers."), + gettext_noop("Specify 0 to have this value determined as a fraction of \"shared_buffers\"."), GUC_UNIT_BLOCKS }, &transaction_buffers, @@ -2942,7 +2942,7 @@ struct config_int ConfigureNamesInt[] = { {"wal_buffers", PGC_POSTMASTER, WAL_SETTINGS, gettext_noop("Sets the number of disk-page buffers in shared memory for WAL."), - gettext_noop("Specify -1 to have this value determined as a fraction of shared_buffers."), + gettext_noop("Specify -1 to have this value determined as a fraction of \"shared_buffers\"."), GUC_UNIT_XBLOCKS }, &XLOGbuffers, @@ -3068,7 +3068,7 @@ struct config_int ConfigureNamesInt[] = {"log_min_duration_sample", PGC_SUSET, LOGGING_WHEN, gettext_noop("Sets the minimum execution time above which " "a sample of statements will be logged." - " Sampling is determined by log_statement_sample_rate."), + " Sampling is determined by \"log_statement_sample_rate\"."), gettext_noop("Zero logs a sample of all queries. -1 turns this feature off."), GUC_UNIT_MS }, diff --git a/src/bin/pg_archivecleanup/pg_archivecleanup.c b/src/bin/pg_archivecleanup/pg_archivecleanup.c index 5a124385b7..c25348bcb8 100644 --- a/src/bin/pg_archivecleanup/pg_archivecleanup.c +++ b/src/bin/pg_archivecleanup/pg_archivecleanup.c @@ -270,7 +270,7 @@ usage(void) " clean up\n")); printf(_(" -?, --help show this help, then exit\n")); printf(_("\n" - "For use as archive_cleanup_command in postgresql.conf:\n" + "For use as \"archive_cleanup_command\" in postgresql.conf:\n" " archive_cleanup_command = 'pg_archivecleanup [OPTION]... ARCHIVELOCATION %%r'\n" "e.g.\n" " archive_cleanup_command = 'pg_archivecleanup /mnt/server/archiverdir %%r'\n")); diff --git a/src/bin/pg_basebackup/pg_createsubscriber.c b/src/bin/pg_basebackup/pg_createsubscriber.c index 6b793931f3..e804b2a63e 100644 --- a/src/bin/pg_basebackup/pg_createsubscriber.c +++ b/src/bin/pg_basebackup/pg_createsubscriber.c @@ -523,7 +523,7 @@ connect_database(const char *conninfo, bool exit_on_error) res = PQexec(conn, ALWAYS_SECURE_SEARCH_PATH_SQL); if (PQresultStatus(res) != PGRES_TUPLES_OK) { - pg_log_error("could not clear search_path: %s", + pg_log_error("could not clear \"search_path\": %s", PQresultErrorMessage(res)); PQclear(res); PQfinish(conn); @@ -911,7 +911,7 @@ check_publisher(const struct LogicalRepInfo *dbinfo) if (strcmp(wal_level, "logical") != 0) { - pg_log_error("publisher requires wal_level >= \"logical\""); + pg_log_error("publisher requires \"wal_level\" >= \"logical\""); failed = true; } diff --git a/src/bin/pg_dump/pg_backup_db.c b/src/bin/pg_dump/pg_backup_db.c index a02841c405..4d1957a7eb 100644 --- a/src/bin/pg_dump/pg_backup_db.c +++ b/src/bin/pg_dump/pg_backup_db.c @@ -39,7 +39,7 @@ _check_database_version(ArchiveHandle *AH) remoteversion_str = PQparameterStatus(AH->connection, "server_version"); remoteversion = PQserverVersion(AH->connection); if (remoteversion == 0 || !remoteversion_str) - pg_fatal("could not get server_version from libpq"); + pg_fatal("could not get \"server_version\" from libpq"); AH->public.remoteVersionStr = pg_strdup(remoteversion_str); AH->public.remoteVersion = remoteversion; diff --git a/src/bin/pg_rewind/pg_rewind.c b/src/bin/pg_rewind/pg_rewind.c index 1027d235dc..960916a1e8 100644 --- a/src/bin/pg_rewind/pg_rewind.c +++ b/src/bin/pg_rewind/pg_rewind.c @@ -1104,7 +1104,7 @@ getRestoreCommand(const char *argv0) restore_command = pipe_read_line(postgres_cmd->data); if (restore_command == NULL) - pg_fatal("could not read restore_command from target cluster"); + pg_fatal("could not read \"restore_command\" from target cluster"); (void) pg_strip_crlf(restore_command); diff --git a/src/include/libpq/libpq-be-fe-helpers.h b/src/include/libpq/libpq-be-fe-helpers.h index fe50829274..46a90dfb02 100644 --- a/src/include/libpq/libpq-be-fe-helpers.h +++ b/src/include/libpq/libpq-be-fe-helpers.h @@ -142,13 +142,13 @@ libpqsrv_connect_prepare(void) (errcode(ERRCODE_SQLCLIENT_UNABLE_TO_ESTABLISH_SQLCONNECTION), errmsg("could not establish connection"), errdetail("There are too many open files on the local server."), - errhint("Raise the server's max_files_per_process and/or \"ulimit -n\" limits."))); + errhint("Raise the server's \"max_files_per_process\" and/or \"ulimit -n\" limits."))); #else ereport(ERROR, (errcode(ERRCODE_SQLCLIENT_UNABLE_TO_ESTABLISH_SQLCONNECTION), errmsg("could not establish connection"), errdetail("There are too many open files on the local server."), - errhint("Raise the server's max_files_per_process setting."))); + errhint("Raise the server's \"max_files_per_process\" setting."))); #endif } }