res = PQexec(conn, ALWAYS_SECURE_SEARCH_PATH_SQL);
if (PQresultStatus(res) != PGRES_TUPLES_OK)
{
- pg_log_error("could not clear search_path: %s",
+ pg_log_error("could not clear \"search_path\": %s",
PQerrorMessage(conn));
PQclear(res);
PQfinish(conn);
res = PQexec(conn, ALWAYS_SECURE_SEARCH_PATH_SQL);
if (PQresultStatus(res) != PGRES_TUPLES_OK)
{
- pg_log_error("failed to set search_path: %s", PQerrorMessage(conn));
+ pg_log_error("failed to set \"search_path\": %s", PQerrorMessage(conn));
PQclear(res);
PQfinish(conn);
return -1;
ereport(ERROR,
(errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE),
errmsg("WAL level not sufficient for making an online backup"),
- errhint("wal_level must be set to \"replica\" or \"logical\" at server start.")));
+ errhint("\"wal_level\" must be set to \"replica\" or \"logical\" at server start.")));
/*
* OK to update backup counter and session-level lock.
if (XLogArchiveCommand[0] != '\0')
return true;
- arch_module_check_errdetail("%s is not set.",
+ arch_module_check_errdetail("\"%s\" is not set.",
"archive_command");
return false;
}
if (max_worker_processes != BackgroundWorkerData->total_slots)
{
ereport(LOG,
- (errmsg("inconsistent background worker state (max_worker_processes=%d, total_slots=%d)",
+ (errmsg("inconsistent background worker state (\"max_worker_processes\"=%d, total slots=%d)",
max_worker_processes,
BackgroundWorkerData->total_slots)));
return;
* assume nothing happened.
*/
if (XLogSegmentOffset(switchpoint, wal_segment_size) != 0)
- elog(DEBUG1, "write-ahead log switch forced (archive_timeout=%d)",
+ elog(DEBUG1, "write-ahead log switch forced (\"archive_timeout\"=%d)",
XLogArchiveTimeout);
}
if (max_replication_slots == 0)
ereport(ERROR,
(errcode(ERRCODE_CONFIGURATION_LIMIT_EXCEEDED),
- errmsg("cannot start logical replication workers when max_replication_slots = 0")));
+ errmsg("cannot start logical replication workers when \"max_replication_slots\"=0")));
/*
* We need to do the modification of the shared memory under lock so that
if (res->status != WALRCV_OK_TUPLES)
ereport(ERROR,
- errmsg("could not fetch primary_slot_name \"%s\" info from the primary server: %s",
+ errmsg("could not fetch primary slot name \"%s\" info from the primary server: %s",
PrimarySlotName, res->err),
- errhint("Check if primary_slot_name is configured correctly."));
+ errhint("Check if \"primary_slot_name\" is configured correctly."));
tupslot = MakeSingleTupleTableSlot(res->tupledesc, &TTSOpsMinimalTuple);
if (!tuplestore_gettupleslot(res->tuplestore, true, false, tupslot))
elog(ERROR,
- "failed to fetch tuple for the primary server slot specified by primary_slot_name");
+ "failed to fetch tuple for the primary server slot specified by \"primary_slot_name\"");
remote_in_recovery = DatumGetBool(slot_getattr(tupslot, 1, &isnull));
Assert(!isnull);
if (newTotal > (uint64) temp_file_limit * (uint64) 1024)
ereport(ERROR,
(errcode(ERRCODE_CONFIGURATION_LIMIT_EXCEEDED),
- errmsg("temporary file size exceeds temp_file_limit (%dkB)",
+ errmsg("temporary file size exceeds \"temp_file_limit\" (%dkB)",
temp_file_limit)));
}
}
{
{"commit_timestamp_buffers", PGC_POSTMASTER, RESOURCES_MEM,
gettext_noop("Sets the size of the dedicated buffer pool used for the commit timestamp cache."),
- gettext_noop("Specify 0 to have this value determined as a fraction of shared_buffers."),
+ gettext_noop("Specify 0 to have this value determined as a fraction of \"shared_buffers\"."),
GUC_UNIT_BLOCKS
},
&commit_timestamp_buffers,
{
{"subtransaction_buffers", PGC_POSTMASTER, RESOURCES_MEM,
gettext_noop("Sets the size of the dedicated buffer pool used for the subtransaction cache."),
- gettext_noop("Specify 0 to have this value determined as a fraction of shared_buffers."),
+ gettext_noop("Specify 0 to have this value determined as a fraction of \"shared_buffers\"."),
GUC_UNIT_BLOCKS
},
&subtransaction_buffers,
{
{"transaction_buffers", PGC_POSTMASTER, RESOURCES_MEM,
gettext_noop("Sets the size of the dedicated buffer pool used for the transaction status cache."),
- gettext_noop("Specify 0 to have this value determined as a fraction of shared_buffers."),
+ gettext_noop("Specify 0 to have this value determined as a fraction of \"shared_buffers\"."),
GUC_UNIT_BLOCKS
},
&transaction_buffers,
{
{"wal_buffers", PGC_POSTMASTER, WAL_SETTINGS,
gettext_noop("Sets the number of disk-page buffers in shared memory for WAL."),
- gettext_noop("Specify -1 to have this value determined as a fraction of shared_buffers."),
+ gettext_noop("Specify -1 to have this value determined as a fraction of \"shared_buffers\"."),
GUC_UNIT_XBLOCKS
},
&XLOGbuffers,
{"log_min_duration_sample", PGC_SUSET, LOGGING_WHEN,
gettext_noop("Sets the minimum execution time above which "
"a sample of statements will be logged."
- " Sampling is determined by log_statement_sample_rate."),
+ " Sampling is determined by \"log_statement_sample_rate\"."),
gettext_noop("Zero logs a sample of all queries. -1 turns this feature off."),
GUC_UNIT_MS
},
" clean up\n"));
printf(_(" -?, --help show this help, then exit\n"));
printf(_("\n"
- "For use as archive_cleanup_command in postgresql.conf:\n"
+ "For use as \"archive_cleanup_command\" in postgresql.conf:\n"
" archive_cleanup_command = 'pg_archivecleanup [OPTION]... ARCHIVELOCATION %%r'\n"
"e.g.\n"
" archive_cleanup_command = 'pg_archivecleanup /mnt/server/archiverdir %%r'\n"));
res = PQexec(conn, ALWAYS_SECURE_SEARCH_PATH_SQL);
if (PQresultStatus(res) != PGRES_TUPLES_OK)
{
- pg_log_error("could not clear search_path: %s",
+ pg_log_error("could not clear \"search_path\": %s",
PQresultErrorMessage(res));
PQclear(res);
PQfinish(conn);
if (strcmp(wal_level, "logical") != 0)
{
- pg_log_error("publisher requires wal_level >= \"logical\"");
+ pg_log_error("publisher requires \"wal_level\" >= \"logical\"");
failed = true;
}
remoteversion_str = PQparameterStatus(AH->connection, "server_version");
remoteversion = PQserverVersion(AH->connection);
if (remoteversion == 0 || !remoteversion_str)
- pg_fatal("could not get server_version from libpq");
+ pg_fatal("could not get \"server_version\" from libpq");
AH->public.remoteVersionStr = pg_strdup(remoteversion_str);
AH->public.remoteVersion = remoteversion;
restore_command = pipe_read_line(postgres_cmd->data);
if (restore_command == NULL)
- pg_fatal("could not read restore_command from target cluster");
+ pg_fatal("could not read \"restore_command\" from target cluster");
(void) pg_strip_crlf(restore_command);
(errcode(ERRCODE_SQLCLIENT_UNABLE_TO_ESTABLISH_SQLCONNECTION),
errmsg("could not establish connection"),
errdetail("There are too many open files on the local server."),
- errhint("Raise the server's max_files_per_process and/or \"ulimit -n\" limits.")));
+ errhint("Raise the server's \"max_files_per_process\" and/or \"ulimit -n\" limits.")));
#else
ereport(ERROR,
(errcode(ERRCODE_SQLCLIENT_UNABLE_TO_ESTABLISH_SQLCONNECTION),
errmsg("could not establish connection"),
errdetail("There are too many open files on the local server."),
- errhint("Raise the server's max_files_per_process setting.")));
+ errhint("Raise the server's \"max_files_per_process\" setting.")));
#endif
}
}