⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 wal.sgml

📁 PostgreSQL 8.1.4的源码 适用于Linux下的开源数据库系统
💻 SGML
📖 第 1 页 / 共 2 页
字号:
   a smaller checkpoint interval increases the volume of output to the WAL log,   partially negating the goal of using a smaller interval,    and in any case causing more disk I/O.  </para>  <para>   Checkpoints are fairly expensive, first because they require writing   out all currently dirty buffers, and second because they result in   extra subsequent WAL traffic as discussed above.  It is therefore   wise to set the checkpointing parameters high enough that checkpoints   don't happen too often.  As a simple sanity check on your checkpointing   parameters, you can set the <xref linkend="guc-checkpoint-warning">   parameter.  If checkpoints happen closer together than   <varname>checkpoint_warning</> seconds,    a message will be output to the server log recommending increasing    <varname>checkpoint_segments</varname>.  Occasional appearance of such   a message is not cause for alarm, but if it appears often then the   checkpoint control parameters should be increased. Bulk operations such   as large <command>COPY</> transfers may cause a number of such warnings   to appear if you have not set <varname>checkpoint_segments</> high   enough.  </para>  <para>   There will be at least one WAL segment file, and will normally   not be more than 2 * <varname>checkpoint_segments</varname> + 1   files.  Each segment file is normally 16 MB (though this size can be   altered when building the server).  You can use this to estimate space   requirements for <acronym>WAL</acronym>.   Ordinarily, when old log segment files are no longer needed, they   are recycled (renamed to become the next segments in the numbered   sequence). If, due to a short-term peak of log output rate, there   are more than 2 * <varname>checkpoint_segments</varname> + 1   segment files, the unneeded segment files will be deleted instead   of recycled until the system gets back under this limit.  </para>  <para>   There are two commonly used internal <acronym>WAL</acronym> functions:   <function>LogInsert</function> and <function>LogFlush</function>.   <function>LogInsert</function> is used to place a new record into   the <acronym>WAL</acronym> buffers in shared memory. If there is no   space for the new record, <function>LogInsert</function> will have   to write (move to kernel cache) a few filled <acronym>WAL</acronym>   buffers. This is undesirable because <function>LogInsert</function>   is used on every database low level modification (for example, row   insertion) at a time when an exclusive lock is held on affected   data pages, so the operation needs to be as fast as possible.  What   is worse, writing <acronym>WAL</acronym> buffers may also force the   creation of a new log segment, which takes even more   time. Normally, <acronym>WAL</acronym> buffers should be written   and flushed by a <function>LogFlush</function> request, which is   made, for the most part, at transaction commit time to ensure that   transaction records are flushed to permanent storage. On systems   with high log output, <function>LogFlush</function> requests may   not occur often enough to prevent <function>LogInsert</function>   from having to do writes.  On such systems   one should increase the number of <acronym>WAL</acronym> buffers by   modifying the configuration parameter <xref   linkend="guc-wal-buffers">.  The default number of <acronym>WAL</acronym>   buffers is 8.  Increasing this value will   correspondingly increase shared memory usage.  When    <xref linkend="guc-full-page-writes"> is set and the system is very busy,    setting this value higher will help smooth response times during the    period immediately following each checkpoint.  </para>  <para>   The <xref linkend="guc-commit-delay"> parameter defines for how many   microseconds the server process will sleep after writing a commit   record to the log with <function>LogInsert</function> but before   performing a <function>LogFlush</function>. This delay allows other   server processes to add their commit records to the log so as to have all   of them flushed with a single log sync. No sleep will occur if   <xref linkend="guc-fsync">   is not enabled, nor if fewer than <xref linkend="guc-commit-siblings">   other sessions are currently in active transactions; this avoids   sleeping when it's unlikely that any other session will commit soon.   Note that on most platforms, the resolution of a sleep request is   ten milliseconds, so that any nonzero <varname>commit_delay</varname>   setting between 1 and 10000 microseconds would have the same effect.   Good values for these parameters are not yet clear; experimentation   is encouraged.  </para>  <para>   The <xref linkend="guc-wal-sync-method"> parameter determines how   <productname>PostgreSQL</productname> will ask the kernel to force    <acronym>WAL</acronym> updates out to disk.    All the options should be the same as far as reliability goes,   but it's quite platform-specific which one will be the fastest.   Note that this parameter is irrelevant if <varname>fsync</varname>   has been turned off.  </para>  <para>   Enabling the <xref linkend="guc-wal-debug"> configuration parameter   (provided that <productname>PostgreSQL</productname> has been   compiled with support for it) will result in each   <function>LogInsert</function> and <function>LogFlush</function>   <acronym>WAL</acronym> call being logged to the server log. This   option may be replaced by a more general mechanism in the future.  </para> </sect1> <sect1 id="wal-internals">  <title>WAL Internals</title>  <para>   <acronym>WAL</acronym> is automatically enabled; no action is   required from the administrator except ensuring that the   disk-space requirements for the <acronym>WAL</acronym> logs are met,   and that any necessary tuning is done (see <xref   linkend="wal-configuration">).  </para>  <para>   <acronym>WAL</acronym> logs are stored in the directory   <filename>pg_xlog</filename> under the data directory, as a set of   segment files, normally each 16 MB in size.  Each segment is divided into   pages, normally 8 KB each.  The log record headers are described in   <filename>access/xlog.h</filename>; the record content is dependent   on the type of event that is being logged.  Segment files are given   ever-increasing numbers as names, starting at   <filename>000000010000000000000000</filename>.  The numbers do not wrap, at   present, but it should take a very very long time to exhaust the   available stock of numbers.  </para>  <para>   It is of advantage if the log is located on another disk than the   main database files.  This may be achieved by moving the directory   <filename>pg_xlog</filename> to another location (while the server   is shut down, of course) and creating a symbolic link from the   original location in the main data directory to the new location.  </para>  <para>   The aim of <acronym>WAL</acronym>, to ensure that the log is   written before database records are altered, may be subverted by   disk drives<indexterm><primary>disk drive</></> that falsely report a   successful write to the kernel,    when in fact they have only cached the data and not yet stored it   on the disk.  A power failure in such a situation may still lead to   irrecoverable data corruption.  Administrators should try to ensure   that disks holding <productname>PostgreSQL</productname>'s   <acronym>WAL</acronym> log files do not make such false reports.  </para>  <para>   After a checkpoint has been made and the log flushed, the   checkpoint's position is saved in the file   <filename>pg_control</filename>. Therefore, when recovery is to be   done, the server first reads <filename>pg_control</filename> and   then the checkpoint record; then it performs the REDO operation by   scanning forward from the log position indicated in the checkpoint   record.  Because the entire content of data pages is saved in the   log on the first page modification after a checkpoint, all pages   changed since the checkpoint will be restored to a consistent   state.  </para>  <para>   To deal with the case where <filename>pg_control</filename> is   corrupted, we should support the possibility of scanning existing log   segments in reverse order &mdash; newest to oldest &mdash; in order to find the   latest checkpoint.  This has not been implemented yet.   <filename>pg_control</filename> is small enough (less than one disk page)   that it is not subject to partial-write problems, and as of this writing   there have been no reports of database failures due solely to inability   to read <filename>pg_control</filename> itself.  So while it is   theoretically a weak spot, <filename>pg_control</filename> does not   seem to be a problem in practice.  </para> </sect1></chapter><!-- Keep this comment at the end of the fileLocal variables:mode:sgmlsgml-omittag:nilsgml-shorttag:tsgml-minimize-attributes:nilsgml-always-quote-attributes:tsgml-indent-step:1sgml-indent-data:tsgml-parent-document:nilsgml-default-dtd-file:"./reference.ced"sgml-exposed-tags:nilsgml-local-catalogs:("/usr/lib/sgml/catalog")sgml-local-ecat-files:nilEnd:-->

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -