⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 regular.tex

📁 聚类算法全集以及内附数据集
💻 TEX
字号:
\documentclass{article}\pagestyle{empty}\oddsidemargin 2.1mm\textwidth     155mm\topmargin     -12mm\textheight    230mm\def\IR{\mbox{I\kern-0.22em R}}\def\genlistlabel#1{#1\hfill}\newenvironment{genlist}[1]{\begin{list}{}{%\topsep2pt plus1pt minus1pt\itemsep\topsep\advance\topsep\parsep\settowidth\labelwidth{#1}\settowidth\labelsep{.~}%\leftmargin\labelwidth\advance\leftmargin\labelsep\let\makelabel\genlistlabel}}{\end{list}}\begin{document}%-----------------------------------------------------------------------\subsection*{Regularization}Regularizing the shape of a cluster or the sizes of the clusters meansto modify the shape or the sizes in such a way that certain conditionsare satisfied or at least a tendency (of varying strength, as specifiedby a user) towards satisfying these conditions is introduced.%-----------------------------------------------------------------------\subsubsection*{Shape Regularization}The shape of a cluster is represented by its covariance matrix$\mathbf{\Sigma}$. Intuitively, $\mathbf{\Sigma}$ describes a generalellipsoidal shape, which can be obtained, for example, by computing theCholesky decomposition of~$\mathbf{\Sigma}$ and mapping the unit spherewith it. Shape regularization means to modify the covariance matrix,so that a certain relation of the lengths of the major axes of therepresented ellipsoid is obtained or that at least a tendency towardsthis relation is introduced. Since the lengths of the major axes arethe roots of the eigenvalues of the covariance matrix, regularizing itmeans shifting the eigenvalues of~$\mathbf{\Sigma}$. Note that such ashift leaves the eigenvectors unchanged, i.e., the orientation of therepresented ellipsoid is preserved. The clustering programs support twoversion of shape regularization:\begin{genlist}{\tt -H-$h$}\item[\tt -H\,$h$] {\bf Standard Version} \\     Adapt the covariance matrices~$\mathbf{\Sigma}_i$,     $i = 1, \ldots, c$ (where $c$ is the number of clusters),     according to     \[ \mathbf{\Sigma}_i^{\rm (adap)}        = \sigma_i^2 \cdot          \frac{\mathbf{S}_i +h^2 \mathbf{E}}               {\sqrt[m]{|\mathbf{S}_i +h^2 \mathbf{E}|}}        = \sigma_i^2 \cdot          \frac{\mathbf{\Sigma}_i +\sigma_i^2 h^2 \mathbf{E}}               {\sqrt[m]{|\mathbf{\Sigma}_i                         +\sigma_i^2 h^2 \mathbf{E}|}}, \]     where $m$ is the dimension of the data space (i.e.,     the $\mathbf{\Sigma}_i$ are $m \times m$ matrices,     $\mathbf{E}$ is the $m \times m$ unit matrix),     $\sigma_i^2 = \sqrt[m]{|\mathbf{\Sigma}_i|}$     is the equivalent isotropic variance, and     $\mathbf{S}_i = \sigma_i^{-2} \mathbf{\Sigma}_i$     is the covariance matrix scaled to unit determinant.     This regularization shifts all eigenvalues up by the value     of~$\sigma^2 h^2$ and then renormalizes the resulting matrix so     that the determinant of the old covariance matrix is preserved     (the cluster size is kept constant). This regularization tends     to equalize the lengths of the axes of the ellipsoid and thus     introduces a tendency towards spherical clusters. This tendency     is the stronger, the greater the value of~$h$. In the limit, for     $h \to \infty$, the clusters are forced to be exactly spherical.     On the other hand, for $h = 0$ the shape is left unchanged.\item[\tt -H-$r$] {\bf Alternative Version} \\     The standard version always changes the relation of the lengths     of the major axes and thus introduces a general tendency towards     spherical clusters. In the alternative version a limit~$r$,     $r > 1$, for the length ratio of the longest to the shortest major     axis of the (hyper-)ellipsoid is used and only if this limit is     exceeded, the eigenvalues are shifted in such a way that the limit     is satisfied. Formally: let $\lambda_k$, $k = 1, \ldots m$, be the     eigenvalues of the matrix~$\mathbf{\Sigma}_i$. Set     \[ h^2 = \left\{\begin{array}{ll}              0, & \mbox{if~~}\displaystyle                   \frac{\max{}_{k=1}^m \lambda_k}                        {\min{}_{k=1}^m \lambda_k} \le r^2, \\              \displaystyle              \frac{     \max{}_{k=1}^m \lambda_k                    -r^2 \min{}_{k=1}^m \lambda_k}                   {\sigma_i^2 (r^2 -1)},                 & \mbox{otherwise,}              \end{array}\right. \]     and execute the standard version with this value of~$h^2$.\end{genlist}%-----------------------------------------------------------------------\subsubsection*{Size Regularization}The size of a cluster can be described in different ways, for example,by the determinant of its covariance matrix~$\mathbf{\Sigma}_i$,$i = 1, \ldots, c$ ($c$ is the number of clusters), which is ameasure of the clusters squared (hyper-)volume, an equivalent isotropicvariance~$\sigma_i^2$ or an equivalent isotropic radius (standarddeviation)~$\sigma_i$ (equivalent in the sense that they lead to thesame (hyper-)volume). The latter two measures are defined as\[ \sigma_i^2 = \sqrt[m]{|\mathbf{\Sigma}_i|} \qquad\mbox{and}\qquad   \sigma_i = \sqrt{\sigma_i^2} = \sqrt[2m]{|\mathbf{\Sigma}_i|} \]and thus the (hyper-)volume can also be written as$\sigma_i^m = \sqrt{|\mathbf{\Sigma}_i|}$. Size regularization meansto ensure a certain relation between the cluster sizes or at least tointroduce a tendency into this direction. The clustering programssupport different versions of size regularization, in which themeasure that is used to describe the cluster size is specifiedby an exponent~$a$ of the equivalent isotropic radius~$\sigma_i$:\begin{center}\begin{tabular}{ll}$a = 1:$ & size is measured by equivalent isotropic radius, \\$a = 2:$ & size is measured by equivalent isotropic variance, \\$a = m:$ & size is measured by (hyper-)volume.\end{tabular}\end{center}The supported size regularization versions are:\begin{genlist}{\tt -R-$b$:$a$:[-]$s$}\item[\tt -R\,$b$:$a$:$s$] {\bf Standard Version} \\     The equivalent isotropic radius~$\sigma_i$ is adapted according to     \[ \sigma_i^{\rm (adap)}        = \sqrt[a]{s \cdot \frac{\sum_{k=1}^c \sigma_k^a}                                {\sum_{k=1}^c (\sigma_k^a +b)}                     \cdot (\sigma_i^a + b)}        = \sqrt[a]{s \cdot \frac{\sum_{k=1}^c \sigma_k^a}                                {c b +\sum_{k=1}^c \sigma_k^a}                     \cdot (\sigma_i^a + b)}. \]     That is, each size is increased by the value of~$b$ and then the     sizes are renormalized so that the sum of the cluster sizes is     preserved. However, the parameter~$s$ can be used to scale the     sum of the sizes up or down.\item[\tt -R\,$b$:$a$:-$s$] {\bf Simplified Version} \\     The simplified version does not renormalize the sizes. However,     this missing normalization may be mitigated to some degree by     specifying a value of $s$ that is smaller than 1.     The equivalent isotropic radius~$\sigma_i$ is adapted according to     \[ \sigma_i^{\rm (adap)} = \sqrt[a]{s \cdot (\sigma_i^a + b)}. \]     Note that, with $b = 0$, this version can be used to simple scale     the clusters.\item[\tt -R-$b$:$a$:$s$] {\bf Full Size Equalization} \\     In the limit, for $b \to \infty$, the standard version leads to     a full equalization of the cluster sizes. This can be achieved     by specifying a negative first parameter, $b \le 1$. In this case     the equivalent isotropic radius~$\sigma_i$ is adapted according to     \[ \sigma_i^{\rm (adap)}        = \sqrt[a]{\frac{s}{c}\sum_{k=1}^c \sigma_k^a}. \]     That is, all clusters are set to the same size.\item[\tt -R-$r$:$a$:[-]$s$] {\bf Alternative Version} \\     The standard version always changes the relation of the cluster     sizes and thus introduces a general tendency towards clusters of     equal size. In the alternative version a limit~$r$, $r > 1$, for     the size ratio of the largest to the smallest cluster is used and     only if this limit is exceeded, the sizes are changed in such a     way that the limit is satisfied. To achieve this, $b$ is set     according to     \[ b = \left\{\begin{array}{ll}            0, & \mbox{if }\displaystyle                 \frac{\max{}_{k=1}^c \sigma_k^a}                 {\min{}_{k=1}^c \sigma_k^a} \le r, \\            \displaystyle            \frac{   \max{}_{k=1}^c \sigma_k^a                  -r \min{}_{k=1}^c \sigma_k^a}                 {r -1},               & \mbox{otherwise,}            \end{array}\right. \]      and the standard or simplified version (depending on the sign of     the third parameter) is executed with this value of~$b$ and the     values of $a$ and $s$ as given on the command line.\end{genlist}%-----------------------------------------------------------------------\end{document}

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -