Model-based Agglomerative Hierarchical Clustering
Agglomerative hierarchical clustering based on maximum likelihood criteria for Gaussian mixture models parameterized by eigenvalue decomposition.
hc(data, modelName = "VVV", use = "VARS", partition = dupPartition(data), minclus = 1, ...)
data |
A numeric vector, matrix, or data frame of observations. Categorical variables are not allowed. If a matrix or data frame, rows correspond to observations (n) and columns correspond to variables (d). |
modelName |
A character string indicating the model to be used in model-based agglomerative hierarchical clustering.
If |
use |
A character string specifying the type of input variables/data transformation to be used for model-based agglomerative hierarchical clustering.
If |
partition |
A numeric or character vector representing a partition of
observations (rows) of |
minclus |
A number indicating the number of clusters at which to stop the agglomeration. The default is to stop when all observations have been merged into a single cluster. |
... |
Arguments for the method-specific |
Most models have memory usage of the order of the square of the
number groups in the initial partition for fast execution.
Some models, such as equal variance or "EEE"
,
do not admit a fast algorithm under the usual agglomerative
hierarchical clustering paradigm.
These use less memory but are much slower to execute.
The function hc()
returns a numeric two-column matrix in which
the ith row gives the minimum index for observations in each of
the two clusters merged at the ith stage of agglomerative
hierarchical clustering. Several other informations are also returned
as attributes.
The plotting method plot.hc()
draws a dendrogram, which can be based
on either the classification loglikelihood or the merge level (number of
clusters). For details, see the associated help file.
If modelName = "E"
(univariate with equal variances) or
modelName = "EII"
(multivariate with equal spherical
covariances), then underlying model is the same as that for
Ward's method for hierarchical clustering.
Banfield J. D. and Raftery A. E. (1993). Model-based Gaussian and non-Gaussian Clustering. Biometrics, 49:803-821.
Fraley C. (1998). Algorithms for model-based Gaussian hierarchical clustering. SIAM Journal on Scientific Computing, 20:270-281.
Fraley C. and Raftery A. E. (2002). Model-based clustering, discriminant analysis, and density estimation. Journal of the American Statistical Association, 97:611-631.
Scrucca L. and Raftery A. E. (2015). Improved initialisation of model-based clustering using Gaussian hierarchical partitions. Advances in Data Analysis and Classification, 9/4:447-460.
hcE
, ...,
hcVVV
,
plot.hc
,
hclass
,
mclust.options
hcTree <- hc(modelName = "VVV", data = iris[,-5]) hcTree cl <- hclass(hcTree,c(2,3)) table(cl[,"2"]) table(cl[,"3"]) clPairs(iris[,-5], classification = cl[,"2"]) clPairs(iris[,-5], classification = cl[,"3"])
Please choose more modern alternatives, such as Google Chrome or Mozilla Firefox.