mou_EM {deepMOU} | R Documentation |
Mixture of Unigrams by Expectation-Maximization algorithm
Description
Performs parameter estimation by means of the Expectation-Maximization (EM) algorithm and cluster allocation for the Mixture of Unigrams.
Usage
mou_EM(x, k, n_it = 100, eps = 1e-07, seed_choice = 1, init = NULL)
Arguments
x |
Document-term matrix describing the frequency of terms that occur in a collection of documents. Rows correspond to documents in the collection and columns correspond to terms. |
k |
Number of clusters/groups. |
n_it |
Number of iterations for the Expectation-Maximization algorithm. |
eps |
Tolerance level for the convergence of the algorithm. Default is |
seed_choice |
Set seed for reproducible results |
init |
Vector containing the initial document allocations for the initialization of the algorithm. If NULL (default) initialization is carried out via spherical k-means (skmeans). |
Details
Starting from the data given by x
the Mixture of Unigrams is fitted
and k
clusters are obtained.
The algorithm for the parameter estimation is the Expectation-Maximization (EM).
In particular, the function assigns initial values to weights of the Multinomial distribution for each cluster
and inital weights for the components of the mixture. The estimates are obtained with maximum n_it
steps of the
EM algorithm or until the tolerance level eps
is reached; by using the posterior distribution
of the latent variable z, the documents are allocated to the cluster which maximizes the
posterior distribution.
For further details see the references.
Value
A list containing the following elements:
x |
The data matrix. |
clusters |
the clustering labels. |
k |
the number of clusters. |
numobs |
the sample size. |
p |
the vocabulary size. |
likelihood |
vector containing the likelihood values at each iteration. |
pi_hat |
estimated probabilities of belonging to the |
omega |
matrix containing the estimates of the omega parameters for each cluster. |
f_z_x |
matrix containing the posterior probabilities of belonging to each cluster. |
AIC |
Akaike Information Criterion (AIC) value of the fitted model. |
BIC |
Bayesian Information Criterion (BIC) value of the fitted model. |
References
Nigam, K., McCallum, A., Thrun, S., Mitchell, T.: Text classification from labeled and unlabeled documents using EM. Machine learning 39, 103-134 (2000).
Examples
# Load the CNAE2 dataset
data("CNAE2")
# Perform parameter estimation and clustering
mou_CNAE2 = mou_EM(x = CNAE2, k = 2)
# Shows cluster labels to documents
mou_CNAE2$clusters