Title: | Seeded Sequential LDA for Topic Modeling |
---|---|
Description: | Seeded Sequential LDA can classify sentences of texts into pre-define topics with a small number of seed words (Watanabe & Baturo, 2023) <doi:10.1177/08944393231178605>. Implements Seeded LDA (Lu et al., 2010) <doi:10.1109/ICDMW.2011.125> and Sequential LDA (Du et al., 2012) <doi:10.1007/s10115-011-0425-1> with the distributed LDA algorithm (Newman, et al., 2009) for parallel computing. |
Authors: | Kohei Watanabe [aut, cre, cph], Phan Xuan-Hieu [aut, cph] (GibbsLDA++) |
Maintainer: | Kohei Watanabe <[email protected]> |
License: | GPL-3 |
Version: | 1.4.1 |
Built: | 2024-11-07 06:21:15 UTC |
Source: | CRAN |
A corpus object containing 2,000 movie reviews.
https://www.cs.cornell.edu/people/pabo/movie-review-data/
Pang, B., Lee, L. (2004) "A Sentimental Education: Sentiment Analysis Using Subjectivity Summarization Based on Minimum Cuts.", Proceedings of the ACL.
divergence()
computes the regularized topic divergence scores to help users
to find the optimal number of topics for LDA.
divergence( x, min_size = 0.01, select = NULL, regularize = TRUE, newdata = NULL, ... )
divergence( x, min_size = 0.01, select = NULL, regularize = TRUE, newdata = NULL, ... )
x |
a LDA model fitted by |
min_size |
the minimum size of topics for regularized topic divergence.
Ignored when |
select |
names of topics for which the divergence is computed. |
regularize |
if |
newdata |
if provided, |
... |
additional arguments passed to textmodel_lda. |
divergence()
computes the average Jensen-Shannon divergence
between all the pairs of topic vectors in x$phi
. The divergence score
maximizes when the chosen number of topic k
is optimal (Deveaud et al.,
2014). The regularized divergence penalizes topics smaller than min_size
to avoid fragmentation (Watanabe & Baturo, forthcoming).
Returns a singple numeric value.
Deveaud, Romain et al. (2014). "Accurate and Effective Latent Concept Modeling for Ad Hoc Information Retrieval". doi:10.3166/DN.17.1.61-84. Document Numérique.
Watanabe, Kohei & Baturo, Alexander. (2023). "Seeded Sequential LDA: A Semi-supervised Algorithm for Topic-specific Analysis of Sentences". doi:10.1177/08944393231178605. Social Science Computer Review.
perplexity()
computes the perplexity score to help users to chose the
optimal values of hyper-parameters for LDA.
perplexity(x, newdata = NULL, ...)
perplexity(x, newdata = NULL, ...)
x |
a LDA model fitted by |
newdata |
if provided, |
... |
additional arguments passed to textmodel_lda. |
perplexity()
predicts the distribution of words in the dfm based
on x$alpha
and x$gamma
and then compute the sum of disparity between their
predicted and observed frequencies. The perplexity score minimizes when the
chosen values of hyper-parameters such as k
, alpha
and gamma
are
optimal.
Returns a singple numeric value.
Compute the sizes of topics as the proportions of topic words in the corpus.
sizes(x)
sizes(x)
x |
a LDA model fitted by |
a numeric vector in the same lengths as k
.
terms()
returns the most likely terms, or words, for topics based on the
phi
parameter.
terms(x, n = 10)
terms(x, n = 10)
x |
a LDA model fitted by |
n |
number of terms to be extracted. |
Users can access the original matrix x$phi
for likelihood scores.
a character matrix with the most frequent words in each topic.
Implements unsupervised Latent Dirichlet allocation (LDA). Users can run
Seeded LDA by setting gamma > 0
.
textmodel_lda( x, k = 10, max_iter = 2000, auto_iter = FALSE, alpha = 0.5, beta = 0.1, gamma = 0, adjust_alpha = 0, model = NULL, update_model = FALSE, batch_size = 1, verbose = quanteda_options("verbose") )
textmodel_lda( x, k = 10, max_iter = 2000, auto_iter = FALSE, alpha = 0.5, beta = 0.1, gamma = 0, adjust_alpha = 0, model = NULL, update_model = FALSE, batch_size = 1, verbose = quanteda_options("verbose") )
x |
the dfm on which the model will be fit. |
k |
the number of topics. |
max_iter |
the maximum number of iteration in Gibbs sampling. |
auto_iter |
if |
alpha |
the values to smooth topic-document distribution. |
beta |
the values to smooth topic-word distribution. |
gamma |
a parameter to determine change of topics between sentences or
paragraphs. When |
adjust_alpha |
[experimental] if |
model |
a fitted LDA model; if provided, |
update_model |
if |
batch_size |
split the corpus into the smaller batches (specified in
proportion) for distributed computing; it is disabled when a batch include
all the documents |
verbose |
logical; if |
If auto_iter = TRUE
, the iteration stops even before max_iter
when delta <= 0
. delta
is computed to measure the changes in the number
of words whose topics are updated by the Gibbs sampler in every 100
iteration as shown in the verbose message.
If batch_size < 1.0
, the corpus is partitioned into sub-corpora of
ndoc(x) * batch_size
documents for Gibbs sampling in sub-processes with
synchronization of parameters in every 10 iteration. Parallel processing is
more efficient when batch_size
is small (e.g. 0.01). The algorithm is the
Approximate Distributed LDA proposed by Newman et al. (2009). User can
changed the number of sub-processes used for the parallel computing via
options(seededlda_threads)
.
set.seed()
should be called immediately before textmodel_lda()
or
textmodel_seededlda()
to control random topic assignment. If the random
number seed is the same, the serial algorithm produces identical results;
the parallel algorithm produces non-identical results because it
classifies documents in different orders using multiple processors.
To predict topics of new documents (i.e. out-of-sample), first, create a
new LDA model from a existing LDA model passed to model
in
textmodel_lda()
; second, apply topics()
to the new model. The model
argument takes objects created either by textmodel_lda()
or
textmodel_seededlda()
.
Returns a list of model parameters:
k |
the number of topics. |
last_iter |
the number of iterations in Gibbs sampling. |
max_iter |
the maximum number of iterations in Gibbs sampling. |
auto_iter |
the use of |
adjust_alpha |
the value of |
alpha |
the smoothing parameter for |
beta |
the smoothing parameter for |
epsilon |
the amount of adjustment for |
gamma |
the gamma parameter for Sequential LDA. |
phi |
the distribution of words over topics. |
theta |
the distribution of topics over documents. |
words |
the raw frequency count of words assigned to topics. |
data |
the original input of |
call |
the command used to execute the function. |
version |
the version of the seededlda package. |
Newman, D., Asuncion, A., Smyth, P., & Welling, M. (2009). Distributed Algorithms for Topic Models. The Journal of Machine Learning Research, 10, 1801–1828.
require(seededlda) require(quanteda) corp <- head(data_corpus_moviereviews, 500) toks <- tokens(corp, remove_punct = TRUE, remove_symbols = TRUE, remove_number = TRUE) dfmt <- dfm(toks) %>% dfm_remove(stopwords("en"), min_nchar = 2) %>% dfm_trim(max_docfreq = 0.1, docfreq_type = "prop") lda <- textmodel_lda(dfmt, k = 6, max_iter = 500) # 6 topics terms(lda) topics(lda)
require(seededlda) require(quanteda) corp <- head(data_corpus_moviereviews, 500) toks <- tokens(corp, remove_punct = TRUE, remove_symbols = TRUE, remove_number = TRUE) dfmt <- dfm(toks) %>% dfm_remove(stopwords("en"), min_nchar = 2) %>% dfm_trim(max_docfreq = 0.1, docfreq_type = "prop") lda <- textmodel_lda(dfmt, k = 6, max_iter = 500) # 6 topics terms(lda) topics(lda)
Implements semisupervised Latent Dirichlet allocation
(Seeded LDA). textmodel_seededlda()
allows users to specify
topics using a seed word dictionary. Users can run Seeded Sequential LDA by
setting gamma > 0
.
textmodel_seededlda( x, dictionary, levels = 1, valuetype = c("glob", "regex", "fixed"), case_insensitive = TRUE, residual = 0, weight = 0.01, max_iter = 2000, auto_iter = FALSE, alpha = 0.5, beta = 0.1, gamma = 0, adjust_alpha = 0, batch_size = 1, ..., verbose = quanteda_options("verbose") )
textmodel_seededlda( x, dictionary, levels = 1, valuetype = c("glob", "regex", "fixed"), case_insensitive = TRUE, residual = 0, weight = 0.01, max_iter = 2000, auto_iter = FALSE, alpha = 0.5, beta = 0.1, gamma = 0, adjust_alpha = 0, batch_size = 1, ..., verbose = quanteda_options("verbose") )
x |
the dfm on which the model will be fit. |
dictionary |
a |
levels |
levels of entities in a hierarchical dictionary to be used as seed words. See also quanteda::flatten_dictionary. |
valuetype |
|
case_insensitive |
|
residual |
the number of undefined topics. They are named "other" by
default, but it can be changed via |
weight |
determines the size of pseudo counts given to matched seed words. |
max_iter |
the maximum number of iteration in Gibbs sampling. |
auto_iter |
if |
alpha |
the values to smooth topic-document distribution. |
beta |
the values to smooth topic-word distribution. |
gamma |
a parameter to determine change of topics between sentences or
paragraphs. When |
adjust_alpha |
[experimental] if |
batch_size |
split the corpus into the smaller batches (specified in
proportion) for distributed computing; it is disabled when a batch include
all the documents |
... |
passed to quanteda::dfm_trim to restrict seed words based on their term or document frequency. This is useful when glob patterns in the dictionary match too many words. |
verbose |
logical; if |
The same as textmodel_lda()
with extra elements for dictionary
.
Lu, Bin et al. (2011). "Multi-aspect Sentiment Analysis with Topic Models". doi:10.5555/2117693.2119585. Proceedings of the 2011 IEEE 11th International Conference on Data Mining Workshops.
Watanabe, Kohei & Zhou, Yuan. (2020). "Theory-Driven Analysis of Large Corpora: Semisupervised Topic Classification of the UN Speeches". doi:10.1177/0894439320907027. Social Science Computer Review.
Watanabe, Kohei & Baturo, Alexander. (2023). "Seeded Sequential LDA: A Semi-supervised Algorithm for Topic-specific Analysis of Sentences". doi:10.1177/08944393231178605. Social Science Computer Review.
require(seededlda) require(quanteda) corp <- head(data_corpus_moviereviews, 500) toks <- tokens(corp, remove_punct = TRUE, remove_symbols = TRUE, remove_number = TRUE) dfmt <- dfm(toks) %>% dfm_remove(stopwords("en"), min_nchar = 2) %>% dfm_trim(max_docfreq = 0.1, docfreq_type = "prop") dict <- dictionary(list(people = c("family", "couple", "kids"), space = c("alien", "planet", "space"), moster = c("monster*", "ghost*", "zombie*"), war = c("war", "soldier*", "tanks"), crime = c("crime*", "murder", "killer"))) lda_seed <- textmodel_seededlda(dfmt, dict, residual = TRUE, min_termfreq = 10, max_iter = 500) terms(lda_seed) topics(lda_seed)
require(seededlda) require(quanteda) corp <- head(data_corpus_moviereviews, 500) toks <- tokens(corp, remove_punct = TRUE, remove_symbols = TRUE, remove_number = TRUE) dfmt <- dfm(toks) %>% dfm_remove(stopwords("en"), min_nchar = 2) %>% dfm_trim(max_docfreq = 0.1, docfreq_type = "prop") dict <- dictionary(list(people = c("family", "couple", "kids"), space = c("alien", "planet", "space"), moster = c("monster*", "ghost*", "zombie*"), war = c("war", "soldier*", "tanks"), crime = c("crime*", "murder", "killer"))) lda_seed <- textmodel_seededlda(dfmt, dict, residual = TRUE, min_termfreq = 10, max_iter = 500) terms(lda_seed) topics(lda_seed)
Implements Sequential Latent Dirichlet allocation (Sequential LDA).
textmodel_seqlda()
allows the users to classify sentences of texts. It
considers the topics of previous document in inferring the topics of currency
document. textmodel_seqlda()
is a shortcut equivalent to
textmodel_lda(gamma = 0.5)
. Seeded Sequential LDA is
textmodel_seededlda(gamma = 0.5)
.
textmodel_seqlda( x, k = 10, max_iter = 2000, auto_iter = FALSE, alpha = 0.5, beta = 0.1, batch_size = 1, model = NULL, verbose = quanteda_options("verbose") )
textmodel_seqlda( x, k = 10, max_iter = 2000, auto_iter = FALSE, alpha = 0.5, beta = 0.1, batch_size = 1, model = NULL, verbose = quanteda_options("verbose") )
x |
the dfm on which the model will be fit. |
k |
the number of topics. |
max_iter |
the maximum number of iteration in Gibbs sampling. |
auto_iter |
if |
alpha |
the values to smooth topic-document distribution. |
beta |
the values to smooth topic-word distribution. |
batch_size |
split the corpus into the smaller batches (specified in
proportion) for distributed computing; it is disabled when a batch include
all the documents |
model |
a fitted LDA model; if provided, |
verbose |
logical; if |
The same as textmodel_lda()
Du, Lan et al. (2012). "Sequential Latent Dirichlet Allocation". doi.org/10.1007/s10115-011-0425-1. Knowledge and Information Systems.
Watanabe, Kohei & Baturo, Alexander. (2023). "Seeded Sequential LDA: A Semi-supervised Algorithm for Topic-specific Analysis of Sentences". doi:10.1177/08944393231178605. Social Science Computer Review.
require(seededlda) require(quanteda) corp <- head(data_corpus_moviereviews, 500) %>% corpus_reshape() toks <- tokens(corp, remove_punct = TRUE, remove_symbols = TRUE, remove_number = TRUE) dfmt <- dfm(toks) %>% dfm_remove(stopwords("en"), min_nchar = 2) %>% dfm_trim(max_docfreq = 0.01, docfreq_type = "prop") lda_seq <- textmodel_seqlda(dfmt, k = 6, max_iter = 500) # 6 topics terms(lda_seq) topics(lda_seq)
require(seededlda) require(quanteda) corp <- head(data_corpus_moviereviews, 500) %>% corpus_reshape() toks <- tokens(corp, remove_punct = TRUE, remove_symbols = TRUE, remove_number = TRUE) dfmt <- dfm(toks) %>% dfm_remove(stopwords("en"), min_nchar = 2) %>% dfm_trim(max_docfreq = 0.01, docfreq_type = "prop") lda_seq <- textmodel_seqlda(dfmt, k = 6, max_iter = 500) # 6 topics terms(lda_seq) topics(lda_seq)
topics()
returns the most likely topics for documents based on the theta
parameter.
topics(x, min_prob = 0, select = NULL)
topics(x, min_prob = 0, select = NULL)
x |
a LDA model fitted by |
min_prob |
ignores topics if their probability is lower than this value. |
select |
returns the selected topic with the
highest probability; specify by the names of columns in |
Users can access the original matrix x$theta
for likelihood
scores; run max.col(x$theta)
to obtain the same result as topics(x)
.
Returns predicted topics as a vector.