Next Article in Journal
Some Identities Involving Degenerate q-Hermite Polynomials Arising from Differential Equations and Distribution of Their Zeros
Next Article in Special Issue
A Modified Stein Variational Inference Algorithm with Bayesian and Gradient Descent Techniques
Previous Article in Journal
SVC Parameters Optimization Using a Novel Integrated MCDM Approach
Previous Article in Special Issue
Self-Supervised Graph Representation Learning via Information Bottleneck
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Generative Model for Topic Discovery and Polysemy Embeddings on Directed Attributed Networks

1
Hebei Key Laboratory of Optoelectronic Information and Geo-Detection Technology, Hebei GEO University, Shijiazhuang 050031, China
2
Information Engineering College, Hebei GEO University, Shijiazhuang 050031, China
3
Intelligent Sensor Network Engineering Research Center of Hebei Province, Hebei GEO University, Shijiazhuang 050031, China
*
Author to whom correspondence should be addressed.
Symmetry 2022, 14(4), 703; https://doi.org/10.3390/sym14040703
Submission received: 27 February 2022 / Revised: 25 March 2022 / Accepted: 28 March 2022 / Published: 30 March 2022
(This article belongs to the Special Issue Symmetry and Asymmetry Studies on Graph Data Mining)

Abstract

:
Combining topic discovery with topic-specific word embeddings is a popular, powerful method for text mining in a small collection of documents. However, the existing researches purely modeled on the contents of documents and led to discovering noisy topics. This paper proposes a generative model, the skip-gram topical word-embedding model (simplified as steoLC) on asymmetric document link networks, where nodes correspond to documents and links refer to directed references between documents. It simultaneously improves the performance of topic discovery and polysemous word embeddings. Each skip-gram in a document is generated based on the topic distribution of the document and the two word embeddings in the skip-gram. Each directed link is generated based on the hidden topic distribution of the beginning document node. For a document, the skip-grams and links share a common topic distribution. Parameter estimation is inferred and an algorithm is designed to learn the model parameters by combining the expectation-maximization (EM) algorithm with the negative sampling method. Experimental results show that our method generates more useful topic-specific word embeddings and coherent latent topics than the state-of-the-art models.

1. Introduction

With the development of internet technology, an enormous number of online information service platforms generate more and more information with rich content and links, such as paper citation networks and hyperlink networks of the World Wide Web, etc. Many methods model these networked data as attributed networks. In the formalized representation, each node corresponds to a paper or a webpage, which is often associated with a rich set of text attributes formed by word sequence features. An edge represents whether there are associations between the two nodes. In this paper, we mainly focus on directed links, and an undirected link equals two directed links. Summarizing the semantic structures and topology structures hidden in these networked data can help people understand the networked data, which have been exploited in the areas of natural language processing and graph mining. In the former area, traditional topic discovery and word embedding technologies purely mine the data based on the node contents of networked data. While in the area of graph mining, research shows that the network topological structure and node attributes are often strongly correlated with each other. For example, two papers with high correlation topics would have a reference link. Semantic community detection methods are always used, which jointly exploit the two information sources to enhance the learning performance. The following will analyze the two kinds of methods.
Topic discovery is a popular tool for detecting the topic structure of the contents of networked data, such as Probabilistic Latent Semantic Analysis (PLSA) [1] and Latent Dirichlet Allocation (LDA) [2]. Topic models can summarize documents as a mixture of the topics, and the topics can be summarized as a distribution on the large vocabularies of the document sets. However, they suffer from the coherence of discovered topics due to one hot word representation. In parallel with the development of traditional topic discovery, word embedding algorithms represent words as dense distributed vectors [3] to avoid the semantic gaps and sparsity caused by one hot representation. However, the embedding methods just learn local word co-ocurrence information, which loses the global semantic information of the words. Recently, some studies integrate topic discovery with word embedding. They aim to learn more comprehensive word embedding and discover more accurate topics. There are three main kinds of methods, whose advantages and disadvantages are compared in Table 1. The first kind makes use of the pre-trained word embeddings to improve the performance of topic discovery. For example, TopicVec modeled the generative process of words in each document given the topic embedding and the embeddings of the word and its contexts [4]. The GaussianLDA [5,6] learned pre-trained word embeddings, and modeled a multivariate Gaussian distribution with the topic embedding as its mean to generate word embeddings. The second kind utilizes topic discovery to aid word embeddings. For example, Topic Word Embedding (TWE) first detected topics using the LDA model, and then treated each topic as a pseudo-word to learn topic embedding, which was concentrated with word embeddings to get the final word embeddings [7]. Briakou et al. [8] first learned topics from a large corpus and then learned the topic-specific word embeddings spanned by anchor words. The aforementioned two approaches are both two-step processes, and they cannot model the mutual influence of topic discovery and word embeddings. Besides these two kinds of methods, the third method integrates the advantages of topic discovery with word embedding and models their mutual interactions [9,10,11,12]. The Collaborative Language Model (CLM) applied a nonnegative matrix factorization to model both topic discovery and word embeddings [9]. The Joint Topic Word-embedding (JTW) model provided a deep generative model by combining a variational autoencoder with the topic model [10]. Topic Modeling boosted with Sparse Autoencoder (TMSA) modeled the mutual influence of topic discovery and word embedding based topic modeling and an autoencoder [11]. The skip-gram Topical word Embedding (STE) model extended the skip-gram model by considering topic discovery and learned topic-specific word embeddings to solve the problem of polysemy [12]. Compared with other similar methods, the skip-gram embedding (STE) model has two advantages. It not only learns word embeddings and topics in a unified framework, but also explicitly obtains topic-specific word embeddings, thus solving polysemy problems. However, the STE model purely models the contents of documents, which always captains inconsistent top words in the detected topics. This inaccuracy of topic results would further mislead word embedding learning. In the area of graph mining, research has shown that semantic community detection methods on document link networks were able to improve the performance of topic discovery by either contents or links [13,14].
Several semantic community detection methods combine links and contents of the attributed networks to detect communities with common topics [13,14,15,16,17]. Different from community detection methods on the topology network, they are devoted to detecting semantic communities considering the text features of the nodes. Different from topic discovery, they make use of links to improve the performance of traditional topic discovery and discover topics with link patterns [18]. However, these existing methods largely use one hot encoding to represent words and documents. This representation loses many semantics of the documents and words due to the semantic gaps and the sparsity, and also increases the complexity of the algorithm due to the high dimension vector. Community-Enhanced Topic Embedding (CeTe) was the first model that integrated the contents and links for topic embedding and word embedding on attributed networks [19]. The CeTe model represented each word by just one word embedding vector, which did not explicitly model topic-specific word embeddings. In addition, the community detection on links was used in the preprocessing stage, which was not unified with topic discovery on contents and do not fully combine contents and links for topic discovery. This motivates us to design an integrated model for topic-specific embedding learning and semantic community discovery based on links and content, which uses links to improve the integrated models of word embeddings and topic discovery methods.
All in all, the STE model integrated word embeddings with topic discovery to improve the two tasks on the word sequences of documents. However, it ignored the rich links between documents, which were able to complement the semantic fuzziness problem of content-driven topic discovery. The CeTe model was an example that improved topic-specific word embedding on document contents by community detection. However, the community detection on links was not integrated with topic discovery and each word was not represented based on topics. Our model is designed to solve the above problems.
To learn better word embeddings and topics based on attributed networks, we propose a joint probabilistic model named the steoLC model for word embedding learning and topic discovery on document link networks. This model assumes that the topic distributions are decided by both content and links. Topic discovery and word embeddings influence each other. Each skip-gram word pair and each link are generated based on the given topic distributions. The topic distributions are decided by the topic distributions of skip-grams and links, and the more accurate topics are estimated according to links and content. Then word embedding is improved by the topic distribution. This model alleviates the semantic ambiguity of word embeddings caused by the ambiguity of topics and learns more accurate word embeddings. An algorithm for the steoLC model is designed based on the EM algorithm, which iteratively estimates the hidden topic distributions and model parameters. In each iteration, the skip-gram negative sampling method is used to learn the word embeddings.
The contributions of this work are as follows:
  • First, we present a joint probabilistic generative model, called steoLC, for word embedding learning and topic discovery based on directed document link networks. It models the generative process of both the document word pairs and links by word embedding distributions on topics, the topic distributions of documents, the skip-gram distributions on topics, and the link distributions on topics;
  • Second, an algorithm is inferred and designed to estimate the parameters of the steoLC model, which combines the EM algorithm framework with a negative sampling algorithm;
  • Finally, the performance of the steoLC model is tested on four aspects, including the visualization of word embedding on different topics, document classification, computing nearest words, and the evaluation of topic consistency.
The remainder of the paper is organized as follows. Section 2 introduces the generative process of our model in detail. The algorithm for the estimation of model parameters is inferred and designed in Section 3. Experimental results on three data sets are presented in Section 4. The paper summarizes our model and looks forward to future work in Section 5.

2. A Generative Model for Topic Discovery and Word Embeddings on Attributed Networks

We aim to design a joint probabilistic model for topic discovery and word embedding considering the links and contents of an attributed network. It integrates the contents and links to improve the performance of topic discovery, which is further used to learn word embeddings, degrading the representation fuzziness of polysemy. In this section, we present the steoLC model in detail. The graphical representation of the steoLC model is shown in Figure 1.
The steoLC model is composed of two parts. One part generates the skip-grams in a document given the topic distribution, which is similar to the STE model [12]. To get a more accurate topic distribution, our model extends the STE model by combining the links between documents with contents. The other part models the generative process of each link of a document given a particular topic. The following gives a detailed introduction of the two parts.
Component of generating document contents. There are N documents, noted as a set of documents D = { d 1 , , d N } . The i -th document d i has L i words, and its content information is denoted as a word sequence d i = { w 1 , , w L i } . The generative process of document content models for each skip-gram in a document, i.e., contexts w t + j , ( j [ n , n ] , j 0 ) , given each central word w t . For a central word w t , its context words include n words that appear before w t and the n words that follow w t . Each pair of ( w t + j , w t ) is a skip-gram of the document d i . As in the STE model, the probability of ( w t + j , w t ) depends on the topic z of the central word w t and the embeddings of w t as a central word and w t + j as a context word. We assume the document set includes K topics, and each document has a topic distribution on the K topics, noted as p ( z | d ) .
The probability of each skip-gram ( w t + j , w t ) in a document d is computed in the following.
p ( w t + j w t , d ) = z p ( z d ) p ( w t + j w t , z )
The probability of skip-gram < w t + j , w t > , given the topic z, is evaluated by
p ( w t + j w t , z ) = exp ( V w t + j , z · U w t , z ) w exp ( V w , z · U w t , z ) ,
where U w and V w are the word embedding matrix with K × S dimension as the central word and a context, respectively, S is the dimension of the embedding space, and ∧ is the vocabulary number of the document set.
Component of generating document links. This section describes the generative process of links between two documents. Link set C is the set of all possible links between every two documents in the corpus, C = { C 1 , C 2 , , C N } . C i is the link set of the document d i . It is assumed that a document d connects to c since they exist the same topic z. The document d has a topic distribution p ( z | d ) in terms of its links and contents. The probability from d to c, given all the topics, is computed as:
p ( c d ) = z p ( z d ) p ( c z ) .
After finishing the definition of the two components, a complete generative process of links and contents in an attributed network is shown in the following.
For each document d i ( i { 1 , , N } ) :
   (a) Draw a topic z ( z { 1 , , K } ) according to p ( z | d i ) ;
   (b) Draw a link c l ( l 1 , , | C i | ) from documents d i with probability p ( c l | z ) ;
   (c) For each central word, w t ( t { L 1 , , L i } ) in d i :
         Draw a topic z according to p ( z | d i ) ;
         Draw each context w t + j p ( w t + j | w t , z ) .
Optimizating objection. According to the generative process, the likelihood of generating links and contents on an attributed network is defined in the following.
p ( D ) = i = 1 N p ( d i ) p ( C i ) = i = 1 N { ( t = 1 L i j = n , j 0 n p ( w t + j | w t , d i ) ) × l = 1 | C i | p ( c l | d i ) } = i = 1 N ( t = 1 L i j = n , j 0 n z = 1 K p z d i p w t + j w t , z ) × l = 1 | C i | z = 1 K p z d i p c l z
where p ( z | d i ) is shared by link modeling and content modeling, which integrates the document contents and document links by a common probabilistic distribution for more accurate topic discovery.
The above function is difficult to compute, and is always transformed as the log-likelihood as follows:
log p ( D ) = i = 1 N ( t = 1 L i j = n , j 0 n log p ( w t + j | w t , d i ) ) + i = 1 N l = 1 | C i | log p ( c l | d i )
In real applications, content components and link components have different importance. The loglikelihoods of the two components in Equation (5) are assigned with a weight α and a weight 1 α , respectively, as follows:
L = α i = 1 N t = 1 L i j = n , j 0 n log p ( w t + j w t , d i ) + ( 1 α ) i = 1 N l = 1 | C i | log p ( c l d i ) .

3. Parameter Estimating Algorithm

In this section, we introduce how to learn the values of model parameters given a corpus, including the word embeddings U , V , as well as the topic distribution of documents p ( z | d ) , the link distribution on topics p ( c | z ) , and the skip-gram distribution on topics p ( w t + j | w t , z ) . Next, the algorithm of estimating parameters is described and its complexity is analyzed.

3.1. EM Algorithm with Negative Sampling for the setoLC Model

Since the parameters of the steoLC model contain hidden variables z, the EM algorithm is used to maximize the likelihood of Equation (6). In addition, to estimate parameters U , V in Equation (2), the negative sampling algorithm is used. The algorithm framework of the steoLC model combines the EM algorithm with the negative sampling algorithm. It is summarized in Algorithm 1.
Algorithm 1. The algorithm for the steoLC model.
Input: word sequence setDand link setC
Output: U , V , p ( z | d ) , p ( c | z )
1:
Initialize U , V , p ( z | d ) , p ( c | z ) .
2:
for iter=1 to Max_iteration do
3:
    for each document d i in D do
4:
        for each skip-gram < w t + j , w t > in d i  do
5:
           Sample negative instances from distribution P.
6:
           Update p ( w t + j w t , z ) , p ( z d i , w t , w t + j ) by Equations (13) and (7).
7:
           Update U, V using the gradient decent method with Equations (14) and (15).
8:
        for each link c l in C i  do
9:
           Update p ( z d i , c l ) by Equation (8).
10:
    for each document d i in D do
11:
        Update p ( z d i ) using Equation (11).
12:
        for each link c l in d i  do
13:
           Update p ( c l z ) using Equation (12)
The EM algorithm is an iterative algorithm, and each iteration includes an E step and an M step. In the E step, the distributions of hidden variables z, given each skip-gram < w t + j , w t > and given each link < d , c > , are estimated. At the same time, the expectation of the log-likelihood is computed in the E step. In the M step, by maximizing the expectation of the likelihood, the algorithm updates the word embeddings U , V , as well as p ( z | d ) and p ( c | z ) . To update the word embedding matrices, the algorithm iterates over each skip-gram and several negative sampling instances with the gradient descent method. The algorithm is inferred in detail.
In the E-step, the posterior probability distribution of hidden variables given each skip-gram < w t + j , w t > in d i is evaluated by the Bayes rule as follows:
p ( z d i , w t , w t + j ) = p ( w t + j w t , z ) p ( z d i ) z = 1 K p ( w t + j w t , z ) p ( z d i ) .
Further, the posterior probability distribution of hidden variables given each link < d i , c l > from the document d i is computed as:
p ( z d i , c l ) = p ( c l z ) p ( z d i ) z = 1 | K | p ( c l z , ) p ( z d i ) .
Next, the hidden variable posterior distributions obtained from Equations (7) and (8) are used to compute the expectation of the log-likelihood, which is defined as:
Q = α i = 1 N t = 1 L i j = n , j 0 n z = 1 K p ( z d i , w t , w t + j ) × log ( p ( w t + j w t , z ) × p ( z d i ) ) + ( 1 α ) i = 1 N l = 1 | C i | z = 1 K p ( z d i , c l ) × log ( p ( c l z ) × p ( z d i ) )
The topic distribution given a document satisfies the constraint z = 1 K p z d i = 1 . The skip-gram distribution given a topic satisfies the constraint j = n , j 0 n p w t + j w t , z = 1 . The link distribution given a topic satisfies the constraints l = 1 | C i | p ( c l z ) = 1 .
The Lagrange multiplier method is used to estimate model parameters. The Lagrange function combines the expectation of the log-likelihood function in Equation (9) with the three constraint conditions defined as:
F = α i = 1 N t = 1 L i j = n , j 0 n z = 1 K p ( z d , w t , w t + j ) log ( p ( w t + j w t , z ) p ( z d i ) ) + ( 1 α ) i = 1 N l = 1 | C i | z = 1 K p ( z d i , c l ) log ( p ( c l z , ) p ( z d i ) ) + i = 1 N β 1 ( 1 z = 1 K p ( z d i ) ) + z = 1 | K | β 2 ( 1 l = 1 | C i | p ( c l z ) ) + z = 1 K t = 1 L i β 3 ( 1 j = n , j 0 n p ( w t + j w t , z ) )
In order to maximize the Lagrange function, the partial derivatives with respect to p ( z d i ) , p ( c l z ) are computed. The updating equations of the topic distribution of a document and the link distribution for each topic are obtained by making partial derivatives equal to zero. The equations are as follows:
p ( z d i ) = α t = 1 L i j = n , j 0 n p ( z d i , w t , w t + j ) + ( 1 α ) l = 1 | C i | p ( z d i , c l ) z = 1 | K | ( α t = 1 L i j = n , j 0 n p ( z d i , w t , w t + j ) + ( 1 α ) l = 1 | C i | p ( z d i , c l ) )
p ( c l z ) = i = 1 N p ( z d i , c l ) i = 1 N l = 1 | C i | p ( z d i , c l )
To update the word embeddings, U , V , the immediate method is to compute the partial derivatives of U , V in Equation (2). However, w e x p ( V w t + j , z · U w t , z ) is intractible. The negative sampling method is used to compute U , V , in which the probability distribution of p ( w t + j | w t , z ) is computed as follows:
log p ( w t + j w t , z ) log σ ( V w t + j , z · U w t , z ) + i = 1 m E w i P ( log σ ( V w i , z · U w t , z ) )
where σ represents the sigmoid function, σ ( x ) = 1 / ( 1 + exp ( x ) ) , and w i is a negative instance which is sampled from the distribution P ( ) . P ( w ) is a unigram distribution Unigram(w) raised to the 3 4 r d power [12].
The negative sampling algorithm is a gradient algorithm. The gradient of the objective function concerning U and V is as follows:
L U w t , z = w { w t + j } { W n e g } ( ( ξ w w t σ ( V w , z · U w t , z ) ) · V w , z ) × p ( z d i , w t , w t + j ) U w t , z = U w t , z + d × L U w t , z
L V w , z = ( ξ w w t σ ( V w , z · U w t , z ) · U w t , z ) × p ( z d i , w t , w t + j ) , V w , z = V w , z + d × L V w , z ,
where W n e g is the negative instances corresponding to w t , d, d are step lengths, and
ξ w w t = 1 , if w is a word in the context window of w t . 0 , otherwise .

3.2. Complexity Analysis of the Algorithm

The process of steoLC is shown in Algorithm 1. In each iteration, the algorithm first computes the distributions of hidden variables p ( z | d i , w t + j , w t ) , p ( z | d i , c l ) in line 3–9. Since p ( w t + j | w t , z , d i ) affects p ( z | d i , w t + j , w t ) in Equation (7), it is computed before p ( z | d i , w t + j , w t ) in line 7. At the same time, the parameters of word embedding matrices U and V are updated by a gradient decent method in line 7, and the number of the iterations is set as a constant. The number of document words is L 1 + L 2 + + L N , which is simplified as L. Next, all the skip-grams are N × L × n × 2 . The time complexity of line 3–7 is O ( N × L × n × N e g × K ) , where n is the length of left or right window sampling contexts, N e g represents the number of negative samples, and K is the number of topics. The time complexity of P ( z k | d i , c l ) in line 8–9 is O ( C × K ) , where C is the number of links among the documents. The time complexities of p ( z | d i ) and p ( c l | z ) are, respectively, O ( N × K ) and O ( C × K ) . To summarized, the time complexity of the whole algorithm is O ( M a x _ i t e r a t i o n × ( N × L × n × N e g × K + C × K ) ) , where M a x _ i t e r a t i o n is the iteration times.

4. Results

In this section, we present a detailed analysis of the performance of our method. We first introduce several experimental settings, including the dataset, parameter settings, and the baselines used in the experiment. Next, document classification is used to evaluate the quality of document representations on the topic distribution estimated by the steoLC model. Subsequently, we explain whether word embedding results are able to model polysemous words on topics by qualitative analysis. Finally, we evaluate the topic coherence of the steoLC model.

4.1. Experiment Settings

Dataset description. To verify the performance of the steoLC model, we experiment on three public datasets as the CeTe model [19], including the DBLP dataset [20] and two different scale hep-th datasets. The DBLP dataset includes many papers on the computer field, and the five largest categories are used to construct attributed networks. The titles and abstracts are represented as the contents of papers, and the citation relationships correspond to the links of papers. The hep-th includes a large corpus of physics-related papers. The four largest categories form one subdataset, named large-hep. The three smaller categories form the other subdataset were called small-hep. The dataset details are shown in Table 2.
Baselines. The algorithm of the steoLC model is compared with three state-of-the-art methods. The first model is a traditional topic model, i.e., Latent Dirichlet Allocation (LDA) [2], which models topic discovery on one hot word space. The STE model [12] integrates topic discovery with word embedding based on texts of documents and attempts to solve polysemy. The CeTe model also combines topic discovery with word embedding. In addition, it improves the performance by utilizing community detection on links as a pre-processing step of topic embedding modeling. The CeTe model has been compared with three kinds of state-of-the-art methods for specific-topic embedding modeling. If the performance of our algorithm is better than the CeTe model, our algorithm is superior to the three kinds of methods.
Parameter settings and Hardware specifications. In the experiments, the number of topics K is set as 10, such that topical specific word embedding on a fine-grained topic is possible. The iteration number of the document set and the number of gradient descent for U , V are both set as 15. The dimension of the embedding space is 400. The size n of the context window equals 10. For each skip-gram, eight negative instances are sampled. These settings are similar to the ste model. The steoLC model aims to learn word embeddings, document embeddings, and topic embeddings. Thus, the component of content is more important, and the α is set as 0.75 experimentally. We compare the accuracy of detected topics for classification at the small-hep data set by varying α from 0 to 1, and the accuracy on the test is shown in Figure 2. Computer configurations for the experiments are a CPU-Intel i7, with 16GB of internal storage and 1T of Hard Disk, etc.

4.2. Document Classification

To test whether our algorithm has learned accurate semantics for documents, a classification task is used. The document set is randomly divided into 80% as the training set and 20% as the test set [19]. The document representation learned by all the algorithms by Equation (11) is the input of the classifier. We use the l 1 regularized linear SVM one-vs-all classifier. We compare the steoLC model with the LDA model and the STE model. All the models represent a document by the topic distribution of the document. The performance is measured by accuracy, recall, and the F1-Score.
Each model is run ten times, and the mean and standard deviation of the results are computed and shown in Table 3. Our steoLC model performs the best among the three models. Compared with the distributed representation of the steoLC model, the LDA model discovers topics and learns word embeddings using only the contents represented by the one hot word vector, which suffers from sparsity and semantic gaps. Different from our steoLC model combining contents and links, the STE model only considers content information, although it represents words as dense embeddings. This method may captain vague topics. The CeTe model utilizes community detection on links to degrade the fuzziness of topic discovery, but the community detection is not integrated with topic discovery in a unified framework. Our steoLC model constructs a joint model for contents and links in a generative process, which discovers more accurate semantic topics. In addition, it combines topic modeling with word embedding modeling, such that the two tasks are improve each other. We also find that the results on the large-hep dataset are worse than the ones on the small-hep dataset. This is because each category in the large-hep dataset has more documents and is more noisy than each category in the small-hep dataset, which increases the difficulty of distinguishing the pattern of each category on the former dataset compared with the latter one.

4.3. Qalitative Analysis

To illustrate the performance of integrating word embedding with topic discovery in our model, we use the t-SNE algorithm [21] to visualize the vectors of the 500 most frequent words that are learned from the steoLC and STE models in the small-hep dataset in Figure 3. Similar to the STE model [12], the number of topics K is set as 10. The number of outer iterations and inner iterations are both set to 15. The dimension of the embedding vectors is set as 400. For each skip-gram, we set the window size to 10 and sample eight negative instances.
Each node in Figure 3 denotes a topic-specific word vector. The algorithm of the steoLC model divides the whole space into 10 subspaces, and each subspace represents a topic. To illustrate the advantages of the steoLC model in dealing with polysemous words, we show labels for the word “review”, as an example of a polysemous word and the word “comment” as an example of a monosemous word. The labels show both the word and topic index, separated by “#”. In Figure 3b, we find that the steoLC model divides the whole space into six disjoint subspaces, and each subspace corresponds to a topic. We set the topic number as 10, and our model detects just six topics clearly. This illustrates that several topics are difficult to divide, such as Topic 7,5,9, Topics 6,8, and Topics 2,4. This guides us to select a small topic number. Additionally, “review” and “comment” are close to each other in some subspaces, but far apart in others. For example, under Topic 1, the word “review” is far away from the word “comment”, which indicates that the word “review” does not mean “comment” under Topic 1. Under Topic 0, the word “review” is closer to the word “comment”, which indicates that the word “review” means “comment”. On the other hand, Figure 3a illustrates that some word embeddings of different topics from the STE model are mixed together, which means that different topics are difficult to discriminate. However, the steoLC model can discover topics more accurately by combining the contents and links, such that the word embeddings in the same topic are similar and word embeddings in different topics are able to be distinguished.
Table 4 shows the nearest neighbors of some polysemous words according to the skip-gram model, the STE model, and the steoLC model. Cosine similarity is used to compute the similarity between polysemous words and other words. We observe that these similar words, according to word embeddings learned by skip-gram, mix different senses of the given words. For example, the nearest neighbors of “present” are “show” and “recently”, which indicates that the skip-gram model cannot distinguish different meanings of words. In contrast, the STE and steoLC models can distinguish polysemy due to the consideration of topic modeling. Under Topic 1, the most similar words of “present” are “show”, “description”, and “action”, which clearly corresponds to the meaning of appearance. Under Topic 2, they are “recent” and “recently”, clearly referring to ’lately’. In addition, Table 4 shows that similar words from the steoLC model are better than the ones from the STE model. For example, the most similar word to “argue” obtained by the steoLC model under Topic 1, namely “argue# 1”, has more meaning than the ones obtained by the STE model. The reason is that the steoLC model makes use of links to obtain more accurate topic distributions, which further aids in learning accurate word embeddings.

4.4. Topic Coherence Evaluation

Table 5 compares the top words produced by the STE and steoLC models on three topics detected from the small-hep dataset. In Table 5, Topic 1 is about mathematics, Topic 2 is about cosmology, and Topic 3 is about physics. Both the STE and steoLC models produce words with similar themes. However, the STE model discovers fewer meaningful words related to these three topics. Some words from the STE model are not coherent. For example, for Topic 2, STE produces “conference” and “transition”, which are less related to cosmology, while the hot words produced by steoLC are all related to cosmology. This shows that the steoLC model is able to discover more coherent topics than the STE model.
In addition, a topic coherence measure is used to evaluate the models objectively. Coherence of a set of words measures the hanging and fitting together of single words or subsets of them. The method presented in [22] can cover all existing measures and construct new measures, which demonstrates that the coherence measure ( C v ) has the best performance among PMI, NPMI, UMass, etc. The higher the value, the better the coherence. C v combines the indirect cosine measure with the NPMI and the boolean sliding window. For each top word w i of a topic word set, its vector v i is calculated based on context words, and the l -element v i l is the NMPI value with the l -context word w l is as follows:
v i l = N P M I ( w i , w l ) = log p ( w i , w l ) + ϵ p ( w i ) p ( w l ) log ( P ( w i , w l ) + ϵ ) .
Probability, P ( w i , w l ) , is estimated based on word co-occurrence counts derived from virtual documents by a sliding window over Wikipedia. Each window position defines such a document.
Next, topic coherence C v is averaged on the cosine similarity summation of any two top words w i and w j as C v = a v g ( 1 i j N c o s ( v i , v j ) ) , where N is the number of top words of a topic, and a v g ( ) and c o s ( ) are the average function and the cosine similarity function, respectively. C v is realized by tool gensim.
Table 6 shows the coherence values of C v for the STE and steoLC models on the small-hep datasets. We can see that our steoLC model generally improves the coherence of the learned topics. Compared with the STE model, our model incorporates link information to improve the quality of the detected topics.

5. Discussion and Conclusions

A novel joint probabilistic model, the steoLC model, is first provided for topic discovery and polysemy word embedding based on document link networks. An algorithm is then designed to learn the model parameters by the EM algorithm with a negative sampling algorithm. The algorithm of the steoLC model learns the topic distribution on documents, and the embedding distribution on topics for words as central words and context words, as well as the link distribution on topics. This method not only mines topics and word embeddings by integrating contents with links, but also solves the polysemous problem, reducing the fuzzy semantic of topics. Compared with the state-of-the-art methods on several tasks, the experimental results demonstrate the superiority of our model. However, the proposed algorithm is difficult to apply on large-scale data due to the complexity of our probabilistic model. It may be helpful to use a python library for probabilistic models, such as ZhuSuan. In addition, the steoLC model does not model multi-granularity texts in the same representation space, which will affect the performance and explainability of topic discovery and text representations. In the future, we will further use the graph convolutional neural network or the depth generation model to address these problems.

Author Contributions

Methodology, B.C., X.J. and J.G.; investigation, B.C. and L.M.; data curation, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Natural Science Foundation of Hebei Province (No.F201940 3070) and the science and technology research project for universities of Hebei (ZD2020175, ZD2020344) and the Science and Technology Innovation Team Project of Hebei GEO University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is downloaded from https://dblp.uni-trier.de/xml/ and https://www.cs.cornell.edu/projects/kddcup/datasets.html, accessed on 26 February 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hofmann, T. Probabilistic Latent Semantic Analysis. In Proceedings of the Fifteenth Conference on Uncertainty in Artificial Intelligence (UAI’99), Stockholm, Sweden, 30 July–1 August 1999; Laskey, K.B., Prade, H., Eds.; Morgan Kaufmann: Burlington, MA, USA, 1999; pp. 289–296. [Google Scholar]
  2. Blei, D.M.; Ng, A.Y.; Jordan, M.I. Latent Dirichlet Allocation. J. Mach. Learn. Res. 2003, 3, 993–1022. [Google Scholar]
  3. Bengio, Y.; Ducharme, R.; Vincent, P.; Janvin, C. A Neural Probabilistic Language Model. J. Mach. Learn. Res. 2003, 3, 1137–1155. [Google Scholar]
  4. Li, S.; Chua, T.S.; Zhu, J.; Miao, C. Generative Topic Embedding: A Continuous Representation of Documents. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 7–13 August 2016; pp. 666–675. [Google Scholar]
  5. Das, R.; Zaheer, M.; Dyer, C. Gaussian LDA for Topic Models with Word Embeddings. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Beijing, China, 26–31 July 2015. [Google Scholar]
  6. Nguyen, D.Q.; Billingsley, R.; Du, L.; Johnson, M. Improving Topic Models with Latent Feature Word Representations. Trans. Assoc. Comput. Linguist. 2015, 3, 299–313. [Google Scholar] [CrossRef]
  7. Liu, Y.; Liu, Z.; Chua, T.; Sun, M. Topical Word Embeddings. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; Bonet, B., Koenig, S., Eds.; AAAI Press: Cambridge, MA, USA, 2015; pp. 2418–2424. [Google Scholar]
  8. Briakou, E.; Athanasiou, N.; Potamianos, A. Cross-Topic Distributional Semantic Representations Via Unsupervised Mappings. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, 2–7 June 2019; (Long and Short Papers). Burstein, J., Doran, C., Solorio, T., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; Volume 1, pp. 1052–1061. [Google Scholar] [CrossRef]
  9. Xun, G.; Li, Y.; Gao, J.; Zhang, A. Collaboratively Improving Topic Discovery and Word Embeddings by Coordinating Global and Local Contexts. In Proceedings of the 23rd ACM SIGKDD International Conference, Halifax, NS, Canada, 13–17 August 2017. [Google Scholar]
  10. Zhu, L.; He, Y.; Zhou, D. A Neural Generative Model for Joint Learning Topics and Topic-Specific Word Embeddings. Trans. Assoc. Comput. Linguist. 2020, 8, 471–485. [Google Scholar] [CrossRef]
  11. Li, D.; Zhang, J.; Li, P. TMSA: A Mutual Learning Model for Topic Discovery and Word Embedding. In Proceedings of the 2019 SIAM International Conference on Data Mining, SDM 2019, Calgary, AB, Canada, 2–4 May 2019; Berger-Wolf, T.Y., Chawla, N.V., Eds.; SIAM: Philadelphia, PA, USA, 2019; pp. 684–692. [Google Scholar] [CrossRef] [Green Version]
  12. Shi, B.; Lam, W.; Jameel, S.; Schockaert, S.; Lai, K.P. Jointly learning word embeddings and latent topics. In Proceedings of the SIGIR 2017 International Conference, Tokyo, Japan, 7–11 August 2017. [Google Scholar]
  13. Sun, H.; He, F.; Huang, J.; Sun, Y.; Li, Y.; Wang, C.; He, L.; Sun, Z.; Jia, X. Network Embedding for Community Detection in Attributed Networks. ACM Trans. Knowl. Discov. Data 2020, 14, 36:1–36:25. [Google Scholar] [CrossRef]
  14. Chunaev, P. Community detection in node-attributed social networks: A survey. Comput. Sci. Rev. 2020, 37, 100286. [Google Scholar] [CrossRef]
  15. Bothorel, C.; Cruz, J.D.; Magnani, M.; Micenková, B. Clustering Attributed Graphs: Models, Measures and Methods. arXiv 2015, arXiv:1501.01676. [Google Scholar] [CrossRef] [Green Version]
  16. Liu, X.; Wang, W.; He, D.; Jiao, P.; Jin, D.; Cannistraci, C.V. Semi-supervised community detection based on non-negative matrix factorization with node popularity. Inf. Sci. 2017, 381, 304–321. [Google Scholar] [CrossRef]
  17. Falih, I.; Grozavu, N.; Kanawati, R.; Bennani, Y. Community detection in Attributed Network. In Proceedings of the Companion of the The Web Conference 2018 on The Web Conference 2018, Lyon, France, 23–27 April 2018; Champin, P., Gandon, F.L., Lalmas, M., Ipeirotis, P.G., Eds.; ACM: New York, NY, USA, 2018; pp. 1299–1306. [Google Scholar] [CrossRef] [Green Version]
  18. Yang, T.; Jin, R.; Chi, Y.; Zhu, S. Combining Link and Content for Community Detection. In Proceedings of the Acm Sigkdd International Conference on Knowledge Discovery Data Mining, Paris, France, 28 June–1 July 2009. [Google Scholar]
  19. Jin, D.; Huang, J.; Jiao, P.; Yang, L.; He, D.; Fogelman-Soulié, F.; Huang, Y. A Novel Generative Topic Embedding Model by Introducing Network Communities. In Proceedings of the The World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; ACM: New York, NY, USA, 2019; pp. 2886–2892. [Google Scholar] [CrossRef]
  20. Tang, J.; Zhang, J.; Yao, L. ArnetMiner: Extraction and Mining of Academic Social Networks. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data, Las Vegas, NV, USA, 24–27 August 2008; pp. 990–998. [Google Scholar]
  21. Laurens, V.D.M.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  22. RoDer, M.; Both, A.; Hinneburg, A. Exploring the Space of Topic Coherence Measures; ACM: New York, NY, USA, 2015; pp. 399–408. [Google Scholar]
Figure 1. Graphical representation of the steoLC model. Part 1 in the box outlined by dotted lines denotes the component of generating document contents. Part 2 in the box outlined by solid lines denotes the component of generating document links.
Figure 1. Graphical representation of the steoLC model. Part 1 in the box outlined by dotted lines denotes the component of generating document contents. Part 2 in the box outlined by solid lines denotes the component of generating document links.
Symmetry 14 00703 g001
Figure 2. Classification accuracy on the small-hep data set for the steoLC model ( 0 α 1 ).
Figure 2. Classification accuracy on the small-hep data set for the steoLC model ( 0 α 1 ).
Symmetry 14 00703 g002
Figure 3. Visualization of the word embeddings learned by the ste and steoLC models. The polysemous word “review” and the monosemous word “comment” are highlighted for comparison.
Figure 3. Visualization of the word embeddings learned by the ste and steoLC models. The polysemous word “review” and the monosemous word “comment” are highlighted for comparison.
Symmetry 14 00703 g003
Table 1. The SOTA methods for topic discovery and word embeddings.
Table 1. The SOTA methods for topic discovery and word embeddings.
Word Embedding-Based Topic Discovery
Classical ModelsTopicVec [4], GaussianLDA [5], LF-LDA [6]
AdvantagesWord embeddings supplement topic discovery
DisadvantagesWrong word embeddings limit topic discovery
Topic Discovery-Based Word Embeddings
Classical ModelsTWE [7]
AdvantagesPre-trained topics improve word embeddings
DisadvantagesUnaccuracy topics make word embeddings worse
Mutual Model for Topic Discovery and Word Embeddings
Classical ModelsCTM [9], TMSA [11], STE [12]
AdvantagesTopic discovery and word embeddings improve mutually
DisadvantagesRich Links are not modeled
Table 2. Dataset information.
Table 2. Dataset information.
DatasetDocumentsEdgesWordsCategories
DBLP693612,353506,2695
Small-hep39781218,7183
Large-hep11,752134,956622,6424
Table 3. Document classification effect of different methods.
Table 3. Document classification effect of different methods.
DatasetModelAccuracyRecallF1
small-hepLDA0.287 ± 0.0020.287 ± 0.0010.281 ± 0.001
STE0.668 ± 0.0210.647 ± 0.0190.649 ± 0.008
CeTe0.671 ± 0.0070.650 ± 0.0060.654 ± 0.002
steoLC0.698 ± 0.0020.675 ± 0.0020.682 ± 0.001
large-hepLDA0.414 ± 0.0020.411 ± 0.0290.401 ± 0.003
STE0.432 ± 0.0130.432 ± 0.0080.415 ± 0.005
CeTe0.437 ± 0.0050.433 ± 0.0010.417 ± 0.002
steoLC0.469 ± 0.0030.462 ± 0.0030.446 ± 0.001
DBLPLDA0.700 ± 0.0110.457 ± 0.0050.404 ± 0.003
STE0.756 ± 0.0330.729 ± 0.020.725 ± 0.009
CeTe0.790 ± 0.0080.736 ± 0.0050.731 ± 0.003
steoLC0.812 ± 0.0020.756 ± 0.0190.748 ± 0.005
Table 4. The nearest neighbors of some polysemous words.
Table 4. The nearest neighbors of some polysemous words.
ModelWordsSimilar Words
Word2vecarguedemonstrate review consider
STEargue#1
argue#2
prove gauge evidence
introduce consider present
steoLCargue#1
argue#2
demonstrate prove conclude
conjectured consider suggest
Word2vecpresentshow action recently
STEpresent#1
present#2
show description presence
recently recent reviewed
steoLCpresent#1
present#2
show reveals exhibit
recently resent conventional
Table 5. The top words on three topics for the STE and steoLC models.
Table 5. The top words on three topics for the STE and steoLC models.
TopicMethodWord
Topic1STE
steoLC
model formulae surface negative boundary
model formulae surface matrix relation
Topic2STE
steoLC
time cosmological polyhedra conference transition
time cosmological polyhedra infrared domain
Topic3STE
steoLC
quantization supergravity entropy electric create
quantization supergravity entropy geometry calculation
Table 6. Topic coherence evaluation with different numbers of top words.
Table 6. Topic coherence evaluation with different numbers of top words.
ModelT = 5T = 10T = 15
STE0.5100.4560.463
steoLC0.52104960.496
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chai, B.; Ji, X.; Guo, J.; Ma, L.; Zheng, Y. A Generative Model for Topic Discovery and Polysemy Embeddings on Directed Attributed Networks. Symmetry 2022, 14, 703. https://doi.org/10.3390/sym14040703

AMA Style

Chai B, Ji X, Guo J, Ma L, Zheng Y. A Generative Model for Topic Discovery and Polysemy Embeddings on Directed Attributed Networks. Symmetry. 2022; 14(4):703. https://doi.org/10.3390/sym14040703

Chicago/Turabian Style

Chai, Bianfang, Xinyu Ji, Jianglin Guo, Lixiao Ma, and Yibo Zheng. 2022. "A Generative Model for Topic Discovery and Polysemy Embeddings on Directed Attributed Networks" Symmetry 14, no. 4: 703. https://doi.org/10.3390/sym14040703

APA Style

Chai, B., Ji, X., Guo, J., Ma, L., & Zheng, Y. (2022). A Generative Model for Topic Discovery and Polysemy Embeddings on Directed Attributed Networks. Symmetry, 14(4), 703. https://doi.org/10.3390/sym14040703

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop