729
Views
2
CrossRef citations to date
0
Altmetric
Articles

Towards Autoencoding Variational Inference for Aspect-Based Opinion Summary

, &

Figures & data

Figure 1. Aspect-based opinion summary result of one product.

Figure 1. Aspect-based opinion summary result of one product.

Figure 2. Probabilistic graphical model of LDA and JST.

Figure 2. Probabilistic graphical model of LDA and JST.

Figure 3. A general autoencoder system.

Figure 3. A general autoencoder system.

Figure 4. Variational auto encoder model where observable variable x and its correspondent latent z are distributed on Gaussian distribution N(μx|z,Σx|z) and N(μz|x,Σz|x), respectively.

Figure 4. Variational auto encoder model where observable variable x and its correspondent latent z are distributed on Gaussian distribution N(μx|z,Σx|z) and N(μz|x,Σz|x), respectively.

Figure 6. AutoEncoding variational inference for aspect discovery. As illustrated, the yellow block θ and β is corresponded to the document-topic and topic-word distributions which is described in , respectively. Meanwhile, γ and γprior are additional blocks which play an important role in the aspect discovery task.

Figure 6. AutoEncoding variational inference for aspect discovery. As illustrated, the yellow block θ and β is corresponded to the document-topic and topic-word distributions which is described in Figure 2a, respectively. Meanwhile, γ and γprior are additional blocks which play an important role in the aspect discovery task.

Figure 5. Prior distribution matrix.

Figure 5. Prior distribution matrix.

Table 1. Discovered aspects (bold text indicates seed words).

Figure 7. Example of constructing the square loss term. Firstly, every rows in the γ matrix is normalized via softmax function. After that, a submatrix is constructed by choosing only set of words in the normalized matrix γ such that this word must also exists in a given γprior matrix. Finally, the prior loss is computed via the Euclidean distance between this submatrix and γprior.

Figure 7. Example of constructing the square loss term. Firstly, every rows in the γ matrix is normalized via softmax function. After that, a submatrix is constructed by choosing only set of words in the normalized matrix γ such that this word must also exists in a given γprior matrix. Finally, the prior loss is computed via the Euclidean distance between this submatrix and γprior.

Figure 8. Autoencoding variational inference for joint sentiment/topic modeling where θ, β and π are the corresponding latent variables in JST graphical model (). Moreover, the yellow block which is used to mapping document x to latent variable π can be represented as any modern classification deep neural network.

Figure 8. Autoencoding variational inference for joint sentiment/topic modeling where θ, β and π are the corresponding latent variables in JST graphical model (Figure 2b). Moreover, the yellow block which is used to mapping document x to latent variable π can be represented as any modern classification deep neural network.

Table 2. AVIAD and WLDA seedwords.

Table 3. Statistics for IMDB and Yelp datasets.

Table 4. Topics extracted by AVIAD and WLDA.

Table 5. IMDB Topics extracted by AVIJST and JST.

Table 6. Yelp Topics extracted by AVIJST and JST.

Figure 9. Aspect discovery performance on restaurant dataset.

Figure 9. Aspect discovery performance on restaurant dataset.

Figure 10. Joint sentiment/topic modeling performance based on average topic coherent.

Figure 10. Joint sentiment/topic modeling performance based on average topic coherent.

Table 7. Sentiment words discovered.

Table 8. Aspect identification results.

Table 9. Accuracy on test set for IMDB.

Table 10. Accuracy on test set for Yelp.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.